Oh dear. There’s a language processing module called GPT-3 which is really good at generating natural English text, and it was coupled to an experimental medical diagnostic program. You might be able to guess where this is going.
LeCun cites a recent experiment by the medical AI firm NABLA, which found that GPT-3 is woefully inadequate for use in a healthcare setting because writing coherent sentences isn’t the same as being able to reason or understand what it’s saying.
…
After testing it in a variety of medical scenarios, NABLA found that there’s a huge difference between GPT-3 being able to form coherent sentences and actually being useful.
For example…
Um, yikes?
Human language is really hard and messy, and medicine is also extremely complicated, and maybe AI isn’t quite ready for something with the multiplicative difficulty of trying to combine the two.
How about something much simpler? Like steering a robot car around a flat oval track? Sure, that sounds easy.
Roborace is the world first driver-less/autonomous motorsports category.
This is one of their first live-broadcasted events.
This was the second run.
It drove straight into a wall. pic.twitter.com/ss5R2YVRi3
— Ryan (@dogryan100) October 29, 2020
Now I’m scared of both robotic telemedicine and driverless cars.
Or maybe I should just combine and simplify and be terrified of software engineers.