More on this book
Community
Kindle Notes & Highlights
Read between
November 17, 2019 - March 27, 2020
As we’ve seen, there are many ways to accidentally sabotage an AI by giving it faulty or inadequate data. But there’s another kind of AI failure in which we discover that they’ve succeeded in doing what we asked, but what we asked them to do isn’t what we actually wanted them to do.
1. They develop their own ways of solving a problem rather than relying on step-by-step instructions from a programmer. 2. They lack the contextual knowledge to understand when their solutions are not what humans would have preferred.
“I’ve taken to imagining [AI] as a demon that’s deliberately misinterpreting your reward and actively searching for the laziest possible local optima. It’s a bit ridiculous, but I’ve found it’s actually a productive mindset to have,” writes Alex Irpan, an AI researcher at Google.
An increased tension to a very old view of computer programming. The computer will do exactly what you told it to do with the worst possible result.