Jeff Dean, Google's Chief Scientist (and guy whose paycheck depends on you believing AI is amazing), went on The Moonshot Podcast to announce that AI is now better than "average humans" at most tasks. According to Dean, when faced with something completely new, most people suck at it, but AI is "pretty reasonable at most things."
Cool story. Of course, this is coming from someone whose entire career is built on convincing people that Google's AI is revolutionary. It's like asking a car salesman if you really need that extended warranty.
Here's what Dean's not talking about: AI is great at things that have clear answers and lots of training data. Ask ChatGPT to write a generic email or summarize a Wikipedia article? Sure, it's probably better than the average person. But try getting it to navigate office politics, figure out why the coffee machine is making that weird noise, or explain to your kid why their goldfish died without traumatizing them. AI hallucination is still a massive problem. GPT-4 fails at basic math more often than you'd expect. Google's Bard has been caught making up facts in its own launch demo.
The "average human" comparison is doing a lot of heavy lifting here. AI might be better than humans at trivia, pattern recognition, and tasks where failure doesn't matter. But humans are still better at the stuff that actually matters: common sense, dealing with unexpected situations, and not confidently stating that there are two Rs in "strawberry."
Dean admits AI "will fail at a lot of things" and isn't "human expert level" in specialized fields. Which is a fancy way of saying AI is mediocre at most things but confident about it. That's not necessarily better than humans who know when they don't know something.
The timing of these claims is interesting. Google is desperately trying to justify the billions they're spending on AI development while competitors like OpenAI and Anthropic are breathing down their necks. Microsoft's partnership with OpenAI is eating into Google's search dominance. Alphabet's stock has been under pressure from AI competition. Declaring victory over "average humans" is great for investor presentations, less great for users who've tried to get AI to do anything complex. Google's AI search initiatives are their attempt to stay relevant in the AI race.
So yeah, AI might be better than humans at generating generic content and answering well-defined questions. Whether that makes it "better than most people at most tasks" depends on how you define tasks and whether you value accuracy over confidence.