A team of researchers at Facebook’s parent company Meta has come up with a new benchmark to gauge the abilities of AI assistants like OpenAI’s large language model GPT-4.
And judging by current standards, OpenAI’s current crop of AI models are all… still pretty stupid.
The team, which includes “AI godfather” and Meta chief scientist Yann LeCun, came up with an exam called GAIA that’s made up of 466 questions that “are conceptually simple for humans yet challenging for most advanced AIs,” per a yet-to-be-peer-reviewed paper.
Comments are closed.