- Reaction score
- 1,697
IN A NEW INTELLIGENCE TEST, AI GOT *DESTROYED* BY HUMANS.
A team of researchers at Facebook's parent company Meta has come up with a new benchmark to gauge the abilities of AI assistants like OpenAI's large language model GPT-4.
And judging by current standards, OpenAI's current crop of AI models are all... still pretty stupid.
The team, which includes "AI godfather" and Meta chief scientist Yann LeCun, came up with an exam called GAIA that's made up of 466 questions that "are conceptually simple for humans yet challenging for most advanced AIs," per a yet-to-be-peer-reviewed paper.
The results speak for themselves: human respondents were capable of correctly answering 92 percent of the questions, while GPT-4, even equipped with some manually selected plugins, scored a measly 15 percent. OpenAI's recently-released GPT4 Turbo scored less than ten percent, according to the team's published GAIA leaderboard.
A team of researchers at Facebook's parent company Meta has come up with a new benchmark to gauge the abilities of AI assistants like OpenAI's large language model GPT-4.
And judging by current standards, OpenAI's current crop of AI models are all... still pretty stupid.
The team, which includes "AI godfather" and Meta chief scientist Yann LeCun, came up with an exam called GAIA that's made up of 466 questions that "are conceptually simple for humans yet challenging for most advanced AIs," per a yet-to-be-peer-reviewed paper.
The results speak for themselves: human respondents were capable of correctly answering 92 percent of the questions, while GPT-4, even equipped with some manually selected plugins, scored a measly 15 percent. OpenAI's recently-released GPT4 Turbo scored less than ten percent, according to the team's published GAIA leaderboard.
Facebook Researchers Test AI's Intelligence and Find It Is Unfortunately Quite Stupid
A team of researchers at Facebook's parent company Meta has come up with a new benchmark to gauge the intelligence of AI.
futurism.com