Advertisement

Humans Outshine Google And OpenAI AI At Prestigious Math Olympiad Despite Record Scores

Humans outperformed Google and OpenAIs AI models at the International Mathematical Olympiad in Australia.

Humans Outshine Google And OpenAI AI At Prestigious Math Olympiad Despite Record Scores
While the AI achieved gold-level scores, five young competitors scored full marks.

At the International Mathematical Olympiad (IMO) held this month in Queensland, Australia, human participants triumphed over cutting-edge artificial intelligence models developed by Google and OpenAI.

For the first time, these AI models achieved gold-level scores in the prestigious competition. Google announced on Monday that its advanced Gemini chatbot successfully solved five out of six challenging problems. However, neither Google's Gemini nor OpenAI's AI reached a perfect score.

In contrast, five talented young mathematicians under the age of 20 achieved full marks, outperforming the AI models. The IMO, regarded as the world's toughest mathematics competition for students, showcased that human intuition and problem-solving skills still hold an edge over AI in complex reasoning tasks.

This result highlights that while generative AI is advancing rapidly, it has yet to surpass the brightest human minds in all areas of intellectual competition.

"We can confirm that Google DeepMind has reached the much-desired milestone, earning 35 out of a possible 42 points a gold medal score," the US tech giant cited IMO president Gregor Dolinar as saying.

"Their solutions were astonishing in many respects. IMO graders found them to be clear, precise and most of them easy to follow."

Around 10 percent of human contestants won gold-level medals, and five received perfect scores of 42 points.

US ChatGPT maker OpenAI said that its experimental reasoning model had scored a gold-level 35 points on the test.

The result "achieved a longstanding grand challenge in AI" at "the world's most prestigious math competition", OpenAI researcher Alexander Wei wrote on social media.

"We evaluated our models on the 2025 IMO problems under the same rules as human contestants," he said.

"For each problem, three former IMO medalists independently graded the model's submitted proof."

Google achieved a silver-medal score at last year's IMO in the British city of Bath, solving four of the six problems.

That took two to three days of computation -- far longer than this year, when its Gemini model solved the problems within the 4.5-hour time limit, it said.

The IMO said tech companies had "privately tested closed-source AI models on this year's problems", the same ones faced by 641 competing students from 112 countries.

"It is very exciting to see progress in the mathematical capabilities of AI models," said IMO president Dolinar.

Contest organisers could not verify how much computing power had been used by the AI models or whether there had been human involvement, he cautioned.

(With inputs from AFP)

Listen to the latest songs, only on JioSaavn.com