
AIs are getting better at maths problems
Andresr/ Getty Images
Experimental AI Models Achieve Gold-Level Performance in International Math Olympiad
For the first time, experimental AI models from Google DeepMind and OpenAI have achieved a gold-level performance in the International Mathematical Olympiad (IMO). This milestone has significant implications for the future of AI in solving complex scientific and mathematical problems.
The achievement has been hailed as a major breakthrough by the companies involved, but mathematicians are urging caution as the details of the models’ results and their workings have not been fully disclosed.
AI Tackles a Prestigious Competition
The IMO is renowned as one of the most prestigious competitions for young mathematicians, serving as a challenging test for AI systems due to its focus on mathematical reasoning. Last year, Google DeepMind introduced AI systems called AlphaProof and AlphaGeometry, which achieved a silver medal-level performance but were not officially graded by the competition.
Prior to this year’s competition in Queensland, Australia, companies like Google, Huawei, and ByteDance, along with academic researchers, sought permission to have their AI models officially graded by the IMO. The request was granted under the condition that the results would be announced after the closing ceremonies.
AI Success at the Olympiad
OpenAI’s new AI system and Google DeepMind’s Gemini Deep Think both achieved gold medal scores in the IMO, answering challenging mathematical questions with high accuracy. While Google’s previous systems used a computer programming language, Lean, this year’s models operated entirely in natural language, making their solutions more accessible and applicable to broader AI systems.
Both companies utilized reinforcement learning, a training method that allows AI to learn through trial and error, to improve their models’ problem-solving capabilities. Google’s Gemini Deep Think also employed parallel thinking and was trained on a dataset specifically tailored for the IMO.
Future Implications and Challenges
While the progress in AI’s mathematical abilities is promising, mathematicians like Terence Tao and Geordie Williamson emphasize the need for more transparency and detailed research on the models’ performance. Natural language processing in AI systems could offer significant advancements but may also pose challenges in verifying complex proofs.
Both Google and OpenAI plan to make their systems available for testing by mathematicians before releasing them to the public. The potential for these AI models to tackle unsolved scientific problems is promising, opening up new avenues for research and innovation.
As AI continues to evolve and enhance its mathematical capabilities, the collaboration between AI researchers and mathematicians will be crucial in ensuring the accuracy and reliability of AI solutions for complex problems.