logo
#

Latest news with #AlphaProof

Google AI system wins gold medal in International Mathematical Olympiad
Google AI system wins gold medal in International Mathematical Olympiad

The Star

time24-07-2025

  • Science
  • The Star

Google AI system wins gold medal in International Mathematical Olympiad

SAN FRANCISCO: An artificial intelligence system built by Google DeepMind, the tech giant's primary AI lab, has achieved 'gold medal' status in the annual International Mathematical Olympiad, a premier math competition for high school students. It was the first time that a machine – which solved five of the six problems at the 2025 competition, held in Australia this month – reached that level of success, Google said in a blog post Monday. The news is another sign that leading companies are continuing to improve their AI systems in areas such as math, science and computer coding. This kind of technology could accelerate the research of mathematicians and scientists and streamline the work of experienced computer programmers. Two days before Google revealed its feat, an OpenAI researcher said in a social media post that the startup had built technology that achieved a similar score on this year's questions, although it did not officially enter the competition. Both systems were chatbots that received and responded to the questions much like humans. Other AI systems have participated in the International Mathematical Olympiad, or IMO, but they could answer questions only after human experts translated them into a computer programming language built for solving math problems. 'We solved these problems fully in natural language,' Thang Luong, a senior staff research scientist at Google DeepMind, said in an interview. 'That means there was no human intervention – at all.' After OpenAI started the AI boom with the release of ChatGPT in late 2022, the leading chatbots could answer questions, write poetry, summarise news articles, even write a little computer code. But they often struggled with math. Over the past two years, companies such as Google and OpenAI have built AI systems better suited to mathematics, including complex problems that the average person cannot solve. Last year, Google DeepMind unveiled two systems that were designed for math: AlphaGeometry and AlphaProof. Competing in the IMO, these systems achieved 'silver medal' performance, solving four of the competition's six problems. It was the first time a machine reached silver-medal status. Other companies, including a startup called Harmonic, have built similar systems. But systems such as AlphaProof and Harmonic are not chatbots. They can answer questions only after mathematicians translate the questions into Lean, a computer programming language designed for solving math problems. This year, Google entered the IMO with a chatbot that could read and respond to questions in English. This system is not yet available to the public. Called Gemini Deep Think, the technology is what scientists call a 'reasoning' system. This kind of system is designed to reason through tasks involving math, science and computer programming. Unlike previous chatbots, this technology can spend time thinking through complex problems before settling on an answer. Other companies, including OpenAI, Anthropic and China's DeepSeek, offer similar technologies. Like other chatbots, a reasoning system initially learns its skills by analysing enormous amounts of text culled from across the internet. Then it learns additional behaviour through extensive trial and error in a process called reinforcement learning. A reasoning system can be expensive, because it spends additional time thinking about a response. Google said Deep Think had spent the same amount of time with the IMO as human participants did: 4 1/2 hours. But the company declined to say how much money, processing power or electricity had been used to complete the test. In December, an OpenAI system surpassed human performance on a closely watched reasoning test called ARC-AGI. But the company ran afoul of competition rules because it spent nearly US$1.5mil (RM6.3mil) in electricity and computing costs to complete the test, according to pricing estimates. – ©2025 The New York Times Company This article originally appeared in The New York Times.

World's First AI Model Wins Gold At International Math Olympiad. Check Details
World's First AI Model Wins Gold At International Math Olympiad. Check Details

NDTV

time22-07-2025

  • Science
  • NDTV

World's First AI Model Wins Gold At International Math Olympiad. Check Details

Google's artificial intelligence (AI) research arm DeepMind has won a gold medal at the International Mathematical Olympiad (IMO), the world's most prestigious competition for young mathematicians. It is the first time a machine has solved five of the six problems in algebra, combinatorics, geometry, and number theory -- signalling a breakthrough in math capabilities of AI systems that can rival human intelligence. IMO problems are known for their difficulty, and solving them requires a deep understanding of mathematical concepts -- something which the AI models had not been able to achieve up until now. However, an advanced version of Gemini Deep Think managed to ace the competition where 67 contestants, or about 11 per cent, achieved gold-medal scores. "We can confirm that Google DeepMind has reached the much-desired milestone, earning 35 out of a possible 42 points, a gold medal score. Their solutions were astonishing in many respects. IMO graders found them to be clear, precise and most of them easy to follow," said IMO President Dr Gregor Dolinar. Last year, DeepMind's combined AlphaProof and AlphaGeometry 2 systems achieved the silver-medal standard in the competition, but it took two to three days of computation. This year, the advanced Gemini model operated end-to-end in natural language and managed to produce the results within the 4.5-hour competition time limit. The DeepMind team trained the advanced Gemini model on novel reinforcement learning techniques that can leverage more multi-step reasoning, problem-solving and theorem-proving data. "We'll be making a version of this Deep Think model available to a set of trusted testers, including mathematicians, before rolling it out to Google AI Ultra subscribers," DeepMind CEO Demis Hassabis wrote on X (formerly Twitter). Prior to Google, an OpenAI researcher also claimed that the startup had built technology that achieved a similar score on this year's questions, though it did not officially enter the competition. 1/N I'm excited to share that our latest @OpenAI experimental reasoning LLM has achieved a longstanding grand challenge in AI: gold medal-level performance on the world's most prestigious math competition—the International Math Olympiad (IMO). — Alexander Wei (@alexwei_) July 19, 2025 The advancements shown by the AI systems suggest that the technology was less than a year away from being used by mathematicians to crack unsolved research problems at the frontier of the field. "I think the moment we can solve hard reasoning problems in natural language will enable the potential for collaboration between AI and mathematicians," Junehyuk Jung, a math professor at Brown University and visiting researcher in DeepMind AI unit, was quoted as saying by Reuters.

Google DeepMind's AI earns gold medal at International Math Olympiad
Google DeepMind's AI earns gold medal at International Math Olympiad

Business Upturn

time21-07-2025

  • Business
  • Business Upturn

Google DeepMind's AI earns gold medal at International Math Olympiad

Google DeepMind has achieved a major breakthrough in AI performance by reaching gold medal status at the prestigious 2025 International Mathematical Olympiad (IMO), one of the world's toughest math competitions for high school students. The AI system, a new and improved version of DeepMind's Gemini Deep Think, scored 35 out of 42 points, solving five out of six complex math problems with perfect accuracy. This marks a significant jump from last year, when DeepMind's earlier AI models (AlphaProof and AlphaGeometry 2) earned a silver medal with 28 points. The IMO's President, Prof. Dr. Gregor Dolinar, confirmed the achievement, praising the AI's solutions as 'clear, precise,' and often 'easy to follow.' That's a major compliment, given that the Olympiad problems are designed to challenge some of the brightest young minds on the planet. What makes this feat even more impressive is how the AI approached the problems. Unlike earlier systems, which needed human experts to rewrite the questions into computer-friendly formats, Gemini Deep Think worked entirely in natural language, reading and understanding the exact same problems given to human contestants, and solving them within the official 4.5-hour time limit. The model achieved this by using advanced reasoning tools, including something DeepMind calls 'parallel thinking,' which lets the AI explore multiple problem-solving strategies at the same time, similar to how skilled mathematicians think through different angles before committing to a solution. The system was also trained using new reinforcement learning techniques and a curated set of high-quality math solutions to boost its problem-solving abilities. Google plans to make this version of Gemini available first to trusted testers, including professional mathematicians, before rolling it out to Google AI Ultra subscribers. The development comes amid growing competition in the AI space. OpenAI also announced that one of its experimental models hit the same score, 35 out of 42, in the same IMO challenge. With both companies now demonstrating gold-level performance on a human math competition, this signals a major leap forward in AI's ability to reason logically, handle abstract problems, and produce rigorous, formal proofs. Google CEO Sundar Pichai described the progress as 'just astounding,' while DeepMind CEO Demis Hassabis called it 'incredible progress' in AI's ability to handle high-level intellectual tasks. While the AI isn't replacing mathematicians anytime soon, its success opens up new possibilities for AI-assisted research, education, and mathematical discovery, and represents a milestone in machines tackling one of the most human of challenges: creative problem-solving through logic.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store