logo
Google AI system wins gold medal in International Mathematical Olympiad

Google AI system wins gold medal in International Mathematical Olympiad

The Star4 days ago
SAN FRANCISCO: An artificial intelligence system built by Google DeepMind, the tech giant's primary AI lab, has achieved 'gold medal' status in the annual International Mathematical Olympiad, a premier math competition for high school students.
It was the first time that a machine – which solved five of the six problems at the 2025 competition, held in Australia this month – reached that level of success, Google said in a blog post Monday.
The news is another sign that leading companies are continuing to improve their AI systems in areas such as math, science and computer coding. This kind of technology could accelerate the research of mathematicians and scientists and streamline the work of experienced computer programmers.
Two days before Google revealed its feat, an OpenAI researcher said in a social media post that the startup had built technology that achieved a similar score on this year's questions, although it did not officially enter the competition.
Both systems were chatbots that received and responded to the questions much like humans. Other AI systems have participated in the International Mathematical Olympiad, or IMO, but they could answer questions only after human experts translated them into a computer programming language built for solving math problems.
'We solved these problems fully in natural language,' Thang Luong, a senior staff research scientist at Google DeepMind, said in an interview. 'That means there was no human intervention – at all.'
After OpenAI started the AI boom with the release of ChatGPT in late 2022, the leading chatbots could answer questions, write poetry, summarise news articles, even write a little computer code. But they often struggled with math.
Over the past two years, companies such as Google and OpenAI have built AI systems better suited to mathematics, including complex problems that the average person cannot solve.
Last year, Google DeepMind unveiled two systems that were designed for math: AlphaGeometry and AlphaProof. Competing in the IMO, these systems achieved 'silver medal' performance, solving four of the competition's six problems. It was the first time a machine reached silver-medal status. Other companies, including a startup called Harmonic, have built similar systems.
But systems such as AlphaProof and Harmonic are not chatbots. They can answer questions only after mathematicians translate the questions into Lean, a computer programming language designed for solving math problems.
This year, Google entered the IMO with a chatbot that could read and respond to questions in English. This system is not yet available to the public.
Called Gemini Deep Think, the technology is what scientists call a 'reasoning' system. This kind of system is designed to reason through tasks involving math, science and computer programming. Unlike previous chatbots, this technology can spend time thinking through complex problems before settling on an answer.
Other companies, including OpenAI, Anthropic and China's DeepSeek, offer similar technologies.
Like other chatbots, a reasoning system initially learns its skills by analysing enormous amounts of text culled from across the internet. Then it learns additional behaviour through extensive trial and error in a process called reinforcement learning.
A reasoning system can be expensive, because it spends additional time thinking about a response. Google said Deep Think had spent the same amount of time with the IMO as human participants did: 4 1/2 hours. But the company declined to say how much money, processing power or electricity had been used to complete the test.
In December, an OpenAI system surpassed human performance on a closely watched reasoning test called ARC-AGI. But the company ran afoul of competition rules because it spent nearly US$1.5mil (RM6.3mil) in electricity and computing costs to complete the test, according to pricing estimates. – ©2025 The New York Times Company
This article originally appeared in The New York Times.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

AI is replacing search engines as a shopping guide, research suggests
AI is replacing search engines as a shopping guide, research suggests

The Star

time16 hours ago

  • The Star

AI is replacing search engines as a shopping guide, research suggests

Finding products, comparing prices and browsing reviews: Until now, you'd have done most of this in a search engine like Google. But that era appears to be ending thanks to AI, research shows. — Photo: Christin Klose/dpa COPENHAGEN: Three in four people who use AI are turning to the likes of ChatGPT, Gemini and Copilot to get advice and recommendations on shopping and travel instead of using the previous online method of search engines like Google, new research shows. AI-supported online shopping is done at least occasionally by 76% of AI users, with 17% doing so most or even all of the time, according to a study conducted by the market research institute Norstat on behalf of Verdane, a leading European investment company. The changes in consumer search behaviour pose a major challenge not only for search engine providers like Google but also for manufacturers and retailers, who must adapt to maintain their visibility in the AI-driven world. AI chatbots have emerged as powerful tools for tracking down specific products, often providing helpful advice in response to complex and specific queries. Of the survey respondents, 3% are dedicated AI enthusiasts who always use AI tools instead of search engines when shopping online, while 14% said they mostly use AI and 35% do so occasionally. A total of 7,282 people from the UK, Germany, Sweden, Norway, Denmark and Finland aged between 18 and 60 participated in the survey in June. The highest proportion of AI use is in online travel research, at 33%. This is followed by consumer electronics (22%), DIY and hobby supplies (20%), and software or digital subscriptions (19%). However, AI usage is still relatively low in fashion and clothing (13%), cosmetics (12%), and real estate (7%). Among AI tools, ChatGPT is far ahead of its competitors and 86% of AI users regularly use OpenAI's chatbot. This is followed at a considerable distance by Google's Gemini (26% regular users) and Microsoft's Copilot (20%). The Chinese AI bot DeepSeek, which has been the subject of heated debate among AI experts and data protection advocates, appears to have no significant role among consumers in Europe. – dpa

‘It's the most empathetic voice in my life': How AI is transforming the lives of neurodivergent people
‘It's the most empathetic voice in my life': How AI is transforming the lives of neurodivergent people

The Star

timea day ago

  • The Star

‘It's the most empathetic voice in my life': How AI is transforming the lives of neurodivergent people

-For Cape Town-based filmmaker Kate D'hotman, connecting with movie audiences comes naturally. Far more daunting is speaking with others. 'I've never understood how people [decipher] social cues,' the 40-year-old director of horror films says. D'hotman has autism and attention-deficit hyperactivity disorder (ADHD), which can make relating to others exhausting and a challenge. However, since 2022, D'hotman has been a regular user of ChatGPT, the popular AI-powered chatbot from OpenAI, relying on it to overcome communication barriers at work and in her personal life. 'I know it's a machine,' she says. 'But sometimes, honestly, it's the most empathetic voice in my life.' Neurodivergent people — including those with autism, ADHD, dyslexia and other conditions — can experience the world differently from the neurotypical norm. Talking to a colleague, or even texting a friend, can entail misread signals, a misunderstood tone and unintended impressions. AI-powered chatbots have emerged as an unlikely ally, helping people navigate social encounters with real-time guidance. Although this new technology is not without risks — in particular some worry about over-reliance — many neurodivergent users now see it as a lifeline. How does it work in practice? For D'hotman, ChatGPT acts as an editor, translator and confidant. Before using the technology, she says communicating in neurotypical spaces was difficult. She recalls how she once sent her boss a bulleted list of ways to improve the company, at their request. But what she took to be a straightforward response was received as overly blunt, and even rude. Now, she regularly runs things by ChatGPT, asking the chatbot to consider the tone and context of her conversations. Sometimes she'll instruct it to take on the role of a psychologist or therapist, asking for help to navigate scenarios as sensitive as a misunderstanding with her best friend. She once uploaded months of messages between them, prompting the chatbot to help her see what she might have otherwise missed. Unlike humans, D'hotman says, the chatbot is positive and non-judgmental. That's a feeling other neurodivergent people can relate to. Sarah Rickwood, a senior project manager in the sales training industry, based in Kent, England, has ADHD and autism. Rickwood says she has ideas that run away with her and often loses people in conversations. 'I don't do myself justice,' she says, noting that ChatGPT has 'allowed me to do a lot more with my brain.' With its help, she can put together emails and business cases more clearly. The use of AI-powered tools is surging. A January study conducted by Google and the polling firm Ipsos found that AI usage globally has jumped 48%, with excitement about the technology's practical benefits now exceeding concerns over its potentially adverse February, OpenAI told Reuters that its weekly active users surpassed 400 million, of which at least 2 million are paying business users. But for neurodivergent users, these aren't just tools of convenience and some AI-powered chatbotsare now being created with the neurodivergent community in mind. Michael Daniel, an engineer and entrepreneur based in Newcastle, Australia, told Reuters that it wasn't until his daughter was diagnosed with autism — and he received the same diagnosis himself — that he realised how much he had been masking his own neurodivergent traits. His desire to communicate more clearly with his neurotypical wife and loved ones inspired him to build Neurotranslator, an AI-powered personal assistant, which he credits with helping him fully understand and process interactions, as well as avoid misunderstandings. 'Wow … that's a unique shirt,' he recalls saying about his wife's outfit one day, without realising how his comment might be perceived. She asked him to run the comment through NeuroTranslator, which helped him recognise that, without a positive affirmation, remarks about a person's appearance could come across as criticism. 'The emotional baggage that comes along with those situations would just disappear within minutes,' he says of using the app. Since its launch in September, Daniel says NeuroTranslator has attracted more than 200 paid subscribers. An earlier web version of the app, called Autistic Translator, amassed 500 monthly paid subscribers. As transformative as this technology has become, some warn against becoming too dependent. The ability to get results on demand can be 'very seductive,' says Larissa Suzuki, a London-based computer scientist and visiting NASA researcher who is herself neurodivergent. Overreliance could be harmful if it inhibits neurodivergent users' ability to function without it, or if the technology itself becomes unreliable — as is already the case with many AI search-engine results, according to a recent study from the Columbia Journalism Review.'If AI starts screwing up things and getting things wrong,' Suzuki says, 'people might give up on technology, and on themselves." Baring your soul to an AI chatbot does carry risk, agrees Gianluca Mauro, an AI adviser and co-author of Zero to AI. 'The objective [of AI models like ChatGPT] is to satisfy the user,' he says, raising questions about its willingness to offer critical advice. Unlike therapists, these tools aren't bound by ethical codes or professional guidelines. If AI has the potential to become addictive, Mauro adds, regulation should follow. A recent study by Carnegie Mellon and Microsoft (which is a key investor in OpenAI) suggests that long-term overdependence on generative AI tools can undermine users' critical-thinking skills and leave them ill-equipped to manage without it. 'While AI can improve efficiency,' the researchers wrote, 'it may also reduce critical engagement, particularly in routine or lower-stakes tasks in which users simply rely on AI.' While Dr. Melanie Katzman, a clinical psychologist and expert in human behaviour, recognises the benefits of AI for neurodivergent people, she does see downsides, such as giving patients an excuse not to engage with others. A therapist will push their patient to try different things outside of their comfort zone. "I think it's harder for your AI companion to push you," she says. But for users who have come to rely on this technology, such fears are academic. 'A lot of us just end up kind of retreating from society,' warns D'hotman, who says that she barely left the house in the year following her autism diagnosis, feeling overwhelmed. Were she to give up using ChatGPT, she fears she would return to that traumatic period of isolation. 'As somebody who's struggled with a disability my whole life,' she says, 'I need this.' (Editing by Yasmeen Serhan and Sharon Singleton)

People are starting to talk more like ChatGPT
People are starting to talk more like ChatGPT

The Star

time2 days ago

  • The Star

People are starting to talk more like ChatGPT

Artificial intelligence, the theory goes, is supposed to become more and more human. Chatbot conversations should eventually be nearly indistinguishable from those with your fellow man. But a funny thing is happening as people use these tools: We're starting to sound more like the robots. A study by the Max Planck Institute for Human Development in Berlin has found that AI is not just altering how we learn and create, it's also changing how we write and speak. The study detected 'a measurable and abrupt increase' in the use of words OpenAI's ChatGPT favours – such as delve, comprehend, boast, swift, and meticulous – after the chatbot's release. 'These findings,' the study says, 'suggest a scenario where machines, originally trained on human data and subsequently exhibiting their own cultural traits, can, in turn, measurably reshape human culture.' Researchers have known ChatGPT-speak has already altered the written word, changing people's vocabulary choices, but this analysis focused on conversational speech. Researchers first had OpenAI's chatbot edit millions of pages of emails, academic papers, and news articles, asking the AI to 'polish' the text. That let them discover the words ChatGPT favoured. Following that, they analysed over 360,000 YouTube videos and 771,000 podcasts from before and after ChatGPT's debut, then compared the frequency of use of those chatbot-favoured words, such as delve, realm, and meticulous. In the 18 months since ChatGPT launched, there has been a surge in use, researchers say – not just in scripted videos and podcasts but in day-to-day conversations as well. People, of course, change their speech patterns regularly. Words become part of the national dialogue and catch-phrases from TV shows and movies are adopted, sometimes without the speaker even recognising it. But the increased use of AI-favoured language is notable for a few reasons. The paper says the human parroting of machine-speak raises 'concerns over the erosion of linguistic and cultural diversity, and the risks of scalable manipulation.' And since AI trains on data from humans that are increasingly using AI terms, the effect has the potential to snowball. 'Long-standing norms of idea exchange, authority, and social identity may also be altered, with direct implications for social dynamics,' the study says. The increased use of AI-favoured words also underlines a growing trust in AI by people, despite the technology's immaturity and its tendency to lie or hallucinate. 'It's natural for humans to imitate one another, but we don't imitate everyone around us equally,' study co-author Levin Brinkmann tells Scientific American. 'We're more likely to copy what someone else is doing if we perceive them as being knowledgeable or important.' The study focused on ChatGPT, but the words favoured by that chatbot aren't necessarily the same standbys used by Google's Gemini or Perplexity's Claude. Linguists have discovered that different AI systems have distinct ways of expressing themselves. ChatGPT, for instance, leans toward a more formal and academic way of communicating. Gemini is more conversational, using words such as sugar when discussing diabetes, rather than ChatGPT's favoured glucose, for instance. (Grok was not included in the study, but, as shown with its recent meltdown, where it made a series of antisemitic comments – something the company attributed to a problem with a code update – it heavily favours a flippant tone and wordplay.) 'Understanding how such AI-preferred patterns become woven into human cognition represents a new frontier for psycholinguistics and cognitive science,' the Max Planck study says. 'This measurable shift marks a precedent: machines trained on human culture are now generating cultural traits that humans adopt, effectively closing a cultural feedback loop.' – Inc./Tribune News Service

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store