&w=3840&q=100)
AI models of Google and OpenAI win milestone gold at global math contest
The results marked the first time that AI systems crossed the gold-medal scoring threshold at the International Mathematical Olympiad for high-school students. Both companies' models solved five out of six problems, achieving the result using general-purpose "reasoning" models that processed mathematical concepts using natural language, in contrast to the previous approaches used by AI firms.
The achievement suggests AI is less than a year away from being used by mathematicians to crack unsolved research problems at the frontier of the field, according to Junehyuk Jung, a math professor at Brown University and visiting researcher in Google's DeepMind AI unit.
"I think the moment we can solve hard reasoning problems in natural language will enable the potential for collaboration between AI and mathematicians," Jung told Reuters.
OpenAI's breakthrough was achieved with a new experimental model centered on massively scaling up "test-time compute." This was done by both allowing the model to "think" for longer periods and deploying parallel computing power to run numerous lines of reasoning simultaneously, according to Noam Brown, researcher at OpenAI. Brown declined to say how much in computing power it cost OpenAI, but called it "very expensive." To OpenAI researchers, it is another clear sign that AI models can command extensive reasoning capabilities that could expand into other areas beyond math.
The optimism is shared by Google researchers, who believe AI models' capabilities can apply to research quandaries in other fields such as physics, said Jung, who won an IMO gold medal as a student in 2003.
Google's DeepMind AI unit last year achieved a silver medal score using AI systems specialized for math. This year, Google used a general-purpose model called Gemini Deep Think, a version of which was previously unveiled at its annual developer conference in May.
Unlike previous AI attempts that relied on formal languages and lengthy computation, Google's approach this year operated entirely in natural language and solved the problems within the official 4.5-hour time limit, the company said in a blog post.
OpenAI, which has its own set of reasoning models, similarly built an experimental version for the competition, according to a post by researcher Alexander Wei on social media platform X.
He noted that the company does not plan to release anything with this level of math capability for several months.
This year marked the first time the competition coordinated officially with some AI developers, who have for years used prominent math competitions like IMO to test model capabilities.
IMO judges certified the results of those companies, including Google, and asked them to publish results on July 28.
"We respected the IMO Board's original request that all AI labs share their results only after the official results had been verified by independent experts and the students had rightly received the acclamation they deserved," Google DeepMind CEO Demis Hassabis said on X on Monday.
OpenAI, which published its results on Saturday and first claimed gold-medal status, said in an interview that it had permission from an IMO board member to do so after the closing ceremony on Saturday.
The competition on Monday allowed cooperating companies to publish results, Gregor Dolinar, president of IMO's board, told Reuters.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Indian Express
26 minutes ago
- Indian Express
What is Wide Research, Manus' new multi-agent AI tool to take on OpenAI and Google?
Manus, an AI startup with roots in China, has announced a new feature that ropes in scores of AI agents in order to carry out complex, large-scale tasks that require information on hundreds of items. The feature known as Wide Research is designed to make deep, high-volume research effortless, Manus said in its announcement blog post published on Thursday, July 31. It is said to be the company's most significant launch since its debut in March this year. Wide Research is seen as Manus' answer to similar research-focused AI features and tools introduced by US rivals such as OpenAI's Deep Research and Google's Deep Think. It also comes weeks after OpenAI's ChatGPT Agent made its debut. 'With the launch of Wide Research, Manus unlocks a powerful new way for users to tackle complex, large-scale tasks that require information on hundreds of items,' Manus said. 'Wide Research marks a milestone in our exploration of the scaling laws of AI agents, but it's just the beginning. The infrastructure behind it is part of something much bigger we're building,' it added. The newly unveiled feature will be immediately available for paid Pro customers, with plans to roll out to Plus and Basic tier users in the future. Wide Research is the first feature rolled out by Manus that has been built on top of its large-scale virtualization infrastructure and highly efficient agent architecture which reportedly took 'months of optimisation'. 'At its core, Wide Research is a system-level mechanism for parallel processing, and a protocol for agent-to-agent collaboration,' the company said. Wide Research differs from other multi-agent systems such as specialised AI agent managers or coding assistants because the multiple AI agents that carry out the tasks are general-purpose. 'This generality unlocks flexibility: tasks are not constrained to rigid formats or predefined domains. In practice, the only real limit is your imagination,' Manus said. In a video posted on X, Manus co-founder Peak Ji demonstrated how Wide Research can be used to quickly conduct research on 100 different sneakers simultaneously or create 50 different poster designs in a matter of minutes. According to Manus, users will be able to perform deep, high-volume research tasks such as exploring Fortune 500 companies, comparing top MBA programs, or diving into generative AI tools. However, Ji also said that Wide Research is still in the experimental phase and warned that it may come with certain limitations. Manus made a splash in the AI landscape earlier this year with the launch of its general-purpose AI agent that is capable of autonomously carrying out web-based tasks such as creating trip itineraries in response to basic prompts. In June this year, the company released its own AI video generator. Manus' AI features are reportedly wrapped around other large language models like Anthropic's Claude. 'Behind every Manus session runs a dedicated cloud-based virtual machine, allowing users to orchestrate complex cloud workloads — simply by talking to an agent,' the company said. 'From generating tailored rental presentations to safely evaluating cutting-edge open-source projects in a secure sandbox, the Turing-completeness of the virtual machine is what gives Manus its generality — and opens the door to endless creative possibilities,' it added. Manus recently shifted its operations out of China to Singapore, Tokyo and San Mateo, California. It currently does not offer its AI products and tools in China, according to a report by Bloomberg. Recently, the AI startup raised $75 million in a funding round led by US venture firm Benchmark. Its early investors include Chinese tech giants such as Tencent as well as HSG (formerly Sequoia China) and ZhenFund.


Hindustan Times
26 minutes ago
- Hindustan Times
Google Pixel 10 Pro Fold 5G mobile: Price in India, design, camera, performance, and everything to expect
Google is expected to launch its new generation foldable smartphone, the Pixel 10 Pro Fold 5G, in India on August 20, 2025. The foldable will be accompanied by three other models, the Pixel 10, Pixel 10 Pro, and finally the Pixel 10 Pro XL. As launch is just around the corner, several leaks about the foldable have started to circulate, giving us a glimpse of what the Pixel 10 Pro Fold will look like in terms of upgrades and new features. As of now, it is anticipated that the smartphone may look similar to its predecessor, but could get some design upgrades with a slimmer bezel, improved water protection, and more. Let's have a closer look at what the Google Pixel 10 Pro Fold 5G will look like during the launch. Google Pixel 10 Pro Fold 5G is likely to be powered by the Tensor G5 chip in India. Here's everything we know so far.(Google) Google Pixel 10 Pro Fold 5G: Design and display Last year, Google made a significant design revamp with the Pixel 9 Pro Fold, bringing a more attractive look. Now, we expect the Pixel 10 Pro Fold to retain a similar design with a flat design, a satin metal frame, Gorilla Glass Victus protection, and slimmer bezels. The foldable is expected to offer an IP68 rating for protection against dust and water, making it the first foldable smartphone with advanced durability. The smartphone is expected to come in two new Moonstone and Jade colours. For display, the Google Pixel 10 Pro Fold is expected to feature a 6.4-inch cover display and an 8-inch main foldable OLED display. It may offer a 120Hz refresh rate and more than 2700nits peak brightness. Google Pixel 10 Pro Fold 5G: Camera The Google Pixel 10 Pro Fold 5G is expected to feature a triple camera setup that includes a 50MP main camera with a Samsung GN8 sensor, a 10.5MP ultrawide and a 10.8MP telephoto lens with 5x optical zoom. Therefore, the main camera is getting a slight upgrade over its predecessor. The smartphone will also include a dual front-facing camera, 10MP on the cover and 10MP on the main screen. Google Pixel 10 Pro Fold 5G: Performance and battery The Google Pixel 10 Pro Fold 5G is powered by the Tensor G5 chipset, likely to be built with TSMC's 3nm process. The smartphone will likely offer up to 16GB of RAM and up to 1TB of internal storage. The foldable will run on Android 16, which is getting a Material 3 Expressive redesign. For a lasting performance, the foldable will likely be powered by a 5015 mAh battery that may support 23W of wired charging. Google Pixel 10 Pro Fold 5G: Price The Google Pixel 10 Pro Fold 5G is expected to be priced similarly to its predecessor. Last year, the Pixel 9 Pro Fold was launched at a starting price of Rs. 1,72,999. Now, as per leaked US pricing, the new-gen is expected to cost $1799.


Economic Times
an hour ago
- Economic Times
"ChatGPT is not a diary, therapist, lawyer, or friend": LinkedIn user warns against oversharing everything with AI
ChatGPT users are being warned to think twice before typing anything personal into the chatbot. OpenAI CEO Sam Altman recently confirmed that interactions with ChatGPT aren't protected by confidentiality laws. Conversations you assume are private may be stored, reviewed, and even presented in court — no matter how sensitive, emotional or casual they seem.'If you go talk to ChatGPT about your most sensitive stuff and then there's like a lawsuit or whatever, we could be required to produce that, and I think that's very screwed up,' Altman said in an interview on the This Past Weekend podcast. He added, 'We should have, like, the same concept of privacy for your conversations with AI that we do with a therapist or whatever.'But as of now, that legal framework doesn't explained, 'Right now, if you talk to a therapist or a lawyer or a doctor about those problems, there's legal privilege for it. There's confidentiality. We haven't figured that out yet for ChatGPT.'This sharp warning is echoed by Shreya Jaiswal, a Chartered Accountant and founder of Fawkes Solutions, who posted her concerns on LinkedIn. Her message was blunt and alarming. 'ChatGPT can land you in jail. No, seriously. Not even joking,' she to Jaiswal, Altman's own words spell out the legal dangers. 'Sam Altman – the CEO of OpenAI, literally said that anything you type into ChatGPT can be used as evidence in court. Not just now, even months or years later, if needed. There's no privacy, no protection, nothing, unlike talking to a real lawyer or therapist who is sworn to client confidentiality.'She laid out a few scenarios that, while hypothetical, are disturbingly someone types: 'I cheated on my partner and I feel guilty, is it me or the stars that are misaligned?' Jaiswal pointed out how this could resurface in a family court battle. 'Boom. You're in court 2 years later fighting an alimony or custody battle. That chat shows up. And your 'private guilt trip' just became public proof.' Even seemingly harmless curiosity can be risky. 'How do I save taxes using all the loopholes in the Income Tax Act?' or 'How can I use bank loans to become rich like Vijay Mallya?' could be interpreted as intent during a future audit or legal probe. 'During a tax audit or loan default, this could easily be used as evidence of intent even if you never actually did anything wrong,' she warned. In another example, she highlighted workplace risk. 'I'm thinking of quitting and starting my own company. How can I use my current company to learn for my startup?' This, she argued, could be used against you in a lawsuit for breach of contract or intellectual property theft. 'You don't even need to have done anything. The fact that you thought about it is enough.'Jaiswal expressed concern that people have become too casual, even intimate, with AI tools. 'We've all gotten way too comfortable with AI. People are treating ChatGPT like a diary. Like a best friend. Like a therapist. Like a co-founder.''But it's none of those. It's not on your side, it's not protecting you. And legally, it doesn't owe you anything.'She closed her post with a simple piece of advice: 'Let me make this simple – if you wouldn't say it in front of a judge, don't type it into ChatGPT.'And her final thought was one that many might relate to: 'I'm honestly scared. Not because I have used ChatGPT for something I shouldn't have. But because we've moved too fast, and asked too few questions, and continue to do so in the world of AI.'These concerns aren't just theory. In a 2024 bankruptcy case in the United States, a lawyer submitted a legal brief that cited fake court cases generated by ChatGPT. The judge imposed a fine of $5,500 and ordered the lawyer to attend an AI ethics session. — slow_developer (@slow_developer) Similar disciplinary actions were taken against lawyers in Utah and Alabama who relied on fabricated AI-generated incidents have underscored a critical truth: AI cannot replace verified legal research or professional advice. It can mislead, misrepresent, or completely fabricate information — what researchers call "AI hallucinations".Altman also flagged a worrying trend among younger users. Speaking at a Federal Reserve conference, he said, 'There are young people who say, 'I can't make any decision in my life without telling ChatGPT everything that's going on. It knows me. I'm going to do whatever it says.' That feels really bad to me.'He's concerned that blind faith in AI could be eroding people's ability to think critically. While ChatGPT is programmed to provide helpful answers, Altman stressed it lacks context, responsibility, and real emotional advice is straightforward, and it applies to everyone: Don't use ChatGPT to confess anything sensitive, illegal or personal Never treat it as a lawyer, therapist, or financial advisor Verify any factual claims independently Use AI to brainstorm, not to confess And most importantly, don't say anything to a chatbot that you wouldn't be comfortable seeing in court While OpenAI claims that user chats are reviewed for safety and model training, Altman admitted that conversations may be retained if required by law. Even if you delete a conversation, legal demands can override those actions. With ongoing lawsuits, including one from The New York Times, OpenAI may soon have to store conversations indefinitely. For those looking for more privacy, Altman suggested considering open-source models that can run offline, like GPT4All by Nomic AI or Ollama. But he stressed that what's needed most is a clear legal framework.'I think we will certainly need a legal or a policy framework for AI,' he then, treat your chats with caution. Because what you type could follow you — even years later.