
Humans beat AI gold-level score at top maths contest
Sydney (AFP)Humans beat generative AI models made by Google and OpenAI at a top international mathematics competition, despite the programmes reaching gold-level scores for the first time.Neither model scored full marks -- unlike five young people at the International Mathematical Olympiad (IMO), a prestigious annual competition where participants must be under 20 years old.Google said on Monday that an advanced version of its Gemini chatbot had solved five out of the six maths problems set at the IMO, held in Australia's Queensland this month."We can confirm that Google DeepMind has reached the much-desired milestone, earning 35 out of a possible 42 points -- a gold medal score," the US tech giant cited IMO president Gregor Dolinar as saying."Their solutions were astonishing in many respects. IMO graders found them to be clear, precise and most of them easy to follow."Around 10% of human contestants won gold-level medals, and five received perfect scores of 42 points.US ChatGPT maker OpenAI said that its experimental reasoning model had scored a gold-level 35 points on the test.The result "achieved a longstanding grand challenge in AI" at "the world's most prestigious math competition", OpenAI researcher Alexander Wei wrote on social media."We evaluated our models on the 2025 IMO problems under the same rules as human contestants," he said."For each problem, three former IMO medalists independently graded the model's submitted proof."Google achieved a silver-medal score at last year's IMO in the British city of Bath, solving four of the six problems.That took two to three days of computation -- far longer than this year, when its Gemini model solved the problems within the 4.5-hour time limit, it said.The IMO said tech companies had "privately tested closed-source AI models on this year's problems", the same ones faced by 641 competing students from 112 countries."It is very exciting to see progress in the mathematical capabilities of AI models," said IMO president Dolinar.
Contest organisers could not verify how much computing power had been used by the AI models or whether there had been human involvement, he cautioned.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Al Etihad
a few seconds ago
- Al Etihad
Tesla taps Samsung for AI6 Chips in $16.5 billion deal, Elon Musk says
28 July 2025 09:44 SEOUL (AFP)Tesla CEO Elon Musk said Monday that tech giant Samsung Electronics will provide his company with its next-generation AI6 chips, following the South Korean firm's announcement of a $16.5 billion said Monday it had secured an eight-year agreement, without naming the client, describing it only as a "major global company" in a regulatory the deal, the partnership -- effective from last Thursday -- runs through the end of 2033."Samsung's giant new Texas fab will be dedicated to making Tesla's next-generation AI6 chip. The strategic importance of this is hard to overstate," Musk said on X."Samsung agreed to allow Tesla to assist in maximising manufacturing efficiency," he added, calling it a "critical point" in striking the deal."I will walk the line personally to accelerate the pace of progress," he said, noting Samsung's Texas plant was "conveniently located not far from my house".The deal represents about 7.6 percent of Samsung's projected annual sales for 2024, the company declined to confirm the client was Tesla, even after Musk's message, citing agreement is expected to provide a major boost to Samsung, which has faced headwinds in its foundry business, lagging rivals in the race for cutting-edge artificial intelligence Electronics is the flagship unit of South Korea's Samsung Group, by far the largest of the family-run conglomerates that dominate Asia's fourth-largest company said this month that it expected operating profit to fall 56 percent on-year and 31 percent from the previous quarter, citing a slump in its core semiconductor division. Experts have attributed the decline to weaknesses in its foundry operations, which involve contract-based manufacturing of chips designed by other companies.


Al Etihad
a day ago
- Al Etihad
Urgent need for 'global approach' on AI regulation: UN tech chief
27 July 2025 08:55 GENEVA (AFP)The world urgently needs to find a global approach on regulating artificial intelligence, the United Nations' top tech chief said this week, warning that fragmentation could deepen risks and Bogdan-Martin, head of the UN's International Telecommunications Union (ITU) agency, told AFP she hoped that AI "can actually benefit humanity".But as concerns mount over the risks posed by the fast-moving technology -- including fears of mass job losses, the spread of deepfakes and disinformation, and society's fabric fraying -- she insisted that regulation was key."There's an urgency to try to get... the right framework in place," she said, stressing the need for "a global approach".Her comments came after US President Donald Trump this week unveiled an aggressive, low-regulation strategy aimed at ensuring the United States stays ahead of China on more than 90 proposals, Trump's plan calls for sweeping deregulation, with the administration promising to "remove red tape and onerous regulation" that could hinder private sector AI if she had concerns about an approach that urges less, not more, regulation of AI technologies, Bogdan-Martin refrained from commenting."I think there are different approaches," she said."We have the EU approach. We have the Chinese approach. Now we're seeing the US approach. I think what's needed is for those approaches to dialogue," she the same time, she highlighted that "85 percent of countries don't yet have AI policies or strategies".A consistent theme among those strategies that do exist is the focus on innovation, capacity building and infrastructure investments, Bogdan-Martin said."But where I think the debate still needs to happen at a global level is trying to figure out how much regulation, how little regulation, is needed," she who grew up in New Jersey and has spent most of her more than three-decade career at the ITU, insisted the Geneva-based telecoms agency that sets standards for new technologies was well-placed to help facilitate much-needed dialogue on the issue."The need for a global approach I think is critical," she said, cautioning that "fragmented approaches will not help serve and reach all".The ITU chief hailed "mind-blowing" advances within artificial intelligence, with the potential to improve everything from education to agriculture to health care -- but insisted the benefits must be a concerted effort, there is a risk that AI will end up standing for "advancing inequalities", she warned, cautioning against deepening an already dire digital divide worldwide."We have 2.6 billion people that have no access to the internet, which means they have no access to artificial intelligence", Bogdan-Martin pointed out."We have to tackle those divides if we're actually going to have something that is beneficial to all of humanity." Bogdan-Martin, the first woman to serve as ITU secretary-general in the organisation's nearly 160-year history, also stressed the need to get more women into the digital space.


Arabian Post
2 days ago
- Arabian Post
LLMs Fail to Deliver Real Intelligence Despite Huge Investment
The trajectory of large language models like GPT and its counterparts has raised numerous questions in recent months. As companies such as OpenAI continue to pour billions into scaling these models, the fundamental issue of their cognitive limitations remains glaring. The hype surrounding LLMs, though widely praised for their fluency and utility, overlooks a critical flaw in their design. These models may perform tasks that mimic intelligent behaviour but do not actually possess the ability to think, reason, or understand. A growing chorus of AI researchers and experts argues that no amount of funding, data, or compute power will transform LLMs into entities capable of genuine intelligence. Despite ambitious plans from companies like OpenAI to expand the infrastructure behind LLMs to an unimaginable scale, their current model architecture continues to hit the same cognitive wall. At the core of this issue is the realization that LLMs are fundamentally engineered to mimic intelligence rather than to achieve it. OpenAI's recent announcements have been staggering. The company has unveiled plans to deploy up to 100 million GPUs—an infrastructure investment that could exceed $3 trillion. These resources would be used to enhance the size and speed of existing LLMs. Such efforts would consume enormous amounts of energy, rivaling that of entire countries, and generate vast quantities of emissions. The scale of the operation is unprecedented, but so too is the question: What exactly will this achieve? Will adding more tokens to a slightly bigger and faster model finally lead to true intelligence? ADVERTISEMENT The simple answer appears to be no. LLMs are not designed to possess cognition. They are designed to predict, autocomplete, summarise, and assist with routine tasks—but these are functions of performance, not understanding. The biggest misconception in AI development today is the conflation of fluency with intelligence. Proponents of scaling continue to tout that more data, more models, and more compute will unlock something that is fundamentally elusive. But as the limitations of LLMs become increasingly apparent, the vision of artificial general intelligence using current methodologies seems like a pipe dream. The reality of AI's current state is jarring: a vast, burning of resources with little to show for it. Companies like Meta, xAI, and DeepMind are all investing heavily in LLMs, creating an illusion of progress by pushing for bigger and more powerful systems. However, these innovations are essentially 'performance theatre,' with much of the energy and resources funnelled into creating benchmarks and achieving superficial gains in fluency rather than advancing the underlying technology. This raises important questions: Why is there so little accountability for the environmental impact of such projects? Where is the true innovation in cognitive science? LLMs, despite their capacity to accomplish specific tasks effectively, are essentially still limited by their design. The push to scale them further, under the assumption that doing so will lead to breakthroughs in artificial intelligence, ignores the inherent problems that cannot be solved with brute force alone. The architecture behind LLMs—based on pattern recognition and statistical correlation—simply cannot generate the complex, dynamic processes involved in real cognition. Experts argue that the AI community must acknowledge these limitations and pivot toward new approaches. The vast majority of AI researchers now agree that a shift in paradigm is necessary. LLMs, no matter how large or finely tuned, cannot produce the kind of intelligence required to understand, reason, or adapt in a human-like way. To move forward, a radically different model must be developed—one that incorporates cognitive architecture and a deeper understanding of how real intelligence functions. The current momentum in AI, driven by large companies and investors, seems to be propelled by a desire for immediate results and visible performance metrics. But it's crucial to remember that speed means little if it's headed in the wrong direction. Without a rethinking of the very foundations of AI research, the race to scale LLMs will continue to miss the mark. In fact, there's a real risk that the over-emphasis on the scalability of these models could stifle the kind of breakthroughs needed to move the field forward.