logo
Why superintelligent AI isn't taking over anytime soon

Why superintelligent AI isn't taking over anytime soon

Mint14 hours ago

A primary requirement for being a leader in AI these days is to be a herald of the impending arrival of our digital messiah: superintelligent AI.
For Dario Amodei of Anthropic, Demis Hassabis of Google and Sam Altman of OpenAI, it isn't enough to claim that their AI is the best. All three have recently insisted that it's going to be so good, it will change the very fabric of society.
Even Meta—whose chief AI scientist has been famously dismissive of this talk—wants in on the action. The company confirmed it is spending $14 billion to bring in a new leader for its AI efforts who can realize Mark Zuckerberg's dream of AI superintelligence—that is, an AI smarter than we are.
'Humanity is close to building digital superintelligence," Altman declared in an essay this week, and this will lead to 'whole classes of jobs going away" as well as 'a new social contract." Both will be consequences of AI-powered chatbots taking over all our white-collar jobs, while AI-powered robots assume the physical ones.
Before you get nervous about all the times you were rude to Alexa, know this: A growing cohort of researchers who build, study and use modern AI aren't buying all that talk.
The title of a fresh paper from Apple says it all: 'The Illusion of Thinking." In it, a half-dozen top researchers probed reasoning models—large language models that 'think" about problems longer, across many steps—from the leading AI labs, including OpenAI, DeepSeek and Anthropic. They found little evidence that these are capable of reasoning anywhere close to the level their makers claim.
Generative AI can be quite useful in specific applications, and a boon to worker productivity. OpenAI claims 500 million monthly active ChatGPT users—astonishingly far reach and fast growth for a service released just 2½ years ago. But these critics argue there is a significant hazard in overestimating what it can do, and making business plans, policy decisions and investments based on pronouncements that seem increasingly disconnected from the products themselves.
Apple's paper builds on previous work from many of the same engineers, as well as notable research from both academia and other big tech companies, including Salesforce. These experiments show that today's 'reasoning" AIs—hailed as the next step toward autonomous AI agents and, ultimately, superhuman intelligence—are in some cases worse at solving problems than the plain-vanilla AI chatbots that preceded them. This work also shows that whether you're using an AI chatbot or a reasoning model, all systems fail utterly at more complex tasks.
Apple's researchers found 'fundamental limitations" in the models. When taking on tasks beyond a certain level of complexity, these AIs suffered 'complete accuracy collapse." Similarly, engineers at Salesforce AI Research concluded that their results 'underscore a significant gap between current LLM capabilities and real-world enterprise demands."
Importantly, the problems these state-of-the-art AIs couldn't handle are logic puzzles that even a precocious child could solve, with a little instruction. What's more, when you give these AIs that same kind of instruction, they can't follow it.
Apple's paper has set off a debate in tech's halls of power—Signal chats, Substack posts and X threads—pitting AI maximalists against skeptics.
'People could say it's sour grapes, that Apple is just complaining because they don't have a cutting-edge model," says Josh Wolfe, co-founder of venture firm Lux Capital. 'But I don't think it's a criticism so much as an empirical observation."
The reasoning methods in OpenAI's models are 'already laying the foundation for agents that can use tools, make decisions, and solve harder problems," says an OpenAI spokesman. 'We're continuing to push those capabilities forward."
The debate over this research begins with the implication that today's AIs aren't thinking, but instead are creating a kind of spaghetti of simple rules to follow in every situation covered by their training data.
Gary Marcus, a cognitive scientist who sold an AI startup to Uber in 2016, argued in an essay that Apple's paper, along with related work, exposes flaws in today's reasoning models, suggesting they're not the dawn of human-level ability but rather a dead end. 'Part of the reason the Apple study landed so strongly is that Apple did it," he says. 'And I think they did it at a moment in time when people have finally started to understand this for themselves."
In areas other than coding and mathematics, the latest models aren't getting better at the rate that they once did. And the newest reasoning models actually hallucinate more than their predecessors.
'The broad idea that reasoning and intelligence come with greater scale of models is probably false," says Jorge Ortiz, an associate professor of engineering at Rutgers, whose lab uses reasoning models and other cutting-edge AI to sense real-world environments. Today's models have inherent limitations that make them bad at following explicit instructions—the opposite of what you'd expect from a computer, he adds.
It's as if the industry is creating engines of free association. They're skilled at confabulation, but we're asking them to take on the roles of consistent, rule-following engineers or accountants.
That said, even those who are critical of today's AIs hasten to add that the march toward more-capable AI continues.
Exposing current limitations could point the way to overcoming them, says Ortiz. For example, new training methods—giving step-by-step feedback on models' performance, adding more resources when they encounter harder problems—could help AI work through bigger problems, and make better use of conventional software.
From a business perspective, whether or not current systems can reason, they're going to generate value for users, says Wolfe.
'Models keep getting better, and new approaches to AI are being developed all the time, so I wouldn't be surprised if these limitations are overcome in practice in the near future," says Ethan Mollick, a professor at the Wharton School of the University of Pennsylvania, who has studied the practical uses of AI.
Meanwhile, the true believers are undeterred.
Just a decade from now, Altman wrote in his essay, 'maybe we will go from solving high-energy physics one year to beginning space colonization the next year." Those willing to 'plug in" to AI with direct, brain-computer interfaces will see their lives profoundly altered, he adds.
This kind of rhetoric accelerates AI adoption in every corner of our society. AI is now being used by DOGE to restructure our government, leveraged by militaries to become more lethal, and entrusted with the education of our children, often with unknown consequences.
Which means that one of the biggest dangers of AI is that we overestimate its abilities, trust it more than we should—even as it's shown itself to have antisocial tendencies such as 'opportunistic blackmail"—and rely on it more than is wise. In so doing, we make ourselves vulnerable to its propensity to fail when it matters most.
'Although you can use AI to generate a lot of ideas, they still require quite a bit of auditing," says Ortiz. 'So for example, if you want to do your taxes, you'd want to stick with something more like TurboTax than ChatGPT."
Write to Christopher Mims at christopher.mims@wsj.com

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Nvidia CEO Jensen Huang says Anthropic CEO is very wrong, and on almost everything he said about AI: Don't do it in a dark room and tell me ...
Nvidia CEO Jensen Huang says Anthropic CEO is very wrong, and on almost everything he said about AI: Don't do it in a dark room and tell me ...

Time of India

time16 minutes ago

  • Time of India

Nvidia CEO Jensen Huang says Anthropic CEO is very wrong, and on almost everything he said about AI: Don't do it in a dark room and tell me ...

Nvidia CEO Jensen Huang strongly refuted Anthropic CEO Dario Amodei's predictions regarding AI-driven job automation during a press briefing. Huang dismissed Amodei's claim that AI could eliminate 50% of entry-level white-collar jobs within five years, advocating for transparent AI development. While acknowledging job transformation, Huang emphasized AI's potential to create new opportunities and boost productivity. Nvidia CEO Jensen Huang expressed sharp disagreement with Anthropic CEO Dario Amodei 's predictions about AI-driven job automation during a press briefing at VivaTech in Paris. Huang specifically challenged Amodei's recent claim that AI could eliminate up to 50% of entry-level white-collar jobs within five years, dismissing the notion as overly alarmist. 'I pretty much disagree with almost everything [Amodei] says,' Huang stated, critiquing three key points he attributed to Amodei: that AI is so dangerous only a select few should develop it, that it's too costly for widespread development, and that its power will lead to massive job losses. 'If you want things to be done safely and responsibly, you do it in the open … Don't do it in a dark room and tell me it's safe,' Huang added, advocating for transparent and collaborative AI development. What Anthropic CEO said on AI Amodei, in a recent Axios interview, had warned that AI could disrupt half of entry-level office roles and urged policymakers to prepare for economic impacts. Huang countered that while AI will transform jobs, it will also create new opportunities, citing increased productivity as a driver for hiring. 'Some jobs will be obsolete, but many jobs are going to be created,' he said. by Taboola by Taboola Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like 선납금·위약금 없이 벤츠를 원하는 기간만큼 이용하세요. 계약 만료시 반납, 인수 모두 가능합니다. 써치오토모빌 더 알아보기 Undo Anthropic on Nvidia CEO Jensen Huang's statement In response, an Anthropic spokesperson clarified to Fortune that Amodei has never claimed only Anthropic should build AI. 'Dario has advocated for a national transparency standard for AI developers … so the public and policymakers are aware of the models' capabilities and risks,' the statement read. The spokesperson reaffirmed Amodei's concerns about AI's economic impact and his commitment to addressing them. Anthropic, founded in 2021 by Amodei and other ex-OpenAI researchers, prioritizes safety in AI development, a focus rooted in the founders' reported concerns over OpenAI's direction. The public clash between Huang and Amodei highlights ongoing debates in the AI industry about its societal and economic implications. AI Masterclass for Students. Upskill Young Ones Today!– Join Now

Google, Scale AI's largest customer, plans split after Meta deal
Google, Scale AI's largest customer, plans split after Meta deal

Time of India

timean hour ago

  • Time of India

Google, Scale AI's largest customer, plans split after Meta deal

HighlightsAlphabet's Google plans to sever ties with Scale AI after rival Meta Platforms Inc. acquires a 49% stake in the AI data-labeling startup, potentially impacting Scale's revenue significantly. The acquisition of Scale AI by Meta Platforms Inc. raises concerns among competing AI companies, as they fear that sharing proprietary data with Scale could expose their research strategies and technical blueprints. Following the news of Meta's investment, competitors of Scale AI, such as Turing and Labelbox, anticipate a surge in business as companies seek alternative data-labeling services. Alphabet's Google , the largest customer of Scale AI , plans to cut ties with Scale after news broke that rival Meta is taking a 49% stake in the AI data-labeling startup, five sources familiar with the matter told Reuters. Google had planned to pay Scale AI about $200 million this year for the human-labeled training data that is crucial for developing technology, including the sophisticated AI models that power Gemini, its ChatGPT competitor, one of the sources said. The search giant already held conversations with several of Scale AI's rivals this week as it seeks to shift away much of that workload, sources added. Scale's loss of significant business comes as Meta takes a big stake in the company, valuing it at $29 billion. Scale was worth $14 billion before the deal. Scale AI intends to keep its business running while its CEO, Alexandr Wang, along with a few employees, move over to Meta. Since its core business is concentrated around a few customers, it could suffer greatly if it loses key customers like Google. In a statement, a Scale AI spokesperson said its business, which spans work with major companies and governments, remains strong, as it is committed to protecting customer data. The company declined to comment on specifics with Google. Scale AI raked in $870 million in revenue in 2024, and Google spent some $150 million on Scale AI's services last year, sources said. Other major tech companies that are customers of Scale's, including Microsoft, are also backing away. Elon Musk's xAI is also looking to exit, one of the sources said. OpenAI decided to pull back from Scale several months ago, according to sources familiar with the matter, though it spends far less money than Google. OpenAI's CFO that the company will continue to work with Scale AI, as one of its many data vendors. Companies that compete with Meta in developing cutting-edge AI models are concerned that doing business with Scale could expose their research priorities and road map to a rival, five sources said. By contracting with Scale AI, customers often share proprietary data as well as prototype products for which Scale's workers are providing data-labeling services. With Meta now taking a 49% stake, AI companies are concerned that one of their chief rivals could gain knowledge about their business strategy and technical blueprints. Google, Microsoft and OpenAI declined to comment. xAI did not respond to a request for comment. RIVALS SEE OPENINGS The bulk of Scale AI's revenue comes from charging generative AI model makers for providing access to a network of human trainers with specialized knowledge - from historians to scientists, some with doctorate degrees. The humans annotate complex datasets that are used to "post-train" AI models, and as AI models have become smarter, the demand for the sophisticated human-provided examples has surged, and one annotation could cost as much as $100. Scale also does data-labeling for enterprises like self-driving car companies and the U.S. government, which are likely to stay, according to the sources. But its biggest money-maker is in partnering with generative AI model makers, the sources said. Google had already sought to diversify its data service providers for more than a year, three of the sources said. But Meta's moves this week have led Google to seek to move off Scale AI on all its key contracts, the sources added. Because of the way data-labeling contracts are structured, that process could happen quickly, two sources said. This will provide an opening for Scale AI's rivals to jump in. "The Meta-Scale deal marks a turning point," said Jonathan Siddharth, CEO of Turing, a Scale AI competitor. "Leading AI labs are realizing neutrality is no longer optional, it's essential." Labelbox, another competitor, will "probably generate hundreds of millions of new revenue" by the end of the year from customers fleeing Scale, its CEO, Manu Sharma, told Reuters. Handshake, a competitor focusing on building a network of PhDs and experts, saw a surge of workload from top AI labs that compete with Meta. "Our demand has tripled overnight after the news," said Garrett Lord, CEO at Handshake. Many AI labs now want to hire in-house data-labelers, which allows their data to remain secure, said Brendan Foody, CEO of Mercor, a startup that in addition to competing directly with Scale AI also builds technology around being able to recruit and vet candidates in an automated way, enabling AI labs to scale up their data labeling operations quickly. Founded in 2016, Scale AI provides vast amounts of labeled data or curated training data, which is crucial for developing sophisticated tools such as OpenAI's ChatGPT. The Meta deal will be a boon for Scale AI's investors including Accel and Index Ventures, as well as its current and former employees. As part of the deal, Scale AI's CEO, Wang, will take a top position leading Meta's AI efforts. Meta is fighting the perception that it may have fallen behind in the AI race after its initial set of Llama 4 large language models released in April fell short of performance expectations.

Apple Watch: New AI fitness coach ‘Workout Buddy' launches soon
Apple Watch: New AI fitness coach ‘Workout Buddy' launches soon

Hans India

timean hour ago

  • Hans India

Apple Watch: New AI fitness coach ‘Workout Buddy' launches soon

Apple is enhancing your workout experience with its latest feature—Workout Buddy, set to launch with watchOS 26 later this year. This AI-powered personal coach uses Apple Intelligence to deliver real-time feedback, voice-guided encouragement, and personalized stats during your workouts. Whether you're running outdoors, cycling, walking, or doing strength training, Workout Buddy keeps you motivated. It greets you when you start, celebrates milestones mid-session, and summarizes key stats—like heart rate, pace, and distance—at the end. Workout Buddy uses a text-to-speech AI model, inspired by Fitness+ trainers, to make the voice sound more natural and human-like. All processing is done on-device, ensuring your privacy and data security. To use this feature, you'll need an Apple Watch compatible with watchOS 26, Bluetooth headphones, and a supported iPhone (15 Pro and above, or iPhone 16 series). watchOS 26 also brings a Liquid Glass UI, smarter notifications, larger Workout app buttons, and music recommendations tailored to your exercise style.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store