logo
​​​​Meta's chief AI scientist says scaling AI won't make it smarter

​​​​Meta's chief AI scientist says scaling AI won't make it smarter

For years the AI industry has abided by a set of principles known as "scaling laws." OpenAI researchers outlined them in the seminal 2020 paper, "Scaling Laws for Neural Language Models."
"Model performance depends most strongly on scale, which consists of three factors: the number of model parameters N (excluding embeddings), the size of the dataset D, and the amount of compute C used for training," the authors wrote.
In essence, more is more when it comes to building highly intelligent AI. This idea has fueled huge investments in data centers that allow AI models to process and learn from huge amounts of existing information.
But recently, AI experts across Silicon Valley have started to challenge that doctrine.
"Most interesting problems scale extremely badly," Meta's chief AI scientist, Yann LeCun, said at the National University of Singapore on Sunday. "You cannot just assume that more data and more compute means smarter AI."
LeCun's point hinges on the idea that training AI on vast amounts of basic subject matter, like internet data, won't lead to some sort of superintelligence. Smart AI is a different breed.
"The mistake is that very simple systems, when they work for simple problems, people extrapolate them to think that they'll work for complex problems," he said. "They do some amazing things, but that creates a religion of scaling that you just need to scale systems more and they're going to naturally become more intelligent."
Right now, the impact of scaling is magnified because many of the latest breakthroughs in AI are actually "really easy," LeCun said. The biggest large language models today are trained on roughly the amount of information in the visual cortex of a four-year-old, he said.
"When you deal with real-world problems with ambiguity and uncertainty, it's not just about scaling anymore," he added.
AI advancements have been slowing lately. This is due, in part, to a dwindling corpus of usable public data.
LeCun is not the only prominent researcher to question the power of scaling. Scale AI CEO Alexandr Wang said scaling is "the biggest question in the industry" at the Cerebral Valley conference last year. Cohere CEO Aidan Gomez called it the "dumbest" way to improve AI models.
LeCun advocates for a more world-based training approach.
"We need AI systems that can learn new tasks really quickly. They need to understand the physical world — not just text and language but the real world — have some level of common sense, and abilities to reason and plan, have persistent memory — all the stuff that we expect from intelligent entities," he said during his talk Sunday.
Last year, on an episode of Lex Fridman's podcast, LeCun said that in contrast to large language models, which can only predict their next steps based on patterns, world models have a higher level of cognition. "The extra component of a world model is something that can predict how the world is going to evolve as a consequence of an action you might take."

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Here's how Uber's product chief uses AI at work — and one tool he's going to use next
Here's how Uber's product chief uses AI at work — and one tool he's going to use next

Business Insider

timean hour ago

  • Business Insider

Here's how Uber's product chief uses AI at work — and one tool he's going to use next

Uber's chief product officer has one AI tool on his to-do list. In an episode of "Lenny's Podcast" released on Sunday, Uber's product chief, Sachin Kansal, shared two ways he is using AI for his everyday tasks at the ride-hailing giant and how he plans to add NotebookLM to his AI suite. Kansal joined Uber eight years ago as its director of product management after working at cybersecurity and taxi startups. He became Uber's product chief last year. Kansal said he uses OpenAI's ChatGPT and Google's Gemini to summarize long reports. "Some of these reports, they're 50 to 100 pages long," he said. "I will never have the time to read them." He said he uses the chatbots to acquaint himself with what's happening and how riders are feeling in Uber's various markets, such as South Africa, Brazil, and Korea. The CPO said his second use case is treating AI like a research assistant, because some large language models now offer a deep research feature. Kansal gave a recent example of when his team was thinking about a new driver feature. He asked ChatGPT's deep research mode about what drivers may think of the add-on. "It's an amazing research assistant and it's absolutely a starting point for a brainstorm with my team with some really, really good ideas," the CPO said. In April, Uber's CEO, Dara Khosrowshahi, said that not enough of his 30,000-odd employees are using AI. He said learning to work with AI agents to code is "going to be an absolute necessity at Uber within a year." Uber did not immediately respond to a request for comment from Business Insider. Kansal's next tool: NotebookLM On the podcast, Kansal also highlighted NotebookLM, Google Lab's research and note-taking tool, which is especially helpful for interacting with documents. He said he doesn't use the product yet, but wants to. "I know a lot of people who have started using it, and that is the next thing that I'm going to use," he said. "Just to be able to build an audio podcast based on a bunch of information that you can consume. I think that's awesome," he added. Kansal was referring to the "Audio Overview" feature, which summarizes uploaded content in the form of two AIs having a voice discussion. NotebookLM was launched in mid-2023 and has quickly become a must-have tool for researchers and AI enthusiasts. Andrej Karpathy, Tesla's former director of AI and OpenAI cofounder, is among those who have praised the tool and its podcast feature. "It's possible that NotebookLM podcast episode generation is touching on a whole new territory of highly compelling LLM product formats," he said in a September post on X. "Feels reminiscent of ChatGPT. Maybe I'm overreacting."

Scoop: Apple to appeal EU requirement to share info with tech rivals
Scoop: Apple to appeal EU requirement to share info with tech rivals

Axios

time3 hours ago

  • Axios

Scoop: Apple to appeal EU requirement to share info with tech rivals

Apple filed an appeal to the European Commission's specifications around how the company complies with its interoperability requirements, which compel it to share user information with outside developers as part of the Digital Markets Act, per a source familiar with the situation. Why it matters: Apple continues to push back forcefully against requirements imposed by the DMA, which the company has said forces it to give up its intellectual property and compromise user privacy. The deadline to appeal this particular requirement was May 30. How it works: The interoperability requirements are meant to give other device manufacturers and app developers access to Apple features that are normally exclusive to Apple products, including WiFi pairing and notifications on non-Apple smartwatches and headsets. Apple and Meta were the first companies to be hit with DMA fines in April, as Axios first reported. Apple put out a report last December arguing that the DMA's interoperability reports could be abused and expose private user information. What they're saying: "We design our technology to work seamlessly together, so it can deliver the unique experience our users love and expect from our products. The EU's interoperability requirements threaten that foundation, while creating a process that is unreasonable, costly, and stifles innovation," an Apple spokesperson told Axios. "These requirements will also hand data-hungry companies sensitive information, which poses massive privacy and security risks to our EU users." Per the Apple spokesperson, as allowed under the DMA's interoperability requirement, companies including Meta, Google, Garmin and Spotify have requested from Apple user notification content and stored WiFi networks, "giving them the ability to access personal information that even Apple doesn't see." The spokesperson added, "In the end, these deeply flawed rules that only target Apple—and no other company—will severely limit our ability to deliver innovative products and features to Europe, leading to an inferior user experience for our European customers." The other side: Companies that rely on the App Store to distribute their products have welcomed the DMA's requirements of Apple, saying they are long overdue for developers and other businesses to offer customers better features on their apps and devices.

Ads Ruined Social Media. Now They're Coming to AI.
Ads Ruined Social Media. Now They're Coming to AI.

Bloomberg

time3 hours ago

  • Bloomberg

Ads Ruined Social Media. Now They're Coming to AI.

Chatbots might hallucinate and sprinkle too much flattery on their users — 'That's a fascinating question!' one recently told me — but at least the subscription model that underpins them is healthy for our wellbeing. Many Americans pay about $20 a month to use the premium versions of OpenAI's ChatGPT, Google's Gemini Pro or Anthropic's Claude, and the result is that the products are designed to provide maximum utility. Don't expect this status quo to last. Subscription revenue has a limit, and Anthropic's new $200-a-month 'Max' tier suggests even the most popular models are under pressure to find new revenue streams.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store