logo
Inner workings of AI an enigma - even to its creators

Inner workings of AI an enigma - even to its creators

France 2413-05-2025

"People outside the field are often surprised and alarmed to learn that we do not understand how our own AI creations work," Anthropic co-founder Dario Amodei wrote in an essay posted online in April.
"This lack of understanding is essentially unprecedented in the history of technology."
Unlike traditional software programs that follow pre-ordained paths of logic dictated by programmers, generative AI (gen AI) models are trained to find their own way to success once prompted.
In a recent podcast Chris Olah, who was part of ChatGPT-maker OpenAI before joining Anthropic, described gen AI as "scaffolding" on which circuits grow.
Olah is considered an authority in so-called mechanistic interpretability, a method of reverse engineering AI models to figure out how they work.
This science, born about a decade ago, seeks to determine exactly how AI gets from a query to an answer.
"Grasping the entirety of a large language model is an incredibly ambitious task," said Neel Nanda, a senior research scientist at the Google DeepMind AI lab.
It was "somewhat analogous to trying to fully understand the human brain," Nanda added to AFP, noting neuroscientists have yet to succeed on that front.
Delving into digital minds to understand their inner workings has gone from a little-known field just a few years ago to being a hot area of academic study.
"Students are very much attracted to it because they perceive the impact that it can have," said Boston University computer science professor Mark Crovella.
The area of study is also gaining traction due to its potential to make gen AI even more powerful, and because peering into digital brains can be intellectually exciting, the professor added.
Keeping AI honest
Mechanistic interpretability involves studying not just results served up by gen AI but scrutinizing calculations performed while the technology mulls queries, according to Crovella.
"You could look into the model...observe the computations that are being performed and try to understand those," the professor explained.
Startup Goodfire uses AI software capable of representing data in the form of reasoning steps to better understand gen AI processing and correct errors.
The tool is also intended to prevent gen AI models from being used maliciously or from deciding on their own to deceive humans about what they are up to.
"It does feel like a race against time to get there before we implement extremely intelligent AI models into the world with no understanding of how they work," said Goodfire chief executive Eric Ho.
In his essay, Amodei said recent progress has made him optimistic that the key to fully deciphering AI will be found within two years.
"I agree that by 2027, we could have interpretability that reliably detects model biases and harmful intentions," said Auburn University associate professor Anh Nguyen.
According to Boston University's Crovella, researchers can already access representations of every digital neuron in AI brains.
"Unlike the human brain, we actually have the equivalent of every neuron instrumented inside these models", the academic said. "Everything that happens inside the model is fully known to us. It's a question of discovering the right way to interrogate that."
Harnessing the inner workings of gen AI minds could clear the way for its adoption in areas where tiny errors can have dramatic consequences, like national security, Amodei said.
For Nanda, better understanding what gen AI is doing could also catapult human discoveries, much like DeepMind's chess-playing AI, AlphaZero, revealed entirely new chess moves that none of the grand masters had ever thought about.
Properly understood, a gen AI model with a stamp of reliability would grab competitive advantage in the market.
Such a breakthrough by a US company would also be a win for the nation in its technology rivalry with China.
"Powerful AI will shape humanity's destiny," Amodei wrote.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

'Meta Is Redefining Warfare': U.S. Army Adopts AR-AI Headset That Turns Soldiers Into Real-Time Combat Intelligence Hubs
'Meta Is Redefining Warfare': U.S. Army Adopts AR-AI Headset That Turns Soldiers Into Real-Time Combat Intelligence Hubs

Sustainability Times

time10 hours ago

  • Sustainability Times

'Meta Is Redefining Warfare': U.S. Army Adopts AR-AI Headset That Turns Soldiers Into Real-Time Combat Intelligence Hubs

IN A NUTSHELL 🚀 Meta and Anduril Industries partner to bring advanced augmented reality gear to U.S. soldiers, enhancing battlefield capabilities. and partner to bring advanced augmented reality gear to U.S. soldiers, enhancing battlefield capabilities. 🛡️ The integration of AI and AR technology with Anduril's Lattice platform aims to transform soldiers into 'technomancers' with real-time combat intelligence. and technology with Anduril's Lattice platform aims to transform soldiers into 'technomancers' with real-time combat intelligence. 🤝 Zuckerberg and Luckey reunite in a politically charged environment, emphasizing the importance of dual-use technology for national defense. and reunite in a politically charged environment, emphasizing the importance of dual-use technology for national defense. 🌐 Meta expands its national security footprint by aligning its innovations with U.S. defense priorities and countering global tech competition. As the dynamics of modern warfare evolve, companies like Meta and Anduril Industries are at the forefront of integrating cutting-edge technology into military operations. With the announcement of their new partnership, these tech giants aim to revolutionize the way American soldiers interact with the battlefield using augmented reality (AR) and artificial intelligence (AI). This collaboration represents a significant shift in how commercial technology intersects with national defense, setting the stage for a future where soldiers could operate as 'technomancers' equipped with real-time combat intelligence and autonomous control systems. Tech Meets Warfare in a New Era of Battlefield Computing The collaboration between Meta and Anduril Industries introduces a new era in battlefield computing. The focus is on developing wearable technology, such as AR goggles or visors, that will significantly enhance soldiers' situational awareness and decision-making capabilities. Meta's hardware and AI tools will integrate with Anduril's Lattice platform, a sophisticated system that aggregates data from numerous sources to deliver actionable intelligence directly to soldiers on the ground. Through intuitive AR/VR interfaces, soldiers will interact with this comprehensive data in real time, allowing for unprecedented control and perception in combat situations. Palmer Luckey, Anduril's founder, expressed that their mission is to transform soldiers into technomancers, highlighting the transformative potential of these advancements. This initiative, privately funded and not reliant on taxpayer dollars, promises to leverage commercial technology for defense, potentially saving billions in military expenditure. 'Nasa Confirms the Unthinkable': China's Giant Water Diversion Project Will Slow Earth's Rotation and Disrupt Global Timekeeping Zuckerberg, Luckey Reunite in High-Stakes Political Climate This partnership also signifies a reunion between Mark Zuckerberg and Palmer Luckey, who share a history dating back to Luckey's sale of Oculus to Meta (formerly Facebook) in 2014. Their renewed collaboration comes amidst a politically charged environment, with President Trump serving a second term and tech companies navigating closer ties with the government. Zuckerberg's frequent visits to the administration underscore Meta's strategic positioning in the national defense landscape. Luckey's departure from Facebook in 2017, following controversy over his political activities, adds a layer of intrigue to this partnership. However, both leaders emphasize the critical importance of utilizing dual-use technology to bolster national defense capabilities. As they embark on this ambitious venture, the fusion of their expertise promises to deliver groundbreaking advancements in military technology. 'Zuckerberg's Billionaire Odyssey': 5,300-Mile Superyacht Voyage Ends With Helicopter Ski Drop in the Ultimate Tech Tycoon Flex Meta Expands National Security Footprint Meta's collaboration with Anduril is part of a broader strategy to cement its position as a key player in U.S. national security. By aligning its AI and AR technologies with defense initiatives, Meta aims to contribute to U.S. global competitiveness and counter China's growing tech influence. The company's open-source Llama AI models are being pitched to defense agencies, reflecting their commitment to supporting national interests. Alongside Anduril, Meta has submitted a bid for the Army's SBMC Next contract, formerly known as IVAS Next, and is developing the EagleEye headset to enhance soldiers' sensory capabilities. Even if the contract bid is unsuccessful, the companies plan to continue their work, underscoring a long-term commitment to military technology innovation. This partnership exemplifies the increasing engagement of Silicon Valley with U.S. defense priorities, with companies like Meta, OpenAI, and Palantir at the forefront of tech-driven national security solutions. 'Don't Head for the Hills': Experts Reveal the Only Safe Places to Survive a Nuclear Explosion in the U.S. Implications and Future Perspectives The partnership between Meta and Anduril represents a pivotal moment in the intersection of technology and defense. By harnessing the power of AI and AR, these companies are poised to redefine the capabilities of modern soldiers, offering enhanced situational awareness and decision-making tools. Such advancements have the potential to transform the nature of military operations, presenting both opportunities and challenges. As commercial technology becomes increasingly integral to national defense, questions arise about the ethical and strategic implications of these developments. How will these advancements impact the future of warfare, and what measures will be taken to ensure their responsible deployment? As we stand at the cusp of this technological revolution, the dialogue surrounding the role of tech in defense will likely continue to evolve, prompting further exploration and innovation in the years to come. Our author used artificial intelligence to enhance this article. Did you like it? 4.7/5 (22)

AI-generated Pope sermons flood YouTube, TikTok
AI-generated Pope sermons flood YouTube, TikTok

AFP

time14 hours ago

  • AFP

AI-generated Pope sermons flood YouTube, TikTok

An AFP investigation identified dozens of YouTube and TikTok pages that have been churning out AI-generated messages delivered in the pope's voice or otherwise attributed to him since he took charge of the Catholic Church last month. The hundreds of fabricated sermons and speeches, in English and Spanish, underscore how easily hoaxes created using artificial intelligence can elude detection and dupe viewers. "There's natural interest in what the new pope has to say, and people don't yet know his stance and style," said University of Washington professor emeritus Oren Etzioni, founder of a nonprofit focused on fighting deepfakes (archived here and here). "A perfect opportunity to sow mischief with AI-generated misinformation." After AFP presented YouTube with 26 channels posting predominantly AI-generated pope content, the platform terminated 16 of them for violating its policies against spam, deceptive practices and scams, and another for violating YouTube's terms of service (archived here and here). "We terminated several channels flagged to us by AFP for violating our Spam policies and Terms of Service," spokesperson Jack Malon said. The company also booted an additional six pages from its partner program allowing creators to monetize their content (archived here). TikTok similarly removed 11 accounts that AFP pointed out -- with more than 1.3 million combined followers -- citing the platform's policies against impersonation, harmful misinformation and misleading AI-generated content of public figures (archived here, here and here). - 'Chaotic uses' - With names such as "Pope Leo XIV Vision," the social media pages portrayed the pontiff supposedly offering a flurry of warnings and lessons he never preached. Image Screenshot from YouTube taken May 28, 2025 But disclaimers annotating their use of AI were often hard to find -- and sometimes nonexistent. On YouTube, a label demarcating "altered or synthetic content" is required for material that makes someone appear to say something they did not (archived here). But such disclosures only show up toward the bottom of each video's click-to-open description. A YouTube spokesperson said the company has since applied a more prominent label to some videos on the channels flagged by AFP that were not found to have violated the platform's guidelines. Image Screenshot from YouTube taken June 6, 2025, with a newly added synthetic content label outlined by AFP TikTok also requires creators to label posts sharing realistic AI-generated content, though several pope-centric videos went unmarked (archived here). A TikTok spokesperson said the company proactively removes policy-violating content and uses verified badges to signal authentic accounts. Brian Patrick Green, director of technology ethics at Santa Clara University, said the moderation difficulties are the result of rapid AI developments inspiring "chaotic uses of the technology" (archived here). Many clips on the YouTube channels AFP identified amassed tens of thousands of views before being deactivated. On TikTok, one Spanish-language video received 9.6 million views while claiming to show Leo preaching about the value of supportive women. Another, which carried an AI label but still fooled viewers, was watched some 32.9 million times. No video on the pope's official Instagram page has more than 6 million views. Experts say even seemingly harmless fakes can be problematic especially if used to farm engagement for accounts that might later sell their audiences or pivot to other misinformation. The AI-generated sermons not only "corrode the pope's moral authority" and "make whatever he actually says less believable," Green said, but could be harnessed "to build up trust around your channel before having the pope say something outrageous or politically expedient." The pope himself has also warned about the risks of AI, while Vatican News called out a deepfake that purported to show Leo praising Burkina Faso leader Ibrahim Traore, who seized power in a 2022 coup (archived here). AFP also debunked clips depicting the pope, who holds American and Peruvian citizenships, criticizing US Vice President JD Vance and Peru's President Dina Boluarte. "There's a real crisis here," Green said. "We're going to have to figure out some way to know whether things are real or fake."

AI-generated Pope sermons flood YouTube, TikTok
AI-generated Pope sermons flood YouTube, TikTok

France 24

timea day ago

  • France 24

AI-generated Pope sermons flood YouTube, TikTok

An AFP investigation identified dozens of YouTube and TikTok pages that have been churning out AI-generated messages delivered in the pope's voice or otherwise attributed to him since he took charge of the Catholic Church last month. The hundreds of fabricated sermons and speeches, in English and Spanish, underscore how easily hoaxes created using artificial intelligence can elude detection and dupe viewers. "There's natural interest in what the new pope has to say, and people don't yet know his stance and style," said University of Washington professor emeritus Oren Etzioni, founder of a nonprofit focused on fighting deepfakes. "A perfect opportunity to sow mischief with AI-generated misinformation." After AFP presented YouTube with 26 channels posting predominantly AI-generated pope content, the platform terminated 16 of them for violating its policies against spam, deceptive practices and scams, and another for violating YouTube's terms of service. "We terminated several channels flagged to us by AFP for violating our Spam policies and Terms of Service," spokesperson Jack Malon said. The company also booted an additional six pages from its partner program allowing creators to monetize their content. TikTok similarly removed 11 accounts that AFP pointed out -- with over 1.3 million combined followers -- citing the platform's policies against impersonation, harmful misinformation and misleading AI-generated content of public figures. - 'Chaotic uses' - With names such as "Pope Leo XIV Vision," the social media pages portrayed the pontiff supposedly offering a flurry of warnings and lessons he never preached. But disclaimers annotating their use of AI were often hard to find -- and sometimes nonexistent. On YouTube, a label demarcating "altered or synthetic content" is required for material that makes someone appear to say something they did not. But such disclosures only show up toward the bottom of each video's click-to-open description. A YouTube spokesperson said the company has since applied a more prominent label to some videos on the channels flagged by AFP that were not found to have violated the platform's guidelines. TikTok also requires creators to label posts sharing realistic AI-generated content, though several pope-centric videos went unmarked. A TikTok spokesperson said the company proactively removes policy-violating content and uses verified badges to signal authentic accounts. Brian Patrick Green, director of technology ethics at Santa Clara University, said the moderation difficulties are the result of rapid AI developments inspiring "chaotic uses of the technology." Many clips on the YouTube channels AFP identified amassed tens of thousands of views before being deactivated. On TikTok, one Spanish-language video received 9.6 million views while claiming to show Leo preaching about the value of supportive women. Another, which carried an AI label but still fooled viewers, was watched some 32.9 million times. No video on the pope's official Instagram page has more than 6 million views. Experts say even seemingly harmless fakes can be problematic especially if used to farm engagement for accounts that might later sell their audiences or pivot to other misinformation. The AI-generated sermons not only "corrode the pope's moral authority" and "make whatever he actually says less believable," Green said, but could be harnessed "to build up trust around your channel before having the pope say something outrageous or politically expedient." The pope himself has also warned about the risks of AI, while Vatican News called out a deepfake that purported to show Leo praising Burkina Faso leader Ibrahim Traore, who seized power in a 2022 coup. AFP also debunked clips depicting the pope, who holds American and Peruvian citizenships, criticizing US Vice President JD Vance and Peru's President Dina Boluarte. "There's a real crisis here," Green said. "We're going to have to figure out some way to know whether things are real or fake."

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store