
Inner workings of AI an enigma - even to its creators
Even the greatest human minds building generative artificial intelligence that is poised to change the world admit they do not comprehend how digital minds think.
"People outside the field are often surprised and alarmed to learn that we do not understand how our own AI creations work," Anthropic co-founder Dario Amodei wrote in an essay posted online in April.
"This lack of understanding is essentially unprecedented in the history of technology."
Unlike traditional software programs that follow pre-ordained paths of logic dictated by programmers, generative AI (gen AI) models are trained to find their own way to success once prompted.
In a recent podcast Chris Olah, who was part of ChatGPT-maker OpenAI before joining Anthropic, described gen AI as "scaffolding" on which circuits grow.
Olah is considered an authority in so-called mechanistic interpretability, a method of reverse engineering AI models to figure out how they work.
This science, born about a decade ago, seeks to determine exactly how AI gets from a query to an answer.
"Grasping the entirety of a large language model is an incredibly ambitious task," said Neel Nanda, a senior research scientist at the Google DeepMind AI lab.
It was "somewhat analogous to trying to fully understand the human brain," Nanda added to AFP, noting neuroscientists have yet to succeed on that front.
Delving into digital minds to understand their inner workings has gone from a little-known field just a few years ago to being a hot area of academic study.
"Students are very much attracted to it because they perceive the impact that it can have," said Boston University computer science professor Mark Crovella.
The area of study is also gaining traction due to its potential to make gen AI even more powerful, and because peering into digital brains can be intellectually exciting, the professor added.
Mechanistic interpretability involves studying not just results served up by gen AI but scrutinizing calculations performed while the technology mulls queries, according to Crovella.
"You could look into the model...observe the computations that are being performed and try to understand those," the professor explained.
Startup Goodfire uses AI software capable of representing data in the form of reasoning steps to better understand gen AI processing and correct errors.
The tool is also intended to prevent gen AI models from being used maliciously or from deciding on their own to deceive humans about what they are up to.
"It does feel like a race against time to get there before we implement extremely intelligent AI models into the world with no understanding of how they work," said Goodfire chief executive Eric Ho.
In his essay, Amodei said recent progress has made him optimistic that the key to fully deciphering AI will be found within two years.
"I agree that by 2027, we could have interpretability that reliably detects model biases and harmful intentions," said Auburn University associate professor Anh Nguyen.
According to Boston University's Crovella, researchers can already access representations of every digital neuron in AI brains.
"Unlike the human brain, we actually have the equivalent of every neuron instrumented inside these models", the academic said. "Everything that happens inside the model is fully known to us. It's a question of discovering the right way to interrogate that."
Harnessing the inner workings of gen AI minds could clear the way for its adoption in areas where tiny errors can have dramatic consequences, like national security, Amodei said.
For Nanda, better understanding what gen AI is doing could also catapult human discoveries, much like DeepMind's chess-playing AI, AlphaZero, revealed entirely new chess moves that none of the grand masters had ever thought about.
Properly understood, a gen AI model with a stamp of reliability would grab competitive advantage in the market.
Such a breakthrough by a US company would also be a win for the nation in its technology rivalry with China.
"Powerful AI will shape humanity's destiny," Amodei wrote.
"We deserve to understand our own creations before they radically transform our economy, our lives, and our future."

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Hindustan Times
3 hours ago
- Hindustan Times
Why ChatGPT essays still fail to fool experts despite good structure, although they are clear and well structured
The advent of AI has marked the rise of many tools, and ChatGPT is one of the most popular ones. Often used for research and writing, this tool has often been the centre of discussion for its ability to fetch interesting content. However, A new study from the University of East Anglia (UEA) in the UK shows that essays written by real students are still better than those produced by ChatGPT, a popular AI writing tool. Researchers compared 145 essays written by university students with 145 essays generated by ChatGPT to see how well the AI can mimic human writing. The study found that although ChatGPT's essays are clear, well structured, and grammatically correct, they lack something important. The AI essays do not show personal insight or deep critical thinking, which are common in student writing. These missing elements make the AI-generated essays feel less engaging and less convincing. However, the researchers do not see AI only as a threat. They believe tools like ChatGPT can be helpful in education if used properly. Instead of shortcuts to finish assignments, AI should be a tool that supports learning and improves writing skills. After all, education is about teaching students how to think clearly and express ideas. These are things no AI can truly replace. One key difference the researchers looked at was how the writers engage readers. Real student essays often include questions, personal comments, and direct appeals to the reader. These techniques help make the writing feel more interactive and persuasive. On the other hand, ChatGPT's essays tend to avoid questions and personal opinions. They follow academic rules but do not show a clear viewpoint or emotional connection. Professor Ken Hyland from UEA explained that the AI focuses on creating text that is logical and smooth but misses conversational details that humans use to connect with readers. This shows that AI writing still struggles with capturing the personal style and strong arguments that real people naturally use.


Hindustan Times
4 hours ago
- Hindustan Times
You can now schedule tasks with Gemini as Google's powerful new AI feature rivals ChatGPT's capabilities
Google is steadily evolving Gemini into a smarter, more proactive AI assistant that now competes directly with OpenAI's ChatGPT. The tech giant has started rolling out a feature called Scheduled Actions, which lets users automate recurring or timed tasks without repeating commands. Originally previewed during Google I/O, Scheduled Actions is now arriving on both Android and iOS devices. The feature is currently available to subscribers of Google One AI Premium and select Google Workspace business and education plans. With this rollout, Google is pushing Gemini closer to becoming a fully integrated productivity companion. Scheduled Actions let users instruct Gemini to perform specific tasks at set times or intervals. This includes sending daily calendar summaries, weekly content prompts, or even one time reminders. Once scheduled, Gemini handles them automatically in the background with no follow up required. For example, a user might say, 'Send me a summary of today's meetings every morning at 8 AM' or 'Generate weekly blog ideas every Friday at 10 AM.' These tasks run quietly behind the scenes, transforming Gemini from a reactive chatbot into a daily-use productivity tool. The setup process is built to be intuitive, making automation easy for both everyday users and professionals. Within the Gemini app, users can define a task, set the time, and choose the frequency through a clean and accessible interface. Scheduled Actions puts Google in direct competition with the kind of automation ChatGPT users create through Zapier or custom workflows. What gives Gemini a clear edge is its deep integration with Google's suite of apps. Functioning across Gmail, Calendar, Docs, and Tasks, Gemini offers a smooth setup and efficient task execution experience. Since it is built into tools people already use, Gemini can interact directly with information across Google's ecosystem. There is no need for third party services or custom scripts. For users already invested in Google's platform, the experience is more seamless than ChatGPT's dependence on external integrations. Scheduled Actions signals a shift in expectations for how AI assistants should function. Instead of waiting for commands, Gemini can now anticipate and handle repetitive tasks, offering a more personal and assistant like experience. While this may be just the beginning, it is a clear step toward positioning Gemini as a truly productivity first AI assistant. And as Gemini continues to evolve, it may not just catch up to ChatGPT but define the next generation of digital assistance.


Hindustan Times
4 hours ago
- Hindustan Times
Apple WWDC 2025: Expectation of next gen OSes, and an imperative AI question
Next week will be ushered in by the annual Apple Worldwide Developer Conference keynote, traditionally a showcase of the company's software vision, developer focus, and in a continued trend for last year, the question about the direction the tech giant intends to take with artificial intelligence (AI). The hardware announcements, though rare, haven't historically been entirely off the agenda — Apple Silicon and Vision Pro being some examples — and that adds an element of intrigue. One question many will ask in the lead-up is, when is the next-generation Siri finally landing? An answer to that, may define how the market reacts, post the keynote. After all, Siri cannot be seen lagging, as Google Gemini, OpenAI's GPT, and Anthropic's Claude models, make substantive forward leaps, periodically. The big expectation revolves around multiple-layered changes across the OS lines. The iOS for iPhone, iPadOS for the iPad, macOS for the Mac portfolio and watchOS are expected to get significant visual overhaul. The aesthetic changes are likely to be designed to add even greater uniformity across the platforms, with the idea being a switch from the current interface, to something that's closer to what's referred to in design circles as 'glassmorphism'. It is designed to bring a sense of light and space to the interface elements, with the transparency adding a sense of depth. It will be interesting to see to what extent Apple overhauls its core apps, particularly on iOS — Phone, Messages, Notes, Safari and Camera. Some may get more than a new coat of paint, and our estimate would be a higher degree of changes await Safari and Phone, also to keep up with the ecosystem in general. Could Apple shift to a year-based naming scheme, instead of the current generational numbering? Rumours seem to indicate iOS 26 is on the agenda (signifying the year 2026, though curiously, it'll release later in 2025). Updates for iPadOS are expected to be significant this time, a means to an end that is to bring the iPad when paired with a Magic Keyboard, closer to a MacBook-esque user experience. Could we see the introduction of Finder for instance, or some of the macOS staples such as the Preview app for PDF handling, on an iPad too. If that were to happen, we'd have the perfect touchscreen MacBook — and it'll be called the iPad Pro with a Magic Keyboard. The question is, if it'll mean iPadOS adopts two distinct personalities (and interfaces), where a one similar to the current iPadOS is in use without a keyboard, and a Mac-like interface, when a keyboard is? AI question: what does Apple have in store? This is a fork in the road, one that must be navigated carefully, and it'll be intriguing to see which path Apple takes with regard to an inevitable AI question. While expectations would hope for wholesale Apple Intelligence updates, that is very unlikely. Expect Apple to smartly deploy its large language models smartly, by involving developers as well as adding new functionality across devices and software. That is, instead of trying to compete with the rapid pace of evolving the likes of Google, OpenAI, Anthropic and Perplexity have shown, in recent months. The last thing Apple would want is for an AI conversation to come across as unimpressive, particularly to the markets. Using AI models to integrate a live translation feature across Phone and third-party apps such as WhatApp and Zoom, and one that also works on the AirPods, could well be on the agenda. Could Apple open up its foundation models for third-party developers, to build their takes on AI functionality that may complement the pieces of the puzzle Apple already has in place — Writing Tools, AI summaries, and Clean Up, for instance. More to that point, updated and more parameter laden foundation models could be released, with the flexibility of on-device as well as server side processing. In April this year, Google CEO Sundar Pichai during the monopoly trial clearly said that the company expects to have a deal in place with Apple to build Gemini into Apple Intelligence sometime this year — that would be alongside Apple Intelligence's launch partner, OpenAI's GPT models. If and when this integration happens, Gemini on the basis of its incredibly powerful models (some recent updates have reset certain experiential benchmarks) will give Apple Intelligence a significant upgrade in terms of capabilities across the board — a smarter Siri, camera based features that could replicate Gemini Live, and AI based search in Safari. Could Apple and Google announce this at the WWDC keynote on Monday night? The looming shadow of Google's antitrust proceedings, and till the US Justice Department pronounces a verdict, may pose a complication. Could this be shelved till the iPhone keynote, expected in September? Google itself stands to lose a little if Samsung (being by far the largest Android phone OEM) reduces reliance on Gemini which itself received significant updates recently, with the Perplexity partnership. Perplexity, which already has deals in place with Samsung and Motorola to integrate their AI in upcoming phones, could have something to offer Apple as well. It will certainly be more than just a search plug-in, if that partnership has to fructify. Could that mean Perplexity's agentic browser vision intersects with Safari, for instance? Very much a chance of that happening. Case in point, Eddy Cue, Apple's senior vice president of services saying during the Google antitrust trial that, 'We've been pretty impressed with what Perplexity has done, so we've started some discussions with them about what they're doing.' What are the analysts predicting? On Friday, analysts at Wedbush maintained a $270.00 price target on Apple Inc., ahead of the WWDC keynote — they hope Apple will be able to take enough steps with the AI proposition, to be able to monetise it, which will weave into the consistently strong services bundle for the tech giant. Morgan Stanley has set a target price of $235 for Apple, noting significant growth in App Store revenue could add as much as $110 million in the June quarter. In the early trading Friday, Apple stocks held at $204.