logo
Turing Award Goes to 2 Pioneers of Artificial Intelligence

Turing Award Goes to 2 Pioneers of Artificial Intelligence

New York Times05-03-2025

In 1977, Andrew Barto, as a researcher at the University of Massachusetts, Amherst, began exploring a new theory that neurons behaved like hedonists. The basic idea was that the human brain was driven by billions of nerve cells that were each trying to maximize pleasure and minimize pain.
A year later, he was joined by another young researcher, Richard Sutton. Together, they worked to explain human intelligence using this simple concept and applied it to artificial intelligence. The result was 'reinforcement learning,' a way for A.I. systems to learn from the digital equivalent of pleasure and pain.
On Wednesday, the Association for Computing Machinery, the world's largest society of computing professionals, announced that Dr. Barto and Dr. Sutton had won this year's Turing Award for their work on reinforcement learning. The Turing Award, which was introduced in 1966, is often called the Nobel Prize of computing. The two scientists will share the $1 million prize that comes with the award.
Over the past decade, reinforcement learning has played a vital role in the rise of artificial intelligence, including breakthrough technologies such as Google's AlphaGo and OpenAI's ChatGPT. The techniques that powered these systems were rooted in the work of Dr. Barto and Dr. Sutton.
'They are the undisputed pioneers of reinforcement learning,' said Oren Etzioni, a professor emeritus of computer science at the University of Washington and founding chief executive of the Allen Institute for Artificial Intelligence. 'They generated the key ideas — and they wrote the book on the subject.'
Their book, 'Reinforcement Learning: An Introduction,' which was published in 1998, remains the definitive exploration of an idea that many experts say is only beginning to realize its potential.
Psychologists have long studied the ways that humans and animals learn from their experiences. In the 1940s, the pioneering British computer scientist Alan Turing suggested that machines could learn in much the same way.
But it was Dr. Barto and Dr. Sutton who began exploring the mathematics of how this might work, building on a theory that A. Harry Klopf, a computer scientist working for the government, had proposed. Dr. Barto went on to build a lab at UMass Amherst dedicated to the idea, while Dr. Sutton founded a similar kind of lab at the University of Alberta in Canada.
'It is kind of an obvious idea when you're talking about humans and animals,' said Dr. Sutton, who is also a research scientist at Keen Technologies, an A.I. start-up, and a fellow at the Alberta Machine Intelligence Institute, one of Canada's three national A.I. labs. 'As we revived it, it was about machines.'
This remained an academic pursuit until the arrival of AlphaGo in 2016. Most experts believed that another 10 years would pass before anyone built an A.I. system that could beat the world's best players at the game of Go.
But during a match in Seoul, South Korea, AlphaGo beat Lee Sedol, the best Go player of the past decade. The trick was that the system had played millions of games against itself, learning by trial and error. It learned which moves brought success (pleasure) and which brought failure (pain).
The Google team that built the system was led by David Silver, a researcher who had studied reinforcement learning under Dr. Sutton at the University of Alberta.
Many experts still question whether reinforcement learning could work outside of games. Game winnings are determined by points, which makes it easy for machines to distinguish between success and failure.
But reinforcement learning has also played an essential role in online chatbots.
Leading up to the release of ChatGPT in the fall of 2022, OpenAI hired hundreds of people to use an early version and provide precise suggestions that could hone its skills. They showed the chatbot how to respond to particular questions, rated its responses and corrected its mistakes. By analyzing those suggestions, ChatGPT learned to be a better chatbot.
Researchers call this 'reinforcement learning from human feedback,' or R.L.H.F. And it is one of the key reasons that today's chatbots respond in surprisingly lifelike ways.
(The New York Times has sued OpenAI and its partner, Microsoft, for copyright infringement of news content related to A.I. systems. OpenAI and Microsoft have denied those claims.)
More recently, companies like OpenAI and the Chinese start-up DeepSeek have developed a form of reinforcement learning that allows chatbots to learn from themselves — much as AlphaGo did. By working through various math problems, for instance, a chatbot can learn which methods lead to the right answer and which do not.
If it repeats this process with an enormously large set of problems, the bot can learn to mimic the way humans reason — at least in some ways. The result is so-called reasoning systems like OpenAI's o1 or DeepSeek's R1.
Dr. Barto and Dr. Sutton say these systems hint at the ways machines will learn in the future. Eventually, they say, robots imbued with A.I. will learn from trial and error in the real world, as humans and animals do.
'Learning to control a body through reinforcement learning — that is a very natural thing,' Dr. Barto said.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

AI leaders have a new term for the fact that their models are not always so intelligent
AI leaders have a new term for the fact that their models are not always so intelligent

Business Insider

time3 hours ago

  • Business Insider

AI leaders have a new term for the fact that their models are not always so intelligent

Progress is rarely linear, and AI is no exception. As academics, independent developers, and the biggest tech companies in the world drive us closer to artificial general intelligence — a still hypothetical form of intelligence that matches human capabilities — they've hit some roadblocks. Many emerging models are prone to hallucinating, misinformation, and simple errors. Google CEO Sundar Pichai referred to this phase of AI as AJI, or "artificial jagged intelligence," on a recent episode of Lex Fridman's podcast. "I don't know who used it first, maybe Karpathy did," Pichai said, referring to deep learning and computer vision specialist Andrej Karpathy, who cofounded OpenAI before leaving last year. AJI is a bit of a metaphor for the trajectory of AI development — jagged, marked at once by sparks of genius and basic mistakes. "You see what they can do and then you can trivially find they make numerical errors or counting R's in strawberry or something, which seems to trip up most models," Pichai said. "I feel like we are in the AJI phase where dramatic progress, some things don't work well, but overall, you're seeing lots of progress." In 2010, when Google DeepMind launched, its team would talk about a 20-year timeline for AGI, Pichai said. Google subsequently acquired DeepMind in 2014. Pichai thinks it'll take a little longer than that, but by 2030, "I would stress it doesn't matter what that definition is because you will have mind-blowing progress on many dimensions." By then the world will also need a clear system for labeling AI-generated content to "distinguish reality," he said. "Progress" is a vague term, but Pichai has spoken at length about the benefits we'll see from AI development. At the UN's Summit of the Future in September 2024, he outlined four specific ways that AI would advance humanity — improving access to knowledge in native languages, accelerating scientific discovery, mitigating climate disaster, and contributing to economic progress.

OpenAI Wants to get College Kids Hooked on AI
OpenAI Wants to get College Kids Hooked on AI

Gizmodo

time5 hours ago

  • Gizmodo

OpenAI Wants to get College Kids Hooked on AI

AI chatbots like OpenAI's ChatGPT have been shown repeatedly to provide false information, hallucinate completely made-up sources and facts, and lead people astray with their confidently wrong answers to questions. For that reason, AI tools are viewed with skepticism by many educators. So, of course, OpenAI and its competitors are targeting colleges and pushing its services on students—concerns be damned. According to the New York Times, OpenAI is in the midst of a major push to make ChatGPT a fixture on college campuses, replacing many aspects of the college experience with AI alternatives. According to the report, the company wants college students to have a 'personalized AI account' as soon as they step on campus, same as how they receive a school email address. It envisions ChatGPT serving as everything from a personal tutor to a teacher's aide to a career assistant that helps students find work after graduation. Some schools are already buying in, despite the educational world initially greeting AI with distrust and outright bans. Per the Times, schools like the University of Maryland, Duke University, and California State University have all signed up for OpenAI's premium service, ChatGPT Edu, and have started to integrate the chatbot into different parts of the educational experience. It's not alone in setting its sights on higher education, either. Elon Musk's xAI offered free access to its chatbot Grok to students during exam season, and Google is currently offering its Gemini AI suite to students for free through the end of the 2025-26 academic year. But that is outside of the actual infrastructure of higher education, which is where OpenAI is attempting to operate. Universities opting to embrace AI, after initially taking hardline positions against it over fears of cheating, is unfortunate. There is already a fair amount of evidence piling up that AI is not all that beneficial if your goal is to learn and retain accurate information. A study published earlier this year found that reliance on AI can erode critical thinking skills. Others have similarly found that people will 'offload' the more difficult cognitive work and rely on AI as a shortcut. If the idea of university is to help students learn how to think, AI undermines it. And that's before you get into the misinformation of it all. In an attempt to see how AI could serve in a focused education setting, researchers tried training different models on a patent law casebook to see how they performed when asked questions about the material. They all produced false information, hallucinated cases that did not exist, and made errors. The researchers reported that OpenAI's GPT model offered answers that were 'unacceptable' and 'harmful for learning' about a quarter of the time. That's not ideal. Considering that OpenAI and other companies want to get their chatbots ingrained not just in the classroom, but in every aspect of student life, there are other harms to consider, too. Reliance on AI chatbots can have a negative impact on social skills. And the simple fact that universities are investing in AI means they aren't investing in areas that would create more human interactions. A student going to see a tutor, for example, creates a social interaction that requires using emotional intelligence and establishing trust and connection, ultimately adding to a sense of community and belonging. A chatbot just spits out an answer, which may or may not be correct.

YouTube is warning some Premium Lite subscribers about more ads next month, but don't worry
YouTube is warning some Premium Lite subscribers about more ads next month, but don't worry

Android Authority

time5 hours ago

  • Android Authority

YouTube is warning some Premium Lite subscribers about more ads next month, but don't worry

Joe Maring / Android Authority TL;DR YouTube Premium Lite offers a budget-priced paid subscription that removes most ads from YouTube. Exceptions have included things like music videos, and in some markets Google has warned that Shorts may show ads, as well. The company is now sending out notices to more subscribers warning them that ads in Shorts will start appearing at the end of June. YouTube Premium is well worth paying for, giving users ad-free access to maybe the broadest library of content in streaming history. But especially if you get your music fix from another provider (like paying for Spotify Premium), it doesn't make a ton of sense to be paying full price for YouTube Premium and not taking advantage of its YouTube Music access. That's exactly why we were so happy to see Google introduce YouTube Premium Lite, which just focuses on removing (most) ads without worrying about any extras — and does so for a fraction of the price. While Premium Lite removes the vast majority of ads from normal videos, we've known that Google has carved out a series of exceptions. Those consist of 'music content, Shorts, and when you search or browse.' So far, at least in our experience, those have proved to be minimal, and we've found Premium Lite to offer a very reasonable compromise to paying full price. That said, the situation is now changing a bit, and not for the better — at least for Premium Lite subscribers in some regions. Google has recently been sending out emails to Premium Lite users in Germany, according to Deskmodder (via 9to5Google). These advise subscribers that ads in YouTube Shorts will start appearing as of June 30. We've also uncovered TWiT Community user big_D sharing the same message (this time in English). Curious why Google would be sending out notifications about ads we already knew about, and wondering why these messages didn't seem to be targeted at Premium Lite users in all nations, we reached out to Google in the hopes of getting some clarification. And it turns out that there's a simple explanation for all of this. You may recall that when we first began hearing about Premium Lite in testing last fall, it wasn't yet available in the US, instead getting started in Australia, Germany, and Thailand. And it turns out, as Google was still getting its plans for the service together, it hadn't told subscribers in Germany and Thailand that they'd be seeing ads in Shorts. By the time access expanded to the US, ads in Shorts were on the table from the beginning, but Google is only going back now and notifying customers in Germany and Thailand that they're getting them, too. So that's what going on with these emails: Most Premium Lite subscribers already knew about ads for Shorts, and now YouTube's telling the rest of you. Got a tip? Talk to us! Email our staff at Email our staff at news@ . You can stay anonymous or get credit for the info, it's your choice.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store