
AI pioneers who channeled ‘hedonistic' machines win computer science's top prize
Teaching machines in the way that animal trainers mold the behavior of dogs or horses has been an important method for developing artificial intelligence and one that was recognized Wednesday with the top computer science award.
Two pioneers in the field of reinforcement learning, Andrew Barto and Richard Sutton, are the winners of this year's A.M. Turing Award, the tech world's equivalent of the Nobel Prize.
Research that Barto, 76, and Sutton, 67, began in the late 1970s paved the way for some of the past decade's AI breakthroughs. At the heart of their work was channeling so-called 'hedonistic' machines that could continuously adapt their behavior in response to positive signals.
Reinforcement learning is what led a Google computer program to beat the world's best human players of the ancient Chinese board game Go in 2016 and 2017. It's also been a key technique in improving popular AI tools like ChatGPT, optimizing financial trading and helping a robotic hand solve a Rubik's Cube.
But Barto said the field was 'not fashionable' when he and his doctoral student, Sutton, began crafting their theories and algorithms at the University of Massachusetts, Amherst.
'We were kind of in the wilderness,' Barto said in an interview with The Associated Press. 'Which is why it's so gratifying to receive this award, to see this becoming more recognized as something relevant and interesting. In the early days, it was not.'
Google sponsors the annual $1 million prize, which was announced Wednesday by the Association for Computing Machinery.
Barto, now retired from the University of Massachusetts, and Sutton, a longtime professor at Canada's University of Alberta, aren't the first AI pioneers to win the award named after British mathematician, codebreaker and early AI thinker Alan Turing. But their research has directly sought to answer Turing's 1947 call for a machine that 'can learn from experience' — which Sutton describes as 'arguably the essential idea of reinforcement learning.'
In particular, they borrowed from ideas in psychology and neuroscience about the way that pleasure-seeking neurons respond to rewards or punishment. In one landmark paper published in the early 1980s, Barto and Sutton set their new approach on a specific task in a simulated world: balance a pole on a moving cart to keep it from falling. The two computer scientists later co-authored a widely used textbook on reinforcement learning.
'The tools they developed remain a central pillar of the AI boom and have rendered major advances, attracted legions of young researchers, and driven billions of dollars in investments,' said Google's chief scientist Jeff Dean in a written statement.
In a joint interview with the AP, Barto and Sutton didn't always agree on how to evaluate the risks of AI agents that are constantly seeking to improve themselves. They also distinguished their work from the branch of generative AI technology that is currently in fashion — the large language models behind chatbots made by OpenAI, Google and other tech giants that mimic human writing and other media.
'The big choice is, do you try to learn from people's data, or do you try to learn from an (AI) agent's own life and its own experience?' Sutton said.
Sutton has dismissed what he describes as overblown concerns about AI's threat to humanity, while Barto disagreed and said 'You have to be cognizant of potential unexpected consequences.'
Barto, retired for 14 years, describes himself as a Luddite, while Sutton is embracing a future he expects to have beings of greater intelligence than current humans — an idea sometimes known as posthumanism.
'People are machines. They're amazing, wonderful machines,' but they are also not the 'end product' and could work better, Sutton said.
'It's intrinsically a part of the AI enterprise,' Sutton said. 'We're trying to understand ourselves and, of course, to make things that can work even better. Maybe to become such things.'
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


New York Post
25 minutes ago
- New York Post
OpenAI finds more Chinese bad actors using ChatGPT for malicious purposes
Chinese bad actors are using ChatGPT for malicious purposes – generating social media posts to sow political division across the US and seeking information on military technology, OpenAI said. An organized China-linked operation, in one such incident dubbed 'Uncle Spam,' used ChatGPT to generate social media posts that were supportive and critical of contentious topics related to US politics – and then posted both versions of the comments from separate accounts, the company said in a report released Thursday. 'This appears likely designed to exploit existing political divisions rather than to promote a specific ideological stance,' OpenAI wrote in the report, describing what is known as an influence operation. Advertisement 3 A growing number of Chinese bad actors are using ChatGPT for malicious purposes, OpenAI said. REUTERS OpenAI said it followed Meta's lead to disrupt this operation, after the social media conglomerate discovered the actors were posting at hours through the day consistent with a work day in China. The actors also used ChatGPT to make logos for their social media accounts that supported fake organizations – mainly creating personas of US veterans critical of President Trump, like a so-called 'Veterans For Justice' group. These users also tried to request code from ChatGPT that they could use to extract personal data from social media platforms like X and Bluesky, OpenAI said. Advertisement While the number of these operations has jumped, they had relatively little impact as these social media accounts typically had small followings, OpenAI said. Another group of likely Chinese actors used ChatGPT to create polarizing comments on topics like USAID funding cuts and tariffs, which were then posted across social media sites. In the comments of a TikTok video about USAID funding cuts, one of these accounts wrote: 'Our goodwill was exploited. So disappointing.' Advertisement 3 Another group of likely Chinese actors used ChatGPT to create polarizing comments on topics like USAID funding cuts and tariffs. REUTERS Another post on X took the opposite stance: '$7.9M allocated to teach Sri Lankan journalists to avoid binary-gender language. Is this the best use of development funds?' These actors made posts on X appearing to justify USAID cuts as a means of offsetting the tariffs. 'Tariffs make imported goods outrageously expensive, yet the government splurges on overseas aid. Who's supposed to keep eating?' one post said. Advertisement Another read: 'Tariffs are choking us, yet the government is spending money to 'fund' foreign politics.' 3 The operations used ChatGPT to write divisive comments on some of the Trump administration's policies, including USAID funding cuts and tariffs. AFP via Getty Images In another China-linked operation, users posed as professionals based in Europe or Turkey working for nonexistent European news outlets. They engaged with journalists and analysts on social media platforms like X, and offered money in exchange for information on the US economy and classified documents, all while using ChatGPT to translate their requests. OpenAI said it also banned ChatGPT accounts associated with several bad actors who have been publicly linked to the People's Republic of China. These accounts asked ChatGPT for help with software development and for research into US military networks and government technology. OpenAI regularly releases reports on malicious activity across its platform, including reports on fake content for websites and social media platforms and attempts to create damaging malware.
Yahoo
36 minutes ago
- Yahoo
Why two AI leaders are losing talent to startup Anthropic
Why two AI leaders are losing talent to startup Anthropic originally appeared on TheStreet. Many tech companies have announced layoffs over the past few months. In the case of Microsoft, () it's happened more than once. The rise of artificial intelligence has upended the job market in undeniable ways, prompting companies to either completely automate away some positions or scale back their hiring in other areas while increasing their reliance on chatbots and AI agents. 💵💰💰💵 As more and more companies opt for job cuts and shift focus toward implementing an AI-first strategy, questions abound as to which jobs will survive this technology revolution. The number of companies embracing this method includes prominent names such as Shopify and Box. Not every tech company is slashing its workforce, though. AI startup Anthropic isn't slowing down on its hiring. In fact, it is successfully attracting talent for several industry leaders, launching a new battle for AI talent as the industry continues to boom. Founded in 2021, Anthropic is still a fairly new company, although it is making waves in the AI market. Often considered a rival to ChatGPT maker OpenAI, it is best known for producing the Claude family, a group of large language models (LLMs) that have become extremely popular, particularly in the tech describes itself as an AI safety and research company with a focus on creating 'reliable, interpretable, and steerable AI systems.' Most recently, though, it has been in the spotlight after CEO Dario Amodei predicted that AI will wipe out many entry-level white-collar jobs. Even so, Amodei's own company is currently hiring workers for many different areas, including policy, finance, and marketing. But recent reports indicate that Anthropic has been on an engineering hiring spree as well lately, successfully poaching talent from two of its primary competitors. Venture capital firm SignalFire recently released its State of Talent Report for 2025, in which it examined hiring trends in the tech sector. This year's report showed that in an industry dependent on highly skilled engineers, Anthropic isn't just successfully hiring the best talent; it is retaining it. According to SignalFire's data, 80% of the employees hired by Anthropic at least two years remain with the startup. While DeepMind is just behind it with a 78% retention rate, OpenAI trails both with only 78%, despite ChatGPT's popularity among broad ranges of users. As always, the numbers tell the story, and in this case, they highlight a compelling trend that is already shaping the future of AI. The report's authors provide further context on engineers choosing Anthropic over its rivals, stating: 'OpenAI and DeepMind. Engineers are 8 times more likely to leave OpenAI for Anthropic than the reverse. From DeepMind, the ratio is nearly 11:1 in Anthropic's favor. Some of that's expected—Anthropic is the hot new startup, while DeepMind's larger, tenured team is ripe for movement. But the scale of the shift is striking.' More AI News: OpenAI teams up with legendary Apple exec One AI stock makes up 78% of Nvidia's investment portfolio Nvidia, Dell announce major project to reshape AI Tech professionals seeking out opportunities with innovative startups is nothing new. But in this case, all three companies are offering engineers opportunities to work on important projects. This raises the question of what Anthropic more appealing than its peers. AI researcher and senior software engineer Nandita Giri, spoke to TheStreet about this trend, offering insight into why tech workers may be making these decisions. She sees it as being about far more than financial matters.'Anthropic is making serious investments in transparency tooling, scaling laws, and red-teaming infrastructure, which gives technical contributors greater ownership over how systems are evaluated and evolved,' she states. 'Compared to OpenAI and DeepMind both of which are increasingly focused on product cycles Anthropic offers more freedom to pursue deep, foundational research.' However, other experts speculate that it may be more than that. Wyatt Mayham, a lead consultant at Northwest AI, share some insights from his team, stating 'What we've heard from clients is that it's simply easier to work there with less burnout. More worklife balance if you will.' Technology consultant Kate Scott adds that while all three companies are doing important work, she sees this trend as reflecting a shift in the broader industry, one that shows engineers seeking environments 'where organizational purpose and daily execution feel closely aligned,' something that Anthropic seems to be two AI leaders are losing talent to startup Anthropic first appeared on TheStreet on Jun 5, 2025 This story was originally reported by TheStreet on Jun 5, 2025, where it first appeared. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data
Yahoo
36 minutes ago
- Yahoo
Google says its updated Gemini 2.5 Pro AI model is better at coding
Google on Thursday announced an update to its Gemini 2.5 Pro preview model that the company claims is better at certain programming tasks. The company's calling it an "updated preview," building on the upgrade to Gemini 2.5 Pro that Google announced around a month ago. Google says the model will roll out in general availability in a "couple of weeks," and is available starting today in its AI developer platforms AI Studio and Vertex AI and the Gemini app. "[Gemini 2.5 Pro] continues to excel at coding, leading on difficult coding benchmarks," Google writes in a blog post. "It also shows top-tier performance [on] highly challenging benchmarks that evaluate a model's math, science, knowledge, and reasoning capabilities." So what else is new? Google says it addressed feedback from its previous 2.5 Pro release, improving the model's style and structure. Now, 2.5 Pro can be "more creative with better-formatted responses," Google claims. This article originally appeared on TechCrunch at