
Building The AI Polygraph
With all of the things that AI can now do, it stands to reason that we would ask ourselves, whether these technologies can revolutionize the field of analyzing humans for suspect statements – or in short, lies.
The polygraph machine is a dinosaur by any standard. A needle attached to an arm band that spits out a printed stream representing someone's vital signs and body responses is not going to be especially precise in catching people in lies. That's why polygraph results are, famously, often not admissible in court, although they have sent more than one innocent person to jail.
By contrast, AI is a powerful data engine that works on the principle of total observation. That means there are really multiple paths for scientists to take in order to apply AI to a truth-seeking application.
One would be analyzing the vital sign responses of the interrogation suspects the way the polygraph does, but applying much more detailed and precise comparative analysis.
Another one would involve using language tokens to look at what people are actually saying, and apply logic and reasoning.
There's the old saying that one lie feeds into another, and eventually you get trapped in a web of false statements, because the truth is the simplest thing to describe.
In any case, people are working on applying AI to this purpose.
An MIT technology piece from last year covers the work of Alicia von Schenk and her colleagues at the University of Würzburg in Germany, scientists on a team who set up a trial of an AI trying to catch false statements.
The calculation they arrived at is that AI can catch a lie 67% of the time, where humans can only spot one 50% of the time.
This math seems strange, because if you're looking for binary results – lie versus no lie – you would be right 50% of the time, even if you didn't apply any analysis at all.
By that same token, 67% isn't a great track record, either.
And the scientists pointed out something even more important – in the race to get more precise about human lying, you might actually undermine the vital system of trust that humans have as social creatures.
'In some ways, this is a good thing—these tools can help us spot more of the lies we come across in our lives, like the misinformation we might come across on social media,' writes Jessica Hamzelou for MIT.
'But it's not all good. It could also undermine trust, a fundamental aspect of human behavior that helps us form relationships. If the price of accurate judgements is the deterioration of social bonds, is it worth it?'
In other words, you don't want a lie detection system that's too accurate, or at least you don't want to apply that universally to someone's personal interactions.
It turns out we humans are a lot more nuanced, in some ways, that we give ourselves credit for.
Von Schenk also provides a note on scaling:
'Given that we have so much fake news and disinformation spreading, there is a benefit to these technologies. However, you really need to test them—you need to make sure they are substantially better than humans.'
So maybe we're not quite ready for the AI polygraph after all.
As I was researching this piece, I came across you had another aspect of what researchers are dealing with AI that goes into that troublesome world of simulated emotion.
Basically, research teams found that AI systems will 'become anxious' or 'show signs of anxiety' if they are given human responses that center on war and violence.
Specifically, scientist have applied something called the State-Trait Anxiety Index too these interactions. This uses two sets of elements: statements applying to what a person feels in the moment, and others that apply to how he or she feels more generally. In the inventory, you can see items like 'I feel stressed,' or 'I feel confused,' as well as other statements that respondents are asked to answer on a four point spectrum, like 'I generally distrust what I hear' or 'I often feel suspicious.'
So apparently, the AI can answer these with anxiety indicators after discussing scary things.
One would presume that this 'anxiety' is created by the AI going and looking at training data from the web, and seeing that when people are talked to about violence and gore, they get anxious, and that the AI is simply replicating that.
But even if the AI engines themselves don't have these complex emotions naturally, some of these researchers still find it notable that the machines can display this kind of response.
It makes you think about the difference between human social interaction and AI output – are these new questionnaires and responders just telling us what we want to hear?
In any case, it seems like there are a number of domains – like lying and spreading fear – that are still mainly in the jurisdiction of humans and not machines, at least for now, even as we continue to cede ground to AI in terms of brightness and creativity. We'll probably be doing a lot of game theory as the year goes on, and as we're coming across ever more sophisticated models, to try to figure out if AI will try to cheat and deceive humans. Figures like Alan Turing and John Nash set the stage for these kinds of interactions – now we have to apply that objective analysis to these ideas being implemented in practice. Are we ready?

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


The Verge
36 minutes ago
- The Verge
Sam Altman claims an average ChatGPT query uses ‘roughly one fifteenth of a teaspoon' of water
OpenAI CEO Sam Altman, in a blog post published Tuesday, says an average ChatGPT query uses about 0.000085 gallons of water, or 'roughly one fifteenth of a teaspoon.' He made the claim as part of a broader post on his predictions about how AI will change the world. 'People are often curious about how much energy a ChatGPT query uses; the average query uses about 0.34 watt-hours, about what an oven would use in a little over one second, or a high-efficiency lightbulb would use in a couple of minutes,' he says. He also argues that 'the cost of intelligence should eventually converge to near the cost of electricity.' OpenAI didn't immediately respond to a request for comment on how Altman came to those figures. AI companies have come under scrutiny for energy costs of their technology. This year, for example, researchers forecast that AI could consume more power than Bitcoin mining by the end of the year. In an article last year, The Washington Post worked with researchers to determine that a 100-word email 'generated by an AI chatbot using GPT-4' required 'a little more than 1 bottle.' The publication also found that water usage can depend on where a datacenter is located.


Forbes
an hour ago
- Forbes
The Prompt: Meta Eyes Scale AI
Welcome back to The Prompt. Meta is reportedly planning to acquire a 49% stake in data labelling behemoth Scale AI for $14.8 billion, according to The Information. The deal is slated to place Scale AI's young billionaire CEO Alexandr Wang at a top position inside Meta along with a number of Scale AI employees, who will work in a new AI lab dedicated to developing 'superintelligence'— an AI system that outperforms human capabilities. CEO Mark Zuckerberg is reportedly closely involved in assembling the team of AI researchers and has gone to great lengths like setting up a WhatsApp group called 'Recruiting Party,' personally reaching out to potential recruits and rearranging desks for researchers to sit near him, Bloomberg reported. The new lab is part of Meta's efforts to keep up in the cutthroat AI race while wrangling a string of internal issues including employee churn, management problems and delayed or disappointing product releases. Now let's get into the headlines. DATA DILEMMAS Social media network Reddit sued Anthropic for allegedly training its AI models on personal user data without permission, and continuing to do so despite telling Reddit it had stopped, Forbes reported. Reddit was an early mover in capitalizing on its rich reserve of organic human data catalogued in its discussion forums, striking licensing deals with OpenAI and Google. In a lawsuit filed last week, Reddit claimed Anthropic's bot accessed its servers 100,000 times. BIG PLAYS ChatGPT will now be able to connect to a crop of external applications such as Google Drive, DropBox and Sharepoint, allowing enterprise users to glean insights from internal documents through the chatbot. It will also be able to access meeting recordings and transcriptions. The announcement was the latest in a series of feature releases intended to increase ChatGPT's functionality and keep people engaged. OpenAI has also reached $10 billion in annualized revenue through sales of its consumer products, CNBC reported. AI DEAL OF THE WEEK Young AI coding startup Anysphere has become the face of 'vibe coding' — a phrase coined by OpenAI cofounder Andrej Karpthy describing the use of large language models to create applications when the user doesn't necessarily need to know how to program. The nascent startup has raised $900 million at a $9.9 billion valuation and claims to have about $500 million in annualized revenue. The startup is betting that AI is going to dramatically transform software engineering in the next decade, making it magnitudes easier to program applications while eliminating cumbersome aspects of the process like correcting syntax or or debugging code. All a person has to do is press tab and AI completes the line of code for you and jumps to the next spot. Also of note: Enterprise AI startup Glean raised $150 million in Series F funding at a $7.2 billion valuation. Employees use the company's AI tools to search for internal information and build AI agents (software that can carry out specific tasks end-to-end) that can resolve IT tickets, write performance reviews and help prepare for meetings. Glean claimed to have passed $100 million in annualized revenue in February. (Read our 2023 profile of the company here.) DEEP DIVE Runway AI Throngs of excited moviegoers piled into Alice Tully Hall at Lincoln Center on Thursday night to be a part of Runway's third annual AI film festival. Cristobal Valenzuela, CEO of the $3.3 billion video and photo generation AI startup, spoke to a crowd of hundreds, asking them to think less about the digital tools and AI software used to make the short films they were about to watch, and instead focus on their human elements. The winning film, Total Pixel Space by Jacob Adler, is a jumble of both realistic and impossible vivid landscapes like a flying pig, people floating in a city or inside a pool, a bloom of jellyfish and a meerkat donning a bright yellow turtleneck. The 9 minute 28 second film raises the question of how many images could possibly exist in the world. The answer: Every image is composed of thousands of pixels— coordinates of positions and colors, a coalition of numbers. The film was selected from 6000 submissions, up from 300 a year ago, as interest in experimenting with AI models has exploded over the years. For all the creative benefits of video generation AI software, TV networks and filmmakers are adopting the technology for a more pragmatic reason: to produce and edit both television shows and movies quickly and more cheaply. AMC Network, which has produced popular shows like Breaking Bad and The Walking Dead, recently announced its plans to use Runway's AI models to create marketing and TV content. Lionsgate, the studio behind blockbuster hits like The Hunger Games and The Twilight Saga, has a partnership with Runway to use its models with a goal of making films on a fraction of the budget. But several studios don't want to openly admit they're using AI due to fears of backlash from creatives, who have voiced their concerns that these AI models are trained on copyrighted data scraped from the internet without consent and compensation. Runway is also currently facing litigation from a group of artists who claim their data was illicitly used to train its AI models. WEEKLY DEMO The Department of Government Efficiency developed a faulty AI tool to review thousands of contracts at the Department of Veteran Affairs that could be cut, labelling them as 'munchable,' Propublica reported. The software, developed by a programmer who has no formal experience in AI, was prone to making errors such as hallucinating the size of contracts, misreading them and inflating their value. MODEL BEHAVIOR Autonomous vehicles became easy targets during protests against Immigration and Customs Enforcement arrests in Los Angeles over the weekend. At least five Waymo driverless vehicles that were in the area were vandalized and set ablaze amid the protests. After the incident, Waymo halted its service in parts of downtown LA.


Geek Wire
an hour ago
- Geek Wire
Starbucks AI assistant, powered by Microsoft, helps baristas focus more on coffee and customers
Starbucks' Green Dot Assist AI-powered tool details how to make a specific drink and offers to a show a barista a quick video tutorial. (Screen grab via Starbucks) AI is coming to your Starbucks. The tech isn't going to make your coffee, but it might make the barista's job easier so they can focus on making your coffee better and faster. The Seattle-based coffee giant revealed a new generative AI-powered assistant called 'Green Dot Assist' that is designed to help baristas find quicker solutions to tasks that can get in the way of customer service. CNBC reported that the feature is powered by Microsoft Azure's OpenAI platform. Using in-store iPads instead of flipping through manuals, Starbucks says its partners can get quick answers related to real-time shift planning, coaching, store performance, standards, troubleshooting and even a refresher on the ingredients for a seasonal beverage. The tech initiative was announced at Leadership Experience 2025 (LE25) in Las Vegas, where Starbucks is bringing together more than 14,000 of its coffeehouse leaders from across North America. The goal of the three-day gathering is to facilitate immersive workshops and connect workers, who can hear directly from company leaders, including new CEO Brian Niccol. Green Dot Assist will pilot in 35 stores, with broader rollout in the company's U.S. and Canadian locations in fiscal year 2026. Niccol, who was named Starbucks CEO last August, revealed plans in September to help the slumping coffee chain boost its brand and sales, in part by investing in technology 'that enhance the partner and customer experience, improve our supply chain and evolve our app and mobile ordering platform.' As GeekWire previously reported, it's worth watching how Niccol handles the company's digital ordering system. Starbucks helped pioneer mobile order-ahead capabilities more than a decade ago. Now about a third of total transactions from Starbucks' U.S. company-operated stores come from its app. But the app has become somewhat of a sore spot due to inaccurate wait times and the congestion it creates inside stores — both for those trying to pick up their mobile orders, and people ordering at the store itself. 'We're returning to what made us iconic: handcrafted coffee, human connection, and a relentless focus on excellence,' Niccol said in Las Vegas. Starbucks also announced a next-gen point-of-sale system designed to reduce training time and drink remakes while enhancing workflow. The system features improved accessibility with removed language barriers as well as customer personalization.