logo
Untangling safety from AI security is tough, experts say

Untangling safety from AI security is tough, experts say

Axios03-03-2025

Recent moves by the U.S. and the U.K. to frame AI safety primarily as a security issue could be risky, depending on how leaders ultimately define "safety," experts tell Axios.
Why it matters: A broad definition of AI safety could encompass issues like AI models generating dangerous content, such as instructions for building weapons or providing inaccurate technical guidance.
But a narrower approach might leave out ethical concerns, like bias in AI decision-making.
Driving the news: The U.S. and the U.K. declined to sign an international AI declaration at the Paris summit this month that emphasized an "open," "inclusive" and "ethical" approach to AI development.
Vice President JD Vance said at the summit that "pro-growth AI policies" should be prioritized over AI safety regulations.
The U.K. recently rebranded its AI Safety Institute as the AI Security Institute.
And the U.S. AI Safety Institute could soon face workforce cuts.
The big picture: AI safety and security often overlap, but where exactly they intersect depends on perspective.
Experts universally agree that AI security focuses on protecting models from external threats like hacks, data breaches and model poisoning.
AI safety, however, is more loosely defined. Some argue it should ensure models function reliably — like a self-driving car stopping at red lights or an AI-powered medical tool correctly identifying disease symptoms.
Others take a broader view, incorporating ethical concerns such as AI-generated deepfakes, biased decision-making, and jailbreaking attempts that bypass safeguards.
Yes, but: Overly rigid definitions could backfire, Chris Sestito, founder and CEO of AI security company HiddenLayer, tells Axios.
"We can't be flippant and just say, 'Hey, this is just on the bias side and this is on the content side,'" Sestito says. "It can get very out of control very quickly."
Between the lines: It's unclear which AI safety initiatives may be deprioritized as the U.S. shifts its approach.
In the U.K., some safety-related work — such as preventing AI from generating child sexual abuse materials — appears to be continuing, says Dane Sherrets, AI researcher and staff solutions architect at HackerOne.
Sestito says he's concerned that AI safety will be seen as a censorship issue, mirroring the current debate on social platforms.
But he says AI safety encompasses much more, including keeping nuclear secrets out of models.
Reality check: These policy rebrands may not meaningfully change AI regulation.
"Frankly, everything that we have done up to this point has been largely ineffective anyway," Sestito says.
What we're watching: AI researchers and ethical hackers have already been integrating safety concerns into security testing — work that is unlikely to slow down, especially given recent criticisms of AI red teaming in a DEF CON paper.
But the biggest signals may come from AI companies themselves, as they refine policies on whom they sell to and what security issues they prioritize in bug bounty programs.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

76% of workers have seen senior colleagues resist new tech
76% of workers have seen senior colleagues resist new tech

Yahoo

timean hour ago

  • Yahoo

76% of workers have seen senior colleagues resist new tech

This story was originally published on To receive daily news and insights, subscribe to our free daily newsletter. If you're looking for cues as to how your employees will adapt to new technology tools, consider the age profile of the organization's workforce. In a survey of 500 U.S. professionals across various industries conducted by Yooz, a provider of purchase-to-pay automation, more than three-quarters (76%) of all workers witnessed senior colleagues push back against new technology. Furthermore, 55% of millennials (those born 1981 to 1996) said they're 'excited and eager' to try new tools, while only 22% of baby boomers (those born 1946 and 1964) said the same. More than a third (35%) of the baby boomers said they feel cautious or annoyed, or prefer sticking to the old system, when new technology is plugged in. Not a single Gen Z survey participant (those born 1997 and 2012) selected that response. At the same time, about a quarter of Gen Z employees have refused to use a new workplace tool, more than any of the other generations, which Yooz characterized as Gen Z not being 'afraid to say no.' About AI specifically, 35% of Gen Z workers said they 'love' new tools, versus 13% of Boomers. Still, 40% of employees overall said they find AI helpful but unreliable. 'The takeaway: skepticism is still widespread, but younger employees see clear value,' Yooz wrote in its survey report. The report said organizations 'need to manage rollouts carefully: leverage the enthusiasm of younger adopters to build momentum, but also address the concerns of veteran staff who may need more reassurance to get on board.' The key to winning over anyone reluctant to embrace AI is building trust through real-world use cases and support, showcasing quick wins such as an AI tool that saves everyone time on a tedious task, Yooz wrote. Among employees, the most commonly cited need with regard to new technologies is better training on AI tools. More than half (52%) of those polled said their company takes a 'learn-as-you-go' approach, providing only some basic training or documentation. Relatively few employees said their employer provides 'a lot' of training on new tools. This embedded content is not available in your region. And, almost half (48%) of the employees said better training for all would be among the most effective ways to help a company adopt new technology more effectively. The research also delved into the question of who should drive decisions to implement new workplace technology. While younger employees are craving clear direction from the C-suite, at the same time, there is 'a call for more bottom-up input in tech decisions,' according to the survey report. More than a third (35%) of Gen Z respondents said new workplace tools should be chosen by leadership with input from younger employees. Additionally, more than a quarter of Millennial and Gen Z respondents said adoption would improve if leadership embraced change faster and more visibly. A sizable minority (28%) of Gen Z employees said they feel older employees are actively holding back innovation at their company. Yooz advised a collaborative approach to the issue, pairing 'tech-savvy younger employees with veteran staff to share knowledge and encourage cross-generational support during rollouts.' Yooz partnered on the research with Pollfish, a third-party survey platform, and it was conducted in February 2025. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

Vista CEO expects AI Could Replace 60% of Bankers
Vista CEO expects AI Could Replace 60% of Bankers

Bloomberg

timean hour ago

  • Bloomberg

Vista CEO expects AI Could Replace 60% of Bankers

Get a jump start on the US trading day with Matt Miller, Katie Greifeld and Sonali Basak on "Bloomberg Open Interest." President Trump and China's Xi Jinping spoke by phone says they cleared up disputes surrounding rare earth exports. Meanwhile, the payrolls report also looms as job cuts pile up around the world... and billionaire Robert Smith sounds the alarm about AI's impact on the workforce. And Dan Mendelson the CEO of Morgan Health joins Bloomberg Open Interest to talk about Trump's order to lower drug prices. (Source: Bloomberg)

OpenAI CEO Sam Altman says AI is ready for entry-level jobs—but unbothered Gen Z have made it their new work friend
OpenAI CEO Sam Altman says AI is ready for entry-level jobs—but unbothered Gen Z have made it their new work friend

Yahoo

timean hour ago

  • Yahoo

OpenAI CEO Sam Altman says AI is ready for entry-level jobs—but unbothered Gen Z have made it their new work friend

Billionaire OpenAI CEO Sam Altman reveals that AI can already perform tasks of junior-level employees—and the ability for it to work days at a time is just around the corner. With fellow tech leaders like Nvidia's Jensen Huang saying those who fail to embrace the technology will be replaced, some Gen Zers are catching on. If you're in desperate need of an intern, there's good news: there may soon be an abundance of them. But they might not be able to fetch you a coffee. OpenAI CEO Sam Altman admitted this week that AI agents—AI-powered systems that can complete job-related tasks with other software tools—can now effectively do the same work as entry-level employees. 'Today (AI) is like an intern that can work for a couple of hours but at some point it'll be like an experienced software engineer that can work for a couple of days,' Altman said on a panel with Snowflake CEO Sridhar Ramaswamy. In the coming months, AI agents will only get exponentially better, Altman said—to the point where their skills are just as good as an experienced software engineer. They're anticipated to operate continuously for days on end, without pause. 'I would bet next year that in some limited cases, at least in some small ways, we start to see agents that can help us discover new knowledge, or can figure out solutions to business problems that are very non-trivial,' the 40-year-old AI CEO added. Fortune reached out to Altman for comment. While this may seem like a grim reality for some workers, the future of human employees' success may depend on following the advice of tech CEOs like Nvidia's Jensen Huang. He predicted those who fail to embrace AI might be the next employee to get the pink slip. 'You're not going to lose your job to an AI, but you're going to lose your job to someone who uses AI,' he said at the Milken Institute's Global Conference last month. Generative AI may be eclipsing the skills of entry-level workers—like conducting research or developing PowerPoints. Some Gen Z have already seen the writing on the wall, and begun embracing the technology more than other age groups. About 51% of Gen Z now view generative AI just like a co-worker or as a friend, according to a recent survey from That's compared to just over 40% of millennials and 35% of Gen X or baby boomers who feel the same way. Altman has gone even further to say that many young people (including millennials) are turning to AI for far more than just internet sleuthing: '(It's a) gross oversimplification, but like older people use ChatGPT as a Google replacement. Maybe people in their 20s and 30s use it as like a life advisor, and then, like people in college use it as an operating system,' Altman said at Sequoia Capital's AI Ascent event earlier this month. 'And there's this other thing where (young people) don't really make life decisions without asking ChatGPT what they should do,' he added. Not all tech leaders have been as upbeat about the future, and have instead used their public appearances to highlight fears of an AI-driven job market reckoning. According to Anthropic CEO Dario Amodei, AI could eliminate half of all entry-level white-collar jobs within five years. Unemployment could skyrocket to 10% to 20%, he told Axios. To put that into context, it's currently at around 4%. Researchers at his company added that the next decade will be 'pretty terrible' for humans as desk jobs are automated, they told tech podcaster Dwarkesh Patel in an interview. This comes as the latest model of Claude—Anthropic's generative AI—can now reportedly code autonomously for nearly seven hours. This story was originally featured on

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into the world of global news and events? Download our app today from your preferred app store and start exploring.
app-storeplay-store