
Exclusive: New Claude Model Triggers Stricter Safeguards at Anthropic
Today's newest AI models might be capable of helping would-be terrorists create bioweapons or engineer a pandemic, according to the chief scientist of the AI company Anthropic.
Anthropic has long been warning about these risks—so much so that in 2023, the company pledged to not release certain models until it had developed safety measures capable of constraining them.
Now this system, called the Responsible Scaling Policy (RSP), faces its first real test.
On Thursday, Anthropic launched Claude 4 Opus, a new model that, in internal testing, performed more effectively than prior models at advising novices on how to produce biological weapons, says Jared Kaplan, Anthropic's chief scientist. 'You could try to synthesize something like COVID or a more dangerous version of the flu—and basically, our modeling suggests that this might be possible,' Kaplan says.
Accordingly, Claude 4 Opus is being released under stricter safety measures than any prior Anthropic model. Those measures—known internally as AI Safety Level 3 or 'ASL-3'—are appropriate to constrain an AI system that could 'substantially increase' the ability of individuals with a basic STEM background in obtaining, producing or deploying chemical, biological or nuclear weapons, according to the company. They include beefed-up cybersecurity measures, jailbreak preventions, and supplementary systems to detect and refuse specific types of harmful behavior.
To be sure, Anthropic is not entirely certain that the new version of Claude poses severe bioweapon risks, Kaplan tells TIME. But Anthropic hasn't ruled that possibility out either.
'If we feel like it's unclear, and we're not sure if we can rule out the risk—the specific risk being uplifting a novice terrorist, someone like Timothy McVeigh, to be able to make a weapon much more destructive than would otherwise be possible—then we want to bias towards caution, and work under the ASL-3 standard,' Kaplan says. 'We're not claiming affirmatively we know for sure this model is risky … but we at least feel it's close enough that we can't rule it out.'
If further testing shows the model does not require such strict safety standards, Anthropic could lower its protections to the more permissive ASL-2, under which previous versions of Claude were released, he says.
This moment is a crucial test for Anthropic, a company that claims it can mitigate AI's dangers while still competing in the market. Claude is a direct competitor to ChatGPT, and brings in some $1.4 billion in annualized revenue. Anthropic argues that its RSP thus creates an economic incentive for itself to build safety measures in time, lest it lose customers as a result of being prevented from releasing new models. 'We really don't want to impact customers,' Kaplan told TIME earlier in May while Anthropic was finalizing its safety measures. 'We're trying to be proactively prepared.'
But Anthropic's RSP—and similar commitments adopted by other AI companies—are all voluntary policies that could be changed or cast aside at will. The company itself, not regulators or lawmakers, is the judge of whether it is fully complying with the RSP. Breaking it carries no external penalty, besides possible reputational damage. Anthropic argues that the policy has created a 'race to the top' between AI companies, causing them to compete to build the best safety systems. But as the multi-billion dollar race for AI supremacy heats up, critics worry the RSP and its ilk may be left by the wayside when they matter most.
Still, in the absence of any frontier AI regulation from Congress, Anthropic's RSP is one of the few existing constraints on the behavior of any AI company. And so far, Anthropic has kept to it. If Anthropic shows it can constrain itself without taking an economic hit, Kaplan says, it could have a positive effect on safety practices in the wider industry.
Anthropic's new safeguards
Anthropic's ASL-3 safety measures employ what the company calls a 'defense in depth' strategy—meaning there are several different overlapping safeguards that may be individually imperfect, but in unison combine to prevent most threats.
One of those measures is called 'constitutional classifiers:' additional AI systems that scan a user's prompts and the model's answers for dangerous material. Earlier versions of Claude already had similar systems under the lower ASL-2 level of security, but Anthropic says it has improved them so that they are able to detect people who might be trying to use Claude to, for example, build a bioweapon. These classifiers are specifically targeted to detect the long chains of specific questions that somebody building a bioweapon might try to ask.
Anthropic has tried not to let these measures hinder Claude's overall usefulness for legitimate users—since doing so would make the model less helpful compared to its rivals. 'There are bioweapons that might be capable of causing fatalities, but that we don't think would cause, say, a pandemic,' Kaplan says. 'We're not trying to block every single one of those misuses. We're trying to really narrowly target the most pernicious.'
Another element of the defense-in-depth strategy is the prevention of jailbreaks—or prompts that can cause a model to essentially forget its safety training and provide answers to queries that it might otherwise refuse. The company monitors usage of Claude, and 'offboards' users who consistently try to jailbreak the model, Kaplan says. And it has launched a bounty program to reward users for flagging so-called 'universal' jailbreaks, or prompts that can make a system drop all its safeguards at once. So far, the program has surfaced one universal jailbreak which Anthropic subsequently patched, a spokesperson says. The researcher who found it was awarded $25,000.
Anthropic has also beefed up its cybersecurity, so that Claude's underlying neural network is protected against theft attempts by non-state actors. The company still judges itself to be vulnerable to nation-state level attackers—but aims to have cyberdefenses sufficient for deterring them by the time it deems it needs to upgrade to ASL-4: the next safety level, expected to coincide with the arrival of models that can pose major national security risks, or which can autonomously carry out AI research without human input.
Lastly the company has conducted what it calls 'uplift' trials, designed to quantify how significantly an AI model without the above constraints can improve the abilities of a novice attempting to create a bioweapon, when compared to other tools like Google or less advanced models. In those trials, which were graded by biosecurity experts, Anthropic found Claude 4 Opus presented a 'significantly greater' level of performance than both Google search and prior models, Kaplan says.
Anthropic's hope is that the several safety systems layered over the top of the model—which has already undergone separate training to be 'helpful, honest and harmless'—will prevent almost all bad use cases. 'I don't want to claim that it's perfect in any way. It would be a very simple story if you could say our systems could never be jailbroken,' Kaplan says. 'But we have made it very, very difficult.'
Still, by Kaplan's own admission, only one bad actor would need to slip through to cause untold chaos. 'Most other kinds of dangerous things a terrorist could do—maybe they could kill 10 people or 100 people,' he says. 'We just saw COVID kill millions of people.'
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
40 minutes ago
- Yahoo
An OpenAI exec says she was diagnosed with breast cancer and that ChatGPT has helped her navigate it
Kate Rouch, OpenAI's chief marketing officer, said she was diagnosed with breast cancer. Rouch said she is expected to make a full recovery and urged other women to prioritize their health. She said she leaned on OpenAI's ChatGPT to navigate her treatment. Kate Rouch, the chief marketing officer at OpenAI, shared on Friday that she was diagnosed with invasive breast cancer weeks after assuming the role, which she called her "dream job," in December. In a thread posted on X, Rouch said she was sharing her story to help other women, adding, "We can't control what happens to us--but we can choose how we face it. My biggest lesson: no one fights alone." Prior to joining OpenAI as the company's first CMO, Rouch was CMO at Coinbase and, before that, spent over a decade at Meta, including as vice president, global head of brand and product marketing. Rouch said she started treatment right around the Super Bowl in February, when OpenAI aired its first-ever ad, and that she has since gone through 13 rounds of chemotherapy while leading OpenAI's marketing team. She wrote that she is expected to make a full recovery. "It has been the hardest season of life — for me, for my husband, and for our two young children," Rouch said, adding she has been supported by OpenAI "at every step." "Silicon Valley can be brutal and transactional. And yet — I've never felt more held," she said, adding that "people showed up in incredible and unexpected ways." Rouch also said OpenAI's ChatGPT has helped her navigate her diagnosis and treatment, including by explaining cancer in a way that is age-appropriate for her kids, helping her manage the side effects of chemo, and creating custom meditations. "Experiencing our work as a patient has made OpenAI's mission feel more personal and important," she said. Rouch said she was sharing her story to encourage other women to "prioritize their health over the demands of families and jobs." "A routine exam saved my life. It could save yours, too," she said. Business Insider reached out to OpenAI for comment. Kevin Weil, the chief product officer at OpenAI, expressed support for Rouch in a reply to her thread. "We love you @kate_rouch!" he wrote. "Proud of you for telling your story and for being so full of fight." Read the original article on Business Insider
Yahoo
an hour ago
- Yahoo
Tensions between the US and China have delayed Apple's AI rollout in China, FT reports
The Financial Times claims that the Cyberspace Administration of China (CAC) is delaying the introduction of AI services by Apple Inc. (NASDAQ:AAPL) and Alibaba in China. A wide view of an Apple store, showing the range of products the company offers. Their February deal to incorporate AI technologies into iPhones in China is on hold due to geopolitical tensions resulting from U.S. President Donald Trump's trade war. Consumer-facing AI technologies need regulatory permission, and the CAC has not yet given its approval to these applications. The business is facing diminishing iPhone sales in China due to increased local competition, particularly from Huawei, which has integrated DeepSeek's AI models into its handsets. The firm is at a disadvantage to its AI-enabled Android rivals due to its lack of sophisticated AI technologies, such as ChatGPT-powered features and Apple Inc. (NASDAQ:AAPL)'s postponed "Apple Intelligence." More pressure is added by Trump's recent pronouncement of a 25% tariff on iPhones sold in the US that are not produced locally. New software improvements are anticipated during Apple Inc. (NASDAQ:AAPL)'s WWDC event, which takes place from June 9–13. While we acknowledge the potential of AAPL as an investment, our conviction lies in the belief that some AI stocks hold greater promise for delivering higher returns and have limited downside risk. If you are looking for an extremely cheap AI stock that is also a major beneficiary of Trump tariffs and onshoring, see our free report on the best short-term AI stock. READ NEXT: 10 High-Growth EV Stocks to Invest In and 13 Best Car Stocks to Buy in 2025. Disclosure. None. Sign in to access your portfolio

Business Insider
an hour ago
- Business Insider
An OpenAI exec says she was diagnosed with breast cancer and that ChatGPT has helped her navigate it
Kate Rouch, the chief marketing officer at OpenAI, shared on Friday that she was diagnosed with invasive breast cancer weeks after assuming the role, which she called her "dream job," in December. In a thread posted on X, Rouch said she was sharing her story to help other women, adding, "We can't control what happens to us--but we can choose how we face it. My biggest lesson: no one fights alone." Prior to joining OpenAI as the company's first CMO, Rouch was CMO at Coinbase and, before that, spent over a decade at Meta, including as vice president, global head of brand and product marketing. Rouch said she started treatment right around the Super Bowl in February, when OpenAI aired its first-ever ad, and that she has since gone through 13 rounds of chemotherapy while leading OpenAI's marketing team. She wrote that she is expected to make a full recovery. "It has been the hardest season of life — for me, for my husband, and for our two young children," Rouch said, adding she has been supported by OpenAI "at every step." "Silicon Valley can be brutal and transactional. And yet — I've never felt more held," she said, adding that "people showed up in incredible and unexpected ways." Rouch also said OpenAI's ChatGPT has helped her navigate her diagnosis and treatment, including by explaining cancer in a way that is age-appropriate for her kids, helping her manage the side effects of chemo, and creating custom meditations. "Experiencing our work as a patient has made OpenAI's mission feel more personal and important," she said. Rouch said she was sharing her story to encourage other women to "prioritize their health over the demands of families and jobs." "A routine exam saved my life. It could save yours, too," she said. Business Insider reached out to OpenAI for comment. Kevin Weil, the chief product officer at OpenAI, expressed support for Rouch in a reply to her thread. "We love you @kate_rouch!" he wrote. "Proud of you for telling your story and for being so full of fight."