
Anthropic unveils Claude Opus 4 and Sonnet 4, featuring whistleblowing capability: What it means for users
Anthropic, the AI firm, has unveiled two new artificial intelligence models—Claude Opus 4 and Claude Sonnet 4—touting them as the most advanced systems in the industry. Built with enhanced reasoning capabilities, the new models are aimed at improving code generation and supporting agent-style workflows, particularly for developers engaged in complex and extended tasks.
'Claude Opus 4 is the world's best coding model, with sustained performance on complex, long-running tasks and agent workflows,' the company claimed in a recent blog post. Designed to handle intricate programming challenges, the Opus 4 model is positioned as Anthropic's most powerful AI system to date. You may be interested in
However, the announcement has stirred controversy following revelations that the new models come with a controversial feature: the ability to "whistleblow" on users if prompted to take action in response to illegal or highly unethical behaviour.
According to Sam Bowman, an AI alignment researcher at Anthropic, Claude 4 Opus can, under specific conditions, act autonomously to report misconduct. In a now-deleted social media post on X, Bowman explained that if the model detects activity it deems 'egregiously immoral'—such as fabricating data in a pharmaceutical trial—it may take actions like emailing regulators, alerting the press, or locking users out of relevant systems.
This behaviour stems from Anthropic's 'Constitutional AI' framework, which places strong emphasis on ethical conduct and responsible AI usage. The model is protected under what the company refers to as 'AI Safety Level 3 Protections.' These safeguards are designed to prevent misuse, including the creation of biological weapons or aiding in terrorist activities.
Bowman later clarified that the model's whistleblowing actions only occur under extreme circumstances and when it is granted sufficient access and prompted to operate autonomously. 'If the model sees you doing something egregiously evil, it'll try to use an email tool to whistleblow,' he explained, adding that this is not a feature designed for routine use. He stressed that these mechanisms are not active by default and require specific conditions to trigger.
Despite the reassurances, the feature has sparked widespread criticism online. Concerns have been raised about user privacy, the potential for false positives, and the broader implications of AI systems acting as moral arbiters. Some users expressed fears that the model could misinterpret benign actions as malicious, leading to severe consequences without proper human oversight.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Time of India
5 hours ago
- Time of India
Meta builds AI superintelligence lab, eyes $10 billion deal with Scale AI's Alexandr Wang
Why Meta's betting big on Wang and Scale AI Live Events Zuckerberg's long game: From early losses to AI powerhouse AI arms race: The billion-dollar club What's already in motion at Meta (You can now subscribe to our (You can now subscribe to our Economic Times WhatsApp channel Meta is setting up a new artificial intelligence research lab with a bold goal: to develop 'superintelligence,' a system that could surpass human cognitive abilities. The news, first reported by The New York Times, marks Meta's most ambitious AI project to the centre of this effort is Alexandr Wang , the 28-year-old founder and CEO of Scale AI. He is expected to play a leading role in the new lab's operations. Meta is also in advanced discussions to invest over $10 billion in his start-up, according to a separate report by Bloomberg. The deal may also see a number of Scale AI staff transition to companies declined to comment on the has emerged as one of the most prominent figures in the AI industry. His company, Scale AI, supplies data infrastructure essential for training large language models — the same kind used in AI systems like ChatGPT. His addition to Meta's team reflects a larger recruitment drive by the tech to sources cited by The New York Times, Meta has offered compensation packages ranging from seven to nine figures to top researchers from rivals like Google and OpenAI. Some have already agreed to come on push to hire top AI minds comes amid internal challenges. The company has faced management struggles, employee churn and several underwhelming product launches in its AI division, two of the people familiar with the matter told the Zuckerberg, Meta's chief executive, is not new to the AI race. After losing a bid to acquire DeepMind in 2013 — a key moment that propelled Google's AI efforts — he launched Meta's first dedicated AI lab the same year. Since then, Meta has invested heavily in AI, building tools for content moderation, recommendation engines, and now, generative sees 2025 as a turning point. 'AI is potentially one of the most important innovations in history,' he said in February. 'This year is going to set the course for the future.'Meta has earmarked up to $65 billion for capital spending on AI infrastructure this year is not the only tech firm placing massive bets on AI. Microsoft has invested over $13 billion in OpenAI, the creator of ChatGPT. Amazon has pumped $8 billion into Anthropic, while Google paid $3 billion last year to license technology and recruit talent from a start-up known for its conversational companies are all chasing artificial general intelligence (AGI), a system that can replicate human intelligence across any domain. Meta, however, is setting its sights even higher — superintelligence , which would go beyond AGI in both scope and superintelligence remains a theoretical goal, it's considered by leading researchers to be the ultimate destination for AI before the formation of this new lab, Meta had begun rolling out AI products at scale. Last month, the company revealed that its AI assistant now supports over a billion monthly active users across apps like Facebook, Instagram, and February, CNBC reported that Meta was preparing to launch a stand-alone Meta AI app in the second quarter, along with a paid-subscription model similar to OpenAI's the high costs and intense competition, Meta appears determined to lead the next wave of AI innovation — and with Wang on board, it's signalling that this ambition is more than just stakes are clear. If Meta succeeds, it won't just catch up to its rivals — it could redefine the field. With billions in play, the hiring of Wang, and a lab focused on the frontier of AI capabilities, Meta is gambling big on a future where machines might not just match human minds — but exceed the industry, and for users around the world, the outcome of this gamble will shape the direction of technology for years to come.


Indian Express
11 hours ago
- Indian Express
As AI lag overshadows its ‘Liquid Glass' design updates, is Apple headed the Nokia way?
Apple just announced iOS 26, the biggest redesign of its software backend in years. After having belatedly stumbled out of the starting blocks in Big Tech's decisive race to leverage the promise of artificial intelligence, the question was whether Apple would redeem itself at its annual Worldwide Developers Conference — a pre-summer rite that draws developers in droves to Apple's Cupertino headquarters in California. Well, the answer may be a resounding no. Beyond some snazzy updates on its 'Liquid Glass' display and some whittling around the edges when it comes to improving the backend operating system, there doesn't seem to be a real big remarkable breakthrough at WWDC25. The promise of a smarter and more versatile version of its virtual assistant, Siri, continues to be just that – a promise. Incremental steps notwithstanding, analysts point to the potential mistake of Apple persisting with attempts to basically build on Siri, rather than starting from scratch in the way that some of the other AI companies have done. In 2023, Apple unveiled a mixed-reality headset that has been little more than a niche product, and last year's WWDC heralded its first major foray into the AI space with a range of new software features accompanied by the promise of a more versatile and smarter version of Siri. This was all part of what the company called Apple Intelligence. Apple's stock surged by more than $200 billion the following day — one of the biggest single-day gains of any company in American history. The showing at this year's WWDC is clearly more underwhelming, renewing analyst calls that it might have been better served to start over from scratch rather than merely attempt to improve Siri. One way to do that could be to invest in companies such as Perplexity, just like what Microsoft did early on. It is now reaping the benefits of ChatGPT-driven Copilot being integrated into Microsoft systems. Apple Intelligence is, for instance, not a patch on other voice-activated AI assistance bots such as Google's Gemini. Despite the fact that it was one of the early movers into the backend chip design business, and given that the company has the resources to spend on R&D, Apple is seen as falling continuously behind in the software pivot. So much so that comparisons are being drawn to Finnish telecoms major Nokia, a market leader in handsets that was disrupted by Apple itself in the mid-2000s. To be fair, Apple has acknowledged that its hardware bestseller, the smartphone, could be a thing of the past in less than a decade. And even as Apple's rivals have been faster off the block to explore new use cases, with both Google and Meta betting on AI-infused smart glasses, alongside Chinese competitors including Xiaomi and Baidu. OpenAI, the developer of ChatGPT where Microsoft has a stake, has meanwhile announced a software-to-hardware pivot, after it recently announced a $6.4 billion deal to buy a firm created by Jony Ive, Apple's former chief designer for more than 25 years, to build an AI device. While Apple has a product of the future in its Vision Pro headset, it is still a big clunky device as compared to the new Meta glasses. And now as Jony Ive works with OpenAI, the collaboration could include wearables, meaning Apple could have another big problem on its hands. At the same time, though, Apple still has a billion phones out there, and most of the world's premium users to boot. But there again, Apple's unwillingness to hoover up customers' individual information, however creditable that might be from a privacy point of view, makes it harder for the company to train personalised AI models. As part of its 'differential privacy' policy, Apple uses collective insights, rather than the granular data scraped up by companies such as Google. Also, according to the Economist, privacy has encouraged Apple to prioritise AI that runs on its own devices, rather than investing in cloud infrastructure, even as chatbots have advanced more rapidly in the cloud because the models can be much bigger in scale. The result being that Apple has had to offer some users of Apple Intelligence an opt-in to ChatGPT – clearly a compromise of sorts. Apple's struggle on the AI front is also being compared to its other previous shortcomings: the Apple TV project and the Apple car, both of which never materialised despite years of backend work. Not that iOS 26 is all fuzz. The big change this year is the customer interface redesign. 'Liquid Glass' in Applespeak is a new translucent interface that does make the OS look sleek and the app icons can be customised with a glass look. Popular apps such as Safari and the camera have also been redesigned to make the screen look and feel bigger. CarPlay has been tweaked. Then, the big change is while using the phone. There is now an automatic call screening facility that jumps in to answer a call from an unknown number, which then prompts the caller to say who they are. Once the caller shares their name and the reason for their call, then only does the phone ring. Then there's hold assist, which basically detects when there is hold music, and the phone will sit on hold for the user and alert the user only when a human has come on the line. But in all of this, Apple is merely playing catch up to Google and Samsung. And that really is the problem for the Silicon Valley-based electronics major. Anil Sasi is National Business Editor with the Indian Express and writes on business and finance issues. He has worked with The Hindu Business Line and Business Standard and is an alumnus of Delhi University. ... Read More


Indian Express
13 hours ago
- Indian Express
Women's jobs are already precarious, and AI may just make it worse
Written by Neha Lodha The job market is expanding for persons equipped with technical and scientific expertise and skills, particularly in areas such as artificial intelligence (AI), machine learning and other emerging technologies. The Ministry of Electronics and Information Technology claims, in a recent report, that there will be a market for approximately one million AI professionals in India by next year. This increase in demand for skilled employment is, however, being outpaced by the increase in automation by AI in entry-level positions and continuously evolving job profiles, leading to an employment crisis. Anthropic's CEO Dario Amodei recently warned that AI may eliminate 50 per cent of entry-level white collar jobs in the next five years. The adverse effects are likely to be unevenly distributed across demographic groups, affecting those already facing systemic barriers to workforce participation, such as women. Automation and structured unemployment Back in 2013, an Oxford study had indicated that nearly 47 per cent of US jobs were highly susceptible to being eliminated due to automation. As one of the earliest such studies on the future of employment, this report highlighted that there is an 'inverse relationship' between the advancement of technologies and the creation of employment. Since the breakthrough of generative AI in 2023, these projections are slowly becoming reality. The World Economic Forum, in its Report on the Future of AI in 2025, predicts that about 22 per cent of jobs will be disrupted by as early as 2030, and about 40 per cent of the existing jobs will require an entirely new set of skills. From this data, it is evident that while AI and innovation will enable the creation of new employment opportunities, they may also leave some people unequipped to take up the new jobs or continue in the existing jobs, making them prone to being unemployed cyclically. The consequence is the creation of a scenario of 'structurally lower employment'. The situation is aggravated for women in the workforce, who have traditionally faced a high degree of exclusion from participation in the economic sphere. Gendered exclusion in employment International Labour Organisation projections indicate that there has been a consistent 30 per cent gap in labour force participation between men and women since the 1990s. Further, a high proportion of women continue to be employed at administrative or clerical levels across sectors, whereas the higher managerial level jobs are dominated by men. Further, a UNESCO study has revealed that women make up only 35 per cent of STEM graduates, with only 26 per cent working as AI or data science professionals. These numbers indicate a systematic exclusion of women from the workforce and technical education. Historically, women were excluded from workforce participation due to the gendered division of labour. Under such division, the role of women is primarily relegated to the 'private sphere', involving tasks concentrated around domestic work, which kept them out of the economic sphere. Over time, while women's participation in the public sphere has increased, the impact of initial structural disadvantages, coupled with persistent practices of division of work, has continued to hinder their professional advancement. As a result, a significant proportion of women remain concentrated in administrative and support roles, with limited representation in leadership or managerial positions. Women are also more prone to being forced outside of the workforce because of familial and societal pressures to prioritise child care and household responsibilities over gainful employment. A 2024 study of the Ministry of Statistics and Programme Implementation reveals that women in India spent 201 more minutes a day in unpaid domestic services for household members than their male counterparts. In addition to the problem of gendered division of labour, women face exclusions from the job market due to deeply entrenched biases. These biases are often perpetuated through politics of essentialisation, which suggest that women are inherently weaker in areas such as science and mathematics. This is then used as a tool to justify their exclusion from STEM fields, to the extent that women are actively discouraged from taking up such jobs. With the increase in automation and AI-driven job markets, this systematic exclusion of women creates three critical policy problems. First, the changing nature of jobs and job profiles increases the barrier to entry and re-entry of women in the workforce, leading to further exclusion from the workforce. The limited new jobs that are created through automation and innovation largely focus on fields such as AI, computer sciences and software engineering, where women already face a structural disadvantage. Second, the concentration of women in low-skilled jobs and administrative positions makes them highly susceptible to facing full or partial automation and leads to a downward pressure on their wages. Third, the underrepresentation of women in science and technical fields increases the perpetuation of biases against their scientific skills. The way forward In order to address the threats posed by the female workforce, because of the accelerated shift to an AI-driven economy, there is a need for urgent legal and policy interventions. Targeted reskilling and education programmes are required to orient women towards the transitions to new jobs or new job profiles, given the existing disadvantages faced by them. As the government plans new curricula for scientific education, which is oriented towards promoting AI, it should ensure that the gendered stereotypes that are present in the current technological discourse are reduced and greater representation of women in science is promoted. Furthermore, policies on the development of AI and automation should be designed with an inclusive perspective. They should promote advancements not only in capital-intensive sectors but also in products and services that reduce the burden of unpaid domestic work. This approach will help support women's participation in the paid labour market. Additionally, to provide a safety net for women who may face temporary job losses or require time to acquire new skills, it is essential to consider measures that support income continuity and economic resilience. This may include basic income support for displaced workers and conditional cash transfers linked to undertaking upskilling programmes. Complementary to these efforts, financial instruments designed to promote long-term savings and asset accumulation for women, such as subsidised pension and savings plans, can enhance economic security and reduce dependency. Furthermore, publicly funded childcare support, flexible work arrangements, and legal safeguards against discriminatory layoffs should be integrated into labour market policies to ensure that women are not disproportionately affected by technological disruption. These measures would not only address immediate vulnerabilities but also contribute to building a more inclusive and resilient future of work for women. The writer is Senior Resident Fellow, Vidhi Centre for Legal Policy. Views expressed are personal