logo
Anthropic's Claude AI models can end 'harmful' conversations

Anthropic's Claude AI models can end 'harmful' conversations

The Hindua day ago
Anthropic has said that their Claude Opus 4 and 4.1 models will now have the ability to end conversations that are 'extreme cases of persistently harmful or abusive user interactions.' The AI firm announced the move in a blog was to maintain welfare of the AI models which showed signs of distress when users insisted on continuing such conversations even when refused by Claude.
The models will end the chat only in rare 'extreme edge cases,' Anthropic said, like when 'requests for sexual content involving minors and attempts to solicit information that would enable large-scale violence or acts of terror.'
The firm said that during pre-deployment testing of Claude Opus 4, the AI model self-reported and behavioural preferences showed they were in 'apparent distress' when engaged in such conversations.
Claude has also been 'directed not to use this ability in cases where users might be at imminent risk of harming themselves or others.'
Once the AI model ends the chat, the user will not be able to send new messages in the same chat. However, older chats will remain and the user will be able to start a new conversation immediately.
Users can still go back to the same chat and edit and retry previous messages so as to not lose important chats by creating new branches from the chat that was ended.
Anthropic said that the feature is still being tested and can change based on user feedback.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

🧠 Neural Dispatch: Anthropic tokens, Perplexity's Chrome play and using the Ray-Ban Meta AI glasses
🧠 Neural Dispatch: Anthropic tokens, Perplexity's Chrome play and using the Ray-Ban Meta AI glasses

Hindustan Times

time13 minutes ago

  • Hindustan Times

🧠 Neural Dispatch: Anthropic tokens, Perplexity's Chrome play and using the Ray-Ban Meta AI glasses

ALGORITHM In this edition, we chat about Google's offer for a free AI Pro subscription for students in India, what exactly is the status with Windsurf (at least at the time of writing this), and Meta's Prometheus project and what it means for the still vague superintelligence conversations. Neural Dispatch(HT Print) Google's free AI, but what does free mean? The coding battles just got a new alignment drop. Anthropic has raised the bar in a way that could force AI rivals including OpenAI to scramble. The company announced a 5X increase in Claude's context window for Claude Sonnet 4, which is now capable of processing up to 1 million tokens for a query — essentially, an ability to digest an equivalent of thousands of pages of text in a single go. Current models available from Anthropic. Why does this matter? Because context length is quickly becoming the new arms race metric in AI, and even more relevant for developers. For context, OpenAI's GPT-4o supports a maximum of 128K tokens, Google's Gemini 1.5 Pro goes up to 2M tokens (in certain settings). With this, Anthropic has planted its flag with a practical, scaled-up version. Longer contexts mean more reliable recall for enterprise tasks, since everyone is talking about AI agents and all that. Perhaps more relevant for legal firms that are using AI tools to go through reams of paper, developers for codebase analysis, and more than one line of work that has research-heavy workflows. Anthropic claims of maintaining speed and accuracy at this scale, which has historically been the bottleneck, will certainly be tested. A key takeaway? Context windows (consumers may not realise this enough) are as crucial as raw model smarts or the hardware infrastructure. Claude is already known for safety-tuned responses (more than others, the focus is clear). Now, Anthropic now looks like it's playing on the foot front. Tredence's Milky Way I have some opinions about Agentic AI, and perhaps you do too. For now, enterprises are only too happy to dabble with the idea, and so be it. A few weeks ago, I'd written about how continuous-learning agentic systems form the 'Milky Way' of possibilities for business decision-making. A use-case from Milky Way. That emerged from my interaction with Soumendra Mohanty, who is chief strategy officer at data science company Tredence. They are better placed than most to decode the changes the AI space is bringing to the workplace, more so because they design AI agent workflows for enterprises, across multiple verticals including healthcare, banking and telecom. Now, they've launched its that they say are designed to act like digital co-workers. The name? Milky Way. So, what does this mean in practice? Unlike traditional AI assistants that sit around waiting for prompts, these agents by the way they are structured, they supposedly can reason, collaborate, and execute certain tasks on their own. Tredence says that in pilot deployments, companies in retail, consumer packaged goods, telecommunications, and healthcare reported 5X improvement in time-to-insight and 50% reduction in analytics costs. The bigger shift here is philosophical. As Tredence's CEO Shub Bhowmick puts it, 'the real challenge isn't building smarter models, it's building systems that understand context, adapt to complexity, and drive meaningful outcomes'. That's rare pragmatism in an era where AI bosses tend to drive the hype. The reality is, not every organisation will benefit from using AI agents, and not every role will be better off by firing the human and putting a machine in its place. What's Perplexity going for? Perplexity has made an unsolicited $34.5 billion bid for Google Chrome. Right after, OpenAI also hinted at some interest. I'll focus on the original move at this time — why is an AI startup trying to buy the world's most dominant browser, and quoting a price that actually exceeds Perplexity's own valuation several times over? If you are wondering where the money will come from, the company insists multiple 'large investment funds' will finance that transaction. At first glance, it seems absurd. But dig deeper, and you see Perplexity's playbook. Chrome stopped being just a browser many years ago, and with its approximately 3 billion user base, its potentially the largest distribution channel that a tech company can have (the Apple iPhones get close, in terms of numbers and demographic of users). For an AI company like Perplexity, even the dream of owning that scale all points to an advantage that none of their rivals would have. Aligning perfectly with their AI browser aspirations, with Comet. Here's what I've been thinking (but this certainly hasn't kept me up at night) — what if Google flips the script? Imagine a reverse Uno, with Google offering to buy Perplexity instead, absorbing its fast-growing AI-driven search platform. That could help Google hedge against AI competition, and Perplexity's own rise. Right now, Perplexity's bid is unlikely to succeed, but it's a shot across the bow — a way for Perplexity to signal ambition, spark conversation, but absolutely unlikely to bait Google into making a move. PROMPT This week, we'll explore how to use Ray-Ban Meta smart glasses for real-time object identification and language translation — features that turn a wearable into a live AI assistant for the world around you. Ray-Ban Meta can be called smart glasses or AI glasses, whichever rolls of your tongue easier. These sunglasses, perhaps the ideal AI wearable which many of us are discovering, combine Meta's AI assistant with an embedded camera and microphone. This allows the wearer to ask questions about what you're seeing, and more. Ray-Ban and Meta launch AI-powered smart glasses in India with camera, voice assistant, and social media features; prices start at ₹29,900 (Vishal Mathur / HT Photo) To get started, tap and hold the side of the glasses or say 'Hey Meta' to activate the assistant. You can then ask the glasses to describe objects, people, or scenes in front of you. Three standout use cases are emerging — object recognition, on-the-fly translation, and travel assistance. For object recognition, simply look at an item and ask, 'What is this?' The glasses will identify it, whether it's a type of plant, a landmark, a nice car or a household item. For translations, look at a menu, street sign, or product packaging and ask, 'Translate this into English (or another language).' The system uses AI vision and large language models to provide spoken translations through the glasses' speakers. Travel assistance is where these two features overlap. Imagine pointing at a metro map in Paris or a restaurant board in Tokyo — the glasses can identify the location and translate the text in real time, removing friction in navigating foreign environments. To get this to work, put on the Ray-Ban Meta glasses and ensure they're connected to the Meta AI app on your phone > activate the assistant by voice or touch > ask contextual questions like 'What is this landmark?' or 'Translate this sign into Hindi.' > listen to the response. In some cases, you'll find more information in the Meta AI app. Keep in mind: Translation and object recognition features may still find some limitations depending on the language and region settings, and for this to work as it should, your phone should have an active 5G/4G or Wi-Fi connection. Only then can queries be framed conversationally, and not as isolated queries or commands. THINKING "I think it's not a super useful term. If I think one definition that people like is, you know, when it's doing like a significant amount of the work in the world, although that keeps changing because people do new jobs. Another that some people use is when it can do like really good AI research on its own. I think the point of all of this is it doesn't really matter and it's just this continuing exponential of model capability that we'll rely on for more and more things. Maybe another one that people like is when it discovers like an important new mathematical theorem or an important new piece of science. I would expect that we're maybe like two-ish years away from something like that" Sam Altman, OpenAI CEO, discussing AGI on CNBC's "Squawk Box," August 8, 2025. The context: Altman's pivot away from AGI (or Artificial General Intelligence) terminology comes at a particularly telling moment. It was in this June, he'd declared that we were "entering the AGI age'. Now, just days after a GPT-5 model release which didn't match the hype, and GPT Go subscription that begins its journey from India, the OpenAI CEO is distancing himself from the very concept that has defined his company's mission (and in a way, somewhat justified its eye-watering $500 billion valuation). From, we are just a GPT-5 release away from AGI, to again being two or so years away, tells me one clear thing — science isn't aboard the hype train that AI bosses seem to be on. ChatGPT's current pricing structure in India. A rhetoric shift that coincides suspiciously with GPT-5's lukewarm reception, with criticism being that this long-awaited model is delivering only incremental (while retaining typical AI foibles) and not revolutionary improvements over its predecessor (there was even an outcry on social media that OpenAI return the GPT-4o model, which they eventually had to). From a corporate perspective, this may be bookmarked as a strategic recalibration of expectations, but for any of us who are watching closely, this is no such thing. A reality check: Altman's newfound skepticism about AGI as a "super useful term" reveals a fundamental problem, one that in any other industry, would have probably seen some exits from a company. For years, OpenAI has raised billions by positioning AGI as its North Star, yet now acknowledges there are "multiple definitions being used by different companies and individuals." Somehow, people defining AGI is a problem. This admission exposes how the AI industry has been operating in a definitional vacuum, where goal posts can be moved whenever convenient, and never at the risk of valuations or funding. When your latest model falls short of revolutionary impact, simply redefine what you were aiming for in the first place. The shift from binary AGI thinking to "levels of progress" is signifying not just semantic gymnastics, but also defensive repositioning in the face of diminishing returns. GPT-5's modest improvements suggest we may be at the doorstep of an AI plateau, where simply throwing more compute and data at transformer architectures is yielding marginal gains. Altman seems to be preemptively managing expectations by arguing that the concept itself is flawed. We've seen this before, a classic tech industry playbook, be it chips, smartphones or operating systems. When a supposedly revolutionary breakthrough doesn't materialise the way it should, quickly pivot to a narrative about continuous improvement and refinement. If AGI becomes a distraction promoted by those who need to keep raising funding, we may be witnessing a potential bubble deflation in real-time. It may be prudent for seemingly gullible corporations worldwide, who have taken the words and promises of AI companies at face value, and started replacing humans who have families to take care of, with AI 'copilots'. If ever a pack of cards was staring us in the face, this is it.

OpenAI heats up AI race with new India offering
OpenAI heats up AI race with new India offering

Time of India

time4 hours ago

  • Time of India

OpenAI heats up AI race with new India offering

OpenAI CEO Sam Altman BENGALURU: Global AI players are sharpening their India strategies as they compete for one of the world's fastest-growing markets. OpenAI on Tuesday introduced ChatGPT Go, a Rs 399 per-month subscription plan built for India, directly pitting itself against rivals Perplexity, Google, and Anthropic's Claude, which have all been pushing their premium offerings in the country. The new plan, powered by GPT-5, gives users higher message limits, image generation, file uploads, and expanded memory. It also marks the first time OpenAI subscriptions can be paid via UPI, widening accessibility in a mobile-first market. OpenAI continues to price ChatGPT Plus at Rs 1,999 a month and ChatGPT Pro at Rs 19,900. Perplexity, which partnered with Bharti Airtel last month to provide its services for free to 360 million subscribers, charges Rs 1,660 per month for its Pro plan and Rs 16,600 for its Max tier. Google's Gemini is available in India at Rs 1,950 a month for Gemini Pro and Rs 24,500 for Gemini Ultra. Anthropic's Claude is priced at Rs 1,415 a month for Claude Pro and Rs 8,300 for Claude Max. More for less India's growing relevance is not lost on OpenAI. CEO Sam Altman described the country as the company's second-largest market and said it may soon overtake the US. He also committed to increasing OpenAI's focus on India, including plans to visit. Sponsored Links Sponsored Links Promoted Links Promoted Links You May Like American Investor Warren Buffett Recommends: 5 Books For Turning Your Life Around Blinkist: Warren Buffett's Reading List Undo by Taboola by Taboola The company recently partnered with the govt's IndiaAI Mission to launch OpenAI Academy, aimed at providing AI education in regional languages and supporting startups, educators, and nonprofits. The contrasting approaches highlight the high-stakes battle for India's AI mindshare. OpenAI is leaning on affordability and local payment rails to broaden adoption, Perplexity is opting for scale through telecom distribution, while Google and Anthropic are maintaining global price positioning. With millions of students, professionals, and creators experimenting with generative AI daily, India has become a critical proving ground for how consumer AI services will scale. Stay informed with the latest business news, updates on bank holidays , public holidays , current gold rate and silver price .

Bug bounty hunting writes new income code for techies
Bug bounty hunting writes new income code for techies

New Indian Express

time4 hours ago

  • New Indian Express

Bug bounty hunting writes new income code for techies

'There was a server leak. If someone had the owner's mobile number, they could get the OTP and control the car remotely,' Shine explained. He reported the issue. Shine, who is also the Kerala chapter lead of ASRG (Automotive Security Research Group), said this kind of reporting is part of what is known as responsible disclosure. That means there is no reward, but the information is shared for the safety of users as a public service. He has earned recognition too. Toyota and Maruti have both assigned CVE (Common Vulnerabilities and Exposures) IDs to Shine for spotting a critical bug that gave unauthorised rootshell access -- a level of control only the car owner should have. 'I used to do bug bounty full-time. Now I focus on the automotive domain,' he said. Bug bounty is not limited to websites and apps anymore. A new frontier is AI security to make sure AI systems don't go rogue. Vishnuraj, from Mattannur, is on the frontlines of this. He works in Berlin with Schwarz Corporate Solutions as an AI red teamer -- a role where experts try to break AI systems to expose vulnerabilities before hackers do. His work has helped identify 10 security flaws in systems like Anthropic's Claude, Google's Bard, OpenAI's ChatGPT, and Gemini. Through this, he has earned over 12,000 Euros.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store