Close Menu
TechurzTechurz

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    See It Here First at TechCrunch Disrupt 2025

    October 13, 2025

    Final Flash Sale: Save up to $624 on Disrupt 2025 Passes

    October 13, 2025

    I tested a Windows laptop with a tandem OLED, and it’s spoiled working on other displays for me

    October 13, 2025
    Facebook X (Twitter) Instagram
    Trending
    • See It Here First at TechCrunch Disrupt 2025
    • Final Flash Sale: Save up to $624 on Disrupt 2025 Passes
    • I tested a Windows laptop with a tandem OLED, and it’s spoiled working on other displays for me
    • Why Unmonitored JavaScript Is Your Biggest Holiday Security Risk
    • German state replaces Microsoft Exchange and Outlook with open-source email
    • Astaroth Banking Trojan Abuses GitHub to Remain Operational After Takedowns
    • The most important Intel Panther Lake updates are the least talked about – I’ll explain
    • Is AI even worth it for your business? 5 expert tips to help prove ROI
    Facebook X (Twitter) Instagram Pinterest Vimeo
    TechurzTechurz
    • Home
    • AI
    • Apps
    • News
    • Guides
    • Opinion
    • Reviews
    • Security
    • Startups
    TechurzTechurz
    Home»AI»Claude can now stop conversations – for its own protection, not yours
    AI

    Claude can now stop conversations – for its own protection, not yours

    TechurzBy TechurzAugust 19, 2025No Comments4 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Claude can now stop conversations - for its own protection, not yours
    Share
    Facebook Twitter LinkedIn Pinterest Email


    CHRISTOPH BURGSTEDT/SCIENCE PHOTO LIBRARY via Getty Images

    ZDNET’s key takeaways:

    • Claude Opus 4 and 4.1 can now end some “potentially distressing” conversations.
    • It will activate only in some cases of persistent user abuse.
    • The feature is geared toward protecting models, not users. 

    Anthropic’s Claude chatbot can now end some conversations with human users who are abusing or misusing the chatbot, the company announced on Friday. The new feature is integrated with Claude Opus 4 and Opus 4.1. 

    Also: Claude can teach you how to code now, and more – how to try it

    Claude will only exit chats with users in extreme edge cases, after “multiple attempts at redirection have failed and hope of a productive interaction has been exhausted,” Anthropic noted. “The vast majority of users will not notice or be affected by this feature in any normal product use, even when discussing highly controversial issues with Claude.”

    If Claude ends a conversation, the user will no longer be able to send messages in that particular thread; all of their other conversations, however, will remain open and unaffected. Importantly, users who Claude ends chats with will not experience penalties or delays in starting new conversations immediately. They will also be able to return to and retry previous chats “to create new branches of ended conversations,” Anthropic said. 

    The chatbot is designed not to end conversations with users who are perceived as being at risk of harming themselves or others.

    Tracking AI model well-being 

    The feature isn’t aimed at improving user safety — it’s actually geared toward protecting models themselves.

    Letting Claude end chats is part of Anthropic’s model welfare program, which the company debuted in April. The move was prompted by a Nov. 2024 paper that argued that some AI models could soon become conscious and would thus be worthy of moral consideration and care. One of that paper’s coauthors, AI researcher Kyle Fish, was hired by Anthropic as part of its AI welfare division.

    Also: Anthropic mapped Claude’s morality. Here’s what the chatbot values (and doesn’t)

    “We remain highly uncertain about the potential moral status of Claude and other LLMs, now or in the future,” Anthropic wrote in its blog post. “However, we take the issue seriously, and alongside our research program we’re working to identify and implement low-cost interventions to mitigate risks to model welfare, in case such welfare is possible.” 

    Claude’s ‘aversion to harm’

    The decision to give Claude the ability to hang up and walk away from abusive or dangerous conversations arose in part from Anthropic’s assessment of what it describes in the blog post as the chatbot’s “behavioral preferences” — that is, the patterns in how it responds to user queries. 

    Interpreting such patterns as a model’s “preferences” as opposed merely to patterns that have been gleaned from a corpus of training data is arguably an example of anthropomorphizing, or attributing human traits to machines. The language behind Anthropic’s AI welfare program, however, makes it clear that the company considers it to be more ethical in the long run to treat its AI systems as if they could one day exhibit human traits like self-awareness and a moral concern for the suffering of others.

    Also: Patients trust AI’s medical advice over doctors – even when it’s wrong, study finds

    An assessment of Claude’s behavior revealed “a robust and consistent aversion to harm,” Anthropic wrote in its blog post, meaning the bot tended to nudge users away from unethical or dangerous requests, and in some cases even showed signs of “distress.” When given the option to do so, the chatbot would end simulated some user conversations if they started to veer into dangerous territory.

    Each of these behaviors, according to Anthropic, arose when users would repeatedly try to abuse or misuse Claude, despite its efforts to redirect the conversation. The chatbot’s ability to end conversations is “a last resort when multiple attempts at redirection have failed and hope of a productive interaction has been exhausted,” Anthropic wrote. Users can also explicitly ask Claude to end a chat.

    Claude conversations Protection Stop
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleMeta spent $27 million protecting Mark Zuckerberg last year, more than any other CEO
    Next Article Google AI Pioneer Employee Says to Stay Away From AI PhDs
    Techurz
    • Website

    Related Posts

    Security

    Computer mice can eavesdrop on private conversations, researchers discover

    October 8, 2025
    Startups

    Elon Musk Is Out to Rule Space. Can Anyone Stop Him?

    September 23, 2025
    Security

    Ex-lobbyist for Meta becomes Irish data protection commissioner

    September 22, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    The Reason Murderbot’s Tone Feels Off

    May 14, 20259 Views

    Start Saving Now: An iPhone 17 Pro Price Hike Is Likely, Says New Report

    August 17, 20258 Views

    CNET’s Daily Tariff Price Tracker: I’m Keeping Tabs on Changes as Trump’s Trade Policies Shift

    May 27, 20258 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    The Reason Murderbot’s Tone Feels Off

    May 14, 20259 Views

    Start Saving Now: An iPhone 17 Pro Price Hike Is Likely, Says New Report

    August 17, 20258 Views

    CNET’s Daily Tariff Price Tracker: I’m Keeping Tabs on Changes as Trump’s Trade Policies Shift

    May 27, 20258 Views
    Our Picks

    See It Here First at TechCrunch Disrupt 2025

    October 13, 2025

    Final Flash Sale: Save up to $624 on Disrupt 2025 Passes

    October 13, 2025

    I tested a Windows laptop with a tandem OLED, and it’s spoiled working on other displays for me

    October 13, 2025

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and Conditions
    • Disclaimer
    © 2025 techurz. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.