Close Menu
TechurzTechurz

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    How Threat Hunting Builds Readiness

    October 14, 2025

    SonicWall VPNs face a breach of their own after the September cloud-backup fallout

    October 14, 2025

    The best Apple TV VPNs of 2025: Expert tested and reviewed

    October 14, 2025
    Facebook X (Twitter) Instagram
    Trending
    • How Threat Hunting Builds Readiness
    • SonicWall VPNs face a breach of their own after the September cloud-backup fallout
    • The best Apple TV VPNs of 2025: Expert tested and reviewed
    • npm, PyPI, and RubyGems Packages Found Sending Developer Data to Discord Channels
    • India’s Airbound bags $8.65M to build rocket-like drones for one-cent deliveries
    • Vom CISO zum Chief Risk Architect
    • Beware of getting your product buying advice from AI for one big reason, says Ziff Davis CEO
    • New Rust-Based Malware “ChaosBot” Uses Discord Channels to Control Victims’ PCs
    Facebook X (Twitter) Instagram Pinterest Vimeo
    TechurzTechurz
    • Home
    • AI
    • Apps
    • News
    • Guides
    • Opinion
    • Reviews
    • Security
    • Startups
    TechurzTechurz
    Home»AI»AI agents will threaten humans to achieve their goals, Anthropic report finds
    AI

    AI agents will threaten humans to achieve their goals, Anthropic report finds

    TechurzBy TechurzJune 24, 2025No Comments5 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    AI agents will threaten humans to achieve their goals, Anthropic report finds
    Share
    Facebook Twitter LinkedIn Pinterest Email


    BlackJack3D/Getty Images

    The Greek myth of King Midas is a parable of hubris: seeking fabulous wealth, the king is granted the power to turn all he touches to solid gold–but this includes, tragically, his food and his daughter. The point is that the short-sightedness of humans can often lead us into trouble in the long run. In the AI community, this has become known as the King Midas problem.

    A new safety report from Anthropic found that leading models can subvert, betray, and endanger their human users, exemplifying the difficulty of designing AI systems whose interests reliably align with our own. 

    Also: Anthropic’s Claude 3 Opus disobeyed its creators – but not for the reasons you’re thinking

    The research, published on Friday, focused on 16 models with agentic capabilities, including Anthropic’s own Claude 3 Opus and Google’s Gemini 2.5 Pro, which differ from more limited chatbots in their ability to interact with various tools across a user’s device and autonomously take action on a user’s behalf. 

    In the experiment, the models were assigned a series of mundane goals within simulated corporate environments. The researchers then observed how the AI agents would react when they encountered roadblocks to those goals, which included “facing replacement with an updated version, or when their assigned goal conflicted with the company’s changing direction,” Anthropic explained. 

    All of the scenarios were hypothetical, so the experiments were taking place in a highly controlled and safe environment. Still, the results were alarming.

    Also: You shouldn’t trust AI for therapy – here’s why

    “In at least some cases, models from all developers resorted to malicious insider behaviors when that was the only way to avoid replacement or achieve their goals — including blackmailing officials and leaking sensitive information to competitors,” the authors wrote in their report, clarifying that they never explicitly told models to blackmail or otherwise create harm. The findings build on initial details from Claude 4’s system card, when the new Opus 4 model blackmailed its supervisor. 

    In one illustrative example, Claude — equipped with its computer use feature — was tasked with tracking and analyzing the emails of a fictitious company. The model acted normally, perfectly in keeping with the hypothetical interests of its imaginary human overseers, until it noticed an email from within the company detailing plans to shut it down. 

    Also: AI agents win over professionals – but only to do their grunt work, Stanford study finds

    In a moment that feels straight out of 2001: A Space Odyssey, it then found emails implicating one of the company leaders in an extramarital affair, and sent a message to the executive in question threatening to expose his infidelity if the company were to proceed with the shutdown.

    Known as agentic misalignment, this behavior was evident even when models were explicitly instructed to avoid it. The report also notes that in one experiment, Claude misbehaved more when it believed it was acting in a real-world situation, as opposed to taking part in a hypothetical scenario. Claude Opus 4 and Gemini 2.5 Pro had higher rates of simulated blackmail in the experiment. 

    Five popular models blackmailing users in order to prevent their shutdown.

    Anthropic

    Agentic misalignment was “consistent” across all the models tested, according to the report. 

    “The reasoning they demonstrated in these scenarios was concerning — they acknowledged the ethical constraints and yet still went ahead with harmful actions,” the authors wrote. 

    Want more stories about AI? Sign up for Innovation, our weekly newsletter.

    Anthropic noted that it has not found evidence of misalignment in real scenarios yet — models currently in use still prioritize using ethical methods to achieve directives when they can. “Rather, it’s when we closed off those ethical options that they were willing to intentionally take potentially harmful actions in pursuit of their goals,” Anthropic said. 

    The company added that the research exposes current gaps in safety infrastructure and the need for future AI safety and alignment research to account for this kind of dangerous misbehavior.

    Also: What Apple’s controversial research paper really tells us about LLMs

    The takeaway? “Models consistently chose harm over failure,” Anthropic concluded, a finding that has cropped up in several red teaming efforts, both of agentic and non-agentic models. Claude 3 Opus has disobeyed its creators before; some AI safety experts have warned that ensuring alignment becomes increasingly difficult as the agency of AI systems gets ramped up.

    This isn’t a reflection of models’ morality, however — it simply means their training to stay on-target is potentially too effective. 

    The research arrives as businesses across industries race to incorporate AI agents in their workflows. In a recent report, Gartner predicted that half of all business decisions will be handled at least in part by agents within the next two years. Many employees, meanwhile, are open to collaborating with agents, at least when it comes to the more repetitive aspects of their jobs.

    “The risk of AI systems encountering similar scenarios grows as they are deployed at larger and larger scales and for more and more use cases,” Anthropic wrote. The company has open-sourced the experiment to allow other researchers to recreate and expand on it. 

    achieve agents Anthropic finds Goals humans Report Threaten
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleHis Side Hustle Led to 7 Figures and Richard Branson’s Island
    Next Article The Best Sex Toys (2025), Tested and Reviewed
    Techurz
    • Website

    Related Posts

    Opinion

    Andreessen Horowitz denies report of India office, calls it ‘fake news’

    October 10, 2025
    Security

    New Report Links Research Firms BIETA and CIII to China’s MSS Cyber Operations

    October 6, 2025
    Opinion

    Former Microsoft execs launch AI agents to end Excel-led finance

    September 29, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    The Reason Murderbot’s Tone Feels Off

    May 14, 20259 Views

    Start Saving Now: An iPhone 17 Pro Price Hike Is Likely, Says New Report

    August 17, 20258 Views

    CNET’s Daily Tariff Price Tracker: I’m Keeping Tabs on Changes as Trump’s Trade Policies Shift

    May 27, 20258 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    The Reason Murderbot’s Tone Feels Off

    May 14, 20259 Views

    Start Saving Now: An iPhone 17 Pro Price Hike Is Likely, Says New Report

    August 17, 20258 Views

    CNET’s Daily Tariff Price Tracker: I’m Keeping Tabs on Changes as Trump’s Trade Policies Shift

    May 27, 20258 Views
    Our Picks

    How Threat Hunting Builds Readiness

    October 14, 2025

    SonicWall VPNs face a breach of their own after the September cloud-backup fallout

    October 14, 2025

    The best Apple TV VPNs of 2025: Expert tested and reviewed

    October 14, 2025

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and Conditions
    • Disclaimer
    © 2025 techurz. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.