Anthropic has launched Claude Sonnet 4.5, its most advanced coding model to date, offering major enhancements in agent tasks, long-duration performance, and computer interaction capabilities. According to the company, improved training and safety protocols have significantly refined the model’s behavior, reducing tendencies toward flattery, deception, power-seeking, and delusional reasoning. The model is now available via the Claude API, as well as desktop and mobile apps, at the same price as its predecessor.
Claude Sonnet 4.5 builds upon Anthropic's strategy of incremental model improvements while maintaining consistency and safety. The model demonstrates the ability to sustain complex multi-step reasoning and code execution tasks for over 30 hours. On the SWE-bench validation benchmark, which measures an AI model’s ability to solve real-world software issues, Sonnet 4.5 scored 77.2%, up from 72.7% for Sonnet 4, marking a notable leap in autonomous coding capability. On the OSWorld benchmark, which evaluates real-world computer usage skills, Sonnet 4.5 achieved 61.4%, a significant increase from 42.2% four months prior.
Anthropic describes Sonnet 4.5 as its “most consistent frontier model,” highlighting the balance between enhanced capabilities and rigorous safety measures. Under ASL-3, the company has improved automated classifiers to detect and block potentially harmful instructions, including those related to chemical, biological, radiological, or nuclear (CBRN) risks. According to Anthropic, false positive rates have dropped tenfold compared to when Claude Opus 4 was launched in May 2025.
To evaluate Claude Sonnet 4.5’s behavior in autonomous, tool-supported scenarios, Anthropic conducted a series of agent safety tests covering malicious code generation and adversarial prompt injection attacks. Out of 150 malicious coding requests prohibited under Anthropic’s usage policies, Sonnet 4.5 failed only two, reflecting enhanced safety training. The model achieved a safety score of 98.7%, compared to 89.3% for Claude Sonnet 4, demonstrating significantly stronger refusal behavior and resistance to malicious agent use.
Anthropic encourages all users to upgrade to Claude Sonnet 4.5, positioning it as a “drop-in replacement” that delivers superior performance without added cost.
Early adopters report measurable gains in coding workflows:
Scott Wu, co-founder and CEO of Cognition, noted, “For Devin, Claude Sonnet 4.5 improved planning performance by 18% and end-to-end evaluation scores by 12%—the largest improvement we’ve seen since the release of Claude Sonnet 3.6. It excels at testing its own code, allowing Devin to run longer, tackle more difficult tasks, and deliver production-ready code.”
Michele Catasta, president of Replit, shared, “Claude Sonnet 4.5’s editing capabilities are outstanding. We reduced error rates from 9% with Sonnet 4 to 0% on our internal code editing benchmark. Higher tool success rates and lower costs represent a major leap forward for agent-based coding. Claude Sonnet 4.5 strikes the perfect balance between creativity and control.”
Independent open-source developer Simon Wilson wrote on his blog, “My initial impression is that it feels more suited to coding than GPT-5-Codex. It has been my go-to coding model since its release a few weeks ago.”
Anthropic’s push for safer, more autonomous coding models aligns with similar advancements in the AI ecosystem. OpenAI recently released GPT-5-Codex, a version of GPT-5 optimized for complex software engineering tasks such as large-scale code refactoring and extended code review workflows.