Back to List
Anthropic Addresses Claude Code Quality Degradation Reports and Implements Fixes for Sonnet and Opus Models
Industry NewsAnthropicClaudeAI Engineering

Anthropic Addresses Claude Code Quality Degradation Reports and Implements Fixes for Sonnet and Opus Models

Anthropic has released a postmortem addressing recent user reports regarding the degradation of Claude's performance across specific tools, including Claude Code, the Claude Agent SDK, and Claude Cowork. The investigation identified three distinct technical issues occurring between March and April 2026: an intentional but poorly received reduction in reasoning effort to manage latency, a session-clearing bug that caused repetitive behavior and memory loss, and a system prompt change aimed at reducing verbosity that inadvertently harmed coding quality. While the API remained unaffected, these issues impacted Sonnet 4.6, Opus 4.6, and Opus 4.7. Anthropic has since reverted the problematic changes and fixed the bugs as of April 20 (v2.1.116), emphasizing their commitment to maintaining model intelligence over speed.

Hacker News

Key Takeaways

  • Three Distinct Issues Identified: The perceived degradation was caused by a change in reasoning effort, a session-clearing bug, and a system prompt instruction to reduce verbosity.
  • Specific Tools Affected: Issues were limited to Claude Code, the Claude Agent SDK, and Claude Cowork; the core API and inference layer were not impacted.
  • Models Impacted: The performance dips affected Sonnet 4.6, Opus 4.6, and Opus 4.7 across different timeframes.
  • Full Resolution: All identified issues were resolved as of April 20 with the release of version 2.1.116.

In-Depth Analysis

Reasoning Effort and Latency Trade-offs

On March 4, Anthropic attempted to address UI latency issues where the interface appeared frozen by changing the default reasoning effort from "high" to "medium." While this was intended to improve the user experience by reducing wait times, it resulted in a noticeable drop in intelligence for Sonnet 4.6 and Opus 4.6. Following user feedback indicating a preference for higher intelligence over speed, Anthropic reverted this change on April 7. The company acknowledged that prioritizing lower latency at the expense of reasoning quality was the "wrong tradeoff."

Technical Bugs and Prompting Side Effects

Two additional technical factors contributed to the degradation. On March 26, a feature designed to clear old thinking from idle sessions to improve resumption speed introduced a bug. This bug caused the system to clear thinking every turn, making the models appear forgetful and repetitive. Furthermore, an April 16 update to the system prompt intended to reduce verbosity negatively impacted coding quality when combined with other prompt adjustments. This specific issue affected the latest models, including Opus 4.7. Both the bug and the prompt changes were corrected and reverted by April 20.

Investigation Challenges and Aggregate Effects

Anthropic noted that because these three changes occurred on different schedules and affected different segments of traffic, the resulting feedback appeared as broad and inconsistent degradation. The investigation began in early March but was complicated by the difficulty of distinguishing these specific technical failures from the normal variation in user feedback. The company has reaffirmed that they never intentionally degrade models and are implementing changes to prevent similar regressions in the future.

Industry Impact

This incident highlights the delicate balance AI providers must maintain between model "intelligence" (reasoning effort) and operational performance (latency). For the AI industry, it serves as a case study in how minor optimizations—such as reducing verbosity or clearing session cache—can have significant, unintended consequences on the quality of complex tasks like coding. Anthropic's transparent postmortem underscores the importance of user feedback loops in identifying non-obvious regressions that automated testing might miss, particularly when those regressions are tied to UI-specific implementations rather than the underlying API.

Frequently Asked Questions

Question: Was the Claude API affected by these quality issues?

No. Anthropic confirmed that the API and inference layer remained unaffected throughout this period; the issues were isolated to Claude Code, the Claude Agent SDK, and Claude Cowork.

Question: Which Claude models were impacted by the performance degradation?

The issues affected Sonnet 4.6, Opus 4.6, and Opus 4.7, depending on the specific technical change and the timeframe.

Question: How has Anthropic resolved these issues?

As of April 20 (v2.1.116), Anthropic has reverted the reasoning effort to "high," fixed the session-clearing bug, and removed the system prompt instructions that were harming coding quality.

Related News

Langfuse: An Open Source LLM Engineering Platform for Observability and Prompt Management
Industry News

Langfuse: An Open Source LLM Engineering Platform for Observability and Prompt Management

Langfuse has emerged as a comprehensive open-source engineering platform specifically designed for Large Language Model (LLM) applications. Originating from the Y Combinator W23 cohort, the platform provides a robust suite of tools including LLM observability, metrics tracking, evaluation frameworks, and prompt management. It also features a dedicated playground and dataset management capabilities. Langfuse is built with broad compatibility in mind, offering seamless integration with industry-standard tools such as OpenTelemetry, Langchain, the OpenAI SDK, and LiteLLM. By focusing on the critical infrastructure needs of AI developers, Langfuse aims to streamline the lifecycle of LLM application development from initial testing to production monitoring.

OpenMetadata: A Unified Platform for Data Discovery, Observability, and Governance Solutions
Industry News

OpenMetadata: A Unified Platform for Data Discovery, Observability, and Governance Solutions

OpenMetadata has emerged as a comprehensive open-source solution designed to streamline how organizations manage their data ecosystems. By providing a unified metadata platform, it addresses the critical needs of data discovery, observability, and governance. The platform is built upon a centralized metadata repository that serves as a single source of truth, complemented by advanced features such as deep column-level lineage and tools for seamless team collaboration. As data environments become increasingly complex, OpenMetadata aims to simplify the management of data assets by integrating these essential functions into a cohesive framework, allowing teams to better understand, monitor, and control their data lifecycle through a standardized metadata approach.

U.S. Soldier Charged with Insider Trading on Polymarket Using Classified Military Information
Industry News

U.S. Soldier Charged with Insider Trading on Polymarket Using Classified Military Information

Gannon Ken Van Dyke, a U.S. Army soldier, has been indicted for allegedly using classified government information to profit from bets on the prediction market platform Polymarket. According to the U.S. Attorney's Office for the Southern District of New York, Van Dyke participated in the planning of 'Operation Absolute Resolve,' a military mission to capture Nicolás Maduro. He is accused of leveraging his access to sensitive details regarding the timing and outcome of this operation to place illegal wagers. The charges include commodities fraud, wire fraud, theft of nonpublic government information, and making unlawful monetary transactions. This case marks a significant legal action against insider trading within decentralized prediction markets involving national security secrets.