Claude Sonnet 4.6 Launches with Opus-Level Performance at Sonnet Pricing

Claude Sonnet 4.6 Launches with Opus-Level Performance at Sonnet Pricing

Anthropic launched Claude Sonnet 4.6, scoring 79.6% on SWE-bench—nearly matching Opus 4.6—at 40% lower output cost. With across-the-board upgrades in coding, agents, and long-context reasoning, this model is rewriting the cost-performance equation in AI.

Anthropic officially launched Claude Sonnet 4.6 on February 18. The model significantly outperforms its predecessor Sonnet 4.5 across nearly every dimension—coding, agentic tasks, and long-context reasoning—while delivering near-Opus 4.6 performance at Sonnet pricing. It's now the default model for Free and Pro users on claude.ai and Claude Cowork.

VentureBeat called it a 'tectonic repricing event in AI.' When you can access Opus-tier capabilities at significantly lower cost, the market implications are hard to overstate.

1. Benchmarks: Nearly Matching Opus 4.6

Benchmark performance comparison table for Claude Sonnet 4.6, Opus 4.6, and competing models
Claude Sonnet 4.6 benchmark comparison table

Sonnet 4.6 scored 79.6% on SWE-bench Verified, just 1.2 percentage points behind Opus 4.6's 80.8%. On OSWorld-Verified, which measures computer use capabilities, Sonnet 4.6 hit 72.5%—virtually identical to Opus 4.6's 72.7%.

More notably, Sonnet 4.6 actually surpasses Opus in some areas. In design quality evaluation (GDPval-AA), Sonnet 4.6 reached Elo 1633 versus Opus 4.6's 1606. In agentic financial analysis, it scored 63.3% against 60.1%. This isn't just a 'cheaper alternative'—it genuinely outperforms the flagship in specific domains.

2. Coding Performance: Preferred by 70% of Claude Code Users

Chart showing Claude Sonnet models' OSWorld computer use benchmark scores over time
Sonnet model progression on OSWorld benchmark

Coding improvements stand out the most. According to Anthropic, 70% of Claude Code users preferred Sonnet 4.6 over the previous Sonnet 4.5, while 59% chose it over the top-tier Opus 4.5. Beyond code generation, practical improvements include reduced over-engineering, fewer false success claims, and lower hallucination rates.

Replit praised its 'remarkable performance-to-cost ratio,' Cursor noted 'noticeable improvements across the board,' and GitHub confirmed 'outstanding performance in complex code modifications.'

3. Pricing and Training Data: Unmatched Value

Sonnet 4.6's base pricing is $3/MTok input and $15/MTok output—identical to Sonnet 4.5. For long-context usage beyond 200K tokens, pricing rises to $6 input and $22.50 output, but the Batch API brings it down to just $1.50 input and $7.50 output. Prompt caching reads cost only $0.30/MTok. Compared to Opus 4.6 ($5 input, $25 output), that's 40% cheaper at base output rates and 70% cheaper via Batch API.

The training data cutoff is also noteworthy. Sonnet 4.6 was trained with data through January 2026—more recent than Opus 4.6. It also supports a 1M-token context window (beta), enabling processing of large codebases and lengthy documents.

In Closing: More Options at a Reasonable Price

Claude Sonnet 4.6 delivers near-Opus performance at a reasonable price with faster speed. It's well-suited for Claude Code's upcoming agent teams feature, large codebase analysis, and complex agentic workflows—broadening the practical options available. The model enables a flexible approach: default to Sonnet 4.6 for cost-efficient high performance, and reserve Opus for when it's truly needed—a welcome update for users across the board.

Menu