LLMsVideo TranslationImage GenerationVideo Generation
AI News

Independent coverage of the latest AI tool updates, releases, and comparisons.

Categories

  • AI LLMs
  • AI Video Translation
  • AI Image Generation
  • AI Video Generation

Company

  • About
  • Contact

Resources

  • Sitemap
  • AI Glossary
  • Tool Comparisons
  • Facts / Grounding
  • llms.txt
  • XML Sitemap
© 2026 AI News. Independent editorial coverage. Not affiliated with any AI company.
AI LLMs

Claude Sonnet 4.6 Matches Opus at One-Fifth the Cost — Users Preferred It 70% of the Time

Anthropic's Sonnet 4.6 ships with 1M token context, adaptive thinking, and web search tools. Internal testing showed users preferred it over Sonnet 4.5 roughly 70% of the time.

MJ

Maya Johnson

Tuesday, February 17, 2026·3 min read

Anthropic released Claude Sonnet 4.6 on February 17, 2026 — twelve days after Opus 4.6. The pricing stays at $3/$15 per million tokens while delivering near-Opus performance: users preferred Sonnet 4.6 over Sonnet 4.5 approximately 70% of the time, and 59% preferred it to Opus 4.5, according to Anthropic.

1M Context Window for Everyone

Sonnet 4.6 joins Opus 4.6 as the second Claude model with a 1M token context window (in beta). At $3/$15 per million tokens, it makes million-token context economically viable for production workloads — something that was previously reserved for the $5/$25 Opus tier.

Max output is 64K tokens. The model supports adaptive thinking, letting Claude dynamically decide how much reasoning to apply. Knowledge cutoff is reliable through August 2025, with training data extending to January 2026.

Web Search and Context Compaction

Two new capabilities shipped with Sonnet 4.6. Web search and fetch tools with dynamic filtering let Claude access real-time information and filter results based on relevance. Context compaction (beta) automatically summarizes earlier parts of long conversations to fit within context limits — effectively enabling infinite conversation length.

Both features also work on Opus 4.6, but they matter more on Sonnet because of the price point. Web-connected Claude at $3/$15 is accessible to startups and individual developers, not just enterprise teams.

Insurance Benchmark: 94% Accuracy

Anthropic highlighted a specific computer use benchmark: 94% accuracy on insurance processing tasks. While narrow, this signals where Claude is finding enterprise adoption — industries with document-heavy, form-filling workflows where reliable automation directly reduces costs.

Default Model Status

Sonnet 4.6 is now the default model for Free and Pro plans on claude.ai and Claude Cowork. This means most Claude users interact with Sonnet 4.6 by default, making it arguably the most widely used frontier model by active users.

Where It Sits in the Market

The LLM landscape at the time of launch: GPT-5.3 Codex had shipped two weeks earlier as OpenAI's best coding model. Gemini 3.1 Pro was leading reasoning benchmarks with 94.3% on GPQA Diamond. Claude Sonnet 4.6 carved out the middle ground — not the absolute best at any single benchmark, but the best value across all of them.

Our Take

Sonnet 4.6 is the model most developers should default to. The 1M context at $3/$15 is industry-leading value, and the 70% preference rate over Sonnet 4.5 suggests genuine improvement, not incremental polish. Adaptive thinking means you're not paying for deep reasoning on simple questions. And web search makes it genuinely useful for tasks requiring current information. If you're building with Claude, start here.

FAQ

What's the model ID for Claude Sonnet 4.6? The API model ID is claude-sonnet-4-6. It's available through the Anthropic API, Amazon Bedrock, and Google Cloud Vertex AI.

How does Sonnet 4.6 compare to Opus 4.6? Sonnet 4.6 delivers approximately 85-90% of Opus 4.6's performance at 60% of the cost ($3/$15 vs $5/$25). Opus 4.6 still leads on the hardest tasks and has 128K max output vs Sonnet's 64K, but for most production use cases Sonnet is the better choice.

Does Sonnet 4.6 support the 1M context window? Yes, Sonnet 4.6 supports a 1M token context window in beta, matching Opus 4.6. This is a significant upgrade from the 200K context of previous Claude models.

What is context compaction? Context compaction is a beta feature that automatically summarizes earlier parts of long conversations, allowing effectively infinite conversation length while staying within the model's context window. It works server-side, requiring no developer implementation.

Tools Mentioned

Claude (Anthropic)Safe, helpful AI assistant with extended context and reasoning
$20/mo (Pro)
GPT (OpenAI)Industry-leading large language models powering ChatGPT
$20/mo (ChatGPT Plus)
Gemini (Google)Google's multimodal AI model family
$19.99/mo (Advanced)

More in AI LLMs

AI LLMs

Meta Launches Muse Spark — Its First Closed-Source Model Targets 'Personal Superintelligence'

Meta Superintelligence Labs unveils Muse Spark with dual modes, 58% on Humanity's Last Exam, and multimodal reasoning. Breaking with tradition, the model is not open-source.

Alex Chen·Apr 8, 2026
AI LLMs

OpenAI, Anthropic, and Google Unite to Combat AI Model Copying From China

The three biggest Western AI labs are sharing information through the Frontier Model Forum to prevent Chinese competitors from extracting their models' capabilities.

Sarah Mueller·Apr 7, 2026
← Back to all news