Singularity Thread Tracks Anthropic’s 1M-Token Context Rollout Without a Long-Context Premium

Original: 1M context is now generally available for Opus 4.6 and Sonnet 4.6. No more long context price increase in the API View original →

Read in other languages: 한국어日本語
LLM Mar 16, 2026 By Insights AI (Reddit) 2 min read Source

What Anthropic changed

On March 13, 2026, a post in r/singularity surfaced Anthropic’s announcement that Claude Opus 4.6 and Sonnet 4.6 now support a full 1 million token context window under general availability. The operational change is not just the raw window size. Anthropic says standard pricing now applies across the whole range, with no separate long-context premium. The post lists Opus 4.6 at $5 / $25 per million tokens and Sonnet 4.6 at $3 / $15, using the same rates whether the request is small or near the top of the window.

That pricing change is why the Reddit thread drew attention. Expanding context is useful, but removing the special surcharge is what makes it easier to treat long-context workflows as default rather than exceptional.

What general availability includes

Anthropic’s release note adds several concrete platform changes. Standard account rate limits now apply even at 1M context. Media limits expand from 100 to 600 images or PDF pages per request. Requests above 200K tokens no longer need a beta header. Anthropic also says that Opus 4.6 with 1M context is now included in Claude Code for Max, Team, and Enterprise users, so long-running coding sessions can keep more of their earlier work instead of compacting aggressively.

Availability is not limited to Anthropic’s own platform. The announcement says 1M context is available on Claude Platform directly and through Amazon Bedrock, Google Cloud Vertex AI, and Microsoft Foundry. That matters for teams whose procurement or security policies depend on specific cloud channels.

The performance argument behind the rollout

Anthropic pairs the pricing story with a quality claim: Opus 4.6 scores 78.3% on MRCR v2 at that context length and, in the company’s framing, maintains retrieval quality across the full window. The post gives several customer examples to justify why that matters in practice: synthesizing many papers and codebases in one research run, reviewing long negotiation histories, handling incident-response context without repeated compaction, and letting code agents keep large diffs or long execution traces intact.

Those examples are still vendor-selected, but they point to a real workflow shift. The question is no longer whether million-token contexts are technically possible. It is whether they are affordable enough, rate-limited sanely enough, and accurate enough to replace chunking, summarization, and repeated context resets in production systems.

Why the Reddit post mattered

The Reddit thread reached 415 points and 47 comments at crawl time, and the title itself focused on the pricing angle. That emphasis is rational. For developers and tool builders, long context only changes architecture once the economics stop punishing the common case. Anthropic’s announcement pushes 1M context from “premium special mode” closer to “normal operating envelope,” especially for coding and agent workflows.

Primary source: Anthropic announcement. Community discussion: r/singularity.

Share: Long

Related Articles

LLM Hacker News 1d ago 2 min read

Anthropic says 1M context is now generally available for Opus 4.6 and Sonnet 4.6 with standard pricing, no long-context premium, and media limits expanded to 600 images or PDF pages. Hacker News treated the announcement as a practical deployment story rather than a simple spec bump.

Comments (0)

No comments yet. Be the first to comment!

Leave a Comment

© 2026 Insights. All rights reserved.