Anthropic made 1M-token context generally available for Opus 4.6 and Sonnet 4.6, removed the long-context premium, and raised media limits to 600 images or PDF pages. Use it for retrieval-heavy and codebase-scale workflows that previously needed beta headers or special long-context pricing.

The headline change is that 1M context is now GA for both Opus 4.6 and Sonnet 4.6. In the Anthropic launch, the company says developers can “load entire codebases, large document sets, and long-running agents,” and the linked blog post adds that standard pricing now applies “across the full 1M window” with no long-context premium.
This also simplifies integration. Anthropic’s API changes says there is “no beta header required in the API,” while the blog summary says requests above 200K tokens are now supported automatically and rate limits stay consistent across the full window. For teams that had built special-case paths for long prompts, that removes both a pricing branch and a header gate.
For Claude Code users, Anthropic is treating long context less like a premium mode and more like the default operating environment. The Claude Code default says “Opus 4.6 1M is now the default Opus model” for Max, Team, and Enterprise plans; Pro and Sonnet users can still opt in with /extra-usage.
That plan change matters because usage accounting changed with it. In normal plan limits, Anthropic says 1M context now “counts against your normal plan limits, no extra usage required.” Practitioner notes in session behavior and config note point to the immediate workflow effect: “fewer compactions,” longer-running sessions, and less need to start over. The same Claude Code docs also expose model-selection and extended-window aliases, while config note adds that CLAUDE_CODE_AUTO_COMPACT_WINDOW can tune when auto-compaction kicks in.
Anthropic is pairing the rollout with a specific retrieval benchmark claim: the benchmark claim says Opus 4.6 scores 78.3% on MRCR v2 at 1M tokens, “highest among frontier models.” The chart in [img:0|MRCR chart] shows Sonnet 4.6 at 65.1% at 1M, versus 36.6% for GPT-5.4 and 25.9% for Gemini 3.1 Pro on the same graphic.
The more consequential engineering detail is that Anthropic is no longer asking developers to pay extra to test whether those retrieval numbers translate into real workloads. pricing context calls out “flat pricing across the full context window to 1M tokens,” and price reaction highlights that the 1M tier now uses the same base Opus pricing rather than a separate long-context premium. That makes retrieval-heavy code search, large-spec analysis, and long agent traces easier to evaluate in production settings instead of behind a special cost model.
A solo developer wired Claude into emulators and simulators to inspect 25 Capacitor screens daily and file bugs across web, Android, and iOS. The writeup is a solid template for unattended QA, but it also shows where iOS tooling and agent reliability still crack.
releaseOpenClaw shipped version 2026.3.22 with ClawHub, OpenShell plus SSH sandboxes, side-question flows, and more search and model options, then followed with a 2026.3.23 patch. Teams get a broader plugin surface, but should patch quickly and review plugin trust boundaries as the ecosystem grows.
releaseCursor shipped Instant Grep, a local regex index built from n-grams, inverted indexes, and Bloom filters that drops large-repo searches from seconds to milliseconds. Faster candidate retrieval shortens the coding-agent loop, especially when ripgrep-style scans become the bottleneck.
breakingChatGPT now saves uploaded and generated files into an account-level Library that can be reused across conversations from the web sidebar or recent-files picker. It removes repetitive re-uploading and makes past PDFs, spreadsheets, and images part of a persistent working context.
breakingEpoch AI says GPT-5.4 Pro elicited a publishable solution to one 2019 conjecture in its FrontierMath Open Problems set, with a formal writeup planned. Treat it as an early milestone worth reproducing, not blanket evidence that frontier models can already automate math research.
1 million context window: Now generally available for Claude Opus 4.6 and Claude Sonnet 4.6.
1 million context window: Now generally available for Claude Opus 4.6 and Claude Sonnet 4.6.
Opus 4.6 scores 78.3% on MRCR v2 at 1 million tokens, highest among frontier models. Load entire codebases, large document sets, and long-running agents. Media limits expand to 600 images or PDF pages per request.
i've been surprised to see how quickly Claude is getting better at long context retrieval + reasoning. also, we just announced flat pricing across the full context window to 1M tokens. it's a good time to explore long context when building with Claude. x.com/RLanceMartin/s…