Back to Blog

Did Claude's 1M Context Window Defeat Context Rot?

6 min read

Did Claude's 1M Context Window Defeat Context Rot?

Anthropic's Opus 4.6 just shipped a 1 million token context window that actually works — and the benchmark data backs it up. On the MRCR v2 eight-needle test, Opus 4.6 scores 78.3% at 1M tokens with only a 14% drop from 256K. For context, Gemini 3.1 Pro hits 26% and GPT-5.4 hits 36%. This isn't just a bigger window — it's the first time a large context window doesn't come with a massive performance penalty.

What Is Context Rot and Why Should You Care?

Context rot is what happens when an AI model's performance degrades as you feed it more tokens. You start a session, everything's sharp, and by the time you're at 100K-120K tokens, the model starts forgetting things, making mistakes, and generally falling apart.

Last summer, Chroma published a study that made this painfully clear. They tested multiple frontier models across increasing context lengths, and the pattern was the same everywhere: massive performance drop-offs as input tokens increased. The consensus was pretty straightforward — once you hit 100K-120K tokens, clear your session and start fresh, or you're going to have a bad time.

This meant those big context window numbers on the spec sheet were basically marketing. Sure, you could theoretically use 200K tokens. But past 100K, you were getting diminished returns that made it pointless.

How Does Opus 4.6 Compare on the Eight-Needle Test?

The eight-needle test (MRCR v2) is the gold standard for measuring how well a model handles long context. Here's how it works: you have a massive conversation with the model where you ask it to do similar things repeatedly — say, write eight poems about dogs at different points across the conversation. Then you ask it to reproduce a specific one. "Give me the third poem about dogs." The model has to find the right needle among eight nearly identical ones buried in up to 1M tokens of text.

Here are the numbers at 1M tokens:

  • Opus 4.6: 78.3%
  • GPT-5.4: 36%
  • Gemini 3.1 Pro: 26%
  • Sonnet 4.5: 18.5%
  • Opus 4.5: 27.1% (at 128K tokens — it couldn't even reach 1M)

Opus 4.6 doesn't just win — it more than doubles every other model on the board. And the jump from Opus 4.5 to 4.6 is even more striking: a 5x increase in context length AND a nearly 3x improvement in effectiveness. That's not an incremental upgrade.

How Much Does Performance Actually Drop at 1M Tokens?

The drop from 256K to 1M tokens is roughly 14%. That's it. Over 750,000 additional tokens, you're losing about 14 percentage points of accuracy.

If you extrapolate that, you're looking at roughly a 2% drop in effectiveness per 100,000 tokens. That's a useful rule of thumb for planning how to manage your context window.

Compare that to what Chroma found last summer with other models — performance falling off a cliff past 100K tokens. The difference here is that Opus 4.6 degrades gradually and predictably instead of cratering.

How Should This Change Your Context Window Management?

This is the practical question. If context rot is less of an issue now, should you stop clearing your session as aggressively?

Here's how I'd think about it:

  • If you can clear at 200K-256K, do it. Why take any degradation at all if you don't need to? Starting fresh gives you peak performance.
  • If you need more context, you actually have room now. Working with a huge codebase? Running a long agentic session? You no longer have to do hacky things to keep your token count artificially low. The model will still perform well at 500K, 700K, even 1M tokens.
  • The autocompact buffer is still 33K tokens. That hasn't changed. But the ceiling above it is now much, much higher.

The bottom line: you have real wiggle room now. The old rule of "clear at 100K-120K or suffer" is dead for Opus 4.6. You still want to be smart about it, but you're not walking a tightrope anymore.

What About Pricing and Availability?

A few things worth noting:

  • Available on Max, Teams, and Enterprise plans inside Claude Code. You need to be on one of these to access the 1M window.
  • No more pricing multiplier. Previously, there was a cost multiplier past 200K tokens in the API. That's gone. Whether you're at 9,000 tokens or 900,000 tokens, it's the same price per token.
  • Media support jumped too. You can now process up to 600 images or PDF pages (up from 100) and it still performs well across that full range.

The pricing change alone is significant. Before, even if you wanted to use the full context window, the cost made it impractical for a lot of use cases. That barrier is gone.

Why Is This a Bigger Deal Than Loops or Sub-Agents?

Anthopic has shipped a lot of features recently — loops, beats, sub-agents, all sorts of agentic capabilities. But a 1M context window that actually retains performance is the foundational upgrade that makes everything else better.

Every agentic workflow benefits from a model that can hold more context without falling apart. Every large codebase analysis gets more reliable. Every long research session stays sharper for longer. This isn't a feature — it's the infrastructure that every other feature depends on.

Frequently Asked Questions

Does the 1M context window work on the free plan?

No. The 1M context window is available on the Max plan, Teams, and Enterprise plans. The free and Pro tiers don't have access to the full 1M window.

Should I stop clearing my context window entirely?

No — but you can be less aggressive about it. If you can clear at 200K-256K without losing important context, that's still optimal. But if you need to push further, Opus 4.6 handles it far better than any previous model. A 2% drop per 100K tokens is manageable.

Is the eight-needle test a good real-world benchmark?

Yes. It simulates exactly the kind of problem you hit with large codebases — lots of similar code doing similar things, and you need the model to find the exact right piece. It tests both retrieval accuracy and the ability to distinguish between similar items, which is directly relevant to coding and research tasks.

Does Sonnet 4.6 also get the 1M context window?

Yes, both Opus 4.6 and Sonnet 4.6 have the 1M context window. However, Sonnet 4.5 scored only 18.5% on the eight-needle test at 1M tokens compared to Opus 4.6's 78.3%, so the raw context retention is significantly weaker on Sonnet.

What's the autocompact buffer and has it changed?

The autocompact buffer in Claude Code is still 33,000 tokens. This is the amount of recent context that Claude Code keeps fully detailed before summarizing older content. The 1M context window increases the total available space, but the autocompact mechanism works the same way.


If you want to go deeper into context window management and Claude Code optimization, join the free Chase AI community for templates, prompts, and live breakdowns. And if you're serious about building with AI, check out the paid community, Chase AI+, for hands-on guidance on how to make money with AI.