EDIT: Don't think Pro has access to it, a typical prompt just hit the context limit.
The removal of extra pricing beyond 200k tokens may be Anthropic's salvo in the agent wars against GPT 5.4's 1M window and extra pricing for that.
Personally, I’m on a 6M+ line codebase and had no problems with the old window. I’m not sending it blindly into the codebase though like I do for small projects. Good prompts are necessary at scale.
I'm not an expert but maybe this explains context rot.
The stats claim Opus at 1M is about like 5.4 at 256k -- these needle long context tests don't always go with quality reasoning ability sadly -- but this is still a significant improvement, and I haven't seen dramatic falloff in my tests, unlike q4 '25 models.
p.s. what's up with sonnet 4.5 getting comparatively better as context got longer?
My employer only pays for GitHub copilot extension
What is OpenAIs response to this? Do they even have 1M context window or is it still opaque and "depends on the time of day"
Normally buying the bigger plan gives some sort of discount.
At Claude, it's just "5 times more usage 5 times more cost, there you go".
They would probably implement _diminishing_-value pricing if pure pricing efficiency was their only concern.
"put high level description of the change you are making in log.md after every change"
works perfectly in codex but i just cant get calude to do it automatically. I always have to ask "did you update the log".
(And, yeah, I'm all Claude Code these days...)
If the chat client is resending the whole conversation each turn, then once you're deep into a session every request already includes tens of thousands of tokens of prior context. So a message at 70k tokens into a conversation is much "heavier" than one at 2k (at least in terms of input tokens). Yes?
dimitri-vs•7h ago
> Standard pricing now applies across the full 1M window for both models, with no long-context premium. Media limits expand to 600 images or PDF pages.
For Claude Code users this is huge - assuming coherence remains strong past 200k tok.
MikeNotThePope•1h ago
No vibes allowed: https://youtu.be/rmvDxxNubIg?is=adMmmKdVxraYO2yQ
ogig•1h ago
boredtofears•1h ago
grafmax•1h ago
chrisweekly•1h ago
saaaaaam•48m ago
SequoiaHope•1h ago
MikeNotThePope•1h ago
ashdksnndck•1h ago
tudelo•56m ago
dimitri-vs•1h ago
scwoodal•1h ago
steve-atx-7600•1h ago
SkyPuncher•1h ago
For me, it's less about being able to look back -800k tokens. It's about being able to flow a conversation for a lot longer without forcing compaction. Generally, I really only need the most recent ~50k tokens, but having the old context sitting around is helpful.
hombre_fatal•54m ago
Now you have to compact and you don’t know what will survive. And the built-in UI doesn’t give you good tools like deleting old messages to free up space.
I’ll appreciate the 1M token breathing room.
roygbiv2•31m ago
garciasn•3m ago
saaaaaam•49m ago
maskull•37m ago
ricksunny•33m ago
stevula•29m ago
quux•28m ago
nujabe•26m ago
Unless you’re using a text editor as an IDE you probably have already
furyofantares•28m ago
1) No longer found the dumb zone
2) No longer feared compaction
Switching to Opus for stupid political reasons, I still have not had the dumb zone - but I'm back to disliking compaction events and so the smaller context window it has, has really hurt.
I hope they copy OpenAI's compaction magic soon, but I am also very excited to try the longer context window.
iknowstuff•3m ago
mgambati•3m ago
twodave•3m ago
That said, 120k is pleeenty if you’re just building front-end components and have your API spec on hand already.
islewis•1h ago
hagen8•1h ago
a_e_k•1h ago
(Note that I'm using it in more of a hands-on pair-programming mode, and not in a fully-automated vibecoding mode.)
chatmasta•1h ago