What else is people using ? Haiku 4.5 ?
So we can maybe vibe, depending what service we use.
They made pro have many times more value than paying per token and then they made max again have 25x more tokens than pro on the $200 plan.
It’s a bit like being offered rice at $1 per grain (pay per token) or a tiny bag of rice for $20 (pro) or a truck load for $200. That’s the pricing structure right now.
So while i agree you can’t easily exceed the quota on the big plans it’s a little crazy how they’ve tiered pricing. I hope no one out there’s paying per token!
Some companies are. Yes, for Claude Code. My co used to be like that as it's an easy ramp up instead of giving devs who might not use it that much a $150/mo seat; if you use it enough you can have a seat and save money, but if you're not touching $150 in credits a month just use the API. Oxide also recommends using API pricing. [0]
0: https://gist.github.com/david-crespo/5c5eaf36a2d20be8a3013ba...
Man that cracks me up.
But what about when LLMs go down and a good chunk of a whole generation won't even know how to think, when the remote system goes down? (Is the ability to think "the engine" of self and agency in this metaphor?)
We are building a wildly irresponsible context to exist in.
There is a very interesting thing happening right now where the "llm over promisers" are incentivized to over promise for all the normal reasons -- but ALSO to create the perception that the "next/soon" breakthrough is only going to be applicable when run on huge cloud infra such that running locally is never going to be all that useful ... I tend to think that will prove wildly wrong and that we will very soon arrive at a world where state of art LLM workloads should be expected to be massively more efficiently runnable than they currently are -- to the point of not even being the bottleneck of the workflows that use these components. Additionally these workloads will be viable to run locally on common current_year consumer level hardware ...
"llm is about to be general intelligence and sufficient llm can never run locally" is a highly highly temporary state that should soon be falsifiable imo. I don't think the llm part of the "ai computation" will be the perf bottleneck for long.
I've often thought that local power generation (via solar or wind) could be (or could have been) a viable alternative to national grid supply.
Imagination, either the first or last thing to die in 2075.
This should be a given for any service that others rely on, but sadly this is seldom the case.
Claude user base believes in Sunday PM work sessions
I was also fortunate to be using Claude at that exact moment (for personal reasons), which meant I could immediately see the severity of the outage.
Spent it in bloody Figma instead :(
"You have reached the messages quota for your account. It will reset in 2 hours, or you can upgrade now"
Either I have perfect timing for reaching my quota limits, or some product monetization manager deserves a raise.Only curious as a developer and dev op. It's all quite interesting where and how things go wrong especially with large deployments like Anthropic.
"I can't change the laws of physics!"
we let you run compute across any cloud so if a region/cloud experiences an issue, users are served on surviving regions/clouds/on-prem.
Cheers, Doron
delaminator•1h ago
sgt•1h ago
onionisafruit•1h ago