frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Every AI Subscription Is a Ticking Time Bomb for Enterprise

https://www.thestateofbrand.com/news/ai-subscription-time-bomb
51•mooreds•1h ago

Comments

jqpabc123•48m ago
It is "bait and switch" --- done on an industrial scale.
returnInfinity•46m ago
Brad Gerstner confirmed that tokens aren't being sold at a loss. Whatever the formula, API + Subscription split, the companies are making a profit on net token sale.

They maybe running at loss after all the salaries and stock comp, but tokens are in profit now.

iLoveOncall•38m ago
He's an interested party. His investments are worth a lot more if he says that tokens are sold at a profit. I don't understand how anyone would trust him?
wqaatwt•33m ago
There are plenty of various providers on OpenRouter serving very large Chinese models like GLM for a fraction of what OpenAI/Anthropic. Presumably they are making a profit.

It’s unlikely that Claude is proportionally that bigger and more expensive to serve so profit margins on inference must be pretty decent

sarchertech•13m ago
Do we know they are making a profit though? They could be subsidizing use to build market share the same way. They might not have billions, but at the volumes they are selling maybe they’ve got the cash to do it.

Even if they are “profitable” how many Uber drivers are “profitable” because they aren’t correctly calculating asset depreciation. Maybe these guys are doing the same thing.

Maybe it’s a lot of people who already had GPUs for crypto mining, and they’ve moved over to this, so that if they need to grow and buy new GPUs the costs would dramatically grow.

mvanbaak•8m ago
also, it's very much possible that the chinese companies get heavy investments from the state. Since it's very hard to get this info we have no idea wether they really make a profit or not.
mpalmer•35m ago
This is the sort of uncritical thinking that inflates bubbles in the aggregate.
wqaatwt•32m ago
Compared to the inference prices for open models it’s highly unlikely OpenAI/Anthropic are not making decent amounts of money from inference.

How many times bigger could Opus be than GLM or Kimi, it’s certainly not proportional to the price

m0llusk•31m ago
That isn't enough. Over time the need for growth and increasing profits will squeeze existing margins.
riddlemethat•26m ago
Open source models apply pressures on the low end of the market. The paid models are so much better that they can charge based on value for enterprises.
rglullis•11m ago
Have you used any of the recent models? My experience with GLM 5.1 does not make me miss Opus at all.
hypercube33•18m ago
I think for a while this is possible - the models definitely aren't as efficient as they can be as we've seen a lot of promising papers over the last year about how people are changing pieces and parts to do more with less. None of it has come to market yet that I'm aware of so for now it's just a hope I suppose but things like Opus definitely burn a ton of compute to be the leader in benchmarks but the gaps are closing.
ainch•22m ago
Tokens can be sold at profit, but 70% of compute expenditure goes to R&D and model training[0]. Inference needs to cover all of that as well as being profitable in a vacuum.

[0] https://epoch.ai/data-insights/openai-compute-spend

utopiah•17m ago
It's like witnessing a rocket using the most powerful engine on Earth then once it escaped orbit turn off the engine and said "It is flying without power!".

Yes, sure, right now it is ... but that's NOT how it got here.

There are trillions invested to recoup and at most billions in sales. It doesn't add up to tokens making a profit any time soon.

cryo32•3m ago
They aren't being sold at a loss but they aren't being sold at enough to cover the current losses and the costs. The losses are being passed around in some fucked up circular funding mess which will inevitably collapse into a debt crisis at some point.
542458•37m ago
I’ve said this before on HN, but there are two things that make me optimistic that we won’t see a big rug pull where price-to-capability ratio skyrockets relative to today:

* People keep finding ways of cramming more intelligence into smaller models, meaning that a given hardware spec delivers more model capability over time. I remember not that long ago when cutting edge 70B parameter models could kinda-sorta-sometimes write code that worked. Versus today, when Qwen 27BA3B (1/23 of the active parameters!) is actually *fun* to vibe code with in a good harness. It’s not opus smart, but the point is you don’t need a trillion parameters to do useful things.

* Hardware will continue to improve and supply will catch up to demand, meaning that a dollar will deliver more hardware spec over time. Right now the industry is massively supply constrained, but I don’t see any reason that has to continue forever. Every vendor knows that memory quality and memory bandwidth and the new metrics of note, and I expect to start seeing products that reflect that in a few years.

I hope that one day we’ll look back on the current model of “accessing AI through provider APIs” the same way we now look back on “everyone connecting to the company mainframe.”

ainch•16m ago
The price for a given level of capability will fall, but the frontier has recently been getting more expensive. If you compare GPT-5 to GPT-5.5 on the Artificial Analysis benchmark, it's ~4x more expensive, but achieves a higher score. Claude 4.7 is also more expensive than predecessors because of a tokenizer change.

As the AI labs become more reliant on enterprise adoption, it makes sense to push capabilities at a cost that makes sense for businesses. Even if it prices out consumers or hobbyists.

garrickvanburen•6m ago
I agree.

Between: more efficient models - tuned for the task at hand, the ability to run those models in-house, or even at the edges, plus Google and Microsoft are well positioned to stay ambivalent as they’ve got lots of products to sell and whether or not LLMs are part of the portfolio mix is completely dependent on enterprise customer demand.

Anthropic/OpenAI have a number of aggressive downward pressures on their pricing.

gizmodo59•29m ago
Inference is profitable. Companies lose money because:

1. Training is expensive. Not just compute but getting the data, researchers salaries etc 2. You have to keep producing new models to ensure people use your inference and there seems to be no end to this. So they have to pour more billions to keep the cycle going on 3. People salary and other admin cost are not that high compared to 1 and 2.

atq2119•25m ago
Inference at per-token pricing is profitable.

The article's point is that if you're relying on flat fee subscriptions, a rude awakening may be coming. That seems plausible to me. Issues around token quotas are a frequent topic on HN.

fg137•21m ago
So? How does it change the equation?

Nobody is going to charge "inference price" for model usage.

einrealist•22m ago
Those price increases will increase the pressure to use cheaper / free models (commoditization), thus cutting into the revenue projections of the frontier model vendors. Its going to be exciting to see what happens to these huge investments and valuations.
fg137•16m ago
> increase the pressure to use cheaper / free models

Not necessarily. Many factors go into what models are available at enterprise level. If you look around, not many companies (everywhere around the world) use DeepSeek models even though they are significantly cheaper.

ghusto•20m ago
TL;DR to save you time:

1. GenAI companies are making a loss in order to gain adoption and later lock-in

2. ???

3. They're going to cash-in soon and start milking you now that business critical systems rely on GenAI

The "???" denotes a complete failure to offer compelling arguments that link 1 and 3.

Sharlin•19m ago
I think I'm going to puke if I see one more "It's not X. It's Y." phrase or the word "load-bearing" used metaphorically.
baal80spam•15m ago
Managers in my org love using it in "their" Slack messages.
fg137•17m ago
Why does the author assume that enterprises use subscriptions?

Many companies use models deployed on Azure/Bedrock etc are already paying based on usage (often with discounts).

stego-tech•9m ago
Not SMBs and SMEs. Big Enterprises would generally be using API buckets or Enterprise-specific consumption models via sales teams and contracts, but most companies would default to subscription tiers - either due to shadow IT paying out of pocket for subscriptions to duck corporate IT, or because they’re too small to negotiate rates and API buckets, or because their IT teams lack the skills needed for the same.

Remember that enthusiasts leaning on API keys and large enterprises are the exception, not the norm, and even some large customers may lean on subscriptions for at-scale adoption and wait for teams to report hitting usage caps before buying more token buckets. Subscriptions are predictable, reliable, and above all else a contractable way to acquire service.

Truth be told, this has been my red flag in orgs and with peers elsewhere for several years, now. Those orgs leaning on subscriptions are in for a nasty surprise within a year or two (like the author, I predict sooner than later), especially if those subscriptions power internal processes instead of AI buckets.

Hell, this is why I think there’s a sudden focus on the “Forward Deployed Engineer” nonsense role: helping organizations migrate from subscriptions to token buckets for processes so the bill shock doesn’t send them running away screaming.

rvcdbn•14m ago
Article is mistaken these subs are not available to businesses. Companies are paying much closer to API prices. The strategy is to get you accustomed to infinite tokens on your personal sub and bet that behavior transfers to work.
PKop•10m ago
> is not a rounding error. It is

Who said it was?

> Pull out the napkin. This matters.

The article wouldn't exist if you didn't think it mattered, just tell us why.

> the question is not whether they got a good deal. The question is

Who said that was the question?

> This Is Not One Company's Problem

Who said it was?

Stop telling us what thing aren't, just speak like a normal human and convey your own thoughts. It's an insult to your audience to throw constant AI slop at them.

> thousands of companies have woven AI subscriptions deep into their operations. Marketing teams draft copy through ChatGPT Plus.

Yea I bet you do..

jeswin•1m ago
Since we can't reliably detect AI generated crap, I think it makes sense to penalize their submission. I say this as a generally pro-AI person.

Native all the way, until you need text

https://justsitandgrin.im/posts/native-all-the-way-until-you-need-text/
102•dive•1h ago•59 comments

I don't think AI will make your processes go faster

https://frederickvanbrabant.com/blog/2026-05-15-i-dont-think-ai-will-make-your-processes-go-faster/
58•TheEdonian•1h ago•36 comments

Apple Silicon costs more than OpenRouter

https://www.williamangel.net/blog/2026/05/17/offline-llm-energy-use.html
59•datadrivenangel•1h ago•34 comments

Every AI Subscription Is a Ticking Time Bomb for Enterprise

https://www.thestateofbrand.com/news/ai-subscription-time-bomb
52•mooreds•1h ago•31 comments

Zerostack – A Unix-inspired coding agent written in pure Rust

https://crates.io/crates/zerostack/1.0.0
463•gidellav•15h ago•235 comments

Prolog Basics Explained with Pokémon

https://unplannedobsolescence.com/blog/prolog-basics-pokemon/
58•birdculture•2d ago•8 comments

Mozilla to UK regulators: VPNs are essential privacy and security tools

https://blog.mozilla.org/netpolicy/2026/05/15/mozilla-to-uk-regulators-vpns-are-essential-privacy...
373•WithinReason•7h ago•147 comments

A nicer voltmeter clock

https://lcamtuf.substack.com/p/a-nicer-voltmeter-clock
242•surprisetalk•14h ago•29 comments

Colossus: The Forbin Project

https://en.wikipedia.org/wiki/Colossus:_The_Forbin_Project
138•doener•2d ago•47 comments

Hosting a website on an 8-bit microcontroller

https://maurycyz.com/projects/mcusite/
165•zdw•12h ago•14 comments

Moving away from Tailwind, and learning to structure my CSS

https://jvns.ca/blog/2026/05/15/moving-away-from-tailwind--and-learning-to-structure-my-css-/
592•mpweiher•1d ago•330 comments

OpenAI and Government of Malta partner to roll out ChatGPT Plus to all citizens

https://openai.com/index/malta-chatgpt-plus-partnership/
237•bookofjoe•17h ago•281 comments

How Diamonds Are Made

https://diamond.jaydip.me/
22•lemonberry•1d ago•2 comments

Playing Atari ST Music on the Amiga with Zero CPU

https://arnaud-carre.github.io/2026-05-15-ym-fast-emu/
72•z303•5h ago•24 comments

SANA-WM, a 2.6B open-source world model for 1-minute 720p video

https://nvlabs.github.io/Sana/WM/
364•mjgil•1d ago•142 comments

Mado: Fast Markdown linter written in Rust

https://github.com/akiomik/mado
13•nateb2022•2d ago•2 comments

Illusions of understanding in the sciences

https://link.springer.com/article/10.1007/s42113-026-00271-1
65•sebg•2d ago•31 comments

Twilight of the Velocipede: Typesetting Races Before the Age of Linotype

https://publicdomainreview.org/essay/twilight-of-the-velocipede/
30•benbreen•16h ago•1 comments

We've made the world too complicated

https://user8.bearblog.dev/the-world-is-too-complicated/
341•James72689•1d ago•331 comments

Roman Letters

https://romanletters.org/
53•diodorus•2d ago•10 comments

The Third Hard Problem

https://mmapped.blog/posts/48-the-third-hard-problem
107•surprisetalk•3d ago•51 comments

Accelerando (2005)

https://www.antipope.org/charlie/blog-static/fiction/accelerando/accelerando.html
308•eamag•1d ago•176 comments

Frontier AI has broken the open CTF format

https://kabir.au/blog/the-ctf-scene-is-dead
396•frays•1d ago•414 comments

Why did Clovis toolmakers choose difficult quartz crystal?

https://phys.org/news/2026-04-clovis-toolmakers-difficult-quartz-crystal.html
33•PaulHoule•2d ago•21 comments

Halt and Catch Fire

https://unstack.io/halt-and-catch-fire
164•ScottWRobinson•19h ago•81 comments

MCP Hello Page

https://www.hybridlogic.co.uk/blog/2026/05/mcp-hello-page
116•Dachande663•15h ago•36 comments

Unknowable Math Can Help Hide Secrets

https://www.quantamagazine.org/how-unknowable-math-can-help-hide-secrets-20260511/
61•Xcelerate•3d ago•13 comments

δ-mem: Efficient Online Memory for Large Language Models

https://arxiv.org/abs/2605.12357
226•44za12•1d ago•58 comments

A molecule with half-Möbius topology

https://www.science.org/doi/10.1126/science.aea3321
103•bryanrasmussen•4d ago•7 comments

Self-Distillation Enables Continual Learning [pdf]

https://arxiv.org/abs/2601.19897
79•teleforce•12h ago•20 comments