frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Interop 2025: A Year of Convergence

https://webkit.org/blog/17808/interop-2025-review/
1•ksec•9m ago•0 comments

JobArena – Human Intuition vs. Artificial Intelligence

https://www.jobarena.ai/
1•84634E1A607A•13m ago•0 comments

Concept Artists Say Generative AI References Only Make Their Jobs Harder

https://thisweekinvideogames.com/feature/concept-artists-in-games-say-generative-ai-references-on...
1•KittenInABox•17m ago•0 comments

Show HN: PaySentry – Open-source control plane for AI agent payments

https://github.com/mkmkkkkk/paysentry
1•mkyang•19m ago•0 comments

Show HN: Moli P2P – An ephemeral, serverless image gallery (Rust and WebRTC)

https://moli-green.is/
1•ShinyaKoyano•28m ago•0 comments

The Crumbling Workflow Moat: Aggregation Theory's Final Chapter

https://twitter.com/nicbstme/status/2019149771706102022
1•SubiculumCode•33m ago•0 comments

Pax Historia – User and AI powered gaming platform

https://www.ycombinator.com/launches/PMu-pax-historia-user-ai-powered-gaming-platform
2•Osiris30•34m ago•0 comments

Show HN: I built a RAG engine to search Singaporean laws

https://github.com/adityaprasad-sudo/Explore-Singapore
1•ambitious_potat•39m ago•0 comments

Scams, Fraud, and Fake Apps: How to Protect Your Money in a Mobile-First Economy

https://blog.afrowallet.co/en_GB/tiers-app/scams-fraud-and-fake-apps-in-africa
1•jonatask•39m ago•0 comments

Porting Doom to My WebAssembly VM

https://irreducible.io/blog/porting-doom-to-wasm/
1•irreducible•40m ago•0 comments

Cognitive Style and Visual Attention in Multimodal Museum Exhibitions

https://www.mdpi.com/2075-5309/15/16/2968
1•rbanffy•42m ago•0 comments

Full-Blown Cross-Assembler in a Bash Script

https://hackaday.com/2026/02/06/full-blown-cross-assembler-in-a-bash-script/
1•grajmanu•47m ago•0 comments

Logic Puzzles: Why the Liar Is the Helpful One

https://blog.szczepan.org/blog/knights-and-knaves/
1•wasabi991011•58m ago•0 comments

Optical Combs Help Radio Telescopes Work Together

https://hackaday.com/2026/02/03/optical-combs-help-radio-telescopes-work-together/
2•toomuchtodo•1h ago•1 comments

Show HN: Myanon – fast, deterministic MySQL dump anonymizer

https://github.com/ppomes/myanon
1•pierrepomes•1h ago•0 comments

The Tao of Programming

http://www.canonical.org/~kragen/tao-of-programming.html
2•alexjplant•1h ago•0 comments

Forcing Rust: How Big Tech Lobbied the Government into a Language Mandate

https://medium.com/@ognian.milanov/forcing-rust-how-big-tech-lobbied-the-government-into-a-langua...
3•akagusu•1h ago•0 comments

PanelBench: We evaluated Cursor's Visual Editor on 89 test cases. 43 fail

https://www.tryinspector.com/blog/code-first-design-tools
2•quentinrl•1h ago•2 comments

Can You Draw Every Flag in PowerPoint? (Part 2) [video]

https://www.youtube.com/watch?v=BztF7MODsKI
1•fgclue•1h ago•0 comments

Show HN: MCP-baepsae – MCP server for iOS Simulator automation

https://github.com/oozoofrog/mcp-baepsae
1•oozoofrog•1h ago•0 comments

Make Trust Irrelevant: A Gamer's Take on Agentic AI Safety

https://github.com/Deso-PK/make-trust-irrelevant
7•DesoPK•1h ago•4 comments

Show HN: Sem – Semantic diffs and patches for Git

https://ataraxy-labs.github.io/sem/
1•rs545837•1h ago•1 comments

Hello world does not compile

https://github.com/anthropics/claudes-c-compiler/issues/1
35•mfiguiere•1h ago•20 comments

Show HN: ZigZag – A Bubble Tea-Inspired TUI Framework for Zig

https://github.com/meszmate/zigzag
3•meszmate•1h ago•0 comments

Metaphor+Metonymy: "To love that well which thou must leave ere long"(Sonnet73)

https://www.huckgutman.com/blog-1/shakespeare-sonnet-73
1•gsf_emergency_6•1h ago•0 comments

Show HN: Django N+1 Queries Checker

https://github.com/richardhapb/django-check
1•richardhapb•1h ago•1 comments

Emacs-tramp-RPC: High-performance TRAMP back end using JSON-RPC instead of shell

https://github.com/ArthurHeymans/emacs-tramp-rpc
1•todsacerdoti•1h ago•0 comments

Protocol Validation with Affine MPST in Rust

https://hibanaworks.dev
1•o8vm•2h ago•1 comments

Female Asian Elephant Calf Born at the Smithsonian National Zoo

https://www.si.edu/newsdesk/releases/female-asian-elephant-calf-born-smithsonians-national-zoo-an...
5•gmays•2h ago•1 comments

Show HN: Zest – A hands-on simulator for Staff+ system design scenarios

https://staff-engineering-simulator-880284904082.us-west1.run.app/
1•chanip0114•2h ago•1 comments
Open in hackernews

Why Everybody Is Losing Money On AI

https://www.wheresyoured.at/why-everybody-is-losing-money-on-ai/
85•speckx•5mo ago

Comments

candiddevmike•5mo ago
This is a really well written article and contains references to back up the claims made. This part was mind blowing though:

> Cursor sends 100% of their revenue to Anthropic, who then takes that money and puts it into building out Claude Code, a competitor to Cursor. Cursor is Anthropic's largest customer. Cursor is deeply unprofitable, and was that way even before Anthropic chose to add "Service Tiers," jacking up the prices for enterprise apps like Cursor.

pimeys•5mo ago
Interesting to see what will happen if Cursor goes down...
xnx•5mo ago
People will move to one of the Cursor alternatives that are as good or better?
pimeys•5mo ago
But how will the market react is the bigger question...
boron1006•5mo ago
I don’t think it’s necessarily bad to be unprofitable but definitely weird to be sending 100% of your revenue to what is essentially your main competitor
dmonitor•5mo ago
It's even weirder from Anthropic's standpoint. Your #1 customer is buying all your product to resell it at loss.
joenot443•5mo ago
What are the odds that Microsoft acquires Cursor eventually, folding those users into a VS Code Premium of sorts?
rcarmo•5mo ago
Microsoft has an arguably better and more generally useful solution in GitHub Copilot.
ciconia•5mo ago
> total revenue: $4B > compute for training models: -$3B > compute for running models: -$2B > employee salaries: -$700M

Though not really representative of what users of said models may experience financially, at this point the question should be raised: if AI compute is 7x more expensive than developer salaries, what's the point? I thought the whole idea was to save money on human resources...

sejje•5mo ago
Someday (probably), a model will be trained once that is better than a human at coding, and it will only need trained once.

It can then be used indefinitely for the cost of inference, which is cheap and will continue getting cheaper.

dpritchett•5mo ago
This sounds a whole lot like Pascal's wager (or Roko's basilisk, if you prefer) for trillionaires.
eu•5mo ago
so the bubble will burst at some point…
FL33TW00D•5mo ago
"Please don't waste your breath saying "costs will come down." They haven't been, and they're not going to."

Yes, every new technology has always stayed exorbitantly priced in perpetuity.

warkdarrior•5mo ago
The first mobile phone, the Motorola DynaTAC 8000X, was launched in 1984 for $3,995 (more than $12k in 2025 dollars). So we should expect a 12x cost reduction in LLMs over 40 years.
floren•5mo ago
Adjusted for inflation the Model T cost about $25k. A new car doesn't cost $2k today. Are LLMs phones, or cars?
undefuser•5mo ago
IBM 3380 Model K introduced in 1987, has 7.5 GB of storage and costed about $160000 to $170000, or adjusted for inflation it is $455000 in 2025 US dollars, that's $60666/GB. A Solidigm D5-P5336 drive that can store 128 TB costs about $16500 in 2025 US dollars, that is $0.129/GB. That's a 470279x price reduction in slightly less than 40 years. So what is likely going to happen to LLM pricing? No one knows and both your example as well as mine doesn't mean anything.
leptons•5mo ago
Computers do get more powerful, but the price for a decent system has been about the same for a long time. $1500 got me a good system 10 years ago, but I'd still be paying $1500 for a good system today.
cwmma•5mo ago
He isn't saying they won't ever come down, he's saying they will not be coming down any time soon due to structural factors he discusses in the article.
cmrdporcupine•5mo ago
There has to be a name for the fallacy where people in our profession imagine that everything in technology follows Moore's law -- even when it doesn't.

We're standing here with a kind of survivorship bias because of all the technologies we use daily that did cost reduce and make it. Plenty did not. We just forget about them.

umbauk•5mo ago
I mean, we know everyone is losing money on AI. I thought from the title it was going to explain why. As in, why are they choosing to lose all that money?
candiddevmike•5mo ago
Irrational exuberance
tristor•5mo ago
> As in, why are they choosing to lose all that money?

Executive egos and market hype

hall0ween•5mo ago
The market hype is real. Check-signers at businesses expect LLMs to have the ability the AI CEOs talk about in their interviews and conferences but don’t exist (and are no where near existing).
GuB-42•5mo ago
The obvious reason is that they think it will pay off in the future. Google didn't start profitable, it is now one of the most profitable companies in the world.

Those who invest in money-losing AI believe that it will be the next Google and that profits will come.

Or alternatively, they hope to sell to the greater fool.

mewpmewp2•5mo ago
I'm not making any claims as to whether AI will become profitable or when, but if there's a new tech that has high potential or is highly desirable, I think it's expected that initially money will be lost.

Simply because strategically if there's high long term potential, it initially makes sense to put more money in than you get out of.

Not saying that AI is this, but if you determined that you have a golden goose that laid out 10 trillion USD worth of eggs when it got 10 years old, how much would you pay for it in the auction, and what would you have to show for it for the initial 9 years?

Now what if the golden goose scaled to 10 trillion each year linearly? First years people sound of mind would overpay for what it makes.

pphysch•5mo ago
The issue is we're moving past the "initially" phase and people are starting to suspect that the $10T golden goose is mythical. GPT-5??
mewpmewp2•5mo ago
I think it's more complex than that. You need to get really specific to calculate the potential value. It's entirely possible that there's 30 use-cases where it's very valuable, it's possible there's 80 use-cases where it's not valuable, and it's unclear how these use-cases are going to balance out in the future. To calculate whether all of this is over or undervalued would require analyzing and understanding all of those use cases and their impact very carefully. There's a lot of direct and indirect value, presumably no one is capable of currently calculating all of that anywhere near accuracy so people are making intuition based guesses on whatever data they can get hold of, but again - not so clear.

I personally think that there's many levels of innovation still to come in terms of robotics, APIs/Frameworks/Coding languages/Infra specifically for LLMs to provide easier and more foolproof ways to code and do things otherwise. I think it's far from played out. I personally think that a lot of potential is still untapped.

patall•5mo ago
Than why don't they raise prices? If AI developers were only worth 200k a year, nobody would pay X times those salaries and development would be cheaper. Similar, if none of the AI coding companies had free offerings, they would have less inference cost or more revenue. Yet they have the feeling that they need to offer those, likely because of competition. The article paints it as if the big companies are a factor of 2 away from profitability. Would absolutely nobody use AI if their tokens were double the price? I highly doubt that.
tim333•5mo ago
There is also the issue that the investors in present geese may find they don't get the golden egg laying one and it actually gets raised by some other goose farmer, like investing in Digg when Facebook would end up with the eggs.
therein•5mo ago
Meanwhile, I stopped using AI months ago. Have no subscriptions to any of the AI services anymore. Life goes on, quality of life is pretty good, haven't suffered in any way nor do I feel like I'm missing out.
mewpmewp2•5mo ago
Yeah, but I have also dreamed of living in the woods, being completely self sustainable blissfully. It doesn't mean there aren't capitalists out there looking to produce and sell more, and people out there looking to buy.
watwut•5mo ago
The difference is that choice to live out in the woods costs you. Choice to not have a phone costs you. A choice to not pay ai at this point ... does not cost you unless you live in special situation.
ffsm8•5mo ago
Not getting a phone didn't really cost you either for the first 5-10 yrs.

But the people that didn't definitely had a harder time adjusting when it got increasingly annoying to live without a smartphone.

It's ultimately a choice you can make, but it definitely also comes with consequences - especially if your dayjob is software - as this is an industry that loves to discriminate against people that aren't aboard the hype train and don't have "10 yrs of experience in d̵o̵c̵k̵e̵r̵ LLMs"

uncircle•5mo ago
You forget that the vast majority of software engineering in the real world is boring tech, not the new hotness that’s being peddled on Hacker News.

There is a large market for Java, C++ and COBOL engineers to this day, despite all startups on here are talking about React and Rust. There will still be a large need for actual engineers that use their meat brain and are not paid by line committed for the foreseeable future. Not everyone is writing junior-tier boilerplate that benefits from LLMs.

ffsm8•5mo ago
Ok I didn't forget. As a matter of fact, most LLMs are outright banned at my workplace.

However, the writing is on the wall and it's likely going to become one of the bullet points you'll be expected to have significant experience in when changing jobs. And I doubt that's gonna take 10 yrs

watwut•5mo ago
I really dont think analogy works at all. If you did not had phone phone first 15 years of their existence, there was very little to pay. Nowhere near anything close to "living in the woods".

And no one was forcing phones on you first 5-10 years of it existing. There was advertisement and competition like with any other product. You was certainly not getting free phones and there was no real top down push to have them like we see with ai.

qcnguy•5mo ago
Or unless you work for Coinbase, where people who refused to use AI got fired.

Think that's rare? Nope. It's coming everywhere. Most companies are at the stage of trying to monitor AI usage and encourage it, but eventually it'll stop being optional.

watwut•5mo ago
I doubt it will become as usual. It scream innefective management in the first place and is not defense of ai at all.

Company trying to get rid if innefective people would make sense, but if you measure it by ai usage all that happens is that your ai usage will go up - and price of using it along with it.

There are always irrational managers enjoying power trips. But the norms normally dont become that irrational

qcnguy•5mo ago
It's rational and will become standard.

Imagine you had a job doing ordinary database backed web app written in Java, and you found you had a coworker who wrote all their code in Notepad. They also refused to run linters or open code reviews, viewing it all as modern nonsense. Would you find that acceptable? Would you be surprised if a few months later that guy got let go for performance reasons? No.

Developers are given a lot degree of freedom, but in return are expected to use that freedom responsibly to deliver as much value as they can for the company. People refusing to use powerful tools had better have a watertight explanation for why. Mere negative vibes aren't good enough.

m_a_g•5mo ago
The cost can be significantly reduced immediately and drastically if OpenAI or Anthropic were to choose to do so.

By simply stopping the training of new models, profitability can be achieved on the same day.

With the existing models, we have already substantial use cases, and there are numerous unexplored improvements beyond the LLM, tailored specifically to the use case.

crooked-v•5mo ago
Ah, but see, those existing uses cases allow for merely finite profit, instead of the infinitely growing profit that late stage capitalism demands.
antiloper•5mo ago
Current frontier models are not good enough because they still suffer from major hallucinations, sycophancy, and context drift. So there has to be at least (and I have no reason to believe it will be the last, GPT-5 demonstrates that the transformer architectures are hitting diminishing returns) one more training cycle.
ten_hands•5mo ago
This only works if all the AI companies collude to stop training at the same time, since the company that trains the last model will have a massive market advantage. That not only seems extremely unlikely but is almost certainly illegal.
palata•5mo ago
> By simply stopping the training of new models, profitability can be achieved on the same day.

But then they stop being up-to-date with... the world, right?

km3r•5mo ago
> At this point, it's becoming obvious that it is not profitable to provide model inference, despite Sam Altman recently saying that OpenAI was.

Except the authors own provided data says it cost them $2B in inference costs to generate $4B in revenue. Yes training costs push it negative, but this is like tech growth 101, debt now to grow faster leads to larger potential upsides in the future.

ASinclair•5mo ago
Training costs keep exploding and several companies are providing frontier models. They'll have to continue shoveling tons of money into training just to stay in place with respect to the competition. So you can't just ignore training costs.
orbital-decay•5mo ago
DeepSeek alone demonstrated a massive reduction in training costs, and there's a ton of low-hanging fruits nobody even started to use.
palata•5mo ago
> Yes training costs push it negative

But training will have to stay forever, right? Otherwise the LLM will be stuck with outdated information...

qcnguy•5mo ago
It's not just training. Look at all the other costs in The Information's graph. Salaries, general admin, data licensing - all huge costs.
golergka•5mo ago
> OpenAI spent 50% of its revenue on inference compute costs alone

This means that they operate existing models with very healthy 50% profit margin, that’s excellent unit economics actually. Losing money by investing more into R&D that you make is not the same as burning it by selling a dollar for 90 cents.

hall0ween•5mo ago
I’m confused. If 50% of the revenue goes to inference, that means the other 50% goes into research?
toddmorey•5mo ago
My understanding is it means half of what a subscriber pays is spent just on the compute required as you chat with the models. Which leaves the other have to be divided up among salaries, marketing, R&D, etc.
habinero•5mo ago
Revenue minus compute is not profit lol. You still have to pay salaries and rent on your buildings, which are usually the biggest expenses.
toddmorey•5mo ago
50% margins would be actually low and concerning for a saas business. What makes software an attractive business is how well it scales. The standard for saas has been at least 80% or higher margins.

Most all saas accounts require lengthy and generous free trials and boy AI compute throws a bit of a hand grenade into that PLG strategy. It’s just new economics that the industry will have to figure out.

davidcbc•5mo ago
You're ignoring the fact that there are more expenses than just inference. Salaries, infrastructure, training new models (a requirement for the model to stay up to date with the changing world)

You can't just eliminate all the costs except inference

soneca•5mo ago
I think the most interesting part is that, in AI, software does not have zero marginal cost anymore. You can’t build once and scale to billions just investing in infrastructure.

Still, companies like OpenAI and Twitter are doing just that. Thus losing money.

Will AI evolve to be again as regular software or will the business model of tech AI become closer to what traditional non-tech companies are?

How the WalMart of AI will look like?

Does SaaS with very high prices and very thin margins even work as a scalable business model?

gdbsjjdn•5mo ago
I think Ed hits on an interesting point about the new user who spends $4 on a TODO file. Current LLM users are very enthusiastic about finding different models for different use cases and evaluating the cost-benefit of those models. But the average end user doesn't give a shit. If LLMs are going to "eat the world" they need to either be a lot better in the median case (bad prompts, bad model selection) or they need to be so cost-effective that you can farm out your query to an ensemble and choose the result dialogue-tree-style.
exe34•5mo ago
> If LLMs are going to "eat the world" they need to either be a lot better in the median case (bad prompts, bad model selection) or they need to be so cost-effective that you can farm out your query to an ensemble and choose the result dialogue-tree-style.

LLMs have been around for two years. it took decades before the PC really took hold.

gdbsjjdn•5mo ago
Segways existed for a long time and they never took off. Zeppelins too. Not every technology automatically gets good just because time passes.
BobbyJo•5mo ago
1) Chat GPT is nearly 3 years old, and LLMs were around before that.

2) Yes, they still have some time to fit the market better, but that doesn't change what they'll need to do to fit the market better.

wood_spirit•5mo ago
The IBM PC was an overnight success. It was less than a decade after the first “PCs” and it was the hockey stick moment. I remember x86 clones being seemingly everywhere in just a year or two
exe34•5mo ago
Pretty much everybody I know is using an LLM for something. some are even using it for things they shouldn't be using it for.
palata•5mo ago
> LLMs have been around for two years. it took decades before the PC really took hold.

But virtually everybody has been using LLMs already. How long would it have taken for the PC if everybody had had the opportunity to use one for more than a year?

exe34•5mo ago
so everybody is using it and it's not good enough for everybody to use yet? why is everybody using it?
palata•5mo ago
I guess "everybody is using it" is not the same as "everybody depends on it" or "it would be economically viable for everybody to pay for what it costs"?

Whereas the PC was clearly economically viable.

exe34•5mo ago
It was clearly so in hindsight. at the time it was a toy for affluent people. it didn't really start affecting people's everyday lives until the 90s.
simonw•5mo ago
> Please don't waste your breath saying "costs will come down." They haven't been, and they're not going to.

Cost to run a million tokens through GPT-3 Da-Vinci in 2022: $60

Cost to run a million tokens through GPT-5 today: $1.25

cwmma•5mo ago
yes but due to reasoning models the same query uses VASTLY more tokens today then a couple years ago
simonw•5mo ago
Sure, if you enable reasoning for your prompt. A lot of prompts don't need that.
orbital-decay•5mo ago
In most use cases the main cost is always input, not output. Agentic workflows, on the other hand, do eat up a ton of tokens on multiple calls. Which can usually be optimized but nobody cares.
simonw•5mo ago
With multiple calls an important factor to consider is token caching, where repeat inputs are discounted.

This is particularly important if you constantly replay the previous conversation and grow it with each subsequent prompt.

GPT-5 offers a 90% discount on these cached tokens! That's a huge saving for this kind of pattern.

jjfoooo4•5mo ago
I think what we'll eventually see is frontier models getting priced dramatically more expensive (or rate limited), and more people getting pickier about what they send to frontier models vs cheaper, less powerful ones. This is already happening to some extent, with Opus being opt-in and much more restricted than Sonnet within Claude Code.

An unknown to me: are the less powerful models cheaper to serve, proportional to how much less capable they are than frontier models? One possible explanation for why e.g. OpenAI was eager to retire GPT 4 is that those older models are still money losers.

simonw•5mo ago
Everything I've seen makes me suspect that models have continually got more efficient to serve.

The strongest evidence is that the models I can run on my own laptop got massively better over the last three years, despite me keeping the same M2 64GB machine without upgrading it.

Compare original LLaMA from 2023 to gpt-oss-20b from this year - same hardware, huge difference.

The next clue is the continuing drop in API prices - at least prior to the reasoning rush of the last few months.

One more clue: o3. OpenAI's o3 had a 80% price drop a few months ago which I believe was due to them finding further efficiencies in serving that model at the same quality.

My hunch is that there are still efficiencies to be wrung out here. I think we'll be able to tell if that's not holding if API prices stop falling over time.

realz•5mo ago
If developers and enterprises can host their own OSS/fine-tuned models, why will they pay Anthropic or OpenAI?
simonw•5mo ago
Because hosting a really GOOD model requires hardware that costs tens of thousands of dollars.

It's much cheaper to timeshare that hardware with other users than to buy and run it yourself.

realz•5mo ago
That may be true for independent devs or startups. Larger companies have enough demand to justify a few A/H100s.
jjfoooo4•5mo ago
Why do you think OpenAI wanted to get rid of GPT-4 etc so aggressively?

I suppose there's a distinction between new less capable models, I can see why those would be more efficient. But maybe the older frontier models are less efficient to serve?

simonw•5mo ago
Definitely less efficient to serve. They used to charge $60/million input tokens for GPT-3 Da Vinci. They charge $1.25/million for GPT-5.

Plus I believe they have to keep each model in GPU memory to serve it, which means that any GPU serving an older model is unavailable to serve the newer ones.

Eddy_Viscosity2•5mo ago
Are these the costs (what the supplier pays) or the prices (what the consumer pays)?
simonw•5mo ago
The price OpenAI charge users of their API.
Lariscus•5mo ago
The price of a token doesn't necessarily reflect the true cost of running a model. After Claude Opus 4 released the price of OpenAIs o3 tokens where slashed practically over night.[0] If you think this happened because inference cost went down, I have a bridge to sell to you.

[0] https://venturebeat.com/ai/openai-announces-80-price-drop-fo...

simonw•5mo ago
Sell me that bridge then, because I believe OpenAI's staff who say it was because inference costs went down: https://twitter.com/TheRealAdamG/status/193244032829380632

Generally I'm skeptical of the idea that any of the major providers are selling inference at a loss. Obviously they're losing money when you include the cost of research and training, but every indication I've seen is that they're not keen to sell $1 for 80 cents.

If you want a hint at the real costs of inference look to the companies that sell access to hosted open source models. They don't have any research costs to cover so their priority is to serve as inexpensively as possible while still turning a profit.

Or take a good open weight model and price out what it would cost to serve at scale. Here's someone who tried that recently: https://martinalderson.com/posts/are-openai-and-anthropic-re...

rcarmo•5mo ago
s/Cost/Price/g
smeeth•5mo ago
Another day, another person not getting discounted cash flow.

Models trained in 2025 don’t ship until 2026/7. That means the $3bn in 2025 training costs show up as expense now, while the revenue comes later. Treating that as a straight loss is just confused.

OAI’s projected $5bn 2025 loss is mostly training spend. If you don’t separate that out with future revenues, you’re misreading the business.

And yes, inference gross margins are positive. No idea why the author pretends they aren’t.

bionhoward•5mo ago
Is this really surprising given how VC funded capitalism works? Spend money to build amazing technology and gain market share, then eventually flip into extraction mode.

Yes, a pullback will kill some weaker companies, but not the ones with sufficient true fans. Plus, we’re talking about a wide-ranging technological revolution with unknown long term limits and economics, you don’t just give up because you’re afraid to spend some money.

I don’t want to pay Anthropic, because I don’t trust them, but I will absolutely pay cursor, because I trust them, and I doubt I’m alone. My cursor usage goes to GPT-5, too, so it’s definitely not 100% Anthropic, even if I’m the only idiot using GPT5 on Cursor

It’s fun to innovate. Making money is a happy byproduct of value creation. Isn’t the price of success always paid in advance, anyway? Why would winning AI tech companies pack it up and stop crushing it over the long term just because they’re afraid to lose someone else’s money in the short term? Wouldn’t capitulation guarantee losses moreso than continued effort?

panosv•5mo ago
What about Google? Anyone has any insights on their unit economics since they own the models and the infrastructure (which is also custom TPUs)? Are they doing better or are they in the same money losing business?
seanalltogether•5mo ago
It feels like Google should be able to come up with a revenue figure for search ai results right? How many people do a search but don't click on any links because they just read the ai blurb, but advertisers are still charged for being visible on the page.
tim333•5mo ago
It must be hard for them to figure how much of their revenue is down to AI and how much to other stuff like search. They certainly make a lot of revenue and it would be foolish for them to ignore AI and have OpenAI and Perplexity eat their lunch.
vb-8448•5mo ago
Nice read, but I'd add an objection here: even if models don't improve any more, and they raise the standard subscription to 100$/month, I'd still buy it (and a lot of other people, I guess) because I'd extract far more value from it.
patall•5mo ago
That's also what I do not get. The companies are unprofitable because of competition, not because what they do cannot be profitable.
leptons•5mo ago
If it costs more to produce a result than a customer is willing to pay, then the company will either be unprofitable (sell at a loss) or just close up shop. The cost for running LLMs is much higher than what customers are likely to want to pay, and that has nothing to do with competition from other LLM companies, it's a result of high cost of cutting-edge hardware, infrastructure, and the massive amount of electricity it consumes - so much electricity that tech companies are now building power plants to power them (which are very expensive to build). It's a massive cost, and all for the hope that people will continue to accept AI slop.
vb-8448•5mo ago
According to the data in the post, the cost of running the model for open AI in 2024 was 2B and if we strip out all training/research costs they had a loss of about 1B, peanuts. They can raise a little bit the standard subscription prices and turn profitable.

The bulk of the cost was model training and research(~4B). They are forced to train new models and improve existing one because of the market and competition.

davidcbc•5mo ago
> if we strip out all training/research costs

You can't just strip out those costs. You have to train new models or the information in the model will be out of date.

vb-8448•5mo ago
It depends, if you have to retrain existing models using new data(basically updating the cutoff date) it will cost you much less than today ... and I'm pretty sure today they are aiming for speed and "intelligence improvement" and not training efficiency.

The only scenario where the training cost won't decrease is in case the limit of the scaling law is not yet reached, or they discover new approaches. But from what we have seen this year, it doesn't seem to be the case.

PS: Plus, we are still not speaking about the elephant in the room: ads. Today's revenues are basically from API usage and subscription, but we all know that at some point ads will come in, and the revenues will increase.

BobbyJo•5mo ago
Does that get them to the TAM they need to justify current valuations though? I'd guess not.
vb-8448•5mo ago
Obviously not, but my point is that they aren't losing money because it's intrinsically non-profitable on mass scale, but because of the market in this specific period.
bbreier•5mo ago
Does controversy cause articles to slide on HN? I noticed that this had more points in less time than several articles ranked above it, which surprises me a bit

e.g. at time of writing a post about MentraOS has 11 points in 1 hour compared to this article's 51 in 53 minutes, but this is ranked 58th to Mentra's 6

bbreier•5mo ago
It has dropped to 100th while the MentraOS post remains at 5th. Is HN pushing negative PR for AI down the ranks?
wood_spirit•5mo ago
Some people must be flagging it

Dang, can we chat about collaborative filtering bubbles please?

uncircle•5mo ago
Yes. Too large a comment/vote ratio causes articles to be defrontpaged, as well as some domains. Moderators do a lot of manipulation behind the scenes to keep what they deem ‘hot-button topics’ down in 3rd page or shadow-banned altogether. Then there’s the user flagging system that is abused whenever some popular article goes against the grain.

That’s why I use https://news.ycombinator.com/active to find the interesting topics instead of having to rely on the strict moderation algorithm

deeviant•5mo ago
I don't understand these posts. Do people not understand how venture capital works?

The majority of these companies know they are burning money, but more than that knew they would be losing money at this point and beyond. That is the play, the thesis is: AI will dominate nearly everything in the near future, the play is to own a piece of that. Investors are willing to risk their investment for a chance of getting a piece of the pie.

Posts that flail around yelling companies 'losing money', without addressing the central premise are just wasting words.

In short, do you think AI is not going to dominate nearly everything? Great, talk about that. If you do believe is, then talk about something other than the completely reasonable and expected state of investors and companies fighting for a piece of the pie.

As a somewhat related tangent, people seem to not understand the likely cost trajectory of model training/inference costs:

* Models will reach a 'good enough' point where further training will be mostly focused on adding recent data. (For specific market segments, not saying that we'll have a universal model anytime soon, but we'll soon have one that is 'good enough' at c++, might already be there).

* Model architecture and infrastructure will improve and adapt. I work for a company that was among the first use deep learning to control real-time kinetic processes in production scenarios, our first production hardware was a nvidia Jetson, we had a 200ms time budget for inference, and our first model took over 2000! We released our product, running under 200ms, *using the same hardware* the only difference was improvements in the cuDNN library and some other drive updates and some domain specific improves on our YOLO implementation. Long story short, yes inference costs are huge, but they are also massively disruptable.

* Hardware will adapt. Nvidia cash machine will continue, right now nvidia hardware is optimized for balance between training and inference, where TPUs, the newer ones are more tilted towards inference. I would be surprized if other hardware companies don't force Nvidia to give the more inference based solution and 2-3x cost savings at time point in the next 5 years. And for all I know, perhaps a hardware startup will disrupt Nvidia, it would be one of the most lucrative hardware plays on the planet.

Focusing inference cost is a deadend to understanding the trajectory of AI, understanding the *capability* of AI is the answer to understanding it's place in the future.

serjester•5mo ago
The big labs have 50+% margins on serving the models, the training is where they lose money. But every new model boosts OpenAI's revenue growth which is unheard of at their size (300+% YoY). Therefore it's completely reasonable to keep doubling down and making bigger bets.

Most people miss that they have almost a billion free users that are waiting to be monetized. Google makes 400B a year and it's crazy to think OpenAI can't achieve some percentage of that. Why would you slow down and let Google catch up for the sake of short term profitability.

jjfoooo4•5mo ago
The article claims otherwise:

> In fact, even if you remove the cost of training models from OpenAI's 2024 revenues (provided by The Information), OpenAI would still have lost $2.2 billion fucking dollars.

Incipient•5mo ago
You also need to remove the research cost, and probably a bit of the people cost.

The issue however is, can an AI company actually go "yep. We're done this is a good as it gets!"?

I don't believe they can do that, so removing training cost is kind of a moot point.

Jimmc414•5mo ago
Cursor burning cash to subsidize Anthropic's losses to subsidize Amazon's compute investments is their problem, not mine.

The people writing all of these "AI is unprofitable" pieces are doing financial journalism similar to analyzing the dot-com bubble by looking at pets.com's burn rate. The infra overspend was real as well as the bankruptcies, but it existentially foolish for a business to ignore the behavioral shift that was taking place.

I have to constant remind myself to stop arguing and evangelizing about AI. There is a growing crowd who insists that AI is a waste of money and that AI cannot do things I'm already doing on a daily basis.

Every minute spent explaining to AI skeptics is a minute not spent actually capitalizing on the asymmetry. They don't want to hear it anyway and I have little incentive to convince anyone otherwise.

The companies bleeding money to serve AI below cost prices won't last, but thats all more the reason use them now while they're cheap.

uludag•5mo ago
I think the main fear is that these products will become so enshitified and engrained into everywhere that, looking back, we'll be wishing we didn't depend so much on the technology. For example, the Overton window around social media has shifted so much to the point that it's pretty normal to hear views that social media is a net negative to society and we'd be better off without it.

Obviously the goal of these companies is to generate as much profit as possible as soon as possible. They will turn the tables eventually. The asymmetry will go in the opposite direction, maybe to the extend that one takes advantage of the current asymmetry.

Jimmc414•5mo ago
I don't disagree with anything you've said.
watwut•5mo ago
But back then, you was better off not depending on pet.com. If products of one of those vaporware companies became important in your process, your company went down along with.

Those were companies that had multiple expensive IT restructurings one after another, each making them more innefective and then either run out of cash or barely made it.

It worked well for companies that were choosing smart.

username135•5mo ago
Amazon was unprofitable for years (like over a decade), famously. I don't see any difference with AI companies.

There is clearly some kind of market for this technology. It will eventually be profitable either through some technology breakthrough that allows creating/processing tokens cheaper or by finding a cost structure consumers can live with.

The cat is already out the bag. This technology isn't going away.

Ekaros•5mo ago
To succeed like Amazon you actually have to have cash-flow in first place. The comparison only works when company can stop investing and immediately make profit. And I do not think any of the AI companies are at that point. Or could continue to run for any length of time if they choose to.
tuatoru•5mo ago
Of course they are losing money. If they are not losing money, they are not investing fast enough.

Standard market share logic.