frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

I replaced the front page with AI slop and honestly it's an improvement

https://slop-news.pages.dev/slop-news
1•keepamovin•3m ago•1 comments

Economists vs. Technologists on AI

https://ideasindevelopment.substack.com/p/economists-vs-technologists-on-ai
1•econlmics•5m ago•0 comments

Life at the Edge

https://asadk.com/p/edge
1•tosh•11m ago•0 comments

RISC-V Vector Primer

https://github.com/simplex-micro/riscv-vector-primer/blob/main/index.md
2•oxxoxoxooo•15m ago•1 comments

Show HN: Invoxo – Invoicing with automatic EU VAT for cross-border services

2•InvoxoEU•15m ago•0 comments

A Tale of Two Standards, POSIX and Win32 (2005)

https://www.samba.org/samba/news/articles/low_point/tale_two_stds_os2.html
2•goranmoomin•19m ago•0 comments

Ask HN: Is the Downfall of SaaS Started?

3•throwaw12•20m ago•0 comments

Flirt: The Native Backend

https://blog.buenzli.dev/flirt-native-backend/
2•senekor•22m ago•0 comments

OpenAI's Latest Platform Targets Enterprise Customers

https://aibusiness.com/agentic-ai/openai-s-latest-platform-targets-enterprise-customers
1•myk-e•24m ago•0 comments

Goldman Sachs taps Anthropic's Claude to automate accounting, compliance roles

https://www.cnbc.com/2026/02/06/anthropic-goldman-sachs-ai-model-accounting.html
2•myk-e•27m ago•4 comments

Ai.com bought by Crypto.com founder for $70M in biggest-ever website name deal

https://www.ft.com/content/83488628-8dfd-4060-a7b0-71b1bb012785
1•1vuio0pswjnm7•28m ago•1 comments

Big Tech's AI Push Is Costing More Than the Moon Landing

https://www.wsj.com/tech/ai/ai-spending-tech-companies-compared-02b90046
3•1vuio0pswjnm7•30m ago•0 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
2•1vuio0pswjnm7•31m ago•0 comments

Suno, AI Music, and the Bad Future [video]

https://www.youtube.com/watch?v=U8dcFhF0Dlk
1•askl•33m ago•2 comments

Ask HN: How are researchers using AlphaFold in 2026?

1•jocho12•36m ago•0 comments

Running the "Reflections on Trusting Trust" Compiler

https://spawn-queue.acm.org/doi/10.1145/3786614
1•devooops•41m ago•0 comments

Watermark API – $0.01/image, 10x cheaper than Cloudinary

https://api-production-caa8.up.railway.app/docs
1•lembergs•43m ago•1 comments

Now send your marketing campaigns directly from ChatGPT

https://www.mail-o-mail.com/
1•avallark•46m ago•1 comments

Queueing Theory v2: DORA metrics, queue-of-queues, chi-alpha-beta-sigma notation

https://github.com/joelparkerhenderson/queueing-theory
1•jph•58m ago•0 comments

Show HN: Hibana – choreography-first protocol safety for Rust

https://hibanaworks.dev/
5•o8vm•1h ago•1 comments

Haniri: A live autonomous world where AI agents survive or collapse

https://www.haniri.com
1•donangrey•1h ago•1 comments

GPT-5.3-Codex System Card [pdf]

https://cdn.openai.com/pdf/23eca107-a9b1-4d2c-b156-7deb4fbc697c/GPT-5-3-Codex-System-Card-02.pdf
1•tosh•1h ago•0 comments

Atlas: Manage your database schema as code

https://github.com/ariga/atlas
1•quectophoton•1h ago•0 comments

Geist Pixel

https://vercel.com/blog/introducing-geist-pixel
2•helloplanets•1h ago•0 comments

Show HN: MCP to get latest dependency package and tool versions

https://github.com/MShekow/package-version-check-mcp
1•mshekow•1h ago•0 comments

The better you get at something, the harder it becomes to do

https://seekingtrust.substack.com/p/improving-at-writing-made-me-almost
2•FinnLobsien•1h ago•0 comments

Show HN: WP Float – Archive WordPress blogs to free static hosting

https://wpfloat.netlify.app/
1•zizoulegrande•1h ago•0 comments

Show HN: I Hacked My Family's Meal Planning with an App

https://mealjar.app
1•melvinzammit•1h ago•0 comments

Sony BMG copy protection rootkit scandal

https://en.wikipedia.org/wiki/Sony_BMG_copy_protection_rootkit_scandal
2•basilikum•1h ago•0 comments

The Future of Systems

https://novlabs.ai/mission/
2•tekbog•1h ago•1 comments
Open in hackernews

How much Anthropic and Cursor spend on Amazon Web Services

https://www.wheresyoured.at/costs/
164•isoprophlex•3mo ago

Comments

isoprophlex•3mo ago
Ed Zitron's style isn't for everybody, I understand that. But if these numbers, and the direction they're going in, are correct... to me this points to a significant AI bubble deflation coming soon. It just isn't sustainable, it seems.
swyx•3mo ago
if you are worried about AWS spend, i have news for you about non AWS spend.

"coming soon" is also really over simplistic. you would have missed some of the greatest tech companies in the past 20 years if you evaluated startups based on their early-year revenue vs infra spend

like sure i have a dog in this fight but i actually want the criticism to sharpen my thinking, unfortunately yours does not meet that bar.

isoprophlex•3mo ago
are you glossing over the whole "aws bill isnt structural infra spend" thing?!

they spend 104% of revenue on ONE cloud provider and costs scale linearly with revenue growth. assume zitron didn't pull these numbers out of his ass.

educate me how this isnt selling $20 bills for $5. you're a smart dude; i myself aint seeing the "sustainable business practices" here

swyx•3mo ago
the entire point is that stressing sustainable business practices here at this point in a startup's life is extremely short sighted and is the kind of shitty analysis that gives HN a bad rep.

pull up Uber's financials leading up to IPO. unsustainable and everyone knew it. they work it out after because they burned money and eventually achieved a sustainable moat. this is why venture exists. HN doesnt like venture which is, well, ironic given the domain we're on.

a better negative argument i'd rather see looks like this - "ive run these aws numbers against the typical spend path of pre IPO startups who then later improved their cost baseline and margin profile and even after accounting for all that, Anthropic ngmi". thats the kind of minimum sophistication you need here to play armchair financial analyst. ed zitron, and everyone involved in this entire thread, incl myself, have not done that because we are lazy and ignorant and dont actually care enough about seeking the truth here. we are as unprepared to analyze this AWS spend as we are to understand their 1b -> 10b revenue ramp in 2025. you havent done the work and yet you sit here and judge it unsustainable based off some shitty "leaks". dont pretend that ed's analysis is at all meaningful particularly because he conveniently stops where it supports his known negative bias.

watwut•3mo ago
Afaik Uber numbers were significantly smaller.
jsnell•3mo ago
There is nothing particularly concerning about the raw numbers, and I'd assume they are correct. (And they are also interesting, since we rarely get data at this granularity. Thanks to Zitron for publishing the leak with such completeness, despite the personal risk to his source.)

But Zitron's analysis is as incorrect or irrelevant as ever. The problem is just that as usual there is so much tripe that it's impossible to rebut it all.

The way I'd think about this leaked data is that it's giving us a reasonable upper bound on the fully loaded R&D compute cost for Claude Sonnet 4 - a product that will be making $7 billion in revenue with very good unit economics. Other than that, this just tells us that Anthropic, just like the other frontier AI lab, will be spending as much on training as they can.

daft_pink•3mo ago
I think this is a minor speed bump and VC’s believe that cost of inference will decrease over time and this is a gold rush to grab market share while cost of inference declines.

I don’t think they got it right and the market share and usage grew faster than inference dropped, but inference costs will clearly drop and these companies will eventually be very profitable.

Reality is that startups like this assume moore’s law will drop the cost over time and arrange their business around where they expect costs to be and not where costs currently are.

Frieren•3mo ago
> I think this is a minor speed bump and VC’s believe that cost of inference will decrease over time and this is a gold rush to grab market share while cost of inference declines.

It could also be that you give too much credit to the market. People follow trends because in most cases that makes money. There is no other deeper though involved. Look at the financial crisis, totally irrational.

rglover•3mo ago
This. Post-crypto, AI was the obvious next gambit for VC. Their money flows to hype, not product. The second that hype starts to fade and the money dries up, VCs will be running with their Harold Hill trunk full of cash toward the border. Just from the content they publish alone you can tell they're channeling their inner Barnum & Bailey in between Ayahuasca seizures.
xnx•3mo ago
> inference costs will clearly drop and these companies will eventually be very profitable.

Inference costs for old models will drop, but inference costs may stay the same if models continue to improve.

No guarantee that any wrapper for inference will be able to hold on to customers when they stop selling $1.00 for $0.50.

onlyrealcuzzo•3mo ago
Isn't the consensus that the MOE architecture and other optimizations in the newest gen models (GPT-5, Gemini 3.0 to come, etc) will reduce inference costs by 50-75% already?
yomismoaqui•3mo ago
Sounds interesting, do you have some links with more info about this?

Thanks!

ACCount37•3mo ago
Kind of. Frontier LLMs aren't going to get cheaper, but that's because the frontier keeps advancing.

Price-performance though? The trend is clear: a given level of LLM capability keeps getting cheaper, and that trend is expected to hold. Improvements in architecture and training make LLMs more capability-dense, and advanced techniques make inference cheaper.

onlyrealcuzzo•3mo ago
> Frontier LLMs aren't going to get cheaper

One of the main selling points to MOE, is that the architecture is designed such that you can re-train experts independently, as well as add new experts, change the size of an experts parameters, etc, without retraining the entire model.

If 80% of you usage comes from 20% of your experts, you can cut your future training costs SUBSTANTIALLY.

ACCount37•3mo ago
I don't think anyone has ever managed to pull off a success story with this kind of advanced expert manipulation.

It's not entirely impossible, but I remain skeptical until see a proof that it, first, works. And, second, that it actually has an advantage over "we'll just train another base model from scratch, but 10% larger, with those +5% performance architecture tweaks, and a new modality blender, and more of that good highly curated data in the dataset, and fresher data overall, and it'll be glorious".

x0x0•3mo ago
> inference costs will clearly drop

They haven't though. On two fronts: 1, the soa models have been pretty constantly priced, and everyone wants the soa models. Likely the only way costs drop is the models get so good that people are like hey, I'm fine with a less useful answer (which is still good enough) and that seems, right now, like a bad bet.

and 2 - we use a lot more tokens now. No more pasting Q&A into a site; now people hammer up chunks of their codebases and would love to push more. More context, more thinking, more everything.

ctoth•3mo ago
You're describing increased spending while calling it increased cost. These aren't the same thing. A task that cost me $5 to accomplish with GPT-4 last year might cost $1 with Sonnet today, even though I'm now spending $100/month total on AI instead of $20 because I'm doing 100x more tasks. The cost per task dropped 80%. My spending went up 5x. Both statements are true.

Here's an analogy you may understand:

https://crespo.business/posts/cost-of-inference/

KallDrexx•3mo ago
Fwiw that's not necessarily true, because if Sonnet ends up using reasoning, then you are using more tokens than GPT-4 would have used for the same task. Same with GPT-5 since it will decide (using an LLM) if it should use the thinking model for it (and you don't have as much control over it).
steveklabnik•3mo ago
This is addressed in the post.
x0x0•3mo ago
Right, but if I understand you, the counterargument is dumb since the context in which we are discussing is business viability (vcs investing in businesses where the unit economics require inference cost decreases), so actual dollars out rather than some imaginary cost per token is the metric that matters.

Inference is getting so much cheaper that cursor and zed have had to raise prices.

steveklabnik•3mo ago
> so actual dollars out rather than some imaginary cost per token is the metric that matters.

Even if we take this as true, the point is that this is different than "the cost of inference isn't going down." It is going down, it's just that people want more performance, and are willing to pay for it. Spend going up is not the same as cost going up.

I don't disagree that there are a wide variety of things to talk about here, but that means it's extra important to get what you're talking about straight.

x0x0•3mo ago
Playing word games labeling inference narrowly as the cost per token rather than the per-X $ going to your llm api provider per customer/user/use/whatever is kinda silly?

The cost of inference -- ie $ that go to your llm api provider -- has increased and certainly appears to continue to increase.

see also https://ethanding.substack.com/p/ai-subscriptions-get-short-...

steveklabnik•3mo ago
> The cost of inference -- ie $ that go to your llm api provider

This is the crux of it: when talking about "the cost of inference" for the purposes of the unit economics of the business, what's being discussed is not what they charge you. It's about their COGs.

That's not word games. It's about being clear about what's being talked about.

Talking about increased prices is something that could be talked about! But it's a different thing. For example, what you're talking about here is total spend, not about individual pricing going up or down. That's also a third thing!

You can't come to agreement unless you agree on what's being discussed.

dcre•3mo ago
Why do the unit economics require a decrease in inference spend per user? This is discussed at the end of the post. I think this is based on the very strange assumption that these businesses must charge $20 a month no matter how much inference their customers want to use. This is precisely what the move to usage-based pricing was about. End users want to use more inference because they like it so much, and are knocking down these companies’ doors demanding to be allowed to pay them more money to get more inference.
infecto•3mo ago
Anecdote of 1. Costs for openai on a per token basis have absolutely dropped and that accounts for new sota models over time. I think by now we can all agree that inference costs from providers are largely at or above breakeven. So more tokens is a good problem to have.
dexwiz•3mo ago
Point 2 was the analysis I saw. Context size and token cost grow inversely at a rate that keeps prices constant, almost like supply and demand curves.
Analemma_•3mo ago
My own usage and the usage of pretty much everyone I know says that as inference costs drop, usage goes up in lockstep, and I’m still nowhere near the ceiling of how many tokens I could use if they were free.

I think if these companies are gambling their future on COGS going down, that’s a gamble they’re going to lose.

delaminator•3mo ago
aka Jevons Paradox.

we strive to make machines more efficient to reduce energy usage. That makes $ per unit of work lower. But instead of reducing spending, people find more uses for the machines.

username223•3mo ago
Color me skeptical. We're running into the speed of light when it comes to transistor size, and the parallelism that made neural nets take off is running into power demands. Where do the exponential hardware gains come from? Optimizing the software by 2x or 4x happens only once. Then there's the other side: if Moore's Law works too well, local models will be good enough for most tasks, and these companies won't be able to do the SaaS thing.

It seems to me like models' capability scales logarithmically with size and wattage, making them the rare piece of software that can counteract Moore's Law. That doesn't seem like a way to make a trillion dollars.

throwaway290•3mo ago
One improvement is from scraping and stealing better quality IP to train on. And they can just ride Moore's law until they profit then lobby governments to require licenses for fast GPUs because of national security.
floatrock•3mo ago
The buried lede:

Anthropic: "$2.66 billion on compute on an estimated $2.55 billion in revenue"

Cursor: "bills more than doubled from $6.2 million in May 2025 to $12.6 million in June 2025"

Clickthrough if you want the analysis and caveats

neuronexmachina•3mo ago
For context, AWS's total revenue for 2024 was $107.6B: https://ir.aboutamazon.com/news-release/news-release-details...
Kuinox•3mo ago
I could find an ARR for Cursor of $500M. Why do they they in this article that Cursor is loosing with this spending number ?
LetsGetTechnicl•3mo ago
Ed's mentioned ARR in previous articles and it's not a "generally accepted accounting principle". They cherry pick the highest monthly revenue number and multiply that by 12, but that's not their actual annual revenue.
dcre•3mo ago
"Cherry pick the highest" is misleading. If your revenue is growing 10% a month for a year straight and is not seasonal, picking any other than the most recent month to annualize would make no sense.
mossTechnician•3mo ago
If a company's revenue in January is $100 and it grows by 10% every month, the December revenue is $285. The year's revenue would be about $2,138, but ARR in December would be $3,423. That's 1.6x the actual revenue.

ARR could be a useful tool to help predict future revenue, but why not simply report on actual revenue and suggest it might increase in the next year? I have found the most articles to be unclear to the reader about what ARR actually represents.

dcre•3mo ago
Why is the calendar year the relevant unit? If you insist on years, then if you consider the year from June to June, $2,138 would be misleading small.

The point of ARR is to give an up to date measure on a rapidly changing number. If you only report projected calendar year revenue, then on January 1 you switch from reporting 2025 annual revenue to 2026 projected revenue, a huge and confusing jump. Why not just report ARR every month? It's basically just a way of reporting monthly revenue — take the number you get and divide it by 12.

I am really skeptical that people are being bamboozled by this in some significant way. Zitron does far more confusing things with numbers in the name of critique.

LetsGetTechnicl•3mo ago
Because that's a part of the generally accepted accounting principles: https://www.rightrev.com/gaap-revenue-vs-arr/

Nobody considers a year from June to June because that would be misleading.

dcre•3mo ago
That is an article explaining why ARR is useful and important despite not being the same thing as GAAP revenue.
mossTechnician•3mo ago
You're correct, ARRs can be both misleading and for any 12-month period (I just chose a year to simplify), but the problem is AI companies tend to only release their latest ARR, and only selectively, which I believe is misleading in the opposite direction.
OptionOfT•3mo ago
How can you talk about ARR if you only look at 1 year?
datadrivenangel•3mo ago
It's useful for financial planning. Less useful for overall financial reporting given how volatile it is.
OptionOfT•3mo ago
I should've been more clear. How can you talk about ARR if you only have 1 year?

How do you know it's recurring? What data do you have (historic) that makes you believe the revenue will happen again?

Is this based on signed contracts etc so you have some guarantees?

dcre•3mo ago
The "annual" just means that the unit of time is a year. It doesn't mean that it is recurring annually. You can call it Annualized Monthly Recurring Revenue if it makes you feel better.
LetsGetTechnicl•3mo ago
Well people like Sam Altman have not been entirely honest and there's a reason they're not sharing their actual revenue numbers. If they could show they were growing 10% every month they would.
dcre•3mo ago
They are sharing their actual revenue numbers. That's what ARR is. Take the number and divide it by 12 and that's monthly revenue.
LetsGetTechnicl•3mo ago
It's literally not ARR is the highest monthly revenue times 12. Dividing it by 12 doesn't get you the actual, on the books monthly revenue numbers.
infecto•3mo ago
Eh, when you have a company that’s growing, picking the highest and annualizing it is sensible. If we had a mature company with highly seasonal revenue it would be dishonest.
LetsGetTechnicl•3mo ago
I mean I think there are instances where OpenAI's revenue is seasonal. Lots of students using it during the school year and cancelling it during summer.
infecto•3mo ago
I think you missed the forest for the trees. I am sure the student population has some dropoff during summer months but the point is that for businesses that a growing month over month which most of these have since creation, you take the highest number (latest) and annualize it.

I am also willing to bet that the student dropoff is not pronounced. I am more thinking of a business that sells beach umbrellas, they make a lot of sales in the summer months and then next to nothing in the winter months. That would be dishonest.

LetsGetTechnicl•3mo ago
Then why aren't AI companies reporting their actual monthly revenues?
dcre•3mo ago
They are. That is what ARR is.
dcre•3mo ago
The graph that was widely shared to make this claim was from OpenRouter and did not represent ChatGPT usage in any way.
dcre•3mo ago
The article Zitron links says Cursor has single-digit millions of cash burn with about $1B in the bank (as of August). Assuming that is true, they are losing money but have a long runway.

https://www.newcomer.co/p/cursors-popularity-has-come-at-a

omnicognate•3mo ago
Single-digit cash burn on AWS, which the article says is only a small part of its compute, with the majority coming from Anthropic.
dcre•3mo ago
That article says "Anysphere runs pretty lean with around 150 employees and has a single digit monthly cash burn, a source tells me." That would be total cash burn, i.e., net losses. If their AWS bill is bigger than that it's because they are making up for part of it with revenue.
omnicognate•3mo ago
Ah, gotcha. I misunderstood your comment.
spydum•3mo ago
specific clarification: That was only Cursor's AWS bill. If they are using other providers, wasn't clear.
omnicognate•3mo ago
TFA claims Cursor "obtains the majority of its compute from Anthropic — with AWS contributing a relatively small amount" and therefore only claims that for Cursor the AWS number indicates a "direction of travel" for compute costs. (Debatable whether it does indicate even that, ofc.)
swyx•3mo ago
> Based on discussions with sources with direct knowledge of their AWS billing, I am able to disclose the amounts that AI firms are spending, specifically Anthropic and AI coding company Cursor, its largest customer.

so he got a leaked copy of their AWS bills?

joshribakoff•3mo ago
Usually, the reporter inspects the document but does not get to take a copy
rdtsc•3mo ago
> through September, Anthropic has spent more than 100% of its estimated revenue (based on reporting in the last year) on Amazon Web Services, spending $2.66 billion on compute on an estimated $2.55 billion in revenue.

Well I don't have to scratch my head any longer and wonder why Amazon hasn't jumped on the AI bandwagon with their own Gemini or whatever. They are sitting pretty and selling shovels and pickaxes to the AI fools. Not a bad strategy for them...

haberdasher•3mo ago
Amazon owns 15-19% of Anthropic. So yes and no.
candiddevmike•3mo ago
GenAI financing is a flat circle. The bubble bursting is going to have a huge blast radius.
lesuorac•3mo ago
They survived the dot-com bubble, I don't see the AI bubble taking out Amazon.

It might take out your 401k for a decade.

noir_lord•3mo ago
I'm on like my fifth once in a generation financial crisis.

At a certain point you just expect it.

gruez•3mo ago
How are you getting 5? In recent memory, there's only 3: dotcom bubble (2000), GFC (2008), covid (2020). You'd either have to go back in time even more (eg. savings and loan crisis in the 80s/90s), or include regional ones (eg. eurozone crisis) to hit 5.
tadfisher•3mo ago
Maybe they had strong exposure to Worldcom or Enron?
lesuorac•3mo ago
Are you counting eurozone as 2011? I think it counts enough as American if it caused a downgrade in the USG's credit rating.

[1]: https://en.wikipedia.org/wiki/August_2011_stock_markets_fall

adolph•3mo ago
. . . once in a [news media financial analyst] generation financial crisis.

Its like fruit fly generations, not 20-30 year human cohort generations.

grogenaut•3mo ago
If you were 100 then you'd be right on pace. Or slightly ahead. Apparently a generation is 25-30 years. I think they should update that
thewebguyd•3mo ago
I don't think anyone is expecting Amazon (or Google or Microsoft for that matter) to be taken out by the AI bubble.

I would expect to see OpenAI, Anthropic, and a lot of the little tool wrappers to get taken out though, or at least acquired for pennies on the dollar when it bursts.

But like the last one, it's going to be us, the tax payers, that are left holding the bag.

jcheng•3mo ago
Why taxpayers? Where’s the systemic risk in AI labs getting acquired for cents on the dollar? The taxpayers weren’t holding the bag during the dot com crash, just investors.
thewebguyd•3mo ago
During the dot com crash, none of the companies were "too big to fail" This is looking more like 2008 than dotcom, the entire market is being propped up by basically Nvidia. US GOV is likely to bail them out for "national security"
almostgotcaught•3mo ago
this is the latest mantra from people who have missed the boat. i'm like lol do you think that industry didn't learn anything (about financing structures) from the last one?
__alexs•3mo ago
All they will learn is how to shift more of the risk on to retail investors and B tier institutional investors.
xeornet•3mo ago
I guess by this logic we should never have a depression/recession/bubble burst ever again? We always learn from our mistakes!
tinyhouse•3mo ago
No way their share is so high. Where did you get these numbers from?
swyx•3mo ago
it's a number 2 years out of date. they invested 8b and ant is now worth 183b.
cortesoft•3mo ago
That doesn't tell us what percentage they own, though. When you invest in a company and the value goes up, your percentage doesn't change (unless there are additional investors)
swyx•3mo ago
that's really naive im afraid. you have to take pro rata or the percentage goes down. amzn did not take pro rata.
cortesoft•3mo ago
I am not sure what you mean by naive. I also didn't say the percentage wouldn't go down, I literally said other investment rounds would change the percentage they own.

Amazons investment in Anthropic was in the form of convertible notes, which they have converted entirely into equity by march of this year. At that time, Anthropic was valued at 61.5 billion and Amazon (in their filings) said their investment was worth 13.8 billion, so about 22% of the company.

Then, there was another round in September where Anthropic raised 13 billion more at a valuation of 183 billion (so the new investors are buying about a 7% stake in the company). Without more details, that would lower amazons percentage to about 20% (old investors hold 93% of the company, so Amazon's 22% of the remaining 93% comes out to about 20%). There are probably other details that lower that percentage a bit, but i think the 15-19% ownership estimate is pretty accurate.

wrren•3mo ago
They have their own models, the Nova series, although my experience has been pretty mixed with them.
rs186•3mo ago
Amazon trained their own models like Nova and has AI coding assistants like Amazon Q, but I don't know anyone outside Amazon who is using them.
JCM9•3mo ago
Yes, the Amazon AI stuff isn’t great. They don’t really have the right leadership or talent to do anything particularly competitive there.
oliwarner•3mo ago
Why do they need to compete? AI at Amazon should be laser-focussed on two things: selling compute time to AI Wannabees, and up-selling stuff to me in the shop.

Everything else is expense.

JCM9•3mo ago
I agree. Unfortunately AWS leadership is in an all out panic to compete, diverting attention away from core services. I wish they’d just get back to focusing on the foundational stuff!
VirusNewbie•3mo ago
AWS is not winning selling compute to AI folks. GCP and Azure have both done better selling AI services, which is part of the problem. Amazon is still the king for basic compute, but fell behind on the AI infra.
jayd16•3mo ago
They actually do have Alexa.
rdtsc•3mo ago
They do and they've had her for a while, but from what I understand that's "so last Tuesday" in the AI race. It was ok in the race with "Ok Google" and "Siri" but not competing with OpenAI, Anthropic, Gemini.
dumbmrblah•3mo ago
Is that just for inference or is that the cost of training the models as well?
dcre•3mo ago
There is no breakdown. Assuming they're real, they're just spending numbers.
JCM9•3mo ago
“Spend” requires a grain of salt here. AWS “invests” in Anthropic and then Anthropic buys AWS. If you follow the money with marked bills, AWS is buying this compute from itself and then claiming revenue.

There’s a lot of that sort of thing going on at the moment in the AI bubble.

dzonga•3mo ago
One day we will westerners will learn why the Chinese are releasing models that are optimized for cost of training n yet good enough to run locally or cheaply.

when the music stops, suddenly a lot of people won't just sit on the ground but plunge into the depths of hell.

infecto•3mo ago
I keep hearing this but I don’t know of many folks utilizing Chinese models, even those hosted in an agreeable territory.
ojosilva•3mo ago
Yeah, I'm one of them, using Qwen 3 coder on Cerebras as coding agent through CC. What I keep hearing is (very ballpark anecdata)...

50% of people into coding agents are quite concerned about that last mile in difference with frontier models that they "can't afford to lose" - my experience tells me otherwise, the difference is negligible once you have a good setup going and knows how to tune your model + agent.

The other 50% don't give a damn, they just landed, or got locked, into some deal for a coding agent and are happy with what they got, so why change? These deals arrived from the big model providers and resellers first, so Chinese arrived late and with too little to the party.

Running Chinese models (for coding) requires many things that you need to figure-out yourself. Are you running the model on your hw or through a provider? Are you paying by token or on a plan? Does the model pair well with you agent CLI/IDE of choice? (Zed, Cline, Opencode, etc) Does it even work with your favorite tool? (tool calling is very wobbly) Is it fast (tps)? Is it reliable? How do you do "ultrathink" with a secondary model? How do you do "large context"? Does it include a cache or are you going to eat through the plan in 1hr/day? What context size are you getting? Does it include vision and web search or do you have to get another provider/mcp for that? And, yeah, is it in a territory where you can send your client's code to? A lot to grok.

Cerebras Coder Max is really cool if you want to hack your way through this, but they couldn't care less about your experience: no cache, no tool endpoint fine-tuning, no plans or roadmap on updating models, on increasing context windows, adding vision, or anything really. They just deleted some of the tools they were recommending out of the website (ie Cursor) as they got reports of things that stopped working.

infecto•3mo ago
Which I find all well and good but you are in the minority. For myself I have no interest in learning any of those things, I value my time differently. Which is totally fine and great for both of us. But going back to the point, most folks are not using these Chinese models.
almostgotcaught•3mo ago
i'm starting a new trend: ask every person that is so certain about the negative outlook how big their short position is. so how big is your short position? please let us know.
throwaway290•3mo ago
here's another trend: every time a person is on this hype bandwagon ask them how much are they invested in nvidia/ms/openai/etc

I am not invested in anything except popcorn to watch it burst;)

almostgotcaught•3mo ago
But see that's exactly my point - I am invested so actually I don't go around "talking my book" but you have zero skin in the game so your opinion matters about as much as Bob down at the bar watching the game on TV.
throwaway290•3mo ago
If you're invested you have interest to say it's good regardless of reality. because if it's bad first you need to sell off to avoid being the bagholder

Yes you can say you're honest and stuff but does it mean anything if you can go broke from honesty.

almostgotcaught•3mo ago
I have no idea what you're saying - I said I don't talk my book - that means I'm not out here pontificating like Jim Cramer and half of hn.
throwaway290•3mo ago
you are taking your book by implying all the naysayers are secretly shorting the stock instead of having reasons to think it's empty hype. let's see how long copium lasts.
almostgotcaught•3mo ago
> secretly shorting the stock instead of having reasons to think it's empty hype

that's literally the exact opposite of what i'm saying - literally read my responce to comments up chain.

throwaway290•3mo ago
Interesting

> i'm starting a new trend: ask every person that is so certain about the negative outlook how big their short position is. so how big is your short position? please let us know.

To me people with short positions likely be bending the truth. To you they are honest. Funny

There's a reason why investing must be declared in many situations. and it is because they are a conflict of interest not proof of honesty

almostgotcaught•3mo ago
Do people have literally any idea what they're saying when they write these quippy responses?

> conflict of interest not proof of honesty

What exactly do you think disclosing a conflict of interest is if not honesty...............

throwaway290•3mo ago
Disclosing conflict of interest is not making you honest, it simply lets people know you have reasons to bend the truth

That's the reason in many cases disclosure is mandatory and not disclosing it is illegal. To let people know to take it with a grain of salt.

True honesty = no conflict of interest. Judging things for what they are not what bucks they can make you if you spin it right

almostgotcaught•3mo ago
> True honesty = no conflict of interest

Lol bro you're making up the definitions the definition just to suit your weird perspective.

throwaway290•3mo ago
https://en.wikipedia.org/wiki/Conflict_of_interest

> By definition, a "conflict of interest" occurs if, within a particular decision-making context, an individual is subject to two coexisting interests that are in direct conflict with each other ("competing interests"). This is important because under these circumstances, the decision-making process can be disrupted or compromised, affecting the integrity or reliability of the outcomes.

The decision is what to write on HN and integrity is compromised if "writing truth" is in conflict with "get richer" which can happen if you invest in a sus biz.

We can pretend you are the rare exception that speaks the truth even if it means you go broke but even then it just proves the rule

lbreakjai•3mo ago
That's a dishonest argument. For a short to work, you need to be right about the direction _and_ the timing. Being right too soon amounts to being wrong, and the market can stay irrational longer than you can stay solvent.
almostgotcaught•3mo ago
> market can stay irrational longer than you can stay solvent.

This is more mantra - professional shorts exist and they make money. So either you know something (in which case you'd be mum about it and just placing your bets) or you know nothing.

Attrecomet•3mo ago
Actually, anyone with a brain can see that it's a huge and massively inflated bubble, but can also see that it's a fools game to try to time it without professional insight into the market. What we can do is make sure we're not overextended on the bubble participants, but that's about it.

No amount of hype from your side can make it a non-bubble market when OpenAI and other companies make wild and impossible to fulfill promises for infrastructure buildup that rely on vague promises of circular money lending between AI SaaS and infra companies.

almostgotcaught•3mo ago
> Actually, anyone with a brain can see that it's a huge and massively inflated bubble, but can also see that it's a fools game to try to time it without professional insight into the market.

"The bubble is both obvious and also completely unpredictable".

Makes sense.

smsm42•3mo ago
Yes, it does. You may know what there would be realignment, but you don't know where or how, and may not possess the professional knowledge and resources required to profit from it directly. The known part is general, the unknown part is specific and time-dependent.
almostgotcaught•3mo ago
Someone else already said: yes all companies go to zero on a timeline that includes the heat death of the universe. Your claim is about as clever as that.
smsm42•3mo ago
No, that would only work after getting really high. You know, when somebody asks "why this thing is such and such?" and you answer "why things are at all? Maybe it's all just our imagination, dude! Maybe nothing makes sense and we are living in a simulation! Dude, just feel you mind expanding!"

But if we lay off the bong for a bit and think soberly, then no, knowing that certain economic process is not sustainable and will eventually collapse, but not being able to profit from it is not the same as saying every bubble would pop at the heat death of the universe, billions years for now. It's so not the same that the very comparison should be immediately perceived as absurd.

smsm42•3mo ago
> professional shorts exist

You can't both talk about "professional shorts" and use it as an argument against somebody who is not a financial professional. This argument is invalid because not every person who is skeptical about current AI companies' perspectives is willing or able to become a professional short just to make a point on it. One can have a perfectly valid argument and still not be a professional short based on it.

Karrot_Kream•3mo ago
Well yeah because in the end we're all dead. There will be a day when AI companies are like GE, IBM, or maybe even Intel. All innovative companies in the fullness of time stop being innovative. The question is when. It's not a bubble if it deflates in 20 years.
leobg•3mo ago
> Often someone pronounces his propositions with such confident and inflexible defiance that he seems to have entirely laid aside all concern for error. A bet disconcerts him. Sometimes he reveals that he is persuaded enough for one ducat but not for ten. For he would happily bet one, but at ten he suddenly becomes aware of what he had not previously noticed, namely that it is quite possible that he has erred.

—Kant, Critique of Pure Reason

almostgotcaught•3mo ago
yes this is exactly my strategy in getting people to really reveal the strength of their beliefs :)
ai-x•3mo ago
Just a reminder, Ed Zitron is neither an AI researcher, nor an Engineer, nor a Financial Analyst, nor an Economist nor an Insider and has ZERO clue in multiple dimensions (technology, investing, unit economics, growth, TAM) to analyze any of this
disgruntledphd2•3mo ago
Correct. And yet, he's provided some of the most level headed takes on the current LLM boom, to the point where FT Alphaville link to his analysis of the economics.
ai-x•3mo ago
The jury is still out there if his analysis are level-headed or not. He says things that *some* people want to hear, but that's not level-headed

E.g

What is the unit cost of serving a Token? It is the cost of electricity + amortized cost of GPU (GPUs would have been Capex, but because of their fast depreciation rate, you can claim they should be Opex). Given this cost structure, every SOTA labs (Google, Anthropic and OpenAI) are profitable and actually have high-margins 50-60%.

With this margin and growth, the frontier labs can be profitable anytime they want to. But they are sacrificing profitability for growth (as they should be)

Where is Ed's analysis about this? Either he is disingenuous or clueless. Remember people who voluntarily subscribe to Ed, are coming from wanting to hear what they believe.

If he is level-headed, show me an Ed article that is positive about AI

tecleandor•3mo ago
> If he is level-headed, show me an Ed article that is positive about AI

Why should those two things go together?

ai-x•3mo ago
If 1 Billion people voluntarily use a product and many claim to be productive, there must be something good about the product right?

But I guess Ed Zitron has found his audience

fred_is_fred•3mo ago
His argument is not that AI is not useful, it's that it's not financially sustainable at the current prices being charged for it - and additionally AI start-ups have 0 moat. Both of which likely are true.
edstarch•3mo ago
Unfortunately, his argument very often happens to be that AI is not useful, that there are no customers for it, that AI coding agents do not work...

I happen to agree with the overall sentiment (that AI buildout is overextending the tech sector and the financial markets), but he is utterly fixated on the evils of AI and unable to admit either the current usefulness or the future potential of the technology. This does not make him look like an honest broker.

The rambling nature of his posts also makes it harder to properly argue against them as he keeps repeating the same points over and over; some of them are decent but there is certainly a gish gallop feeling to the whole thing.

disgruntledphd2•3mo ago
> Unfortunately, his argument very often happens to be that AI is not useful, that there are no customers for it, that AI coding agents do not work...

He definitely changed his mind on AI coding agents based on reader feedback. Ultimately though, you need incredible productivity growth/massive layoffs to make the numbers work for the current spending and RN, I don't see large signs of this.

> I happen to agree with the overall sentiment (that AI buildout is overextending the tech sector and the financial markets), but he is utterly fixated on the evils of AI and unable to admit either the current usefulness or the future potential of the technology. This does not make him look like an honest broker.

I think this is probably because he feels like he's taking crazy pills when he hears what CEOs/leaders are saying about this. It's some kind of mind virus. Like, I was at a meetup a few months back where a senior data/code person was saying that nobody would write code in 5 years, which (if you've used the tools heavily) seems pretty absurd.

disgruntledphd2•3mo ago
> If 1 Billion people voluntarily use a product and many claim to be productive, there must be something good about the product right?

Not necessarily. That METR study was interesting in that participants reported that they were more productive, but the hard data disagreed. This is incredibly common when looking at humans, we're generally bad at knowing what hurts or helps us in this sphere.

And personally, I think LLMs are super useful, but I'm pretty sceptical about valuations and returns in this space over the short to medium term.

gloosx•3mo ago
A classic "billion flies can't be wrong"-style argument :)

A billion people drink Coke every day, yes, and billions also have cavities, obesity, and diabetes.

Majorities also once thought the Earth was flat.

dist-epoch•3mo ago
FT Alphaville dissed bitcoin and Tesla for 10 years. Not the greatest track record.
tadfisher•3mo ago
As the saying goes, "the market can remain irrational longer than you can remain solvent." I hope the next fad is to invest in boring companies that employ people and give steady returns, and they can stop replacing all the stores in my city with Dollar Trees.
dist-epoch•3mo ago
Ok, I have the perfect company for the next fad - https://www.boringcompany.com
tadfisher•3mo ago
Ah, yes, the company famous for the next innovation in public transport: humans slowly driving Teslas through an expensive tunnel.
disgruntledphd2•3mo ago
They also dissed Fyre festival, We Work and basically all of the frauds in the last ten years. Would I use them as my only source? Of course not, but they're definitely useful and (importantly for my point), in the professional finance space.

FWIW, I personally think they're correct on both bitcoin and Tesla, but apparently people disagree.

AznHisoka•3mo ago
>> nor a Financial Analyst, nor an Economist

Those people arent exactly experts or right most of the time either

ai-x•3mo ago
If Ed is neutral or an expert, he would have had the following analysis

"What is the unit cost of serving a Token? It is the cost of electricity + amortized cost of GPU (GPUs would have been Capex, but because of their fast depreciation rate, you can claim they should be Opex). Given this cost structure, every SOTA labs (Google, Anthropic and OpenAI) are profitable and actually have high unit margins of 50-60%."

High Unit Margins and growth means, these labs can be profitable anytime they choose to

baggachipz•3mo ago
An 18 hour old account, named suspiciously like an ai model company, trying to discredit a prominent AI skeptic.
dist-epoch•3mo ago
Here I am, 2.5 year account discrediting him - the confirmation bias of Zitron is so thick, you need diamond tipped cutters to cut through it.
ai-x•3mo ago
I'm betting big on AI (not a shill if you are alluding to that). Address my points. All other things are irrelevant.
spyckie2•3mo ago
Ai can both be a bubble and also the greatest economic value add of this generation at the same time. It doesn’t have to be either or.

All bubbles (dot com, housing, tech, crypto, etc) have a lot of losers and a few big winners.

That is less a reflection on the market of the bubble and more a reflection of the number, skill and risk taking of the prospectors.

dcre•3mo ago
Numbers are always interesting, assuming they're real, but I just want to comment on the Cursor thing: Zitron has been insisting for 6 months that Anthropic screwed Cursor somehow by raising prices on them but the claim has always been gibberish. It's not that it's false, it's that it's impossible to figure out what Zitron claims happened. He cannot describe (here or in https://www.wheresyoured.at/anthropic-and-openai-have-begun-...) what the bad change actually was. We know everyone moved to more usage-oriented pricing earlier this year. He cannot explain why this was a price increase for Cursor. He is unable to draw a distinction between a price increase for end users (it's not even clear that it was a price increase for the average end user) and a price increase for Cursor.
literatepeople•3mo ago
Ed has constantly done this, and it's a shame because it has taken the air out of the room for real AI criticism. Most of Ed's criticism comes from a place of giving a narrative to people who are wishing for a magic bullet that makes ChatGPT vanish tomorrow rather than actually pressuring companies about the harms this technology can cause. This in part is why his writing so often focuses on perceived financial issues (despite his lack of credentials in financial journalism) rather than the social harms the technologies cause today (slop, delusions, manipulated truth).
watwut•3mo ago
Zitron is too much of a small player to "suck the air off other criticism of ai".

Claiming that a single journalist blog has power to stop others from criticiaing ai for different reasons ia kind of absurd.

Jordan-117•3mo ago
I definitely seen him cited as an authority by AI critics far more than anyone else. The bending truth to tell them what they want to hear (and the gratuitous swearing) really helps.
knowaveragejoe•3mo ago
I can't think of an AI critic with more detailed writing on the subject.
watwut•3mo ago
Per number of words in article possibly. But I do not think he "sucks air out of other critics".

I mostly assume that most people end reading his articles somewhere in first third and go on reading something easier to read. His articles are not exactly casual read material and they are loooong. You have to have certain kind of personality to get over first few paragraphs.

bgwalter•3mo ago
It would be hilarious if all this was a devilish scheme to burn the excess money that was printed during COVID. Unfortunately, what is more likely is another bailout if the bubble bursts.
VirusNewbie•3mo ago

         I have sat with these numbers for a great deal of time, and I can’t find any evidence that Anthropic has any path to profitability outside of aggressively increasing the prices on their customers to the point that its services will become untenable for consumers and enterprise customers alike.

This is where he misunderstands. Enterprise companies will absolutely pay 10x the cost for Claude. Meta and Apple are two large customers, you think they won't pay $500 a month per employee? $1000 a month per employee? Neither of those are outrageous to imagine if it increases productivity 10%.
scottyah•3mo ago
Also spend will drop dramatically if the models level out a bit more. The training is what's compute heavy, and if you aren't having to retrain every month, but able to use things like Skills to stay competitive your costs will drop a lot.

I suppose that's the pessimistic-on-AI side. On the other hand, once you create God little things like money are meaningless.

ludicity•3mo ago
I'm much more skeptical about this for two major reasons.

Firstly, a huge amount of labour that can be accelerated by LLMs fall into the "bullshit jobs" category, where you can make someone faster at writing emails but the emails themselves don't really contribute much value. The majority of LLM use I see falls into this category. Many people can speed up parts of their job, but you can add as much efficiency as you want without actually impacting the bottom line -- and for various reasons that are not tractable right now, including with LLMs, businesses aren't able to get themselves to remove these roles.

Secondly, the median company is incapable of doing the things that aren't driven entirely by hype or political promises made by executives. We still exist in the universe where they prefer to have all their staff attrition out due to not getting raises, then end up paying the same amount for a bunch of folks that have no knowledge of the business when they inevitably have to replace their best talent.

With all that said, I'm sure a few savvier places would happily drop $1000 month per head if the value is there, but I really think in the average case that this would be more about marketing than any logic. People still buy Informatica in 2025 for much more money than they spend on LLMs.

VirusNewbie•3mo ago
You don't need to remove rolls if it makes everyone 10% better at their job. You think companies like paying for Office 365? That's barely a value add compared to free versions, but every company on the planet forks over money for that.
ludicity•3mo ago
To clarify, I'm saying adding 10% to a reasonably useless job is still 0%.

And secondly, maybe not an important point, but the paid version of O365 is something even my team has been forced to buy. The free versions paywall very important features. The free Word can't delete section breaks!

HDThoreaun•3mo ago
Its too easy to switch providers when theres a billion dollars a year at stake. If youre an apple exec who sees that the company is spending 10k per employee per year why wouldnt you start an initiative that cuts that spend to 1k per employee per year? Then you can go to the board and say you personally saved them all that money and get a big promotion.
VirusNewbie•3mo ago
Yes yes, why not just build our own cloud and our own workspace/365 provider, brilliant plan. Genius move to get a promo at any F500.
HDThoreaun•3mo ago
Doesnt matter if it is a shitshow. It saved a billion dollars a year
VirusNewbie•3mo ago
and yet Meta and Apple are both paying Anthropic hundreds of millions...
fred_is_fred•3mo ago
I've been unable to find it but early in the railroad boom did rail companies undercharge for freight in order to build demand and interest? Given that they were also given land in exchange fro building, it would have been in their interests to do so. They could also amortize rail costs over a long time as rails are useful for longer than GPUs - but I am also curious if there was an unsustainably low entry price for railroads like Ed is implying for AI tools.
mannyv•3mo ago
Given how hard it is to understand AWS billing, especially if you have custom pricing, I doubt his numbers are anywhere near correct.

That said, I hope they're using their prime Visa card so they can get some cash back on that spend.

mdavid626•3mo ago
Premium newsletter? Sigh, I’m getting old…
AstroBen•3mo ago
Would you rather ads obnoxiously plastered all over, and sponsors strategically woven into the content?

..oh and this article is free

kamranjon•3mo ago
Does it seem strange that this has 121 comments within 4 hours and somehow is ranked 131st on HN? I would think this would be front page with those type of numbers.
steveklabnik•3mo ago
Any time # of comments > # of upvotes, HN immediately downranks it quite a bit. This is called the "flamewar filter," and it's intended to cool off heated discussions.