frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Anthropic expands partnership with Google and Broadcom for next-gen compute

https://www.anthropic.com/news/google-broadcom-partnership-compute
128•l1n•4h ago

Comments

mikert89•1h ago
There's no limit to the algorithms. People dont understand yet. They can learn the whole universe with a big enough compute cluster. We built a generalizable learning machine
totaa•1h ago
the question is will we experience resource constraints before we get there? what if the step up to post-scarcity is gated by a compute level just out of our reach?
mikert89•1h ago
human ingenuity will solve this
__loam•59m ago
Or we'll have ecological collapse.
teaearlgraycold•1h ago
Not sure if this is satire.

Edit: What we have built is a natural language interface to existing, textually recorded, information. Transformers cannot learn the whole universe because the universe has not yet been recorded into text.

supliminal•1h ago
It’s more than likely not.
erelong•1h ago
Poe's (c)law?
bryogenic•46m ago
Poe’s (C)law: The more absurd AI-generated content becomes, the more likely people are to believe it is real.
0x3f•1h ago
Based on a glance at their other comments: not satire.
alfalfasprout•1h ago
100% agreed. Sadly, lots of people out there with the "trust me bro, just need more compute". Hopefully we don't consume all the planet's resources trying.
xvector•56m ago
I reevaluated my priors long ago when I saw that scaling laws show no sign of stopping, no sign of plateau.

Strangely some people on HN seem to desperately cling to the notion that it's all going to come to a halt. This is unscientific. What evidence do you have - any evidence - that the scaling laws are due to come to an end?

rishabhaiover•49m ago
I suspect it's not that people do not see the progress, they fail to fully trust laws not truly backed by physics like the transistor laws. We empirically see that scaling works and continue to work.
0x3f•48m ago
All the curves have been levelling off as expected. Not really sure what you're talking about.
solenoid0937•28m ago
They have not, every successful pre-train as of late has had performance increases greater than what the scaling laws predict.
0x3f•5m ago
Those gains are arch based, data quality based, etc. Scaling laws only relate to data volume, holding other factors constant.
FridgeSeal•41m ago
The issue people have isn’t some interpretation of scaling laws, it’s whether the planet’s ecology is goi g to be able to sustain this endeavour.

I shouldn’t have to say this out loud, but if the environment collapses, we will die, and no amount of “just a bit more scaling bro, just think of the gains” will matter.

xvector•30m ago
People's voluntary dietary choices cause far more suffering and ecological damage than AI, and for much less return or economic output. But you tell people to switch to plant based foods and they lose their shit.
skybrian•40m ago
Why should we have strong priors in either direction? Maybe it will keep scaling for decades like Moore's law. Maybe not.
shimman•36m ago
Bro the planet is literally experiencing a climate disaster and you think the solution is to create more systems that are misaligned with the planet's ecosystem for humans?

I guess the great filter is a real thing and not just a thought experiment.

xvector•31m ago
I assure you that voluntary meat consumption because "taste buds go brr" is a much bigger problem than AI that results in actual productivity gains (and potentially solve the very climate crisis you complain about.)
teaearlgraycold•35m ago
I’d like to see something that indicates models are getting better without the need for more training data. I would expect most gains are coming from more and better labeled data. We’re racing towards a complete encyclopedia of human knowledge. If we get there that’s only a drop in the bucket of all knowable things.
lukeschlather•38m ago
Transformers operate on images and a variety of sensor data. They can also operate completely on non-textual inputs and outputs. I don't know what the ceiling on their capabilities is, but the complaint that they only operate on text seems just obviously wrong. There are numerous examples but one is meteorological forecasting which takes in a variety of time series sensor inputs and outputs e.g. time-series temperature maps. https://www.nature.com/articles/s41598-025-07897-4
Eufrat•1h ago
Can someone explain why everything is being marketed in terms of power consumption?
teaearlgraycold•1h ago
Because that’s the limiting factor
Animats•1h ago
Somehow we must be doing this wrong.

"Do you realize that the human brain has been liken to an electronic brain? Someone said and I don't know whether he is right or not, but he said, if the human brain were put together on the basis of an IBM electronic brain, it would take 7 buildings the size of the Empire State Building to house it, it would take all the water of the Niagara River to cool it, and all of the power generated by the Niagara River to operate it." (Sermon by Paris Reidhead, circa 1950s.[1])

We're there on size and power. Is there some more efficient way to do this?

[1] https://www.sermonindex.net/speakers/paris-reidhead/the-trag...

brianjlogan•1h ago
I'd imagine one day there will be a limiting factor of cash to burn as well.
Animats•1h ago
We're getting close. The first big AI bankruptcy can't be far off.
bdangubic•1h ago
Big Gov will bail out the big guys if/when necessary
whimsicalism•1h ago
pretty sure evolution spent more time and energy getting there then we ultimately will
serf•1h ago
kinda complicated though when you consider it fully. Power consumption only measures the environmental impact really, we come up with more clever ways to use the same amount of power daily.

it's kind of like an electrical motor that exists before the strong understanding of lorentz/ohm's law. We don't really know how inefficient the thing is because we don't really know where the ceiling is aside from some loosey theoretical computational efficiency concepts that don't strongly apply to practical LLMs.

to be clear, I don't disagree that it's the limiting factor, just that 'limits' is nuanced here between effort/ability and raw power use.

zozbot234•57m ago
There's at least a decent argument to be made that the limiting factor is actually the physical silicon itself (at least at cutting-edge nodes) not really the power. This actually gives AI labs an incentive to run those specific chips somewhat cooler, because high device temperatures and high input voltages (which you need to push frequencies higher) might severely impact a modern chip's reliability over time.
Eufrat•47m ago
I feel like that’s a bit glib?

Surely, there should be some more critical questions posed by why just buying a bunch of GPUs is a good idea? It just feels like a cheap way to show that growth is happening. It feels a bit much like FOMO. It feels like nobody with the capital is questioning whether this is actually a good idea or even a desirable way to improve AI models or even if that is money well spent. 1 GW is a lot of power. My understanding is that it is the equivalent to the instantaneous demand of a city like Seattle. This is absurd.

It feels like there is some awareness that asking for gigawatts if not terrawatts of compute probably needs more justification than has been proffered and the big banks are already trying to CYA themselves by publishing reports saying AI has not contributed meaningfully to the economy like Goldman Sachs recently did.

NoahZuniga•1h ago
It's more meaningful to most people than FLOPS/other measures of actual computing power.
jeffbee•1h ago
It's easy to think about. Google reported a global average power consumption of 3.7GW in 2024, so you can think of this deal as representing an expansion of something like 10-15% of that 2024 baseline, if you assume 50% capacity utilization.
skybrian•1h ago
I guess gigawatts is how we roughly measure computing capacity at the datacenter scale? Also saw something similar here:

> Costs and pricing are expressed per “token”, but the published data immediately seems to admit that this is a bad choice of unit because it costs a lot more to output a token than input one. It seems to me that the actual marginal quantity being produced and consumed is “processing power”, which is apparently measured in gigawatt hours these days. In any case, I think more than anything this vindicates my original decision not to get too precise. [...]

https://backofmind.substack.com/p/new-new-rules-for-the-new-...

Is it priced that way, though? I assume next-gen TPU's will be more efficient?

brokencode•1h ago
Gigawatts seems like more a statement of the power supply and dissipation of the actual facility.

I’m assuming you can cram more chips in there if you have more efficient chips to make use of spare capacity?

Trying to measure the actual compute is a moving target since you’d be upgrading things over time, whereas the power aspects are probably more fixed by fire code, building size, and utilities.

delichon•34m ago
Measuring data centers in watts is like measuring cars in horsepower. Power isn't a direct measure of performance, but of the primary constraint on performance. When in doubt choose the thermodynamic perspective.
nomel•1h ago
> but the published data immediately seems to admit that this is a bad choice of unit because it costs a lot more to output a token than input one

And, that's silly, because API pricing is more expensive for output than input tokens, 5x so for Anthropic [1], and 6x so for OpenAI!

[1] https://platform.claude.com/docs/en/about-claude/pricing

[2] https://openai.com/api/pricing

twoodfin•45m ago
That these data centers can turn electricity + a little bit of fairly simple software directly into consumer and business value is pretty much the whole story.

Compare what you need to add to AWS EC2 to get the same result, above and beyond the electricity.

zozbot234•29m ago
That's a convenient story, but most consumers' and businesses' use of AI is light enough that they could easily run local models on their existing silicon. Resorting to proprietary AI running in the datacenter would only add a tiny fraction of incremental value over that, and at a significant cost.
cebert•1h ago
I’m surprised Anthropic wanted to partner with Broadcom when they have such a negative reputation with antics such as their VMWare acquisition.
Eufrat•1h ago
I think it’s also important to add the context that Broadcom’s CEO, Hock Tan, went on CNBC in October and had a vacuous conversation with Jim Cramer about their OpenAI “deal” at the time [0]. Nothing of substance was said, it was just endless loops about the opportunity of AI. It is now 6 months later and there has been nary a peep from Broadcom about any updates.

I think Anthropic is a more grounded company than OpenAI because Sam Altman is insane, but it is still playing the same game.

[0] https://www.youtube.com/watch?v=pU2HhJ3jCts

jeffbee•57m ago
Broadcom makes the TPU. If you want TPUs, you are working with Broadcom whether you want to or not.
thundergolfer•52m ago
Broadcom builds the TPU chip. Google designs it. You can’t avoid partnering with Broadcom if you want TPUs in significant volume .
alephnerd•51m ago
Broadcom designs it as well [0], though GCP also works on design as well.

[0] - https://www.broadcom.com/products/custom-silicon

ggm•50m ago
The VMware s/w rental market has no relevance to this deal, any more than the IBM role in data processing in germany in the 1930s had any relevance to their business in Israel in the 60s, or Oracle's failure in the DC market impacts licencing of the database product.

It's just not material. Broadcom make devices they need, and Broadcom want to sell those devices and exclude another VLSI company from selling, so the two have an interest in doing business. That's all there is to it.

About the most you could say is that the lawyers drafting whatever agreement they sign to, will reflect on the contract in regard to future changes of pricing and supply, in the light of what Broadcom did with VMWare licencing costs.

ketzo•40m ago
$19B -> $30B annualized revenue in a month?

Feels like the lede is buried here!

ai-x•35m ago
But But But "AI is a bubble!!!!!!"

At what point would bubble-callers admit that they were completely wrong?

mrcwinn•21m ago
They won’t. They’ll just casually fade away from prior statements. Just like all the software engineers whose first take was that it’s just autocomplete.
baron816•19m ago
I think you can argue that AI is going to explode and take over the economy, and it’s still a bubble.

I think one possible route is that cloud capacity just becomes totally commoditized and none of the hyperscalers will be able to extract the kinds of profit margins that would allow them to make a good return on their investment (model makers will fall victim to this too). Ultimately, what may happen is that market competition for everything explodes since AI and robots can do all the work, prices for everything (goods, services, assets) collapses, and no one is really any richer than anyone else.

Show HN: Ghost Pepper – Local hold-to-talk speech-to-text for macOS

https://github.com/matthartman/ghost-pepper
226•MattHart88•6h ago•105 comments

Netflix Void Model: Video Object and Interaction Deletion

https://github.com/Netflix/void-model
19•bobsoap•3d ago•0 comments

Solod – A Subset of Go That Translates to C

https://github.com/solod-dev/solod
9•TheWiggles•1h ago•0 comments

Launch HN: Freestyle – Sandboxes for Coding Agents

https://www.freestyle.sh/
205•benswerd•9h ago•115 comments

A cryptography engineer's perspective on quantum computing timelines

https://words.filippo.io/crqc-timeline/
330•thadt•10h ago•142 comments

Porting Go's strings package to C

https://antonz.org/porting-go-strings/
33•ingve•3d ago•4 comments

Anthropic expands partnership with Google and Broadcom for next-gen compute

https://www.anthropic.com/news/google-broadcom-partnership-compute
128•l1n•4h ago•56 comments

German police name alleged leaders of GandCrab and REvil ransomware groups

https://krebsonsecurity.com/2026/04/germany-doxes-unkn-head-of-ru-ransomware-gangs-revil-gandcrab/
260•Bender•11h ago•132 comments

Show HN: GovAuctions lets you browse government auctions at once

https://www.govauctions.app/
216•player_piano•9h ago•66 comments

Sam Altman may control our future – can he be trusted?

https://www.newyorker.com/magazine/2026/04/13/sam-altman-may-control-our-future-can-he-be-trusted
849•adrianhon•15h ago•311 comments

Show HN: Hippo, biologically inspired memory for AI agents

https://github.com/kitfunso/hippo-memory
47•kitfunso•4h ago•14 comments

Issue: Claude Code is unusable for complex engineering tasks with Feb updates

https://github.com/anthropics/claude-code/issues/42796
755•StanAngeloff•12h ago•464 comments

After 20 years I turned off Google Adsense for my websites

https://blog.ericgoldman.org/archives/2025/06/after-20-years-i-turned-off-google-adsense-for-my-w...
109•datadrivenangel•2h ago•63 comments

What being ripped off taught me

https://belief.horse/notes/what-being-ripped-off-taught-me/
329•doctorhandshake•12h ago•182 comments

Book review: There Is No Antimemetics Division

https://www.stephendiehl.com/posts/no_antimimetics/
207•ibobev•12h ago•150 comments

HackerRank (YC S11) Is Hiring

1•rvivek•4h ago

AI singer now occupies eleven spots on iTunes singles chart

https://www.showbiz411.com/2026/04/05/itunes-takeover-by-fake-ai-singer-eddie-dalton-now-occupies...
105•flinner•9h ago•116 comments

Show HN: Tusk for macOS and Gnome

https://shapemachine.xyz/tusk/
48•factorialboy•2d ago•16 comments

Sky – an Elm-inspired language that compiles to Go

https://github.com/anzellai/sky
129•whalesalad•10h ago•46 comments

Show HN: Anos – a hand-written ~100KiB microkernel for x86-64 and RISC-V

https://github.com/roscopeco/anos
20•noone_youknow•2d ago•5 comments

The cult of vibe coding is dogfooding run amok

https://bramcohen.com/p/the-cult-of-vibe-coding-is-insane
471•drob518•7h ago•411 comments

Battle for Wesnoth: open-source, turn-based strategy game

https://www.wesnoth.org
388•akyuu•8h ago•99 comments

Agent Reading Test

https://agentreadingtest.com
48•kaycebasques•6h ago•15 comments

The Last Quiet Thing

https://www.terrygodier.com/the-last-quiet-thing
154•coinfused•2d ago•91 comments

Show HN: TTF-DOOM – A raycaster running inside TrueType font hinting

https://github.com/4RH1T3CT0R7/ttf-doom
24•4RH1T3CT0R•6h ago•4 comments

Ask HN: How do you handle marketing as a solo technical founder?

55•lazarkap•4h ago•26 comments

Eighteen Years of Greytrapping – Is the Weirdness Finally Paying Off?

https://nxdomain.no/~peter/eighteen_years_of_greytrapping.html
52•jruohonen•2d ago•5 comments

SOM: A minimal Smalltalk for teaching of and research on Virtual Machines

http://som-st.github.io/
27•tosh•6h ago•0 comments

The team behind a pro-Iran, Lego-themed viral-video campaign

https://www.newyorker.com/culture/infinite-scroll/the-team-behind-a-pro-iran-lego-themed-viral-vi...
92•tantalor•11h ago•111 comments

Adobe modifies hosts file to detect whether Creative Cloud is installed

https://www.osnews.com/story/144737/adobe-secretly-modifies-your-hosts-file-for-the-stupidest-rea...
244•rglullis•8h ago•120 comments