frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The Evolution of the Interface

https://www.asktog.com/columns/038MacUITrends.html
1•dhruv3006•24s ago•0 comments

Azure: Virtual network routing appliance overview

https://learn.microsoft.com/en-us/azure/virtual-network/virtual-network-routing-appliance-overview
1•mariuz•37s ago•0 comments

Seedance2 – multi-shot AI video generation

https://www.genstory.app/story-template/seedance2-ai-story-generator
1•RyanMu•4m ago•1 comments

Πfs – The Data-Free Filesystem

https://github.com/philipl/pifs
1•ravenical•7m ago•0 comments

Go-busybox: A sandboxable port of busybox for AI agents

https://github.com/rcarmo/go-busybox
1•rcarmo•8m ago•0 comments

Quantization-Aware Distillation for NVFP4 Inference Accuracy Recovery [pdf]

https://research.nvidia.com/labs/nemotron/files/NVFP4-QAD-Report.pdf
1•gmays•9m ago•0 comments

xAI Merger Poses Bigger Threat to OpenAI, Anthropic

https://www.bloomberg.com/news/newsletters/2026-02-03/musk-s-xai-merger-poses-bigger-threat-to-op...
1•andsoitis•9m ago•0 comments

Atlas Airborne (Boston Dynamics and RAI Institute) [video]

https://www.youtube.com/watch?v=UNorxwlZlFk
1•lysace•10m ago•0 comments

Zen Tools

http://postmake.io/zen-list
1•Malfunction92•12m ago•0 comments

Is the Detachment in the Room? – Agents, Cruelty, and Empathy

https://hailey.at/posts/3mear2n7v3k2r
1•carnevalem•12m ago•0 comments

The purpose of Continuous Integration is to fail

https://blog.nix-ci.com/post/2026-02-05_the-purpose-of-ci-is-to-fail
1•zdw•15m ago•0 comments

Apfelstrudel: Live coding music environment with AI agent chat

https://github.com/rcarmo/apfelstrudel
1•rcarmo•15m ago•0 comments

What Is Stoicism?

https://stoacentral.com/guides/what-is-stoicism
3•0xmattf•16m ago•0 comments

What happens when a neighborhood is built around a farm

https://grist.org/cities/what-happens-when-a-neighborhood-is-built-around-a-farm/
1•Brajeshwar•16m ago•0 comments

Every major galaxy is speeding away from the Milky Way, except one

https://www.livescience.com/space/cosmology/every-major-galaxy-is-speeding-away-from-the-milky-wa...
2•Brajeshwar•16m ago•0 comments

Extreme Inequality Presages the Revolt Against It

https://www.noemamag.com/extreme-inequality-presages-the-revolt-against-it/
2•Brajeshwar•17m ago•0 comments

There's no such thing as "tech" (Ten years later)

1•dtjb•17m ago•0 comments

What Really Killed Flash Player: A Six-Year Campaign of Deliberate Platform Work

https://medium.com/@aglaforge/what-really-killed-flash-player-a-six-year-campaign-of-deliberate-p...
1•jbegley•18m ago•0 comments

Ask HN: Anyone orchestrating multiple AI coding agents in parallel?

1•buildingwdavid•19m ago•0 comments

Show HN: Knowledge-Bank

https://github.com/gabrywu-public/knowledge-bank
1•gabrywu•25m ago•0 comments

Show HN: The Codeverse Hub Linux

https://github.com/TheCodeVerseHub/CodeVerseLinuxDistro
3•sinisterMage•26m ago•2 comments

Take a trip to Japan's Dododo Land, the most irritating place on Earth

https://soranews24.com/2026/02/07/take-a-trip-to-japans-dododo-land-the-most-irritating-place-on-...
2•zdw•26m ago•0 comments

British drivers over 70 to face eye tests every three years

https://www.bbc.com/news/articles/c205nxy0p31o
37•bookofjoe•26m ago•12 comments

BookTalk: A Reading Companion That Captures Your Voice

https://github.com/bramses/BookTalk
1•_bramses•27m ago•0 comments

Is AI "good" yet? – tracking HN's sentiment on AI coding

https://www.is-ai-good-yet.com/#home
3•ilyaizen•28m ago•1 comments

Show HN: Amdb – Tree-sitter based memory for AI agents (Rust)

https://github.com/BETAER-08/amdb
1•try_betaer•29m ago•0 comments

OpenClaw Partners with VirusTotal for Skill Security

https://openclaw.ai/blog/virustotal-partnership
2•anhxuan•29m ago•0 comments

Show HN: Seedance 2.0 Release

https://seedancy2.com/
2•funnycoding•29m ago•0 comments

Leisure Suit Larry's Al Lowe on model trains, funny deaths and Disney

https://spillhistorie.no/2026/02/06/interview-with-sierra-veteran-al-lowe/
1•thelok•29m ago•0 comments

Towards Self-Driving Codebases

https://cursor.com/blog/self-driving-codebases
1•edwinarbus•30m ago•0 comments
Open in hackernews

Amazon launches Trainium3

https://techcrunch.com/2025/12/02/amazon-releases-an-impressive-new-ai-chip-and-teases-a-nvidia-friendly-roadmap/
204•thnaks•2mo ago

Comments

aaa_aaa•2mo ago
Interesting that in the article, they do not say what the chip actually does. Not even once.
egorfine•2mo ago
Probably because the only task this chip has to perform is to please shareholders hence there is no need to explain anything to us peasant developers.
caminante•2mo ago
Time to go squat on trainium4.com [0]

[0] https://www.godaddy.com/domainsearch/find?domainToCheck=trai...

wmf•2mo ago
Training. It's in the name.
cobolcomesback•2mo ago
Ironically these chips are being targeted at inference as well (the AWS CEO acknowledged the difficulties in naming things during the announcement).
wmf•2mo ago
The same thing happened to AMD and Gaudi. They couldn't get training to work so they pivoted to inference.
delaminator•2mo ago
Perhaps they should do their training on their Inferentia chips and see how that works out?
djmips•2mo ago
The I stands for Inference then?
Kye•2mo ago
Vector math
Symmetry•2mo ago
A bunch of 128x128 systolic arrays at its heart. More details:

https://newsletter.semianalysis.com/p/amazons-ai-self-suffic...

trebligdivad•2mo ago
Nice article; I hate to think what the DC bus bars look like! ~50v at ~25kW/rack; 500A bus bars - I guess split, but still!
nimbius•2mo ago
the real news is: "and teases an Nvidia-friendly roadmap"

The sole reason amazon is throwing any money at this is because they think they can do to AI what they did with logistics and shipping in an effort to slash costs leading into a recession (we cant fire anyone else.) The hubris is magnanimous to say the least.

but the total confidence is very low...so "Nvidia friendly" is face saving to ensure no bridges they currently cross for AWS profit get burned.

ZeroCool2u•2mo ago
I've had to repeatedly tell our AWS account reps that we're not even a little interested in the Trainium or Inferentia instances unless they have a provably reliable track record of working with the standard libraries we have to use like Transformers and PyTorch.

I know they claim they work, but that's only on their happy path with their very specific AMI's and the nightmare that is the neuron SDK. You try to do any real work with them and use your own dependencies and things tend to fall apart immediately.

It was just in the past couple years that it really became worthwhile to use TPU's if you're on GCP and that's only with the huge investment on Google's part into software support. I'm not going to sink hours and hours into beta testing AWS's software just to use their chips.

ecshafer•2mo ago
IMO AWS once you get off the core services is full of beta services. S3, Dynamo, Lambda, ECS, etc are all solid. But there are a lot of services they have that have some big rough patches.
kentm•2mo ago
I'd add SQS to the solid category.

But yes, the less of a core building block the specific service is (or widely used internally in Amazon), the more likely you are to run into significant issues.

nextworddev•2mo ago
Kinesis is decent
zdc1•2mo ago
That's heartening to know. I find running Kafka less pleasant.
rockwotj•2mo ago
Checkout redpanda
hnlmorg•2mo ago
This. 100 times this.
jeffparsons•2mo ago
RDS, Route53, and Elasticache are decent, too. But yes, I've also been bitten badly in the distant past by attempting to rely on their higher-level services. I guess some things don't change.

I wonder if the difference is stuff they dogfood versus stuff they don't?

ozten•2mo ago
A big problem for a when three AWS teams launch the same thing. Lowers confidence in dogfooding the “right” one.
smallmancontrov•2mo ago
Or when your AWS account rep is schmoozing your boss trying to persuade them to use something that is officially deprecated, lol.
raw_anon_1111•2mo ago
Amazon Connect is a solid higher level offering. But only because it is a productized version of Amazon Retail’s call center
phantasmish•2mo ago
I once used one of their services (I forget which, but I think it was there serverless product) that “supported” Java.

… but the official command line tools had show-stopper bugs if you were deploying Java to this service, that’d been known for months, and some features couldn’t be used in Java, and the docs were only like 20% complete.

But this work-in-progress alpha (not even beta quality because it couldn’t plausibly be considered feature complete) counted as “supported” alongside other languages that were actually supported.

(This was a few years ago and this particular thing might be a lot better now, but it shows how little you can trust their marketing pages and GUI AWS dashboards)

nunez•2mo ago
I'm assuming you're talking about Lambda. I don't mess with their default images. Write a Dockerfile and use containerized Lambdas. Saves so many headaches. Still have to deal with RIE though, which is annoying.
nunez•2mo ago
My understanding is that AWS productizes lots of one-offs for customers (like Snowball), so that makes sense
belter•2mo ago
>But there are a lot of services they have that have some big rough patches.

Enlight us...

kentm•2mo ago
Personally, EMR has never shaken off the "scrappy" feeling (sometimes it feels OK if you're using Spark), and it feels even more neglected recently as they seem to want you on AWS Glue or Athena. LakeFormation is... a thing that I'm sure is good in theory if you're using only managed services, but in practice is like taking a quick jaunt on the Event Horizon.

Glue Catalog has some annoying assumptions baked in.

Frankly the entire analytics space on AWS feels like a huge mess of competing teams and products instead of a uniform vision.

weird-eye-issue•2mo ago
True with Cloudflare too. Just stick with Workers, R2, Durable Objects, etc...
plantain•2mo ago
Not even sure about R2 with it's unpredictable latencies.
weird-eye-issue•2mo ago
Hmm is it actually that bad? Keep in mind r2 is only stored in one region which is chosen when the bucket is first created so that might be what you're seeing

But I've never really looked too closely because I just use it for non-latency critical blob storage

BOOSTERHIDROGEN•2mo ago
Lightsail fortunately behave like core services.
htrp•2mo ago
spoiler alert, they don't work without a lot of custom code
mountainriver•2mo ago
Agree, Google put a ton of work into making TPUs usable with the ecosystem. Given Amazon’s track record I can’t imagine they would ever do that.
klysm•2mo ago
There might be enough market pressure right now to make them think about it, but the stock price went up enough from just announcing it so whatever
vachina•2mo ago
Amazon has no interest in making their platform interoperable.
retinaros•2mo ago
https://awsdocs-neuron.readthedocs-hosted.com/en/latest/fram...
ChrisArchitect•2mo ago
Amazon post: https://www.aboutamazon.com/news/aws/trainium-3-ultraserver-...
cmiles8•2mo ago
AWS keeps making grand statements about Trainium but not a single customer comes on stage to say how amazing it is. Everyone I talked to that tries it says there were too many headaches and they moved on. AWS pushes it hard but “more price performant” isn’t a benefit if it’s a major PITA to deploy and run relative to other options. Chips without a quality developer experience isn’t gonna work.

Seems AWS is using this heavily internally, which makes sense, but not observing it getting traction outside that. Glad to see Amazon investing there though.

giancarlostoro•2mo ago
Not just AWS, looks like Anthropic uses it heavily as well. I assume they get plenty of handholding from Amazon though. I'm surprised any cloud provider does not invest drastically more into their SDK and tooling, nobody will use your cloud if they literally cannot.
cmiles8•2mo ago
Well AWS says Anthropic uses it but Anthropic isn’t exactly jumping up and down telling everyone how awesome it is, which tells you everything you need to know.

If Anthropic walked out on stage today and said how amazing it was and how they’re using it the announcement would have a lot more weight. Instead… crickets from Anthropic in the keynote

teruakohatu•2mo ago
> Anthropic isn’t exactly jumping up and down telling everyone how awesome it is, which tells you everything you need to know.

You can’t really read into that. They are unlikely to let their competitors know if they have a slight performance/$ edge by going with AWS tech.

cmiles8•2mo ago
With GCP announcing they built Gemini 3 on TPUs the opposite is true. Anthropic is under pressure to show they don’t need expensive GPUs. They’d be catching up at this point, not leaking some secret sauce. No reason for them to not boast on stage today unless there’s nothing to boast about.
0x457•2mo ago
Yes, but Google benefit from people using their TPUs, while Anthropic gains nothing unless AWS throws money at them for saying it.
bilbo0s•2mo ago
This.

Anthropic is not going to interrupt their competitors if their competitors don't want to use trainium. Neither would you, I, nor anyone else. The only potential is downside. There's no upside potential for them at all in doing so.

From Anthropic's perspective, if the rest of us can't figure out how to make trainium work? Good.

Amazon will fix the difficulty problem with time, but that's time Anthropic can use to press their advantages and entrench themselves in the market.

breppp•2mo ago
I am not sure, I would imagine enthusiastic quotes will lead to huge discounts and in that scale that matters
fishmicrowaver•2mo ago
Striking a deal with a competitor (AZURE) does though.
cobolcomesback•2mo ago
AWS has built 20 data centers in Indiana full of half a million Trainium chips explicitly for Anthropic. Anthropic is using them heavily. The same press announcement that Anthropic has made about Google TPUs is the exact same one they made a year ago about Trainium. Hell, even in the Google TPU press release they explicitly mention how they are still using Trainium as well.
VirusNewbie•2mo ago
Can you link to the press releases? The only one I'm aware of by Anthropic says they will use Tranium for future LLMs, not that they are using them.
cobolcomesback•2mo ago
This is the Anthropic press release from last year saying they will use Trainium: https://www.anthropic.com/news/anthropic-amazon-trainium

This is the AWS press release from last month saying Anthropic is using 500k Trainium chips and will use 500k more: https://finance.yahoo.com/news/amazon-says-anthropic-will-us...

And this is the Anthropic press release from last month saying they will use more Google TPUs but also are continuing to use Trainium (see the last 2 paragraphs specifically): https://www.anthropic.com/news/expanding-our-use-of-google-c...

VirusNewbie•2mo ago
There is no press release saying that they are using 500k trainium chips. You can search on amazon's site.
smallerize•2mo ago
In the subheading of https://www.aboutamazon.com/news/aws/aws-project-rainier-ai-... Unless there are other Project Ranier customers that it's counting.
hustwindmaple•2mo ago
I met a AWS engineer a couple of weeks ago and he said Trainium is actually being used for Anthropic model inference, not for training. Inferentia is basically defected Trainiums chips that nobody wants to use.
IshKebab•2mo ago
> I'm surprised any cloud provider does not invest drastically more into their SDK and tooling

I used to work for an AI startup. This is where Nvidia's moat is - the tens of thousands of man-hours that has gone into making the entire AI ecosystem work well with Nvidia hardware and not much else.

It's not that they haven't thought of this, it's just that they don't want to hire another 1k engineers to do it.

logicchains•2mo ago
>I'm surprised any cloud provider does not invest drastically more into their SDK and tooling, nobody will use your cloud if they literally cannot.

Building an efficient compiler from high-level ML code to a TPU is actually quite a difficult software engineering feat, and it's not clear that Amazon has the kind of engineering talent needed to build something like that. Not like Google which have developed multiple compilers and language runtimes.

phamilton•2mo ago
The inf1/inf2 spot instances are so unpopular that they cost less than the equivalent cpu instances. Exact same (or better) hardware but 10-20% cheaper.

We're not quite seeing that on the trn1 instances yet, so someone is using them.

kcb•2mo ago
Heh, I was looking at an eks cluster recently that was using Cast AI autoscalar. Scratching my head as there was a bunch of inf instances. Then I realized it must be cheap spot pricing.
jauntywundrkind•2mo ago
Amazon aside, interesting future here with NVLink getting more and more folks using it. Intel is also onboard with NVlink. This is like an PCI -> AGP moment, but Nvidia's AGP.

AMD felt like they were so close to nabbing the accelerator future back in HyperTransport days. But the recent version Infinity Fabric is all internal.

There's Ultra Accelerator Link (UALink) getting some steam. Hypothetically CXL should be good for uses like this, using PCIe PHY but lower latency lighter weight; close to ram latency, not bad! But still a mere PCIe speed, not nearly enough, with PCIe 6.0 just barely emerging now. Ideally IMO we'd also see more chips come with integrated networking too: it was so amazing when Intel Xeon's had 100Gb Omni-Path for barely any price bump. UltraEthernet feels like it should be on core, gratis.

wmf•2mo ago
NVLink Fusion sounds like a total trap where you pay to become Jensen's slave. It may make sense for Intel because they're desperate. It's not a good look for AWS to put themselves in the same category.

UltraEthernet feels like it should be on core, gratis.

I've been saying for a while that AMD should put a SolarFlare NIC in their I/O die. They already have switchable PCIe/SATA ports, why not switchable PCIe/Ethernet? UEC might be too niche though.

mlmonkey•2mo ago
Not a single mention of any benchmarks or performance.
pedalpete•2mo ago
They say 4x more, but not 4x faster, 4x more memory, but not 4x more than what!?
vrosas•2mo ago
4x more units, clearly
matrix2596•2mo ago
yea, they "officially" dont release benchmarks even if we asked the AWS reps
landl0rd•2mo ago
Anyone considering using trainium should view this Completely Factual Infomercial: https://x.com/typedfemale/status/1945912359027114310

Pretty accurate in my experience, especially re: the neuron sdk. Do not use.

deepsquirrelnet•2mo ago
Heavens to Betsy, I don’t know if you can hear me, But try supporting these things if you actually want them to be successful. About the 3rd day into trying to roll your own LMI container in sagemaker because they haven’t updated the vLLM version in 6 months and you can’t run a regular sagemaker endpoint because of a ridiculous 60s timeout that was determined to be adequate 8 years ago. I can only imagine the hell that awaits the developer that decides to try their custom silicon.
trebligdivad•2mo ago
The Block floating point (MXFP8/4 stuff) is interesting; the AI stuff is really pushing basic data types that haven't moved for decades.

https://en.wikipedia.org/wiki/Block_floating_point

hackermeows•2mo ago
at some point the cost of transferring will dwarf the cost you pay to NVIDA. I bet that is their bet
smilekzs•2mo ago
Single-chip specs according to:

https://awsdocs-neuron.readthedocs-hosted.com/en/latest/abou...

https://awsdocs-neuron.readthedocs-hosted.com/en/latest/nki/...

Eight NeuronCore-v4 cores that collectively deliver:

    2,517 MXFP8/MXFP4 TFLOPS
    671 BF16/FP16/TF32 TFLOPS
    2,517 FP16/BF16/TF32 sparse TFLOPS
    183 FP32 TFLOPS
HBM: 144 GiB HBM @ 4.9 TB/sec (4 stacks)

SRAM: 32 MiB * 8 = 256 MiB (ignoring 2 MiB * 8 = 16 MiB of PSUM which is not really general-purpose nor DMA-able)

Interconnect: 2560 GB/s (I think bidirectional, i.e. Jensen Math™)

----

At 3nm process node the FLOP/s is _way_ lower than competition. Compare to B200 which does 2250 BF16, x2 FP8, x4 FP4. TPU7x does 2307 BF16, x2 FP8 (no native FP4). HBM also lags behind (vs ~192 GiB in 6 stacks for both TPU7x and B200).

The main redeeming qualities seem to be: software-managed SRAM size (double of TPU7x; GPUs have L2 so not directly comparable) and on-paper raw interconnect BW (double of TPU7x and more than B200).

parkersweb•2mo ago
When they're in short supply can we name them Unobtranium?
regnull•2mo ago
These product might be great, but seriously, who's choosing those names? Trainium, Inferentia? It's like let's just take the words from what they do, and put a little Latin twist on them? I know naming things is one of the great problems in computer science, but really they could come up with something a little better.