frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenAI Is Just Another Boring, Desperate AI Startup

https://www.wheresyoured.at/sora2-openai/
1•speckx•3m ago•0 comments

Mysterious "rogue planet" gobbles 6B tons of gas and dust a second

https://www.cbsnews.com/news/mysterious-rogue-planet-spotted-gobbling-6-billion-tons-of-gas-and-d...
1•Brajeshwar•3m ago•0 comments

We want to hear about which apps aren't working right with macOS Tahoe

https://appleinsider.com/articles/25/09/19/we-want-to-hear-about-which-apps-arent-working-right-w...
1•wslh•5m ago•0 comments

Dora State of AI-Assisted Software Development

https://cloud.google.com/blog/products/ai-machine-learning/announcing-the-2025-dora-report
1•oatsandsugar•5m ago•0 comments

Tensorglobe

https://vantor.com/product/platform/
1•jonbaer•6m ago•0 comments

Conversation Is Over: In AI-powered shopping, dialogue only gets in the way

https://aboard.com/this-conversation-is-over/
1•gbseventeen3331•7m ago•0 comments

From Nothing, Everything

https://aeon.co/essays/how-nothing-has-inspired-art-and-science-for-millennia
1•Brajeshwar•7m ago•0 comments

A Pipeline for Continual Learning Without Catastrophic Forgetting in LLMs

https://arxiv.org/abs/2509.10518
1•PaulHoule•7m ago•0 comments

Pavel Durov: Telegram, Freedom, Censorship [ ] Lex Fridman

https://lexfridman.com/pavel-durov-transcript/
2•hagbard_c•8m ago•0 comments

Open Source Blueprints for Civilization

https://www.opensourceecology.org/
3•vlugorilla•9m ago•0 comments

PEP 810: Explicit Lazy Imports

https://discuss.python.org/t/pep-810-explicit-lazy-imports/104131
1•ayhanfuat•10m ago•0 comments

Show HN: GitHub Integrated CI and Evals for AI Agents

https://agent-ci.com
1•tcdent•10m ago•0 comments

NiceGUI 3.0 – One of the best Python WebUI

https://github.com/zauberzeug/nicegui
1•synergy20•11m ago•0 comments

Divorce Plunged in Kentucky. Equal Custody for Fathers Is a Big Reason Why

https://www.wsj.com/us-news/law/the-equal-custody-experiment-41e1f7a6
2•thelastgallon•12m ago•0 comments

What is the purpose of the SSL-cert-snakeoil.key

https://askubuntu.com/questions/396120/what-is-the-purpose-of-the-ssl-cert-snakeoil-key
1•keepamovin•13m ago•0 comments

Room-Temperature Superconductivity at High-Pressure Conditions

https://arxiv.org/abs/2510.01273
2•P_qRs•14m ago•0 comments

Levels of AI agent autonomy: learning from self-driving cars – AI Native Dev

https://ainativedev.io/news/the-5-levels-of-ai-agent-autonomy-learning-from-self-driving-cars
1•JnBrymn•14m ago•0 comments

Symboltz.jl

https://github.com/hersle/SymBoltz.jl
1•darboux•16m ago•1 comments

Making Python in Zed Fun

https://zed.dev/blog/making-python-in-zed-fun
1•indigodaddy•17m ago•0 comments

Show HN: Enforza: Cloud-managed Linux firewall and NAT gateway

https://enforza.io/
1•enforzaguy•17m ago•1 comments

ESP32-S3 USB-UAC

https://www.youtube.com/watch?v=LhZ9pWwnn6s
1•iamflimflam1•17m ago•0 comments

Flotilla to Gaza had no humanitarian supplies

https://www.jpost.com/israel-news/article-869330
5•pinewurst•18m ago•2 comments

Show HN: TruVideo – On-device detector for Sora, Veo and other GenAI videos

https://apps.apple.com/id/app/truvideo-ai-detector/id6670568062
1•sanjkris•18m ago•0 comments

Local daily commute – handmade in the browser

https://www.commuting.to/
1•ChrisArchitect•19m ago•0 comments

Cancelling Async Rust

https://sunshowers.io/posts/cancelling-async-rust/
7•todsacerdoti•21m ago•0 comments

A Looking Glass Half Empty, Part 1: Just Lookin' for a Hit

https://www.filfre.net/2025/10/a-looking-glass-half-empty-part-1-just-lookin-for-a-hit/
2•doppp•24m ago•0 comments

AI has designed thousands of potential antibiotics. Will any work?

https://www.nature.com/articles/d41586-025-03201-6
2•pykello•25m ago•0 comments

Illinois utility tries using electric school buses for bidirectional charging

https://insideclimatenews.org/news/03102025/illinois-electric-school-bus-vehicle-to-grid-program/
1•thunderbong•27m ago•0 comments

Startup Offer - what's your comp package worth?

https://www.startupoffer.help/
1•rookieluva94•30m ago•0 comments

Stripe AI-Powered Smart Disputes

https://docs.stripe.com/disputes/smart-disputes
1•rendaw•30m ago•0 comments
Open in hackernews

Microsoft CTO says he wants to swap most AMD and Nvidia GPUs for homemade chips

https://www.theregister.com/2025/10/02/microsoft_maia_dc/
77•fork-bomber•1h ago

Comments

okokwhatever•1h ago
somebody wants to buy NVDA cheaper... ;)
synergy20•1h ago
plus someone has no leverage whatsoever other than talking
ambicapter•1h ago
Microsoft, famously resource-poor.
fidotron•1h ago
They have so much money it is harmful to their ability to execute.

Just look at the implosion of the XBox business.

fennecbutt•14m ago
Granted, if everyone had done what the highly paid executives had told them to do, xbox would never have existed.

And I'm guessing that the decline is due to executive meddling.

What is it that executives do again? Beyond collecting many millions of dollars a year, that is.

fidotron•10m ago
They sit around fantasizing about buying Nintendo because that would be the top thing they could achieve in their careers.
rjbwork•1h ago
Guess MSFT needs somewhere else AI adjacent to funnel money into to produce the illusion of growth and future cash flow in this bubblified environment.
balls187•1h ago
> produce the illusion of growth and future cash flow in this bubblified environment.

I was ranting about this to my friends; Wallstreet is now banking on Tech firms to produce the illusion of growth and returns, rather than repackaging and selling subprime mortgages.

The tech sector seems to have a never ending supply of things to spur investment and growth: cloud computing, saas, mobile, social media, IoT, crypto, Metaverse, and now AI.

Some useful, some not so much.

Tech firms have a lot of pressure to produce growth, it's filled with very smart people, and wields influence on public policy. The flip side is the mortage crisis, at least before it collapsed, got more Americans into home ownership (even if they weren't ready for it). I'm not sure the tech sectors meteoric rise has been as helpful (sentiment of locals in US tech hubs suggests a overall feeling of dissatisfaction with tech)

giancarlostoro•1h ago
So similar to Apple Silicon. If this means they'll be on par with Apple Silicon I'm okay with this, I'm surprised they didn't do this sooner for their Surface devices.

Oh right, for their data centers. I could see this being useful there too, brings costs down lower.

CharlesW•1h ago
> So similar to Apple Silicon.

Yes, in the sense that this is at least partially inspired by Apple's vertical integration playbook, which has now been extended to their own data centers based on custom Apple Silicon¹ and a built-for-purpose, hardened edition of Darwin².

¹ https://security.apple.com/blog/private-cloud-compute/ ² https://en.wikipedia.org/wiki/Darwin_(operating_system)

giancarlostoro•1h ago
Yeah, its interesting, years ago I never thought Apple nor Microsoft would do this, but also Google has done this on their cloud as well, so it makes sense.
georgeburdell•50m ago
Vertical integration only works if your internal teams can stay in the race at each level well enough to keep the stack competitive as a whole. Microsoft can’t attract the same level of talent as Apple because their pay is close to the industry median
quadrature•1h ago
Not suprising that the hyperscalers will make this decision for inference and maybe even a large chunk of training. I wonder if it will spur nvidia to work on an inference only accelerator.
edude03•1h ago
> I wonder if it will spur nvidia to work on an inference only accelerator.

Arguably that's a GPU? Other than (currently) exotic ways to run LLMs like photonics or giant SRAM tiles there isn't a device that's better at inference than GPUs and they have the benefit that they can be used for training as well. You need the same amount of memory and the same ability to do math as fast as possible whether its inference or training.

CharlesW•1h ago
> Arguably that's a GPU?

Yes, and to @quadrature's point, NVIDIA is creating GPUs explicitly focused on inference, like the Rubin CPX: https://www.tomshardware.com/pc-components/gpus/nvidias-new-...

"…the company announced its approach to solving that problem with its Rubin CPX— Content Phase aXcelerator — that will sit next to Rubin GPUs and Vera CPUs to accelerate specific workloads."

edude03•49m ago
Yeah, I'm probably splitting hairs here but as far as I understand (and honestly maybe I don't understand) - Rubin CPX is "just" a normal GPU with GDDR instead of HBM.

In fact - I'd say we're looking at this backwards - GPUs used to be the thing that did math fast and put the result into a buffer where something else could draw it to a screen. Now a "GPU" is still a thing that does math fast, but now sometimes, you don't include the hardware to put the pixels on a screen.

So maybe - CPX is "just" a GPU but with more generic naming that aligns with its use cases.

bonestamp2•17m ago
There are some inference chips that are fundamentally different from GPUs. For example, one of the guys who designed Google's original TPU left and started a company (with some other engineers) called groq ai (not to be confused with grok ai). They make a chip that is quite different from a GPU and provides several advantages for inference over traditional GPUs:

https://www.cdotrends.com/story/3823/groq-ai-chip-delivers-b...

AzN1337c0d3r•34m ago
I would submit Google's TPUs are not GPUs.

Similarly, Tenstorrent seems to be building something that you could consider "better", at least insofar that the goal is to be open.

quadrature•31m ago
I'm not very well versed, but i believe that training requires more memory to store intermediate computations so that you can calculate gradients for each layer.
hkt•1h ago
Even just saying this applies downward pressure on pricing: NVIDIA has an enormous amount of market power (~"excess" profit) right now and there aren't enough near competitors to drive that down. The only thing that will work is their biggest _consumers_ investing, or threatening to invest, if their prices are too high.

Long term, I wonder if we're exiting the "platform compute" era, for want of a better term. By that I mean compute which can run more or less any operating system, software, etc. If everyone is siloed into their own vertically integrated hardware+operating system stack, the results will be awful for free software.

startupsfail•52m ago
In that case, it's great that Microsoft is building their silicon. Keeps NVIDIA in check, otherwise these profits would evaporate into nonsense and NVIDIA would lose the AI industry to competition from China. Which, depending if AGI/ASI is possible or not, may or may not be a great move.
harrall•1h ago
Google has been using its own TPU silicon for machine learning since 2015.

I think they do all deep learning for Gemini on ther own silicon.

But they also invented AI as we know it when they introduced transformer architecture and they’ve been more invested in machine learning than most companies for a very long time.

chrismustcode•1h ago
I thought they use GPU for learning and TPU for inference, I’m open to been corrected.
cendyne•1h ago
I've heard its a mixture because they can't source enough in-house compute
xnx•1h ago
Some details here: https://news.ycombinator.com/item?id=42392310
surajrmal•1h ago
The first tpu they made was inference only. Everything since has been used for training. I think that means they weren't using it for training in 2015 but rather 2017 based on Wikipedia.
lokar•40m ago
The first TPU they *announced" was for inference
dekhn•57m ago
no. for internal training most work is done on TPUs, which have been explicitly designed for high performance training.
buildbot•1h ago
Not that it matters, but Microsoft has been doing AI accelerators for a bit too - project Brainwave has been around since 2018 - https://blogs.microsoft.com/ai/build-2018-project-brainwave/
rjzzleep•39m ago
The first revisions were stuff made by qualcomm right? I don't think we have much data on how much customizations they make and where they their IP from, but given how much of the Tensor cores comes from Samsung I think it's safe to say to assume that there is a decent amount coming from some of the big vendors.
Moto7451•34m ago
To be fair that’s a pretty good approach if you look at Apple’s progression from assembled IPs in the first iPhone CPU to the A and M series.
rjzzleep•15m ago
Yeah but at least when it comes to mobile CPUs Apple seemed vastly more competent in how they approached it.
Der_Einzige•16m ago
I'm 99.999% sure that the claim of "all deep learning for Gemini on their own silicon" is not true.

Maybe if you restrict it similarly to the Deepseek paper to "Gemini uses TPU for the final successful training run and for scaled inference" you might be correct, but there's no way that GPUs aren't involved for at minimum comparability and more rapid iteration reasons during the extremely buggy and error prone point of getting to the final training run. Certainly the theoretical and algorithmic innovations that are often being done at Google and do make their way into Gemini also sometimes using Nvidia GPUs.

GCP has a lot of, likely on the order of at least 1 million GPUs in their fleet today (I'm likely underestimating). Some of that is used internally and is made available to their engineering staff. What constitutes "deep learning for gemini" is very up to interpretation.

leshokunin•1h ago
Honestly this would be great for competition. Would love to see them impish in that direction.
bee_rider•1h ago
The most important note is:

> The software titan is rather late to the custom silicon party. While Amazon and Google have been building custom CPUs and AI accelerators for years, Microsoft only revealed its Maia AI accelerators in late 2023.

They are too late for now, they realistically hardware takes a couple generations to become a serious contender and by the time Microsoft has a chance to learn from their hardware mistakes the “AI” bubble will have popped.

But, there will probably be some little LLM tools that do end up having practical value; maybe there will be a happy line-crossing point for MS and they’ll have cheap in-house compute when the models actually need to be able to turn a profit.

surajrmal•1h ago
At this point it will take a lot of investment to catch up. Google relies heavily on specialized interconnects to build massive tpu clusters. It's more than just designing a chip these days. Folks who work on interconnects are a lot more rare than engineers who can design chips.
alephnerd•40m ago
> hardware takes a couple generations to become a serious contender

Not really and for the same reason Chinese players like Biren are leapfrogging - much of the work profile in AI/ML is "embarrassingly parallel".

If you are able to negotiate competitive fabrication and energy supply deals, you can mass produce your way into providing "good enough" performance.

The persona who cares about hardware performance in training isn't in the market for cloud offered services.

cjbgkagh•1h ago
I guess Microsoft’s investment into Graphcore didn’t pay off. Not sure what they’re planning but more of that isn’t going to cut it. At the time (late 2019) I was arguing for either a GPU approach or specialized architecture targeting transformers.

There was a split at MS where the ‘Next Gen’ bayesian was being done in the US and the frequentist work was being shipped off to China. Chris Bishop was promoted to head of MSR Cambridge which didn’t help.

Microsoft really is an institutionally stupid organization so I have no idea on which direction they actually go. My best guess is that it’s all talk.

Den_VR•46m ago
Microsoft lacks the credibility and track record for this to be anything but talk. Hardware doesn’t simply go from zero to gigawatts of infrastructure on talk. Even Apple is better positioned for such a thing.
latchkey•1h ago
It always falls back on the software. AMD is behind, not because the hardware is bad, but because their software historically has played second fiddle to their hardware. The CUDA moat is real.

So, unless they also solve that issue with their own hardware, then it will be like the TPU, which is limited to usage primarily at Google, or within very specific use cases.

There are only so many super talented software engineers to go around. If you're going to become an expert in something, you're going to pick what everyone else is using first.

amelius•59m ago
> The CUDA moat is real.

I don't know. The transformer architecture uses only a limited number of primitives. Once you have ported those to your new architecture, you're good to go.

Also, Google has been using TPUs for a long time now, and __they__ never hit a brick wall for a lack of CUDA.

outside1234•1h ago
For GPUs at least this is pretty obvious. For CPUs it is less clear to me that they can do it more efficiently.
sgerenser•53m ago
When Microsoft talks about “making their own CPUs,” they just mean putting together a large number of off-the-shelf Arm Neoverse cores into their own SoC, not designing a fully custom CPU. This is the same thing that Google and Amazon are doing as well.
qwertytyyuu•1h ago
better late than never to get into to game... right? right....?
johncolanduoni•56m ago
Just like mobile!
alephnerd•49m ago
I've mentioned this before on HN [0][1].

The name of the game has been custom SoCs and ASICs for a couple years now, because inference and model training is an "embarrassingly parallel" problem, and models that are optimized for older hardware can provide similar gains to models that are run on unoptimized but more performant hardware.

Same reason H100s remain a mainstay in the industry today, as their performance profile is well understood now.

[0] - https://news.ycombinator.com/item?id=45275413

[1] - https://news.ycombinator.com/item?id=43383418

philipwhiuk•21m ago
> The name of the game has been custom SoCs and ASICs for a couple years now, because inference and model training is an "embarrassingly parallel" problem, and models that are optimized for older hardware can provide similar gains to models that are run on unoptimized but more performant hardware.

Is anyone else getting crypto flashbacks?

bravetraveler•41m ago
I'll say what my Dad said before he decided to truly live it up:

    wish with one hand and shit in the other, see which fills first
croisillon•32m ago
homemade chips is probably a lot of fun but buying regular Lay's is so much easier
babuloseo•17m ago
The current M$ sure is doing a great job at making people move to alternatives.
floxy•14m ago
On a slightly different tangent, is anyone working on analog machine learning ASICs? Sub-threshold CMOS or something? I mean even at the research level? Using a handful of transistor for an analog multiplier. And get all of the crazy fascinating translinear stuff of Barrie Gilbert fame.

https://www.electronicdesign.com/technologies/analog/article...

https://www.analog.com/en/resources/analog-dialogue/articles...

http://madvlsi.olin.edu/bminch/talks/090402_atact.pdf