frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

I put my whole life into a single database

https://howisfelix.today/
139•lukakopajtic•2h ago•39 comments

Redox OS has adopted a Certificate of Origin policy and a strict no-LLM policy

https://gitlab.redox-os.org/redox-os/redox/-/blob/master/CONTRIBUTING.md
219•pjmlp•4h ago•191 comments

FreeBSD 14.4-Release Announcement

https://www.freebsd.org/releases/14.4R/announce/
31•vermaden•2h ago•2 comments

The Gervais Principle, or the Office According to "The Office" (2009)

https://www.ribbonfarm.com/2009/10/07/the-gervais-principle-or-the-office-according-to-the-office/
103•janandonly•3d ago•28 comments

Ask HN: Remember Fidonet?

43•ukkare•1h ago•25 comments

Practical Guide to Bare Metal C++

https://arobenko.github.io/bare_metal_cpp/#_abstract_classes
26•ibobev•3d ago•4 comments

Two Years of Emacs Solo

https://www.rahuljuliato.com/posts/emacs-solo-two-years
289•celadevra_•12h ago•93 comments

Lotus 1-2-3 on the PC with DOS

https://stonetools.ghost.io/lotus123-dos/
122•TMWNN•3d ago•42 comments

LoGeR – 3D reconstruction from extremely long videos (DeepMind, UC Berkeley)

https://loger-project.github.io
80•helloplanets•6h ago•21 comments

TCXO Failure Analysis

https://serd.es/2026/03/06/TCXO-failure-analysis.html
44•zdw•3d ago•12 comments

Caxlsx: Ruby gem for xlsx generation with charts, images, schema validation

https://github.com/caxlsx/caxlsx
4•earcar•3d ago•0 comments

Building a Procedural Hex Map with Wave Function Collapse

https://felixturner.github.io/hex-map-wfc/article/
527•imadr•19h ago•74 comments

The hidden compile-time cost of C++26 reflection

https://vittorioromeo.com/index/blog/refl_compiletime.html
44•SuperV1234•3d ago•18 comments

Optimizing Top K in Postgres

https://www.paradedb.com/blog/optimizing-top-k
105•philippemnoel•1d ago•12 comments

No, it doesn't cost Anthropic $5k per Claude Code user

https://martinalderson.com/posts/no-it-doesnt-cost-anthropic-5k-per-claude-code-user/
291•jnord•13h ago•211 comments

Is legal the same as legitimate: AI reimplementation and the erosion of copyleft

https://writings.hongminhee.org/2026/03/legal-vs-legitimate/
502•dahlia•21h ago•517 comments

JSLinux Now Supports x86_64

https://bellard.org/jslinux/
341•TechTechTech•20h ago•109 comments

Show HN: Remotely use my guitar tuner

https://realtuner.online/
222•smith-kyle•3d ago•48 comments

Yann LeCun's AI startup raises $1B in Europe's largest ever seed round

https://www.ft.com/content/e5245ec3-1a58-4eff-ab58-480b6259aaf1
108•ottomengis•2h ago•81 comments

Show HN: I Was Here – Draw on street view, others can find your drawings

https://washere.live
43•mrktsm__•7h ago•32 comments

Darkrealms BBS

http://www.darkrealms.ca/
115•TigerUniversity•3d ago•32 comments

macOS Tahoe windows have different corner radiuses

https://lapcatsoftware.com/articles/2026/3/1.html
219•robenkleene•3d ago•163 comments

EVi, a Hard-Fork of Vim

https://codeberg.org/NerdNextDoor/evi
15•todsacerdoti•3h ago•7 comments

Learnings from paying artists royalties for AI-generated art

https://www.kapwing.com/blog/learnings-from-paying-artists-royalties-for-ai-generated-art/
152•jenthoven•10h ago•121 comments

The “JVG algorithm” only wins on tiny numbers

https://scottaaronson.blog/?p=9615
76•jhalderm•12h ago•41 comments

Graphing how the 10k* most common English words define each other

https://wyattsell.com/experiments/word-graph/
88•wyattsell•2d ago•21 comments

Show HN: DenchClaw – Local CRM on Top of OpenClaw

https://github.com/DenchHQ/DenchClaw
127•kumar_abhirup•22h ago•108 comments

DARPA’s new X-76

https://www.darpa.mil/news/2026/darpa-new-x-76-speed-of-jet-freedom-of-helicopter
209•newer_vienna•20h ago•197 comments

Launch HN: Terminal Use (YC W26) – Vercel for filesystem-based agents

107•filipbalucha•20h ago•76 comments

An opinionated take on how to do important research that matters

https://nicholas.carlini.com/writing/2026/how-to-win-a-best-paper-award.html
143•mad•20h ago•36 comments
Open in hackernews

Yann LeCun's AI startup raises $1B in Europe's largest ever seed round

https://www.ft.com/content/e5245ec3-1a58-4eff-ab58-480b6259aaf1
108•ottomengis•2h ago

Comments

abmmgb•2h ago
Not based on true valuation unless h-index has become a valuation metric lol

Academics don’t always make great entrepeneurs

A_D_E_P_T•1h ago
Justifiable.

There are a lot more degrees of freedom in world models.

LLMs are fundamentally capped because they only learn from static text -- human communications about the world -- rather than from the world itself, which is why they can remix existing ideas but find it all but impossible to produce genuinely novel discoveries or inventions. A well-funded and well-run startup building physical world models (grounded in spatiotemporal understanding, not just language patterns) would be attacking what I see as the actual bottleneck to AGI. Even if they succeed only partially, they may unlock the kind of generalization and creative spark that current LLMs structurally can't reach.

10xDev•1h ago
Whether it is text or an image, it is just bits for a computer. A token can represent anything.
A_D_E_P_T•1h ago
Sure, but don't conflate the representation format with the structure of what's being represented.

Everything is bits to a computer, but text training data captures the flattened, after-the-fact residue of baseline human thought: Someone's written description of how something works. (At best!)

A world model would need to capture the underlying causal, spatial, and temporal structure of reality itself -- the thing itself, that which generates those descriptions.

You can tokenize an image just as easily as a sentence, sure, but a pile of images and text won't give you a relation between the system and the world. A world model, in theory, can. I mean, we ought to be sufficient proof of this, in a sense...

firecall•1h ago
It’s worth noting how our human relationship or understanding of our world model changed as our tools to inspect and describe our world advanced.

So when we think about capturing any underlying structure of reality itself, we are constrained by the tools at hand.

The capability of the tool forms the description which grants the level of understanding.

bsenftner•1h ago
There will be no "unlocking of AGI" until we develop a new science capable of artificial comprehension. Comprehension is the cornucopia that produces everything we are, given raw stimulus an entire communicating Universe is generated with a plethora of highly advanceds predator/prey characters in an infinitely complex dynamic, and human science and technology have no lead how to artificially make sense of that in a simultaneous unifying whole. That's comprehension.
chilmers•1h ago
Ironically, your comment is practically incomprehensible.
copperx•1h ago
These two comments above me capture Slashdot in the early 2000s.
rvz•1h ago
A lot more justifiable than say, Thinking Machines at least. But we will "see".

World models and vision seems like a great use case for robotics which I can imagine that being the main driver of AMI.

andy12_•1h ago
I don't understand this view. How I see it the fundamental bottleneck to AGI is continual learning and backpropagation. Models today are static, and human brains don't learn or adapt themselves with anything close to backpropagation. World models don't solve any of these problems; they are fundamentally the same kind of deep learning architectures we are used to work with. Heck, if you think learning from the world itself is the bottleneck, you can just put a vision-action LLM on a reinforcement learning loop in a robotic/simulated body.
zelphirkalt•1h ago
> I don't understand this view. How I see it the fundamental bottleneck to AGI is continual learning and backpropagation. Models today are static, and human brains don't learn or adapt themselves with anything close to backpropagation.

Even with continuous backpropagation and "learning", enriching the training data, so called online-learning, the limitations will not disappear. The LLMs will not be able to conclude things about the world based on fact and deduction. They only consider what is likely from their training data. They will not foresee/anticipate events, that are unlikely or non-existent in their training data, but are bound to happen due to real world circumstances. They are not intelligent in that way.

Whether humans always apply that much effort to conclude these things is another question. The point is, that humans fundamentally are capable of doing that, while LLMs are structurally not.

The problems are structural/architectural. I think it will take another 2-3 major leaps in architectures, before these AI models reach human level general intelligence, if they ever reach it. So far they can "merely" often "fake it" when things are statistically common in their training data.

andy12_•1h ago
> Even with continuous backpropagation and "learning"

That's what I said. Backpropagation cannot be enough; that's not how neurons work in the slightest. When you put biological neurons in a Pong environment they learn to play not through some kind of loss or reward function; they self-organize to avoid unpredictable stimulation. As far as I know, no architecture learns in such an unsupervised way.

https://www.sciencedirect.com/science/article/pii/S089662732...

wiz21c•1h ago
I'm sure that if a car appeared from nowhere in the middle of your living room, you would not be prepared at all.

So my question is: when is there enough training data that you can handle 99.99% of the world ?

jstummbillig•25m ago
> They will not foresee/anticipate events, that are unlikely or non-existent in their training data, but are bound to happen due to real world circumstances. They are not intelligent in that way.

Can you be a bit more specific at all bounds? Maybe via an example?

energy123•1h ago
I don't understand why online learning is that necessary. If you took Einstein at 40 and surgically removed his hippocampus so he can't learn anything he didn't already know (meaning no online learning), that's still a very useful AGI. A hippocampus is a nice upgrade to that, but not super obviously on the critical path.
andy12_•1h ago
That's true. Though could that hippocampus-less Einstein be able to keep making novel complex discoveries from that point forward? Seems difficult. He would rapidly reach the limits of his short term memory (the same way current models rapidly reach the limits of their context windows).
zelphirkalt•1h ago
I guess the sheer amount and also variety of information you would need to pre-encode to get an Einstein at 40 is huge. Every day stream of high resolution video feed and actions and consequences and thoughts and ideas he has had until the age of 40 of every single moment. That includes social interactions, like a conversation and mimic of the other person in combination with what was said and background knowledge about the other person. Even a single conversation's data is a huge amount of data.

But one might say that the brain is not lossless ... True, good point. But in what way is it lossy? Can that be simulated well enough to learn an Einstein? What gives events significance is very subjective.

andsoitis•1h ago
Where does that training data come from?
A_D_E_P_T•1h ago
You could have continual learning on text and still be stuck in the same "remixing baseline human communications" trap. It's a nasty one, very hard to avoid, possibly even structurally unavoidable.

As for the "just put a vision LLM in a robot body" suggestion: People are trying this (e.g. Physical Intelligence) and it looks like it's extraordinarily hard! The results so far suggest that bolting perception and embodiment onto a language-model core doesn't produce any kind of causal understanding. The architecture behind the integration of sensory streams, persistent object representations, and modeling time and causality is critically important... and that's where world models come in.

ben_w•1h ago
> Models today are static, and human brains don't learn or adapt themselves with anything close to backpropagation.

While I suspect latter is a real problem (because all mammal brains* are much more example-efficient than all ML), the former is more about productisation than a fundamental thing: the models can be continuously updated already, but that makes it hard to deal with regressions. You kinda want an artefact with a version stamp that doesn't change itself before you release the update, especially as this isn't like normal software where specific features can be toggled on or off in isolation of everything else.

* I think. Also, I'm saying "mammal" because of an absence of evidence (to my *totally amateur* skill level) not evidence of absence.

10xDev•1h ago
The fact that models aren't continually updating seems more like a feature. I want to know the model is exactly the same as it was the last time I used it. Any new information it needs can be stored in its context window or stored in a file to read the next it needs to access it.
kergonath•1h ago
> The fact that models aren't continually updating seems more like a feature.

I think this is true to some extent: we like our tools to be predictable. But we’ve already made one jump by going from deterministic programs to stochastic models. I am sure the moment a self-evolutive AI shows up that clears the "useful enough" threshold we’ll make that jump as well.

energy123•1h ago
why LLMs (transformers trained on multimodal token sequences, potentially containing spatiotemporal information) can't be a world model?
ForHackernews•1h ago
https://medium.com/state-of-the-art-technology/world-models-...

> One major critique LeCun raises is that LLMs operate only in the realm of language, which is a simple, discrete space compared to the continuous, complex physical world we live in. LLMs can solve math problems or answer trivia because such tasks reduce to pattern completion on text, but they lack any meaningful grounding in physical reality. LeCun points out a striking paradox: we now have language models that can pass the bar exam, solve equations, and compute integrals, yet “where is our domestic robot? Where is a robot that’s as good as a cat in the physical world?” Even a house cat effortlessly navigates the 3D world and manipulates objects — abilities that current AI notably lacks. As LeCun observes, “We don’t think the tasks that a cat can accomplish are smart, but in fact, they are.”

energy123•1h ago
But they don't only operate on language? They operate on token sequences, which can be images, coordinates, time, language, etc.
kergonath•57m ago
It’s an interesting observation, but I think you have it backwards. The examples you give are all using discrete symbols to represent something real and communicating this description to other entities. I would argue that all your examples are languages.
samrus•9m ago
Whats the first L stand for? Thats not just vestogial, their model of the world is formed almost exclusively from language rather than a range of things contributing significantly like for humans.

The biggest thing thats missing is actual feedback to their decisions. They have no "idea of that because transformers and embeddings dont model that yet. And langiage descriptions and image representations of feedback arent enough. They are too disjointed. It needs more

Unearned5161•1h ago
I have a pet peeve with the concept of "a genuinely novel discovery or invention", what do you imagine this to be? Can you point me towards a discovery or invention that was "genuinely novel", ever?

I don't think it makes sense conceptually unless you're literally referring to discovering new physical things like elements or something.

Humans are remixers of ideas. That's all we do all the time. Our thoughts and actions are dictated by our environment and memories; everything must necessarily be built up from pre-existing parts.

A_D_E_P_T•1h ago
Suno is transformer-based; in a way it's a heavily modified LLM.

You can't get Suno to do anything that's not in its training data. It is physically incapable of inventing a new musical genre. No matter how detailed the instructions you give it, and even if you cheat and provide it with actual MP3 examples of what you want it to create, it is impossible.

The same goes for LLMs and invention generally, which is why they've made no important scientific discoveries.

You can learn a lot by playing with Suno.

davidfarrell•1h ago
W Brian Arthur's book "The Nature of Technology" provides a framework for classifying new technology as elemental vs innovative that I find helpful. For example the Huntley-Mcllroy diff operates on the phenomenon that ordered correspondence survives editing. That was an invention (discovery of a natural phenomenon and a means to harness it). Myers diff improves the performance by exploiting the fact that text changes are sparse. That's innovation. A python app using libdiff, that's engineering. And then you might say in terms of "descendants": invention > innovation > engineering. But it's just a perspective.
0x3f•56m ago
Novel things can be incremental. I don't think LLMs can do that either, at least I've never seen one do it.
general1465•1h ago
Here you can see why it is so hard to compete as European startup with US startups - abysmal access to money. Investment of 1B USD in Europe is glorified as largest seed ever, but in USA it is another Tuesday.
weego•1h ago
A billion seed is not an every day event anywhere.
mattmaroon•1h ago
Not at all. A quick google turns up evidence of 4. There may be more but I think probably not many.
s08148692•1h ago
For a foundation AI lab with a world famous AI researcher at the helm though, it's not so impressive. Won't even touch the sides of the hardware costs they'd need to be anywhere near competitive
oceansky•1h ago
A startup getting 1B net worth is so rare that such companies are called unicorns.

As the other commenter pointed out, this is 1B seed.

ArnoVW•1h ago
actually, they raised $1.03 billion at a $3.5 billion valuation.
compounding_it•1h ago
Europeans have free healthcare and retirement. They consider putting their money with long term benefits not just become CEO on Tuesday and declare bankruptcy on Wednesday.
MrBuddyCasino•1h ago
„free“
ExpertAdvisor01•1h ago
Free healthcare and retirement ?
ExpertAdvisor01•1h ago
It is an universal system but definitely not free . In Germany you pay on average 17.5% of your salary for healthcare insurance and 18.6% for retirement . However contribution caps exists . 70k for healthcare and 100k for retirement .
general1465•1h ago
It is not free, we just pay taxes.
ExpertAdvisor01•1h ago
Retirement is the worst. You are basically forced to pay into a unsustainable system ( at least in Germany ). It already has to be subsidized by taxes .
rvz•1h ago
Once again, US companies and VCs are in this seed round. Just like Mistral with their seed round.

Europe again missing out, until AMI reaches a much higher valuation with an obvious use case in robotics.

Either AMI reaches over $100B+ valuation (likely) or it becomes a Thinking Machines Lab with investors questioning its valuation. (very unlikely since world models has a use-case in vision and robotics)

thibaut_barrere•1h ago
It is well enough to attract worthy talents & produce interesting outcomes.
embedding-shape•1h ago
> Europe again missing out

I can't read the article, but American investors investing into European companies, isn't US the one missing out here? Or does "Europe" "win" when European investors invest in US companies? How does that work in your head?

insydian•1h ago
As someone in the tech twitter sphere this is yann and his ideas performing a suplex on LLM based companies. It is completely unfathomable to start an ai research company… Only sell off 20% and have 1 billion for screwing around for a few years.
insydian•1h ago
I liken this to watching a godzilla esque movie. Just grab some popcorn and enjoy the ride.
whyleyc•1h ago
https://archive.is/TEwfi
mentalgear•1h ago
Adds up : We are seeing a clear exodus of both capital and talent from the US - with the current US administration’s shift toward cronyism - and the EU stands as the most compelling alternative with a uniform market of 500 million people and the last major federation truly committed to the rule of law.
drstewart•1h ago
"Exodus of capital" as if OpenAI didn't just raise 115b
ZeroCool2u•1h ago
Regardless of your opinion of Yann or his views on auto regressive models being "sufficient" for what most would describe as AGI or ASI, this is probably a good thing for Europe. We need more well capitalized labs that aren't US or China centric and while I do like Mistral, they just haven't been keeping up on the frontier of model performance and seem like they've sort of pivoted into being integration specialists and consultants for EU corporations. That's fine and they've got to make money, but fully ceding the research front is not a good way to keep the EU competitive.
Brajeshwar•1h ago
There seem to be other news articles mentioning that they are setting up in Singapore as their base. https://www.straitstimes.com/business/ai-godfather-raises-1-...
Signez•1h ago
Hm, Singapour looks more like "one of their base"; they will have offices in Paris, Montréal, Singapour and New York (according to both this article and the interview Yann Le Cun did this morning on France Inter, the most listened radio in France).

Of course, each relevant newspaper on those areas highlight that it's coming to their place, but it really seems to be distributed.

re-thc•1h ago
> they are setting up in Singapore as their base

Europe in general has been tightening up their rules / taxes / laws around startups / companies especially tech and remote.

It's been less friendly. these days.

Signez•1h ago
Yann Le Cun litteraly said this morning on the radio in France that it is headquarted in Paris and will pay taxes in France. Go figure…
kvgr•1h ago
There will be no corporate taxes for a long time, so alls good.
mi_lk•1h ago
Doesn’t he live in New York himself? Although not sure if that matters depending on his role
roromainmain•34m ago
For such companies, France also offers generous R&D tax credits (Crédit Impôt Recherche): companies can recover roughly 30% of eligible R&D expenses incurred in France as a tax credit, which can eventually be refunded (in cash) if the company has no taxable profit.
ttoinou•22m ago
No he said something like “well yes, only for the parts of profits made in France”
stingraycharles•1h ago
That's a Singaporian newspaper, though; not sure if it's objectively their main base, or just one of them
fnands•50m ago
Probably just a satellite office.

Might be to be close to some of Yann's collaborators like Xavier Bresson at NUS

giancarlostoro•33m ago
I didn't really know who he was, so I went and found his wikipedia, which is written like either he wrote it himself to stroke his ego, or someone who likes him wrote it to stroke his ego:

> He is the Jacob T. Schwartz Professor of Computer Science at the Courant Institute of Mathematical Sciences at New York University. He served as Chief AI Scientist at Meta Platforms before leaving to work on his own startup company.

That entire sentence before the remarks about him service at Meta could have been axed, its weird to me when people compare themselves to someone else who is well known. It's the most Kanye West thing you can do. Mind you the more I read about him, the more I discovered he is in fact egotistical. Good luck having a serious engineering team with someone who is egotistical.

timr•27m ago
It's not comparing him to anyone. He has an endowed professorship. This is standard in academia, and you give the name because a) it's prestigious for the recipient and b) it strokes the ego of the donor.
pama•26m ago
You underestimate academia. Any academic that reads these two sentences only focuses on the first one: He has a named chair at Courant. In Germany, being a a Prof is added to your ID card/passport and becomes part of your official name, like knighthood in other countries.
bobwaycott•24m ago
That’s not a comparison to another person. That’s his job title. It is not uncommon for universities to have distinguished chairs within departments named after a notable person—in this case, the founder of NYU’s Department of Computer Science.
fs111•1h ago
https://archive.is/20260310070651/https://www.ft.com/content...
Oras•1h ago
> But this is not an applied AI company.

There is absolutely no doubt about Yann's impact on AI/ML, but he had access to many more resources in Meta, and we didn't see anything.

It could be a management issue, though, and I sincerely wish we will see more competition, but from what I quoted above, it does not seem like it.

Understanding world through videos (mentioned in the article), is just what video models have already done, and they are getting pretty good (see Seedance, Kling, Sora .. etc). So I'm not quite sure how what he proposed would work.

the_real_cher•1h ago
He was suffocated by the corporate aspect Meta I suspect.
_giorgio_•1h ago
I can’t reconcile this dichotomy: most of the landmark deep learning papers were developed with what, by today’s standards, were almost ridiculously small training budgets — from Transformers to dropout, and so on.

So I keep wondering: if his idea is really that good — and I genuinely hope it is — why hasn’t it led to anything truly groundbreaking yet? It can’t just be a matter of needing more data or more researchers. You tell me :-D

boccaff•55m ago
llama models pushed the envelope for a while, and having them "open-weight" allowed a lot of tinkering. I would say that most of fine tuned evolved from work on top of llama models.
oefrha•21m ago
Llama wasn’t Yann LeCun’s work and he was openly critical of LLMs, so it’s not very relevant in this context.

Source: himself https://x.com/ylecun/status/1993840625142436160 (“I never worked on any Llama.”) and a million previous reports and tweets from him.

npn•1h ago
I wish him luck.

Recently all papers are about LLM, it brings up fatigue.

As GPT is almost reaching its limit, new architecture could bring out new discovery.

margorczynski•1h ago
He couldn't achieve at least parity with LLMs during his days at Meta (and having at his disposal billions in resources most probably) but he'll succeed now? What is the pitch?
mkl•1h ago
Seems like it's the second largest seed round anywhere after Thinking Machines Labs? https://news.crunchbase.com/venture/biggest-seed-round-ai-th...

That article is from June 2025 so may be out of date, and the definition of "seed round" is a bit fuzzy.

_giorgio_•1h ago
Thinking Machines looks half-dead already.

The giant seed round proves investors were willing to fund Mira Murati, not that the company had built anything durable.

Within months, it had already lost cofounder Andrew Tulloch to Meta, then cofounders Barret Zoph and Luke Metz plus researcher Sam Schoenholz to OpenAI; WIRED also reported that at least three other researchers left. At that point, citing it as evidence of real competitive momentum feels weak.

sylware•54m ago
If, for even 1s, they get in a position which is threatening, in any way, Big Tech AI (mostly US based if not all), they will be raided by international finance to be dismantled and poached hardcore with some massive US "investment funds" (which looks more and more as "weaponized" international finance!!). Only china is very immune to international finance. Those funds have tens of thousands of billions of $, basically, in a world of money, there is near zero resistance.
itigges22•28m ago
I just saw a post from Yann mentioning that AMI Labs is hiring too!
secondary_op•10m ago
That being sad, Yann LeCun's twitter reposts are below average IQ.