frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Hallucination Is Inevitable: An Innate Limitation of Large Language Models

https://arxiv.org/abs/2401.11817
12•drob518•2h ago

Comments

bell-cot•2h ago
IANAL, nor expert in this space.

But might any such care to comment on the consequences, if this "it is impossible, even in theory, to eliminate LLM hallucinations" result holds up?

hilariously•1h ago
They are lossy statistical prediction machines - to eliminate hallucinations effectively eliminates the lossy part and you might as well just use predicates in a database of facts.
Jiro•1h ago
From that abstract it doesn't sound like they allowed for the possibility that the LLM could be trained to say "I don't know" for some things.
dwallin•1h ago
Yeah, they only “proved” hallucination is inevitable by defining it to be any case where the llm doesn’t provide the “correct” answer. By this definition, an LLM deciding not to answer is also a “hallucination”.
jaclebus81•1h ago
My intuition on this is like training a classifier on four classes: dog, cat, cow and IDK. It feels intuitive to us but really hard to do in practice. In the classifier case, we are leveraging a subset of data to train the model to give correct answers to unseen data. If we want the model to generalize to unseen data we need it to call unseen dog-like things a dog. If not, then all unseen dogs would be IDK. Learning that boundary of "known vs unknown" is very hard. If done poorly, you have a model that cannot abstract to anything that is not in the dataset which is a huge part of what makes these models so impressive. I'm sure there is more to it than this but I does not surprise me that it is an unsolved problem.
thomastjeffery•1h ago
Describing it as a limitation is the problem. Hallucination is the core feature. It's the only thing they do!
whythismatters•1h ago
>Submitted on 22 Jan 2024 (v1), last revised 13 Feb 2025 (this version, v2)
stevefan1999•1h ago
LLM or transformers just merely extracting signals from human text and build a "contextualized" predictor over a long sequence of words sorted by the information (technically it is attention) of each token, then generate sentences that way, one by one into other sequences at a time.

But the biggest problem is, even human itself is subjectable to hallucination. That is called being delusional, or being drugged. So it is inevitable from the first principle.

prewett•1h ago
Humans hallucinate, in the LLM sense, all the time. Did that sign really say that? Nope, I just extrapolated from the first three letters. In the Cambrian Explosion article on HN this morning, I thought the first line said that the earth was desolate. The second line didn't match up with that idea, so I read it again, and the first line said the opposite of what I thought. I particularly hallucinate things into emails from people at work that I disagree with, so much so that I've learned to wait until the next day to reply, and usually I find that they didn't say what I thought they said.
red75prime•1h ago
They prove that no finite amount of training data is enough to extrapolate an adversarially constructed non-continuous function. It's something akin to the no free lunch theorem (NFL).

No one uses the NFL to "prove" that LLMs can't learn to be the best optimizers, because it also proves that people can't be the best optimizers, but we manage somehow, so the theorem is irrelevant.

This is a fallacy of proving too much.

Days Without GitHub Incidents

https://www.dayswithoutgithubincident.com/
97•goalieca•39m ago•25 comments

Removable batteries in smartphones will be mandatory in the EU starting in 2027

https://www.ecopv-eu.com/en/blog-en/replaceable-smartphone-batteries-2027-eu-regulation/
481•rdeboo•3h ago•412 comments

US healthcare marketplaces shared citizenship and race data with ad tech giants

https://techcrunch.com/2026/05/04/us-healthcare-marketplaces-shared-citizenship-and-race-data-wit...
56•ZeidJ•59m ago•8 comments

Stop big tech from making users behave in ways they don't want to

https://economist.com/by-invitation/2026/04/29/stop-big-tech-from-making-users-behave-in-ways-the...
52•andsoitis•1h ago•17 comments

I am worried about Bun

https://wwj.dev/posts/i-am-worried-about-bun/
97•remote-dev•1h ago•42 comments

Does Employment Slow Cognitive Decline? Evidence from Labor Market Shocks

https://www.nber.org/papers/w35117
86•littlexsparkee•2h ago•48 comments

Redis array: short story of a long development process

https://antirez.com/news/164
133•antirez•3h ago•52 comments

GitHub Is Down

https://www.githubstatus.com/incidents/72q3n8yxthcy
356•gen220•2h ago•207 comments

Talking to 35 Strangers at the Gym

https://thienantran.com/talking-to-35-strangers-at-the-gym/
712•thitran•6h ago•357 comments

I tracked 7,700 UK petrol stations every 10 minutes for 3 months

https://www.fuelinsight.co.uk
77•theazureguy•3h ago•35 comments

Pomiferous: The most extensive apples (pommes) database

https://pomiferous.com/
37•Ariarule•3h ago•17 comments

PyInfra 3.8.0 Is Out

https://github.com/pyinfra-dev/pyinfra/releases/tag/v3.8.0
175•wowi42•5h ago•67 comments

GameStop makes $55.5B takeover offer for eBay

https://www.bbc.co.uk/news/articles/cn0p8yled1do
496•n1b0m•8h ago•432 comments

Trillions in Retirement Dollars Flow into Opaque Trusts

https://www.bloomberg.com/news/features/2026-05-03/trillions-in-us-retirement-dollars-flow-into-o...
30•koolhead17•54m ago•1 comments

OpenAI, Google, and Microsoft Back Bill to Fund 'AI Literacy' in Schools

https://www.404media.co/literacy-in-future-technologies-artificial-intelligence-act-adam-schiff-m...
54•cdrnsf•1h ago•46 comments

Alberta voter list leak is a potential public safety disaster

https://globalnews.ca/news/11828244/alberta-voter-list-leak-public-safety-disaster/
60•Teever•2h ago•42 comments

Newton's law of gravity passes its biggest test

https://www.science.org/content/article/newton-s-law-gravity-passes-its-biggest-test-ever
88•pseudolus•5h ago•60 comments

Offenders sentenced up to 10 years for spying on TSMC

https://www.taipeitimes.com/News/front/archives/2026/04/28/2003856358
6•ironyman•11m ago•0 comments

Using “underdrawings” for accurate text and numbers

https://samcollins.blog/underdrawings/
335•samcollins•3d ago•123 comments

DAG Workflow Engine

https://github.com/vivekg13186/Daisy-DAG
38•blobmty•5h ago•29 comments

Why are neural networks and cryptographic ciphers so similar? (2025)

https://reiner.org/neural-net-ciphers
92•jxmorris12•2d ago•29 comments

Texico: Learn the principles of programming without even touching a computer

https://www3.nhk.or.jp/nhkworld/en/shows/texico/
151•o4c•2d ago•11 comments

Trademark violation: Fake Notepad++ for Mac

https://notepad-plus-plus.org/news/npp-trademark-infringement/
517•maxloh•8h ago•227 comments

BYOMesh – New LoRa mesh radio offers 100x the bandwidth

https://partyon.xyz/@nullagent/116499715071759135
445•nullagent•1d ago•147 comments

Heat pump sales rise 17% across Europe in Q1 as energy prices surge

https://www.pv-magazine.com/2026/05/04/heat-pump-sales-rise-17-across-europe-in-q1-as-energy-pric...
11•doener•41m ago•1 comments

How Monero's proof of work works

https://blog.alcazarsec.com/tech/posts/how-moneros-proof-of-work-works
134•alcazar•4h ago•110 comments

DeepClaude – Claude Code agent loop with DeepSeek V4 Pro

https://github.com/aattaran/deepclaude
611•alattaran•20h ago•260 comments

Homebridge 2.0 is here, and it speaks Matter

https://www.theverge.com/tech/922877/homebridge-2-0-matter-update-robot-vacuums
28•Brajeshwar•2h ago•2 comments

1966 Ford Mustang Converted into a Tesla with Working 'Full Self-Driving'

https://electrek.co/2026/05/02/tesla-1966-mustang-ev-conversion-full-self-driving/
47•Brajeshwar•2h ago•35 comments

Discovering hard disk physical geometry through microbenchmarking (2019)

https://blog.stuffedcow.net/2019/09/hard-disk-geometry-microbenchmarking/
150•TapamN•3d ago•6 comments