frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Why are Windows semiannual updates named H1 and H2?

https://devblogs.microsoft.com/oldnewthing/20250805-00/?p=111435
1•dwattttt•31s ago•0 comments

How much is a pension worth?

https://venkii.substack.com/p/how-much-is-a-pension-worth
1•venkii•2m ago•0 comments

Denmark zoo asks for people to donate their pets to feed its predators

https://abcnews.go.com/International/denmark-zoo-asks-people-donate-pets-feed-predators/story?id=124367988
2•noleary•3m ago•0 comments

A New Theme for Emacs

https://github.com/mastro35/sixcolors-theme
1•mastro35•3m ago•1 comments

Ask HN: How Fit3d Works?

1•johngoodworks•6m ago•0 comments

Show HN: A new way to read company 10-Ks

https://www.proread.ai/company10ks
2•kanodiaashu•6m ago•1 comments

Generalization Gap in over‑Parameterized Models

https://www.gojiberries.io/generalization-gap-in-over-parameterized-models/
1•neehao•7m ago•0 comments

GPT-OSS Playground

https://gpt-oss.com/
1•twapi•7m ago•0 comments

Be Careful with Go Struct Embedding

https://mattjhall.co.uk/posts/be-careful-with-go-struct-embedding.html
1•ingve•10m ago•0 comments

Age Assurance on X

https://help.x.com/en/rules-and-policies/age-assurance
2•akyuu•14m ago•0 comments

SoftBrowse – Hide Reels, Explore, and Feed on Instagram (But Keep DMs)

2•softbrowse•15m ago•0 comments

The mystery of Alice in Wonderland syndrome

https://www.bbc.com/future/article/20230313-the-mystery-of-alice-in-wonderland-syndrome
1•amichail•15m ago•0 comments

OpenAI's new open weight (Apache 2) models are good

https://simonwillison.net/2025/Aug/5/gpt-oss/
3•JohnHammersley•16m ago•0 comments

The modern USD account built for global businesses

https://www.slash.com/products/global-usd
1•lhuser123•17m ago•1 comments

China Is Choking Supply of Critical Minerals to Western Defense Companies

https://www.wsj.com/world/asia/china-western-defense-industry-critical-minerals-3971ec51
5•yyyk•17m ago•2 comments

A Turning Point in Colon Cancer: Young People Are Finding It Earlier

https://www.wsj.com/health/healthcare/colon-cancer-screening-young-adults-5900a8a6
1•bookofjoe•20m ago•1 comments

Why Should We Worry About Declining Birth Rates?

https://jacobin.com/2025/07/demography-fertility-population-crisis-longtermism/
3•PaulHoule•25m ago•1 comments

A first look at GPT-OSS-120B's coding ability

https://blog.brokk.ai/a-first-look-at-gpt-oss-120bs-coding-ability/
3•jbellis•26m ago•0 comments

Embracing the Model Context Protocol in practice: An engineering deep-dive

https://xpander.ai/2025/03/31/how-xpander-ai-embraces-the-model-context-protocol-in-practice-an-engineering-deep-dive/
1•mooreds•31m ago•0 comments

AGI Blueprint - 424 pages– Visual Thought AGI Link

https://zenodo.org/records/15867575
3•derekv123•31m ago•1 comments

Canadian Court Rejects Reverse Class Action Lawsuit Against BitTorrent Pirates

https://torrentfreak.com/canadian-court-rejects-reverse-class-action-lawsuit-against-bittorrent-pirates/
4•gslin•32m ago•2 comments

uBlock Origin still works in Chrome 139

2•AuthorizedCust•32m ago•0 comments

ZK Proofs Are Getting Easier but the Airdrop Game Remains Tricky

1•Earlycrow•34m ago•0 comments

User Interfaces in Agentic CLI Tools: What Developers Need

https://thenewstack.io/user-interfaces-in-agentic-cli-tools-what-developers-need/
1•willm•35m ago•0 comments

Writing culture

https://splits.org/blog/writing-culture/
2•exolymph•35m ago•0 comments

Hammer Time: Scientists Have Figured Out Why Hammerheads Love Eating Other Shark

https://www.mentalfloss.com/animals/fish/why-hammerheads-love-eating-sharks
1•imzadi•36m ago•0 comments

What's the "Points" of Agile, Anyway?

https://spin.atomicobject.com/points-in-agile/
2•philk10•38m ago•0 comments

The New York Post Is Expanding to LA, Launching the California Post Next Year

https://nypost.com/2025/08/04/media/start-the-presses-new-york-post-will-expand-to-la-with-launch-of-the-california-post/
5•Bogdanp•38m ago•0 comments

Writing code was never the bottleneck

https://leaddev.com/velocity/writing-code-was-never-the-bottleneck
1•jrs235•39m ago•1 comments

Show HN: I made a playground and editor for generative AI models

https://mitte.ai
1•akoculu•40m ago•0 comments
Open in hackernews

AI hallucinations will be solvable within a year (2024)

https://fortune.com/2024/04/16/ai-hallucinations-solvable-year-ex-google-researcher/
11•rvz•2d ago

Comments

techpineapple•2d ago
I wonder if it would be better to have 1 “perfect” LLM trying to solve problems or 5 intentionally biased LLM’s.
d00mB0t•2d ago
I'm so tired of these rich dweebs pontificating to everyone.
ipv6ipv4•2d ago
There is ample evidence that hallucinations are incurable in the best extant model of intelligence - people.
add-sub-mul-div•1d ago
Someday we'll figure out how to program computers to behave deterministically so that they can complement our human abilities rather than badly impersonate them.
Dylan16807•1d ago
Getting down to the level of a median helpful human with the same knowledge would be a massive step forward.

Getting down to the level of a moderately humble expert taking the time to double check would be almost as good as solving it.

davesmylie•2d ago
Obviously it's well over a year since this article was posted and if anything I've anecdotally noticed hallucinations getting more, not less, common.

Possibly/probably with another years experience with LLMs I'm just more attuned to noticing when they have lost the plot and are making shit up

BoorishBears•1d ago
RL for reasoning definitely introduces hallucinations, and sometimes it introduces a class of hallucinations that feels a lot worse than the classic ones.

I noticed OpenAI's models picked up a tendency to hold strong convictions on completely unknowable things.

"<suggests possible optimization> Implement this change and it will result in a 4.5% uplift in performance"

"<provides code> I ran the updated script 10 times and it completes 30.5 seconds faster than before on average"

It's bad it enough it convinces itself it did things it can't do, but then it goes further and hallucinates insights from the tasks it hallucinated itself doing in the first places!

I feel like lay people aren't ready for that. Normal hallucinations felt passive, like a slip up. To the unprepared, this becomes more like someone actively trying to sell their slip ups.

I'm not sure if it's a form of RL hacking making it through to the final model or what, but even OpenAI seems to have noticed it in testing based on their model cards.

alkyon•1d ago
The more accurate word would be confabulation
Wowfunhappy•1d ago
You lost this battle, sorry. It's not going to happen.

Both terms are "inaccurate" because we're talking about a computer program, not a person. However, at this point "hallucination" has been firmly cemented in public discourse. I don't work in tech, but all of my colleagues know what an AI hallucination is, as does my grandmother. It's only a matter of time until the word's alternate meaning gets added to the dictionary.

peterashford•1d ago
Correct. This is the way language works. It's annoying when you know what words mean but this is the way it is.
alkyon•1d ago
Maybe I lost this battle, but also in science the terminology evolves. If you replace AI hallucination with AI confabulation even your grandmaother would get it right. I also don't agree that both terms are equally inaccurate.
Wowfunhappy•1d ago
> Maybe I lost this battle, but also in science the terminology evolves.

Ah yes, science, where we have fixed stars that move, imaginary numbers that are real, and atoms that can be divided into smaller particles.

alkyon•1d ago
All of these examples are many centuries old, let's compare apples to apples.
alkyon•1d ago
Obviously, hallucination is by definition a perception, so it incorrectly anthropomorphizes AI models. On the other hand, the term confabulation involves filling in gaps with fabrication, exactly what LLMs do (aka bullshitting).
more_corn•1d ago
What an absurd prediction.
DoctorOetker•1d ago
In the article it is argued that the brainfarts could be beneficial for exploration of new ideas.

I don't agree. The "temperature" parameter should be used for this. Confabulation / bluff / hallucination / unfounded guesses are undesirable at low temperatures.

Wowfunhappy•1d ago
> “If you look at the models before they are fine-tuned on human preferences, they’re surprisingly well calibrated. So if you ask the model for its confidence to an answer—that confidence correlates really well with whether or not the model is telling the truth—we then train them on human preferences and undo this.

Now that is really interesting! I didn't realize RLHF did that.