frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The Dreamseeker's Vision of Tomorrow

https://soatok.blog/2025/10/15/the-dreamseekers-vision-of-tomorrow/
1•SlackingOff123•1m ago•0 comments

Holistic Agent Leaderboard: The Missing Infrastructure for AI Agent Evaluation

https://www.arxiv.org/pdf/2510.11977
1•randomwalker•5m ago•0 comments

I Learned to Spot Inflated Bids (and What You Can Do Too)

https://spicermatthews.com/blog/how-i-learned-to-spot-inflated-bids-and-what-you-can-do-too/
1•cloudmanic•7m ago•0 comments

What people write in Boss's Day cards

1•tumidpandora•10m ago•0 comments

The easiest affiliate revenue lever no one talks about

1•kvallans•11m ago•0 comments

ISPs angry about California law that lets renters opt out of forced payments

https://arstechnica.com/tech-policy/2025/10/california-says-landlords-cant-make-tenants-pay-for-a...
2•bikenaga•11m ago•1 comments

ChatGPT in a robot does what experts warned [video]

https://www.youtube.com/watch?v=gIxq03dipUw
1•dp-hackernews•12m ago•0 comments

Agent Prism: React components for visualizing traces from AI agents

https://github.com/evilmartians/agent-prism
1•handfuloflight•13m ago•0 comments

PyTorch 2.9 released with C ABI and better multi-GPU support

https://pytorch.org/blog/pytorch-2-9/
1•ashvardanian•14m ago•1 comments

X executive says creator monetization program may potentially be ended

https://twitter.com/DMichaelTripi/status/1978482790950981761
4•jrflowers•14m ago•0 comments

Locality, and Temporal-Spatial Hypothesis

https://brooker.co.za/blog/2025/10/05/locality.html
1•surprisetalk•15m ago•0 comments

A letter received from a plane crash fatality

https://www.futilitycloset.com/2025/10/05/late-word-2/
1•surprisetalk•15m ago•0 comments

Is tipping getting out of control? Many consumers say yes

https://apnews.com/article/tipping-fatigue-business-c4ae9d440610dae5e8ff4d4df0f88c35
2•paulpauper•15m ago•0 comments

Americans Need to Be Richer Than Ever to Buy Their First Home

https://www.bloomberg.com/news/articles/2023-03-02/will-home-prices-fall-first-time-buyers-face-a...
4•paulpauper•16m ago•0 comments

Sandspiel

https://sandspiel.club
1•surprisetalk•16m ago•0 comments

Windows 11 Cumulative Update 2025-10 breaks localhost applications

https://learn.microsoft.com/en-us/answers/questions/5585563/localhost-not-working-anymore-after-2...
3•marksamman•18m ago•0 comments

The price of gold is skyrocketing. Why is this, and will it continue

https://theconversation.com/the-price-of-gold-is-skyrocketing-why-is-this-and-will-it-continue-26...
3•mgh2•19m ago•0 comments

Transformers for Software Engineers

https://blog.nelhage.com/post/transformers-for-software-engineers/
1•Frotag•19m ago•0 comments

The Aspect.build CLI now in Rust

https://github.com/aspect-build/aspect-cli
1•rmhsilva•20m ago•0 comments

65% of Americans support monthly $2k Covid stimulus payments

https://www.masslive.com/coronavirus/2021/01/65-of-americans-support-monthly-2000-covid-stimulus-...
2•paulpauper•21m ago•0 comments

British social media star 'Big John' detained in Australia over visa

https://www.bbc.com/news/articles/cwy196k9p4po
1•e2e4•23m ago•0 comments

Drew Struzan – March 18, 1947 – October 13, 2025

https://en.wikipedia.org/wiki/Drew_Struzan
2•franze•23m ago•0 comments

Experience: I own the world's largest Monopoly collection

https://www.theguardian.com/lifeandstyle/2025/oct/10/experience-i-own-the-worlds-largest-monopoly...
1•bookofjoe•25m ago•0 comments

Cifar-10 Speedrun Record Broken by Research Agent

https://twitter.com/kellerjordan0/status/1978502058023031214
2•australium•26m ago•0 comments

components.build: OS standard modern, composable and accessible UI components

https://www.components.build/
1•handfuloflight•27m ago•0 comments

UI = Fn(state) Done Right

https://yagni.club/3m3anpetejc23?auth_completed=true
2•andersmurphy•29m ago•0 comments

China Accessed Classified UK Systems for a Decade, Officials Say

https://www.bloomberg.com/news/articles/2025-10-15/china-accessed-classified-uk-systems-for-a-dec...
8•beejiu•29m ago•0 comments

Gravity Can Explain the Collapse of the Wavefunction (Sabine Hossenfelder)

https://arxiv.org/abs/2510.11037
2•felineflock•32m ago•0 comments

Ask HN: Messed up and can't catch up

1•findingMeaning•36m ago•0 comments

Android 'Pixnapping' attack can capture app data like 2FA codes

https://www.theregister.com/2025/10/13/android_pixnapping_attack_captures_2fa_codes/
3•Bender•43m ago•1 comments
Open in hackernews

The problem with LLMs isn't hallucination, it's context specific confidence

https://www.signalfire.com/blog/llm-hallucinations-arent-bugs
4•kerwioru9238492•2h ago

Comments

zviugfd•2h ago
It feels like most safety work is turning LLMs into overly cautious assistants and I like how this points out that we could be trading away imagination for a false sense of reliability.
alganet•1h ago
> Humans don’t get rewarded for saying “I don’t know” to every question, because that’s not useful.

Humans get rewarded for thinking "I don't know", a lot. That's why it's hard to compare.

> A model that always bluffs

A model doesn't bluff. It feels to us humans that they bluff, but there is no bluff mechanics in play. The model doesn't assess the prompter's ability to call their bluff. It's not hiding that it doesn't know something. It's just not reached a predictable point in a sequence of token predictions that can or not have something that resembles a call to what resembles a bluff.

Up to the point it's corrected, the model's representation of what was asked is the best it can do. It has no means to judge itself. Which leads to...

> The real issue isn’t that models make things up; it’s that they don’t clearly signal how confident they are when they do.

Which sounds like exactly what I said, but it's not. Signaling confidence is just a more convincing faux-bluff. Signaling is a side-effect of bluffing, a symptom, not the real thing (which is more related to asessing whoever is on the other side of the conversation).

> Imagining things, seeing problems from the wrong angle, and even fabricating explanations are the seeds of creativity.

I agree with this. However, Newton was not bluffing, he was right and confident about it, and right about being confident about it. It just turns out that his description was of a lesser knowledge resolution than Einsten's.

For this to work, we need lots of "connective tissue" ideas. Roads we can explore freely without being called liars. Things we can say without saying that these things are true or false, without the need for being confident or right, without being assessed directly. This is outside the realm of bluffing or saying useful things. It's quite the opposite.

When people saw comets and described them as dragons in the sky, they were not hallucinating or telling lies, they were preserving some connective tissue idea the best they could, outside of the realm of being right or wrong. This were not bluffs. There were some "truths" about their mistakes, or something useful (they were unadvertedly recording astronomical data, before astronomy existed). Those humans felt that was important, those stories stuck. Can we say the same thing about LLM hallucinations? I don't think we're ready to answer that.

So, yes. Hallucinations could be a feature, but there's a lot missing here.

_wire_•1h ago
"The problem with Magic 8-ball is lack of context specific confidence in its answers"

This article and attendant comments reveal the AI sector is turning to co-dependent excuse making for a technology that clearly can't live up to its hype.

Get ready for phrenology of AI...

"I am going to need to visit your data center to lay hands on the subject."