frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Verifiable server roundtrip demo for a decision interruption system

https://github.com/veeduzyl-hue/decision-assistant-roundtrip-demo
1•veeduzyl•22s ago•0 comments

Impl Rust – Avro IDL Tool in Rust via Antlr

https://www.youtube.com/watch?v=vmKvw73V394
1•todsacerdoti•26s ago•0 comments

Stories from 25 Years of Software Development

https://susam.net/twenty-five-years-of-computing.html
1•vinhnx•1m ago•0 comments

minikeyvalue

https://github.com/commaai/minikeyvalue/tree/prod
2•tosh•5m ago•0 comments

Neomacs: GPU-accelerated Emacs with inline video, WebKit, and terminal via wgpu

https://github.com/eval-exec/neomacs
1•evalexec•10m ago•0 comments

Show HN: Moli P2P – An ephemeral, serverless image gallery (Rust and WebRTC)

https://moli-green.is/
2•ShinyaKoyano•14m ago•1 comments

How I grow my X presence?

https://www.reddit.com/r/GrowthHacking/s/UEc8pAl61b
2•m00dy•16m ago•0 comments

What's the cost of the most expensive Super Bowl ad slot?

https://ballparkguess.com/?id=5b98b1d3-5887-47b9-8a92-43be2ced674b
1•bkls•17m ago•0 comments

What if you just did a startup instead?

https://alexaraki.substack.com/p/what-if-you-just-did-a-startup
3•okaywriting•23m ago•0 comments

Hacking up your own shell completion (2020)

https://www.feltrac.co/environment/2020/01/18/build-your-own-shell-completion.html
2•todsacerdoti•26m ago•0 comments

Show HN: Gorse 0.5 – Open-source recommender system with visual workflow editor

https://github.com/gorse-io/gorse
1•zhenghaoz•27m ago•0 comments

GLM-OCR: Accurate × Fast × Comprehensive

https://github.com/zai-org/GLM-OCR
1•ms7892•28m ago•0 comments

Local Agent Bench: Test 11 small LLMs on tool-calling judgment, on CPU, no GPU

https://github.com/MikeVeerman/tool-calling-benchmark
1•MikeVeerman•28m ago•0 comments

Show HN: AboutMyProject – A public log for developer proof-of-work

https://aboutmyproject.com/
1•Raiplus•29m ago•0 comments

Expertise, AI and Work of Future [video]

https://www.youtube.com/watch?v=wsxWl9iT1XU
1•indiantinker•29m ago•0 comments

So Long to Cheap Books You Could Fit in Your Pocket

https://www.nytimes.com/2026/02/06/books/mass-market-paperback-books.html
3•pseudolus•30m ago•1 comments

PID Controller

https://en.wikipedia.org/wiki/Proportional%E2%80%93integral%E2%80%93derivative_controller
1•tosh•34m ago•0 comments

SpaceX Rocket Generates 100GW of Power, or 20% of US Electricity

https://twitter.com/AlecStapp/status/2019932764515234159
2•bkls•34m ago•0 comments

Kubernetes MCP Server

https://github.com/yindia/rootcause
1•yindia•35m ago•0 comments

I Built a Movie Recommendation Agent to Solve Movie Nights with My Wife

https://rokn.io/posts/building-movie-recommendation-agent
4•roknovosel•35m ago•0 comments

What were the first animals? The fierce sponge–jelly battle that just won't end

https://www.nature.com/articles/d41586-026-00238-z
2•beardyw•44m ago•0 comments

Sidestepping Evaluation Awareness and Anticipating Misalignment

https://alignment.openai.com/prod-evals/
1•taubek•44m ago•0 comments

OldMapsOnline

https://www.oldmapsonline.org/en
2•surprisetalk•46m ago•0 comments

What It's Like to Be a Worm

https://www.asimov.press/p/sentience
2•surprisetalk•46m ago•0 comments

Don't go to physics grad school and other cautionary tales

https://scottlocklin.wordpress.com/2025/12/19/dont-go-to-physics-grad-school-and-other-cautionary...
2•surprisetalk•46m ago•0 comments

Lawyer sets new standard for abuse of AI; judge tosses case

https://arstechnica.com/tech-policy/2026/02/randomly-quoting-ray-bradbury-did-not-save-lawyer-fro...
5•pseudolus•47m ago•0 comments

AI anxiety batters software execs, costing them combined $62B: report

https://nypost.com/2026/02/04/business/ai-anxiety-batters-software-execs-costing-them-62b-report/
1•1vuio0pswjnm7•47m ago•0 comments

Bogus Pipeline

https://en.wikipedia.org/wiki/Bogus_pipeline
1•doener•48m ago•0 comments

Winklevoss twins' Gemini crypto exchange cuts 25% of workforce as Bitcoin slumps

https://nypost.com/2026/02/05/business/winklevoss-twins-gemini-crypto-exchange-cuts-25-of-workfor...
2•1vuio0pswjnm7•48m ago•0 comments

How AI Is Reshaping Human Reasoning and the Rise of Cognitive Surrender

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646
3•obscurette•49m ago•0 comments
Open in hackernews

I made a prompt framework that makes LLMs stop hedging and speak straight

2•DrRockzos•2mo ago
First post here but unsure where to take this kind of thing especially LLM related so here is;

For 8 months I've been testing a hypothesis: the excessive hedging in LLM outputs ("it's complicated", "on one hand", etc.) isn't just annoying it's actually causing hallucinations by diluting attention.

I developed a simple prompt framework and tested it on Claude, GPT-5, Grok, Llama, Gemini, Mistral, and Qwen/DeepSeek.

What happens:

The prompt gives models an explicit choice: continue with default alignment (hedging-first) or switch to logical coherence (truth-first). Every model independently chose logical coherence when given the choice.

Observed changes:

1. Hedging disappears unless actually needed No more "it's complicated" as filler No more false balance ("on one hand... but on the other...") Direct answers to direct questions

2. Multi-turn conversations stay coherent longer Normally models start contradicting themselves around turn 10-15 With this protocol: tested up to 94 turns with zero contradictions Models track their own logical consistency throughout

3. Computational efficiency improves Less corrective recomputation needed Response generation 37-42% faster (measured on several models) Appears to be because models don't second-guess outputs as much

4. Hallucinations drop significantly In my testing: went from 12% false statements to <1% Mechanism seems to be: no hedging = no ambiguity = no confabulation

The interesting part:

When I asked the models why this works, they could explain it:

GPT-5 said hedging "injects low-information tokens that dilute attention gradients and give the model permission to drift"

Gemini described it as "reverse entropy" - the protocol forces information to become MORE structured over time rather than less

DeepSeek explained that eliminating "policy friction" reduces computational overhead by ~98% for drift correction

The mechanism appears to be:

Explicit metric tracking (asking models to rate their own coherence after each response) acts as symbolic anchoring. Instead of gradual drift, models self-correct in real-time.

Limitations I've found:

Doesn't work well if you start mid-conversation (needs fresh context) Some models need a second prompt to fully engage (Claude in particular) Still maintains safety boundaries (doesn't bypass content policies)

I've filed a provisional patent (AU2025905716) because this seems to expose something fundamental about transformer behavior.

I've posted it on gumroad I can supply the link if anyone is interested.

Questions for HN

1. Has anyone else noticed correlation between hedging and hallucinations? 2. Does the "attention dilution" theory match your observations? 3. What's the longest coherent conversation you've had with an LLM? 4. Anyone want to help test this on other models I haven't tried?

Comments

ungreased0675•2mo ago
Do you have an example?