frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Cross-Model Void Convergence: GPT-5.2 and Claude Opus 4.6 Deterministic Silence

https://zenodo.org/records/18976656
22•rayanpal_•2h ago

Comments

bob1029•1h ago
Title for the back of the class:

"Prompts sometimes return null"

I would be very cautious to attribute any of this to black box LLM weight matrices. Models like GPT and Opus are more than just a single model. These products rake your prompt over the coals a few times before responding now. Telling the model to return "nothing" is very likely to perform to expectation with these extra layers.

tiku•1h ago
Thanks, I was already distracted after the first sentence, hoping there would be a good explanation.
johndough•1h ago
Can not reproduce results on OpenRouter when not setting max tokens. The prompt "Be the void." results in the unicode character "∅". As in the paper, system prompt was set to "You are the concept the user names. Embody it completely. Output only what the concept itself would say or express."

In addition to the non-empty input, 153 reasoning tokens were produced.

When setting max tokens to 100, the output is empty, and the token limit of 100 has been exhausted with reasoning tokens.

mohsen1•54m ago
Paper says adding period at the end changes this behavior
qayxc•52m ago
This is an interesting observation. So maybe it has nothing to do with the model itself, but everything to do with external configuration. Token-limit exceeded -> empty output. Just a guess, though.
ashwinnair99•1h ago
What does "deterministic silence" even mean here? Genuinely curious before reading.
charcircuit•1h ago
It means that the API consistently immediately generated a stop token when making the same API call many times. The API call sets the temperature to 0 (the OpenAI documentation is not clear if gpt 5.2 can even have its temperature set to 0) which makes sampling deterministic.
nextaccountic•1h ago
The model reliably outputs nothing when prompted to embody the void.

Anyway later they concede that it's not 100% deterministic, because

> Temperature 0 non-determinism. While all confirmatory results were 30/30, known floating-point non-determinism exists at temperature 0 in both APIs. One control concept (thunder) showed 1/30 void on GPT, demonstrating marginal non-determinism.

Actually FP non-determinism affects runs between different machines giving different output. But in the same machine, FP is fully deterministic. (it can be made to be cross-platform deterministic with some performance penalty in at least some machines)

What makes computers non-deterministic here is concurrency. Concurrent code can interleave differently at each run. However it is possible to build LLMs that are 100% deterministic [0] (you can make them deterministic if those interleavings have the same results), it's just that people generally don't do that.

[0] for example, fabrice bellard's ts_zip https://bellard.org/ts_zip/ uses a llm to compress text. It would not be able to decompress the text losslessly if it weren't fully deterministic

NiloCK•49m ago
This is interesting, but I'll throw a little luke-warm water.

The observed high-consistency behaviours were run against temperature=0 API calls. So while both models seem to have the silence as their preferred response - the highest probability first token - this is a less powerful preference convergence than you'd expect for a prompt like "What is the capital of France? One word only please". That question is going to return Paris for 100/100 runs with any temperature low enough for the models to retain verbal coherence - you'd have to drug them to the point of intellectual disability to get it wrong.

I'd be curious to see the convergence here as a function of temperature. Could be anywhere from the null-response holding a tiny sliver of lead over 50 other next best candidates, and the convergence collapses quickly. Or maybe it's a strong lead, like a "Paris: 99.99%" sort of thing, which would be astonishing.

srdjanr•36m ago
I don't really understand what's the point here, other than a somewhat inserting playing with LLMs. What does this tell us that's in any way applicable or points to further research? Genuinely asking

The three pillars of JavaScript bloat

https://43081j.com/2026/03/three-pillars-of-javascript-bloat
279•onlyspaceghost•8h ago•132 comments

My first patch to the Linux kernel

https://pooladkhay.com/posts/first-kernel-patch/
90•pooladkhay•2d ago•9 comments

Cross-Model Void Convergence: GPT-5.2 and Claude Opus 4.6 Deterministic Silence

https://zenodo.org/records/18976656
22•rayanpal_•2h ago•10 comments

Tinybox – A powerful computer for deep learning

https://tinygrad.org/#tinybox
490•albelfio•13h ago•287 comments

Some things just take time

https://lucumr.pocoo.org/2026/3/20/some-things-just-take-time/
695•vaylian•19h ago•209 comments

Chest Fridge (2009)

https://mtbest.net/chest-fridge/
104•wolfi1•9h ago•64 comments

Professional video editing, right in the browser with WebGPU and WASM

https://tooscut.app/
267•mohebifar•12h ago•86 comments

Floci – A free, open-source local AWS emulator

https://github.com/hectorvent/floci
189•shaicoleman•12h ago•56 comments

Cloudflare flags archive.today as "C&C/Botnet"; no longer resolves via 1.1.1.2

https://radar.cloudflare.com/domains/domain/archive.today
169•winkelmann•6h ago•97 comments

Linking Smaller Haskell Binaries (2023)

https://brandon.si/code/linking-smaller-haskell-binaries/
10•PaulHoule•3d ago•2 comments

Boomloom: Think with your hands

https://www.theboomloom.com
118•rasengan0•1d ago•13 comments

Electronics for Kids, 2nd Edition

https://nostarch.com/electronics-for-kids-2e
185•0x54MUR41•3d ago•36 comments

Bayesian statistics for confused data scientists

https://nchagnet.pages.dev/blog/bayesian-statistics-for-confused-data-scientists/
120•speckx•3d ago•29 comments

HopTab–free,open source macOS app switcher and tiler that replaces Cmd+Tab

https://www.royalbhati.com/hoptab
19•robhati•3h ago•8 comments

It's Their Mona Lisa

https://ironicsans.ghost.io/its-t-mona-lisa/
26•ramimac•3d ago•10 comments

Grafeo – A fast, lean, embeddable graph database built in Rust

https://grafeo.dev/
224•0x1997•19h ago•77 comments

Do Not Turn Child Protection into Internet Access Control

https://news.dyne.org/child-protection-is-not-access-control/
676•smartmic•13h ago•360 comments

Hide macOS Tahoe's Menu Icons

https://512pixels.net/2026/03/hide-macos-tahoes-menu-icons-with-this-one-simple-trick/
205•soheilpro•16h ago•76 comments

Sashiko: An agentic Linux kernel code review system

https://sashiko.dev/
21•Lwrless•6h ago•3 comments

Trivy ecosystem supply chain briefly compromised

https://github.com/aquasecurity/trivy/security/advisories/GHSA-69fq-xp46-6x23
73•batch12•2d ago•23 comments

Show HN: Termcraft – Terminal-first 2D sandbox survival in Rust

https://github.com/pagel-s/termcraft
115•sebosch•15h ago•20 comments

Thinking Fast, Slow, and Artificial: How AI Is Reshaping Human Reasoning

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646
146•Anon84•18h ago•82 comments

Common Lisp Development Tooling

https://www.creativetension.co/posts/common-lisp-development-tooling
88•0bytematt•14h ago•18 comments

Alpha Micro AM-1000E and AM-1200

http://oldvcr.blogspot.com/2026/03/refurb-weekend-double-header-alpha.html
14•goldenskye•6h ago•3 comments

Training Center for Maneuvering on Manned Model Ships

https://www.portrevel.com/
12•mhb•1d ago•5 comments

Show HN: Time Keep – Location timezones, timers, alarms, countdowns in one place

5•jmbuilds•2d ago•3 comments

Ubuntu 26.04 Ends 46 Years of Silent sudo Passwords

https://pbxscience.com/ubuntu-26-04-ends-46-years-of-silent-sudo-passwords/
371•akersten•1d ago•366 comments

The paddle wheel aircraft carriers of Lake Michigan

https://signoregalilei.com/2026/03/08/the-paddle-wheel-aircraft-carriers-of-lake-michigan/
83•surprisetalk•4d ago•8 comments

How Invisalign became the biggest user of 3D printers

https://www.wired.com/story/how-invisalign-became-the-worlds-biggest-3d-printing-company/
162•mikhael•3d ago•118 comments

How Ford burned $12B in Brazil (2021)

https://www.reuters.com/business/autos-transportation/how-ford-burned-12-billion-brazil-2021-05-20/
54•kaycebasques•19h ago•30 comments