frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

US Army raises enlistment age to 42 and removes marijuana waiver requirement

https://armypubs.army.mil/epubs/DR_pubs/DR_a/ARN42922-AR_601-210-000-WEB-1.pdf
15•Jimmc414•26m ago•3 comments

Flighty Airports

https://flighty.com/airports
159•skogstokig•3h ago•53 comments

Goodbye to Sora

https://twitter.com/soraofficialapp/status/2036532795984715896
487•mikeocool•8h ago•376 comments

Show HN: DuckDB community extension for prefiltered HNSW using ACORN-1

https://github.com/cigrainger/duckdb-hnsw-acorn
9•cigrainger•43m ago•0 comments

I wanted to build vertical SaaS for pest control, so I took a technician job

https://www.onhand.pro/p/i-wanted-to-build-vertical-saas-for-pest-control-i-took-a-technician-job...
229•tezclarke•6h ago•95 comments

Show HN: I took back Video.js after 16 years and we rewrote it to be 88% smaller

https://videojs.org/blog/videojs-v10-beta-hello-world-again
266•Heff•10h ago•40 comments

Apple Business

https://www.apple.com/newsroom/2026/03/introducing-apple-business-a-new-all-in-one-platform-for-b...
548•soheilpro•12h ago•325 comments

Tell HN: Litellm 1.82.7 and 1.82.8 on PyPI are compromised

https://github.com/BerriAI/litellm/issues/24512
550•dot_treo•16h ago•393 comments

Arm AGI CPU

https://newsroom.arm.com/blog/introducing-arm-agi-cpu
295•RealityVoid•10h ago•232 comments

Intel Device Modeling Language for virtual platforms

https://github.com/intel/device-modeling-language
9•transpute•3d ago•0 comments

In Edison’s Revenge, Data Centers Are Transitioning From AC to DC

https://spectrum.ieee.org/data-center-dc
100•jnord•3h ago•118 comments

Zero-Cost POSIX Compliance: Encoding the Socket State Machine in Lean's Types

https://ngrislain.github.io/blog/2026-3-25-zerocost-posix-compliance-encoding-the-socket-state-ma...
30•ngrislain•3h ago•16 comments

A Compiler Writing Journey

https://github.com/DoctorWkt/acwj
46•ibobev•4h ago•3 comments

Algorithm Visualizer

https://algorithm-visualizer.org/
42•vinhnx•4d ago•3 comments

An Aural Companion for Decades, CBS News Radio Crackles to a Close

https://www.nytimes.com/2026/03/21/business/media/cbs-news-radio-appraisal.html
39•tintinnabula•3d ago•6 comments

Show HN: Email.md – Markdown to responsive, email-safe HTML

https://www.emailmd.dev/
238•dancablam•11h ago•57 comments

Wine 11 rewrites how Linux runs Windows games at kernel with massive speed gains

https://www.xda-developers.com/wine-11-rewrites-linux-runs-windows-games-speed-gains/
745•felineflock•9h ago•259 comments

What happened to GEM?

https://dfarq.homeip.net/whatever-happened-to-gem/
57•naves•4d ago•26 comments

Hypura – A storage-tier-aware LLM inference scheduler for Apple Silicon

https://github.com/t8/hypura
194•tatef•12h ago•75 comments

Show HN: Gemini can now natively embed video, so I built sub-second video search

https://github.com/ssrajadh/sentrysearch
280•sohamrj•13h ago•77 comments

Hypothesis, Antithesis, synthesis

https://antithesis.com/blog/2026/hegel/
222•alpaylan•12h ago•82 comments

How the world’s first electric grid was built

https://worksinprogress.co/issue/how-the-worlds-first-electric-grid-was-built/
69•zdw•4d ago•18 comments

Oil at $150 will trigger global recession, says boss of financial BlackRock

https://www.bbc.com/news/articles/c9wqrdkx8ppo
47•tartoran•2h ago•17 comments

Missile defense is NP-complete

https://smu160.github.io/posts/missile-defense-is-np-complete/
285•O3marchnative•15h ago•293 comments

Epoch confirms GPT5.4 Pro solved a frontier math open problem

https://epoch.ai/frontiermath/open-problems/ramsey-hypergraphs
429•in-silico•1d ago•624 comments

No Terms. No Conditions

https://notermsnoconditions.com
234•bayneri•12h ago•104 comments

Lago (YC S21) Is Hiring

https://getlago.notion.site/Lago-Product-Engineer-AI-Agents-for-Growth-327ef63110d280cdb030ccf429...
1•AnhTho_FR•10h ago

Data Manipulation in Clojure Compared to R and Python

https://codewithkira.com/2024-07-18-tablecloth-dplyr-pandas-polars.html
105•tosh•2d ago•29 comments

Epic Games to cut more than 1k jobs as Fortnite usage falls

https://www.reuters.com/legal/litigation/epic-games-said-tuesday-that-it-will-lay-off-more-than-1...
299•doughnutstracks•13h ago•458 comments

Nanobrew: The fastest macOS package manager compatible with brew

https://nanobrew.trilok.ai/
194•syrusakbary•16h ago•119 comments
Open in hackernews

Transformers Are Bayesian Networks

https://arxiv.org/abs/2603.17063
32•Anon84•4d ago

Comments

westurner•4d ago
https://news.ycombinator.com/item?id=45256179 :

> Which statistical models disclaim that their output is insignificant if used with non-independent features? Naieve Bayes [...]

Ironic then, because if transformers are Bayesian networks then we're using Bayesian networks for non-independent features.

From "Quantum Bayes' rule and Petz transpose map from the minimum change principle" (2025) https://news.ycombinator.com/item?id=45074143 :

> Petz recovery map: https://en.wikipedia.org/wiki/Petz_recovery_map :

> In quantum information theory, a mix of quantum mechanics and information theory, the Petz recovery map can be thought of as a quantum analog of Bayes' theorem

But there aren't yet enough qubits for quantum LLMs: https://news.ycombinator.com/item?id=47203219#47250262

"Transformer is a holographic associative memory" (2025) https://news.ycombinator.com/item?id=43028710#43029899

malcolmgreaves•3d ago
> Hallucination is not a bug that scaling can fix. It is the structural consequence of operating without concepts.

NNs are as close to continuous as we can get with discrete computing. They’re flexible and adaptable and can contain many “concepts.” But their chief strength is also their chief weakness: these “concepts” are implicit. I wonder if we can get a hybrid architecture that has the flexibility of NNs while retaining discrete concepts like a knowledge base does.

measurablefunc•1h ago
There is nothing continuous on the computer, it's all bit strings & boolean arithmetic. The semantics imposed on the bit strings does not exist anywhere in the arithmetic operations, i.e. there is no arithmetic operation corresponding to something as simple as the color red.
kelseyfrog•1h ago
It sounds like you're saying that if a computer had infinite precision then hallucinations would not occur?
measurablefunc•1h ago
The way neural networks work is that the base neural network is embedded in a sampling loop, i.e. a query is fed into the network & the driver samples output tokens to append to the query so that it can be re-fed back into the network (q → nn → [a, b, c, ...] → q + sample([a, b, c, ...])). There is no way to avoid hallucinations b/c hallucinations are how the entire network works at the implementation level. The precision makes no difference b/c the arithmetic operations are semantically void & only become meaningful after they are interpreted by someone who knows to associated 1 /w red, 2 w/ blue, 3 w/ clouds, & so on & so forth. The mapping between the numbers & concepts does not exist in the arithmetic.
kelseyfrog•1h ago
Oh, I thought that the embedding space of the residual stream was precisely that.
measurablefunc•1h ago
The arithmetic is meaningless, it doesn't matter what you call it b/c on the computer it's all bit strings & boolean arithmetic. You can call some sequence of operations residual & others embeddings but that is all imposed top-down. There is nothing in the arithmetic that indicates it is somehow special & corresponds to embeddings or residuals.
kelseyfrog•1h ago
Ah ok, so if we had such a mapping then models wouldn't hallucinate?
measurablefunc•1h ago
Maybe it's better if you define the terms b/c what I mean by hallucination is that the arithmetic operations + sampling mean that it's all hallucinations. The output is a trajectory of a probabilistic computation over some set of symbols (0s & 1s). Those symbols are meaningless, the only reason they have meaning is b/c everyone has agreed that the number 97 is the ascii code for "a" & every conformant text processor w/ a conformant video adapter will convert 97 (0b1100001) into the display pattern for the letter "a".
kelseyfrog•55m ago
So kind of like if you flip a coin, the sampling means the heads or tails you get isn't real?
AdieuToLogic•42m ago
> NNs are as close to continuous as we can get with discrete computing.

This is incorrect. For example, fuzzy logic[0] can model analog ("continuous") truth beyond discrete digital representations, such as 1/0, true/false, etc.

0 - https://en.wikipedia.org/wiki/Fuzzy_logic

getnormality•1h ago
> Transformers are the dominant architecture in AI, yet why they work remains poorly understood. This paper offers a precise answer: a transformer is a Bayesian network.

Why would being a Bayesian network explain why transformers work? Bayesian networks existed long before transformers and never achieved their performance.

Mithriil•1h ago
Bayesian network is a really general concept. It applies to all multidimensional probability distribution. It's a graph that encodes independence between variables. Ish.

I have not taken the time to review the paper, but if the claim stands, it means we might have another tool to our toolbox to better understand transformers.

warypet•1h ago
I found this earlier today when looking for research and ended up reporting it for citing fake sources. Please correct me if I'm wrong, but I couldn't find "[9] Jongsuk Jung, Jaekyeom Kim, and Hyunwoo J. Choi. Rethinking attention as belief propagation. In International Conference on Machine Learning (ICML), 2022." anywhere else on the internet
measurablefunc•1h ago
It's "vibe" research. Most of it is basically pure nonsense.
kleiba•49m ago
Care to elaborate?
jack_pp•40m ago
I suspect it means it's LLM generated without it being checked
refulgentis•33m ago
The headline theorem, "every sigmoid transformer is a Bayesian network," is proved by `rfl` [1]. For non-Lean people: `rfl` means "both sides are the same expression." He defines a transformer forward pass, then defines a BP forward pass with the same operations, wraps the weights in a struct called `implicitGraph`, and Lean confirms they match. They match because he wrote them to match.

The repo with a real transformer model (transformer-bp-lean) has 22 axioms and 7 theorems. In Lean, an axiom is something you state without proving. The system takes your word for it. Here the axioms aren't background math, they're the paper's claims:

- "The FFN computes the Bayesian update" [2]. Axiom.

- "Attention routes neighbors correctly" [3]. Axiom.

- "BP converges" [4]. Axiom, with a comment saying it's "not provable in general."

- The no-hallucination corollary [5]. Axiom.

The paper says "formally verified against standard mathematical axioms" about all of these. They are not verified. They are assumed.

The README suggests running `grep -r "sorry"` and finding nothing as proof the code is complete. In Lean, `sorry` means "I haven't proved this" and throws a compiler warning. `axiom` also means "I haven't proved this" but doesn't warn. So the grep returns clean while 22 claims sit unproved. Meanwhile the godel repo has 4 actual sorries [6] anyway, including "logit and sigmoid are inverses," which the paper treats as proven. That same fact appears as an axiom in the other repo [7]. Same hole, two repos, two different ways to leave it open.

Final count across all five repos: 65 axioms, 5 sorries, 149 theorems.

Claude (credited on page 1) turned it into "Not an approximation of it. Not an analogy to it. The computation is belief propagation." Building to a 2-variable toy experiment on 5K parameters presented as the fulfillment of Leibniz's 350-year-old dream. Ending signed by "Calculemus."

[1] https://github.com/gregorycoppola/sigmoid-transformer-lean/b...

[2] https://github.com/gregorycoppola/transformer-bp-lean/blob/7...

[3] https://github.com/gregorycoppola/transformer-bp-lean/blob/7...

[4] https://github.com/gregorycoppola/transformer-bp-lean/blob/7...

[5] https://github.com/gregorycoppola/transformer-bp-lean/blob/7...

[6] https://github.com/gregorycoppola/godel/blob/bc1d138/Godel/O...

[7] https://github.com/gregorycoppola/sigmoid-transformer-lean/b...

kurthr•51m ago
Yep, nothing by even a subset of those authors. Closest paper from that Conference:

Rethinking Attention-Model Explainability through Faithfulness Violation Test Yibing Liu, Haoliang Li, Yangyang Guo, Chenqi Kong, Jing Li, Shiqi Wang

https://proceedings.mlr.press/v162/liu22i.html

https://icml.cc/virtual/2022/spotlight/18082