frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Open in hackernews

A Tiny Boltzmann Machine

https://eoinmurray.info/boltzmann-machine
181•anomancer•6h ago

Comments

vanderZwan•5h ago
Lovely explanation!

Just FYI: mouse-scrolling is much too sensitive for some reason (I'm assuming it swipes just fine in mobile contexts, have not checked that). The result is that it jumped from first to last "page" and back whenever I tried scrolling. Luckily keyboard input worked so I could still read the whole thing.

djulo•5h ago
that's soooo coool
nonrandomstring•5h ago
This takes me back. 1990, building Boltzman machines and Perceptrons from arrays of void pointers to "neurons" in plain C. What did we use "AI" for back then? To guess the next note in a MIDI melody, and to recognise the shape of a scored note, minim, crotchet, quaver on a 5 x 9 dot grid. 85% accuracy was "good enough" then.
bwestergard•4h ago
Did the output sound musical?
nonrandomstring•3h ago
For small values of "music"? Really, no. But tbh, neither have more advanced "AI" composition experiments I've encountered over the years, Markov models, linear predictive coding, genetic/evolutionary algs, rule based systems, and now modern diffusion and transormers... they all lack the "spirit of jazz" [0]

[0] https://i.pinimg.com/originals/e4/84/79/e484792971cc77ddff8f...

bbstats•4h ago
anyone got an archived link?
tambourine_man•4h ago
Typo

“They can be used for generating new data that…”

munchler•4h ago
Another typo (or thinko) in the very first sentence:

"Here we introduce introduction to Boltzmann machines"

croemer•4h ago
More typos (LLMs are really good at finding these):

"Press the "Run Simulation" button to start traininng the RBM." ("traininng" -> "training")

"...we want to derivce the contrastive divergence algorithm..." ("derivce" -> "derive")

"A visisble layer..." ("visisble" -> "visible")

nayuki•4h ago
Oh, this is a neat demo. I took Geoff Hinton's neural networks course in university 15 years ago and he did spend a couple of lectures explaining Boltzmann machines.

> A Restricted Boltzmann Machine is a special case where the visible and hidden neurons are not connected to each other.

This wording is wrong; it implies that visible neurons are not connected to hidden neurons.

The correct wording is: visible neurons are not connected to each other and hidden neurons are not connected to each other.

Alternatively: visible and hidden neurons do not have internal connections within their own type.

CamperBob2•3h ago
Alternatively: visible and hidden neurons do not have internal connections within their own type.

I'm a bit unclear on how that isn't just an MLP. What's different about a Boltzmann machine?

Edit: never mind, I didn't realize I needed to scroll up to get to the introductory overview.

What 0xTJ's [flagged][dead] comment says about it being undesirable to hijack or otherwise attempt to reinvent scrolling is spot on.

nayuki•3h ago
> I'm a bit unclear on how that isn't just a multi-layer perceptron. What's different about a Boltzmann machine?

In a Boltzmann machine, you alternate back and forth between using visible units to activate hidden units, and then use hidden units to activate visible units.

> What 0xTJ's [flagged][dead] comment says about it being undesirable to hijack or otherwise attempt to reinvent scrolling is spot on.

The page should be considered a slideshow that is paged discretely and not scrollable continuously. And there should definitely be no scrolling inertia.

sitkack•3h ago
Fun article on David Ackley https://news.unm.edu/news/24-nobel-prize-in-physics-cited-gr...

Do check out his T2 Tile Project.

AstroJetson•2h ago
The key takeaways are that there are lots of people involved with making these breakthroughs.

The value of grad students is often overlooked, they contribute so much and then later on advance the research even more.

Why does America look on research as a waste, when it has move everything so far?

macintux•29m ago
It's more accurate to say that businesspeople consider research a waste in our quarter-by-quarter investment climate, since it generally doesn't lead to immediate gains.

And our current leadership considers research a threat, since science rarely supports conspiracy theorists or historical revisionism.

itissid•3h ago
IIUC, we need gibbs sampling(to compute the weight updates) instead of using the gradient based forward and backward passes with today's NNetworks that we are used to. Any one understand why that is so?
ebolyen•25m ago
Not an expert, but I have a bit of formal training on Bayesian stuff which handles similar problems.

Usually Gibbs is used when there's no directly straight-forward gradient (or when you are interested in reproducing the distribution itself, rather than a point estimate), but you do have some marginal/conditional likelihoods which are simple to sample from.

Since each visible node depends on each hidden node and each hidden node effects all visible nodes, the gradient ends up being very messy, so its much simpler to use Gibbs sampling to adjust based on marginal likelihoods.

pawanjswal•3h ago
Love how this breaks down Boltzmann Machines—finally makes this 'energy-based model' stuff click!
BigParm•3h ago
That font with a bit of margin looks fantastic on my phone specifically. Really nailing the minimalist look. What font is that?
mac9•3h ago
"font-family: ui-sans-serif, system-ui, sans-serif, "Apple Color Emoji", "Segoe UI Emoji", "Segoe UI Symbol", "Noto Color Emoji";"

from the css so odds are it's whatever your browser or OS's default sans font is, in my case it's SF Pro which is an Apple font though it may vary if you use a non Apple device.

nickvec•3h ago
> Here we introduce introduction to Boltzmann machines and present a Tiny Restricted Boltzmann Machine that runs in the browser.

nit: should "introduction" be omitted?

antidumbass•3h ago
The section after the interactive diagrams has no left padding and thus runs off the screen on iOS.
rollulus•3h ago
Now the real question: is it you enjoying that nice page or is it a Boltzmann Brain?

https://en.m.wikipedia.org/wiki/Boltzmann_brain

alganet•2h ago
It doesn't matter.

It's Decartes demon all over again. Problem solved centuries ago. You can skin it however you want, it's the same problem.

nickvec•3h ago
Great site! Would be cool to be able to adjust the speed at which the simulation runs as well.
thingamarobert•1h ago
This is very well made, and so nostalgic to me! My whole PhD between 2012-16 was based on RBMs and I learned so much about generative ML through these models. Research has come so far and one doesn't hear much about them these days but they were really at the heart of the "AI Spring" back then.
tomrod•1h ago
Great read!

One nit, a misspelling in the Appendix: derivce -> derive

oac•43m ago
Nice and clean explanation!

It brings up a lot of memories! Shameless plug: I made a visualization of an RBM being trained years ago: https://www.youtube.com/watch?v=lKAy_NONg3g

dr_dshiv•34m ago
My understanding is that the Harmonium (Smolensky) was the first restricted Boltzmann machine, but maximized “harmony” instead of minimizing “energy.” When Smolensky, Hinton and Rummelhart collaborated, they instead called it “goodness of fit.”

The harmonium paper [1] is a really nice read. Hinton obviously became the superstar and Smolensky wrote long books about linguistics.

Anyone know more about this history?

[1] https://stanford.edu/~jlmcc/papers/PDP/Volume%201/Chap6_PDP8...

Nevermark•25m ago
I mistook the title for "A Tiny Boltzmann Brain"! [0]

My own natural mind immediately solved the conundrum. Surely this was a case where a very small model was given randomly generated weights and then tested to see if it actually did something useful!

After all, the smaller the model, the more likely simple random generation can produce something interesting, relative to its size.

I stand corrected, but not discouraged!

I propose a new class of model, the "Unbiased-Architecture Instant Boltzmann Model" (UA-IBM).

One day we will have quantum computers large enough to simply set up the whole dataset as a classical constraint on a model defined with N serialized values, representing all the parameters and architecture settings. Then let a quantum system with N qubits take one inference step over all the classical samples, with all possible parameters and architectures in quantum superposition, and then reduce the result to return the best (or near best) model's parametesr and architecture in classical form.

Anyone have a few qubits laying around that want to give this a shot? (The irony that everything is quantum and yet so slippery we can hardly put any of it to work yet.

(Sci-fi story premise: the totally possible case of an alien species that evolved one-off quantum sensor, which evolved into a whole quantum sensory system, then a nervous system, and subsequently full quantum intelligence out of the gate. What kind of society and technological trajectory would they have? Hopefully they are in close orbit around a black hole, so the impact of their explosive progress has not threatened us yet. And then one day, they escape their gravity well, and ...)

[0] https://en.wikipedia.org/wiki/Boltzmann_brain

Baby Is Healed with First Personalized Gene-Editing Treatment

https://www.nytimes.com/2025/05/15/health/gene-editing-personalized-rare-disorders.html
70•jbredeche•1h ago•27 comments

I Don't Like NumPy

https://dynomight.net/numpy/
152•MinimalAction•3h ago•65 comments

Improving Naval Ship Acquisition

https://www.construction-physics.com/p/fixing-naval-ship-acquisition
14•Luc•1h ago•6 comments

Launch HN: Tinfoil (YC X25): Verifiable Privacy for Cloud AI

75•FrasiertheLion•3h ago•43 comments

The current state of TLA⁺ development

https://ahelwer.ca/post/2025-05-15-tla-dev-status/
18•todsacerdoti•54m ago•0 comments

Dr. Dobb's Journal Interviews Jef Raskin (1986)

https://computeradsfromthepast.substack.com/p/dr-dobbs-journal-interviews-jef-raskin
13•rbanffy•1h ago•3 comments

A Tiny Boltzmann Machine

https://eoinmurray.info/boltzmann-machine
181•anomancer•6h ago•31 comments

New paradigm for psychology just dropped

https://www.experimental-history.com/p/new-paradigm-for-psychology-just
12•CharlesW•1h ago•9 comments

Show HN: Min.js style compression of tech docs for LLM context

https://github.com/marv1nnnnn/llm-min.txt
123•marv1nnnnn•6h ago•36 comments

Malicious compliance by booking an available meeting room

https://www.clientserver.dev/p/malicious-compliance-by-booking-an
222•jakevoytko•6h ago•215 comments

Harvard Law Paid $27 for a Copy of Magna Carta. It's an Original

https://www.nytimes.com/2025/05/15/world/europe/harvard-law-magna-carta-original.html
57•jgwil2•1h ago•15 comments

Fetii (YC S22) Is Hiring

https://www.ycombinator.com/companies/fetii/jobs/QDjleWs-senior-operations-manager-fetii
1•Mattiommi•2h ago

Onfim's world: Child artists in history

https://resobscura.substack.com/p/onfims-world-medieval-child-artists
21•benbreen•3h ago•4 comments

Human

https://quarter--mile.com/Human
634•surprisetalk•16h ago•249 comments

Coinbase says hackers bribed staff to steal customer data, demanding $20M ransom

https://www.cnbc.com/2025/05/15/coinbase-says-hackers-bribed-staff-to-steal-customer-data-and-are-demanding-20-million-ransom.html
166•gpi•3h ago•174 comments

Lua for Elixir

https://davelucia.com/blog/lua-elixir
152•davydog187•2d ago•58 comments

Show HN: Real-Time Gaussian Splatting

https://github.com/axbycc/LiveSplat
104•markisus•6h ago•44 comments

Internet Artifacts

https://neal.fun/internet-artifacts/
579•mikerg87•2d ago•116 comments

October the First Is Too Late

https://gwern.net/fiction/october
11•adiabatty•39m ago•0 comments

The Fastest Way yet to Color Graphs

https://www.quantamagazine.org/the-fastest-way-yet-to-color-graphs-20250512/
25•GavCo•2d ago•8 comments

Pathfinding

https://juhrjuhr.itch.io/deep-space-exploitation/devlog/945428/9-pathfinding
84•sebg•7h ago•28 comments

A Rust API Inspired by Python, Powered by Serde

https://ohadravid.github.io/posts/2025-05-serde-reflect/
45•lukastyrychtr•4d ago•20 comments

Radiance Contrasts at Possible Lunar Water Ice Exposures Seen by ShadowCam

https://iopscience.iop.org/article/10.3847/PSJ/adb8d1
5•PaulHoule•2d ago•0 comments

Model Organisms Are Not Static

https://www.asimov.press/p/model-organisms-are-not-static
20•mailyk•4h ago•4 comments

AlphaEvolve: A Gemini-powered coding agent for designing advanced algorithms

https://deepmind.google/discover/blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/
955•Fysi•1d ago•252 comments

Wavelet Trees: An Introduction (2011)

https://www.alexbowe.com/wavelet-trees/
34•Tomte•4h ago•9 comments

Working on complex systems: What I learned working at Google

https://www.thecoder.cafe/p/complex-systems
203•0xKelsey•2d ago•110 comments

Demystifying Ruby (1/3): It's all about threads

https://blog.papey.fr/post/07-demystifying-ruby-01/
28•kadrek•5h ago•0 comments

Gateway Books: The lessons of a defunct canon

https://thepointmag.com/examined-life/gateway-books/
14•samclemens•3h ago•13 comments

California sent residents' personal health data to LinkedIn

https://themarkup.org/pixel-hunt/2025/04/28/how-california-sent-residents-personal-health-data-to-linkedin
139•anticorporate•5h ago•78 comments