[0] https://i.pinimg.com/originals/e4/84/79/e484792971cc77ddff8f...
“They can be used for generating new data that…”
"Here we introduce introduction to Boltzmann machines"
"Press the "Run Simulation" button to start traininng the RBM." ("traininng" -> "training")
"...we want to derivce the contrastive divergence algorithm..." ("derivce" -> "derive")
"A visisble layer..." ("visisble" -> "visible")
> A Restricted Boltzmann Machine is a special case where the visible and hidden neurons are not connected to each other.
This wording is wrong; it implies that visible neurons are not connected to hidden neurons.
The correct wording is: visible neurons are not connected to each other and hidden neurons are not connected to each other.
Alternatively: visible and hidden neurons do not have internal connections within their own type.
I'm a bit unclear on how that isn't just an MLP. What's different about a Boltzmann machine?
Edit: never mind, I didn't realize I needed to scroll up to get to the introductory overview.
What 0xTJ's [flagged][dead] comment says about it being undesirable to hijack or otherwise attempt to reinvent scrolling is spot on.
In a Boltzmann machine, you alternate back and forth between using visible units to activate hidden units, and then use hidden units to activate visible units.
> What 0xTJ's [flagged][dead] comment says about it being undesirable to hijack or otherwise attempt to reinvent scrolling is spot on.
The page should be considered a slideshow that is paged discretely and not scrollable continuously. And there should definitely be no scrolling inertia.
Do check out his T2 Tile Project.
The value of grad students is often overlooked, they contribute so much and then later on advance the research even more.
Why does America look on research as a waste, when it has move everything so far?
And our current leadership considers research a threat, since science rarely supports conspiracy theorists or historical revisionism.
Usually Gibbs is used when there's no directly straight-forward gradient (or when you are interested in reproducing the distribution itself, rather than a point estimate), but you do have some marginal/conditional likelihoods which are simple to sample from.
Since each visible node depends on each hidden node and each hidden node effects all visible nodes, the gradient ends up being very messy, so its much simpler to use Gibbs sampling to adjust based on marginal likelihoods.
from the css so odds are it's whatever your browser or OS's default sans font is, in my case it's SF Pro which is an Apple font though it may vary if you use a non Apple device.
nit: should "introduction" be omitted?
It's Decartes demon all over again. Problem solved centuries ago. You can skin it however you want, it's the same problem.
One nit, a misspelling in the Appendix: derivce -> derive
It brings up a lot of memories! Shameless plug: I made a visualization of an RBM being trained years ago: https://www.youtube.com/watch?v=lKAy_NONg3g
The harmonium paper [1] is a really nice read. Hinton obviously became the superstar and Smolensky wrote long books about linguistics.
Anyone know more about this history?
[1] https://stanford.edu/~jlmcc/papers/PDP/Volume%201/Chap6_PDP8...
My own natural mind immediately solved the conundrum. Surely this was a case where a very small model was given randomly generated weights and then tested to see if it actually did something useful!
After all, the smaller the model, the more likely simple random generation can produce something interesting, relative to its size.
I stand corrected, but not discouraged!
I propose a new class of model, the "Unbiased-Architecture Instant Boltzmann Model" (UA-IBM).
One day we will have quantum computers large enough to simply set up the whole dataset as a classical constraint on a model defined with N serialized values, representing all the parameters and architecture settings. Then let a quantum system with N qubits take one inference step over all the classical samples, with all possible parameters and architectures in quantum superposition, and then reduce the result to return the best (or near best) model's parametesr and architecture in classical form.
Anyone have a few qubits laying around that want to give this a shot? (The irony that everything is quantum and yet so slippery we can hardly put any of it to work yet.
(Sci-fi story premise: the totally possible case of an alien species that evolved one-off quantum sensor, which evolved into a whole quantum sensory system, then a nervous system, and subsequently full quantum intelligence out of the gate. What kind of society and technological trajectory would they have? Hopefully they are in close orbit around a black hole, so the impact of their explosive progress has not threatened us yet. And then one day, they escape their gravity well, and ...)
vanderZwan•5h ago
Just FYI: mouse-scrolling is much too sensitive for some reason (I'm assuming it swipes just fine in mobile contexts, have not checked that). The result is that it jumped from first to last "page" and back whenever I tried scrolling. Luckily keyboard input worked so I could still read the whole thing.