https://news.ycombinator.com/item?id=45923326
https://entropytown.com/articles/2025-11-13-world-model-lecu...
I still want to know the definition of a world model, because it seems that what they are doing now is a very basic model of GS (not sure if it is self-developed) and a GS browser.
It is different from video generators, because it also accepts control inputs (ie. keyboard, mouse).
However the definition of "world models" is a bit nebulous. World labs looks kinda like sprites in a 3d world.
Which also leads me to suggest that in no way are they worth $1b, espcially as google are doing to destroy them (possibly meta, but I doubt they are going to be any good anytime soon)
Also, video games, and movie / tv / video content generation.
It also allows for companies to reclaim their supply chains within the country of manufacture (for high income countries)
Model-free paradigms (which we do now without world models) does not actually model what _happens_ when you take an action, they simply model how good or bad an action is. This is highly data inefficient but asymptotically performs much better than model-based RL because you don't have modeling biases.
Model-based RL, where world-models come in, models the transition matrix T(s, a, s') meaning, I'm in state s and I take action a, what is my belief about my new state? By doing this you can do long-term planning, so it's not just useful for robotics and video generation but for reasoning and planning more broadly. It's also highly data efficient, and right now, for robotics, that is absolutely the name of the game.
What you will see is: approximately zero robots, then approximately one crappy robot (once you get performance + reliability to jusssst cross the boundary where you can market it, even at a loss! and people will buy it and put it in their homes). Once that happens you get the magic: data flywheel for robotics, and things start _rapidly_ improving.
Robotics is where it is because it lacks the volume of data we have on the internet. for robotics today it's not only e.g. egocentric video that we need but also _sensor-specific_ and _robot-specific_ data (e.g. robot A has a different build + components than robot B)
darth_avocado•1h ago
giwook•1h ago
iwontberude•1h ago