LeCun's model will fail as the idea of world model is oxymoronic, brains don't need them and the world isn't modeled, all models are wrong, the world is experienced instantaneously in optic flow that's built atop of olfaction.
https://www.eneuro.org/content/7/4/ENEURO.0069-20.2020
Any real AI that veers at control will have to adopt a neurobio path
https://tbrnewsmedia.com/sbus-sima-mofakham-chuck-mikell-des...
That's built paradoxically from unpredictability
> Any real AI that veers at control will have to adopt a neurobio path
Maybe. Or maybe it's a useless distraction. Only time will tell what signals are meaningful.
That being said there have been models which are pretty effective at other things that don’t use language, so maybe it’s a non issue.
Text really hogged all the attention. Media is where AI is really going to shine.
Some of the most profitable models right now are in music, image, and video generation. A lot of people are having a blast doing things they could legitimately never do before, and real working professionals are able to use the tools to get 1000x more done - perhaps providing a path to independence from bigger studios, and certainly more autonomy for those not born into nepotism.
As long as companies don't over-raise like OpenAI, there should be a smooth gradient from next gen media tools to revolutionary future stuff like immersive VR worlds that you can bend like the Matrix or Holodeck.
And I'll just be exceedingly chuffed if we get open source and highly capable world models from the Chinese that keep us within spitting distance of the unicorns.
Fundamentally what AGI is trying to do is to encode ability to logic and reason. Tokens, images, video and audio are all just information of different entropy density that is the output of that logic reasoning process or emulation of logic reasoning process.
No? The Wason selection task has shown that logic and reason are not really core nor essential to human cognition.
It's really verging on speculation, but see chapter 2 of Jaynes 1976 - in particular the section on spatialization and the features of consciousness.
This is wrong. The vast majority of revenue is being generated by text models because they are so useful.
You could say the same thing about AGI. Ultimately capital will realize intelligence is a drawback.
It is the most impressed I've been with an AI experience since the first time I saw a model one-shot material code.
Sure, its an early product. The visual output reminds me a lot of early SDXL. But just look at what's happened to video in the last year and image in the last three. The same thing is going to happen here, and fast, and I see the vision for generative worlds for everything from gaming/media to education to RL/simulation.
“Taking images and turning them into 3D environments using gaussian splats, depth and inpainting. Cool, but that’s a 3D GS pipeline, not a robot brain.”
techblueberry•1h ago
andrewflnr•1h ago
krainboltgreene•44m ago
CrackerNews•1h ago
I feel LeCun is correct that LLMs as of now have limitations where it needs an architectural overhaul. LLMs now have a problem with context rot, and this would hamper with an effective world model if the world disintegrates and becomes incoherent and hallucinated over time.
It'd doubtful whether investors would be in for the long haul, which may explain the behavior of Sam Altman in seeking government support. The other approaches described in this article may be more investor friendly as there is a more immediate return with creating a 3D asset or a virtual simulation.
Fricken•50m ago