This is about first prototypes and scaling is often easier than the basic principle.
A lot of these novel AI accelerators run into problems like that because they're not capable of general purpose computing. A good example of that are the boltzman machines on Dwave's stuff. Yeah it can do that but it can only do that because the machine is only capable of doing QUBO.
rf15•3h ago
danielbln•3h ago
So I hear you, but on the flip side we _should_ be reading a lot about LLMs here, as they have a direct impact on the work that most of us do.
That said, seeing other papers pop up that are not related to transformer based networks is appreciated.
larodi•47m ago
karanveer•2h ago