If this means there’s a 2x-7x speed up available to a scaled diffusion model like Inception Mercury, that’ll be a game changer. It feels 10x faster already…
blurbleblurble•11m ago
Diffusion language models seem poised to smash purely autoregressive models. I'm giving it 1-2 years.
LarsDu88•50m ago
Google is working on a similar line of research. Wonder why they haven't rolled out a GPT40 scaled version of this yet
yjftsjthsd-h•39m ago
Is anyone doing any form of diffusion language models that are actually practical to run today on the actual machine under my desk? There's loads of more "traditional" .gguf options (well, quants) that are practical even on shockingly weak hardware, and I've been seeing things that give me hope that diffusion is the next step forward, but so far it's all been early research prototypes.
Bolwin•29m ago
Based on my experience running diffusion image models I really hope this isn't going to take over anytime soon. Parallel decoding may be great if you have a nice parallel gpu or npu but is dog slow for cpus
refulgentis•1h ago
blurbleblurble•11m ago