Could Endpoint SLMs Replace Cloud LLMs? Would Datacenter Race Shudder to a Halt?
2•aniijbod•37m ago
Yeah, an SLM on an endpoint like a phone will have fresh latency issues as it goes online to fill gaps in its inference engine's knowledge base that a cloud LLM might not have, but cloud LLMs aren't exactly latency-free either, so the latency/performance issue isn't necessarily LLM's winning card.