You LLM people all here serious cases of Dunning Kruger
Close to what, and how are you measuring?
> nobody in the USA would be spending 7 figures on infrastructure for it
Au contraire, if AI had a moat it would pay for itself. They're funneling capital into infrastructure because they know it can't.
edit: Note that you can run it yourself with sufficient resources, or access it from other providers too: https://openrouter.ai/moonshotai/kimi-k2.6/providers
Not sure about coding usage, Google being weird about these things I could see that quota being separate.
Edit: found it.
> We may use your Content to operate, maintain, improve, and develop the Services, to comply with legal obligations, to enforce our policies, and to ensure security. You may opt out of allowing your Content to be used for model improvement and research purposes by contacting us at membership@moonshot.ai. We will honor your choice in accordance with applicable law.
Section 3 of https://www.kimi.com/user/agreement/modelUse?version=v2
So in other words only if you can point to a local law which requires them to comply with the opt out?
Is this the same model?
Unsloth quants: https://huggingface.co/unsloth/Kimi-K2.6-GGUF
(work in progress, no gguf files yet, header message saying as much)
But the files are only roughly 640GB in size (~10GB * 64 files, slightly less in fact). Shouldn't they be closer to 2.2TB?
So am I misunderstanding "Tensor type F32 · I32 · BF16" or is it just tagged wrong?
Kimi 2.5 (which this is based on) is served at $0.44 input / $2 output by a ton of different providers on OpenRouter, 2.6 will certainly be similar.
That's about 11X less than Opus for similar smarts.
This should be so easy to prove if it were true. Yet there is none of it, just vibes.
Still, your other two points are completely valid. The opaqueness of usage quotas is a scam, within a single month for a single model it can differ by more than 2x. And this indeed has been proven.
I wish they did more smaller models. Kimi Linear doesn't really count, it was more of a proof of concept thing.
Private companies will never open up a technological breakthrough to their competitors. It just doesn't make sense. If you want an entire field to advance, you have to open it up.
I do wonder where we go from here.
The test data is purposely difficult to access to reduce the chance of leaking it into the training dataset.
Also discovered that using OpenCode instead of the kimi cli, really hurts the model performance (2.5).
Transcript and HTML here: https://gist.github.com/simonw/ecaad98efe0f747e27bc0e0ebc669...
I mean the prompt was succinct and clear, as always - and it still decided to hallucinate multiple features (animation + controls) beyond the prompt.
It'd also like to point out that to date no drawing was actually good from an actual quality perspective (as in comparative to what a decent designer would throw together)
Theyre always only "good" from the perspective of it being a one shot low effort prompt. Very little content for training purposes.
Might be a configuration or prompt issue. I guess I'll wait and see, but I can't get use out of this now.
EDIT: Wrong comment: they compared it with 4.6, my comment was for the Qwen-3.6 Max release blog post...
irthomasthomas•1h ago
NitpickLawyer•1h ago
cedws•1h ago
osti•53m ago
NitpickLawyer•48m ago
The ~100k hardware is suitable for multi-user, small team usage. That's what you'd use for actual work in reasonable timeframes. For personal use, sure macs could work.
zozbot234•26m ago
BoorishBears•1h ago
irthomasthomas•1h ago
pixel_popping•44m ago