I’m trying to understand what’s in the way. My hypotheses (and I’d love corrections):
1) People optimize for output quality over privacy. 2) Hardware is far behind. 3) The tool people truly want (e.g., “a trustworthy, local‑only browser extension”) have yet to emerge. 4) No one has informed your lawyer about this—for now. 5) Or: adoption is already happening, just not visible.
It’s possible many teams are quietly using Ollama in daily work, and we just don’t hear about it.
codeptualize•18h ago
1. Self hosting
2. Running locally on device
I have tried both, and find myself not using either.
For both the quality is less than the top performing models in my experience. Part of it is the models, part might be the application layer (chatgpt/claude). It would still work for a lot of use cases, but it certainly limits the possibilities.
The other issue is speed. You can run a lot of things even on fairly basic hardware, but the token speed is not great. Obviously you can get better hardware to mitigate that but then the cost goes up significantly.
For self hosting, you need a certain amount of throughput to make it worth it to have GPU's running. If you have spiky usage you are either paying a bunch for idle GPU's or you have horrible cold start times.
Privacy wise: The business/enterprise TOS's of all big model providers give enough privacy guarantees for all or at least most use cases. You can also get your own OpenAI infra on Azure for example, I assume with enough scale you can get even more customized contracts and data controls.
Conclusion: Quality, speed, price, and you are able to use the hosted versions even in privacy sensitive settings.
briansun•7h ago
Have you run into cases where on‑device still makes sense?
1. Data that is contractually/regulatorily prohibited from being sent to any third‑party processor (no exceptions).
2. Very large datasets where throughput can be low (overnights acceptable) but the cost is high for cloud models.
3. Inputs behind a password-wall that hosted assistants/chatgpt/claude can’t reach and can't do agentic things with them.