I’ve been working on Jiss, a project to make open-source LLMs accessible without the massive cloud bills or the headache of managing your own 24/7 infrastructure.
It’s a distributed network that routes inference requests to a pool of volunteer-run workers. If you have an app already using OpenAI, you can just swap the base_url and start running models like Llama 3 or Qwen3.
The idea:
Instead of relying on big providers, Jiss is powered by people donating idle compute. Whether you’re running Ollama on a Mac, a home server, or a workstation, you can connect your local setup to the network and help serve requests. I built this to see if we could create a useful, shared resource for the open-source AI community.
The highlights:
• Drop-in replacement: Use the standard OpenAI SDK—just change the base URL.
• Streaming: Supports real-time token streaming (SSE).
• Open: Anyone can join as a worker using a single script.
It’s still early so I’d love to get your feedback or have you try running a worker!