Hi — I'm Raphael Mansuy. I built edgequake-litellm to provide a low-latency, Rust-backed drop-in replacement for LiteLLM. It exposes the same Python API (`completion()`, `acompletion()`, `stream()`, `embedding()`), supports provider/model routing (OpenAI, Anthropic, Gemini, Mistral, xAI, OpenRouter, Ollama, LM Studio, etc.), and ships as a single ABI3 wheel with zero Python runtime deps.
Quick migration:
```python
import edgequake_litellm as litellm # drop-in alias
```
Why build it? LiteLLM is excellent, but its pure-Python HTTP layer adds SDK overhead. I moved the core into Rust (edgequake-llm) and wrapped it with PyO3 to cut latency and provide a robust, multi-arch wheel. This is v0.1 — P0 compatibility is in place, but I'd love feedback on priorities: provider coverage, proxy features, billing/budgets, or tool-calling parity.
raphaelmansuy•1h ago
Quick migration:
```python import edgequake_litellm as litellm # drop-in alias ```
Why build it? LiteLLM is excellent, but its pure-Python HTTP layer adds SDK overhead. I moved the core into Rust (edgequake-llm) and wrapped it with PyO3 to cut latency and provide a robust, multi-arch wheel. This is v0.1 — P0 compatibility is in place, but I'd love feedback on priorities: provider coverage, proxy features, billing/budgets, or tool-calling parity.
Install:
pip install edgequake-litellm
Repo: https://github.com/raphaelmansuy/edgequake-llm
If you try it, please star the repo and open issues for features you want most — I'm actively iterating. Happy to answer technical questions here.