- Replaces Anthropic models with NVIDIA-NIM models: It acts as middleware between Claude-Code and NVIDIA-NIM allowing unlimited usage upto 40 RPM with a free NVIDIA-NIM api-key.
- Replaces the Claude mobile app with telegram: Give it access to some directories, send it tasks from telegram and watch it work autonomously.
It has features that distinguish it from similar proxies:
- The interleaved thinking tokens generated between tool calls are preserved allowing reasoning models like GLM 4.7 and kimi-k2.5 to take full advantage of thinking from previous turns.
- Fast prefix detection (+4 other optimizations) stops the CLI from sending bash command prefix classification requests to the LLM making it feel blazing fast.
- Built in rate limiting and session concurrency.
The code is modular so that adding other providers or messaging apps is easy. Hope the community likes it, any PRs are welcome.