- Replaces Anthropic models with NVIDIA-NIM models: It acts as middleware between Claude-Code and NVIDIA-NIM allowing unlimited usage upto 40 RPM with a free NVIDIA-NIM api-key.
- Replaces the Claude mobile app with telegram: Give it access to some directories, send it tasks from telegram and watch it work autonomously.
It has features that distinguish it from similar proxies:
- The interleaved thinking tokens generated between tool calls are preserved allowing reasoning models like GLM 4.7 and kimi-k2.5 to take full advantage of thinking from previous turns.
- Fast prefix detection (+4 other optimizations) stops the CLI from sending bash command prefix classification requests to the LLM making it feel blazing fast.
- Built in rate limiting and session concurrency.
The code is modular so that adding other providers or messaging apps is easy. Hope the community likes it, any PRs are welcome.
No comments