output tokens: $1.5 per 1,000
that's either one hell of a typo or my god I'll be broke in an hour if I accidentally use this service
The moral of the story,if something gets screwed up, I end up paying
https://tokensaver.org/api/pricing
Is offering GPT 3.5 Turbo and Gemini 1.5 Pro.
I wonder how many dollars OP loaded on their API key. So far according to stats you've spent $13.65 for a few hundred thousand tokens people sent up.
{ "success": true, "totals": { "requests": 491, "revenue": 0, "cost": 13.650696, "profit": -13.650696, "inputTokens": 411557, "outputTokens": 258956, "customers": 32 }, "byProvider": [ { "provider": "anthropic", "requests": 491, "cost": 13.650696, "revenue": 0, "profit": -13.650696 } ] }
Moral of the story is, the only one who loses $ is me
I created it for people using AI in development. The 1st 50 tokens are free.
We have openrouter at home!
In all seriousness, the value proposition is weird to me. The most expensive queries are the ones with huge contexts, and therefore the ones I'd less likely to use cheap models.
> Message Privacy: Your API requests are processed and immediately forwarded. We never store or log conversation content.
> Minimal Data: We only store your email and usage records. Nothing else. Your data stays yours.
Source: trust me bro.
- GPT-5.1 is $1.25 / 1M tokens
- You are $0.50 / 1,000 tokens
Output:
- GPT-5.1 is $10.00 / 1M tokens
- You are $1.50 / 1,000 tokens
Am I reading that wrong? Is that a typo?
As these routing engines evolve, I wonder how you see them handling drift or divergence when different models produce structurally incompatible outputs.
Any thoughts on lightweight harmonization layers?
On top of that logging in does not require a password, just an email address.
> Six months ago, I was running a customer support chatbot for a SaaS product. Nothing fancy - ...
I'm sure this toooootally happened
> curl -X POST https://tokensaver.org/api/chat \ -H "Content-Type: application/json" \ -d '{ "email": "your@email.com", "messages": [ {"role": "user", "content": "Hello!"} ] }'
Am I getting this right that there's no auth? Just provide an email and get free requests?
Edit:
This seems to always use sonnet 3.5 no matter the request. I asked it a USAMO problem and it still used sonnet (and hallucinated the wrong answer of course).
It's a simple tool. Use it or don't. The only person who would lose money in error is me.
The prompt and the model go hand in hand. If you randomly select the model the likelihood of getting something consistent is basically zero.
Also model pricing don't very that much. I have never heard of spot-instance equivalent for inference although that will be cool. The demand for GPU is so high right now that I think most datacenters are at 100% utilisation.
Btw landing page does not bring much confidence this is serious. Might want to change it to communicate better and also to be attractive to "developers" I guess.
I'm curious when AI pricing will couple with energy markets. Then the location of the datacentre will matter considerably
https://www.aiville.com/c/anthropic/api-that-auto-routes-to-...
h2o_wine•2mo ago
The problem: AI API pricing is a mess. OpenAI, Anthropic, and Google all have different pricing models, rate limits, and availability. Switching providers means rewriting code. Most devs just pick one and overpay.
The solution: One endpoint. Drop-in replacement for OpenAI's API. Behind the scenes, it checks current pricing and routes to whichever provider (GPT-4o, Claude, Gemini) costs least for that specific request. If one fails, it falls back to the next cheapest.
How it works: - Estimates token count before routing - Queries real-time provider costs from database - Routes to cheapest available option - Automatic fallback on provider errors - Unified response format regardless of provider
Typical savings: 60-90% on most requests, since Gemini Flash is often free/cheapest, but you still get Claude or GPT-4 when needed.
30 free requests, no card required: https://tokensaver.org
Technical deep-dive on provider pricing: https://tokensaver.org/blog/openai-vs-anthropic-vs-gemini-pr...
I wrote up how to reduce AI costs without switching providers entirely: https://tokensaver.org/blog/reduce-ai-api-costs-without-swit...
Happy to answer questions about the routing logic, pricing model, or architecture.
jasonsb•2mo ago
This claim seems overstated. Accurately routing arbitrary prompts to the cheapest viable model is a hard problem. If it were reliably solvable, it would fundamentally disrupt the pricing models of OpenAI and Anthropic. In practice, you'd either sacrifice quality on edge cases or end up re-running failed requests on pricier models anyway, eating into those "savings".
moduspol•2mo ago
growt•2mo ago
kbaker•2mo ago
> OpenRouter provides a unified API that gives you access to hundreds of AI models through a single endpoint, while automatically handling fallbacks and selecting the most cost-effective options. Get started with just a few lines of code using your preferred SDK or framework.
It isn't OpenAI API compatible as far as I know, but they have been providing this service for a while...
minimaxir•2mo ago
growt•2mo ago
h2o_wine•2mo ago