Provider dashboards mostly show totals. Opsmeter shows what caused the bill by breaking spend down by endpointTag, promptVersion, and optionally userId — plus latency and success/error rates.
It’s no-proxy: Opsmeter doesn’t sit in your request path. After each LLM call, you send a small telemetry payload to /v1/ingest/llm-request (provider, model, endpointTag, promptVersion, token counts, latency, status). Opsmeter normalizes cost via a provider/model pricing table and surfaces trends + regressions.
Links:
Home: https://opsmeter.io
Docs: https://opsmeter.io/docs
Pricing: https://opsmeter.io/pricing
If you try it and share anonymized screenshots/feedback, I’m happy to help you interpret the results — e.g.
which endpoints drive spend
which prompt versions increased tokens/cost (deploy regressions)
which users (optional) are the biggest cost drivers
suggested budget thresholds (80% warning / 100% exceeded) and alerting setup
Feedback welcome — especially on what you’d want next: staying telemetry-first, and potentially adding an optional gateway mode later.