We make token-based billing easy. Very easy.

Turn raw API usage into customer invoices instantly.

AI provider

Pricing response

POST /api/models/pricing/calculate
{
"model": "gpt-4o",
"provider": "openai",
"prompt_tokens": 10000,
"completion_tokens": 2000,
"cached_tokens": 8000,
"customer_id": "acme"
}
200 OK
{
"provider": "openai",
"model": "gpt-4o",
"total": "$0.00000",
"input_rate": null,
"output_rate": null
}

Billing platform

Generated invoice

Stripe

Invoice

#INV-2026-042

Acme Workspace

Usage metered by token

Prompt tokens10,000$0.02500
Completion tokens2,000$0.02000
Cache reads8,000$0.00400
Subtotal$0.00000
Platform fee$0.00000
Total due$0.00000

Optimizing cost? Then find the cheapest model based on
user retention.

Expensive models are only worth it if they perform on your specific tasks. Build a custom leaderboard to find the cheapest model that actually works.

Private Leaderboard Dashboard (Light)

Use your results
to configure the router.

Connect your production data to build your routing engine that switches models in real-time.

IFcomplexity < 0.5THEN HaikuELSE Opus

Single Endpoint

Baseline
Opus100.0%
Avg Latency0.0s
Cost per 1K queries$0.00

With Router

Haiku0.0%
Sonnet0.0%
Opus0.0%
Avg Latency0.0s
Cost per 1K queries$0.00

Wondering how much you're spending?

Narev has an open source observability tool for LLM and Cloud costs.