Public alpha — open source SDK shipping

Stop guessing what
your AI actually costs.

AI bills are like restaurant bills with no itemization — you only see the total. We break it down per agent, per user, per endpoint. Then tell you where to cut.

Free tier: 100K events/mo. No credit card. Open source SDK (MIT).

acr-waitlist.hhcreativelab.workers.dev/dashboard
Today's spend
$47.20
▼ 38% vs yesterday
Top cost agent
research-bot
→ Switch to Sonnet, save 62%
Cache hit rate
87%
↑ above 85% target
Cost per agent (last 24h)
research-bot
$22.40
writer
$11.80
classifier
$5.20
other (8 agents)
$7.80
The problem

Three reasons your AI bill keeps surprising you.

Bill goes $200 → $1,100 in a week

A prompt change, a runaway loop, a new feature — and you find out only at month-end. By then the damage is done.

No idea which agent is the culprit

Your provider dashboard shows one big number. Was it the research bot? The classifier loop? The customer support agent? You're guessing.

Optimization is gut feel, not data

Should this task run on Haiku or Sonnet? Is prompt caching even working? Without per-call attribution, every change is a coin flip.

What ACR does

Three lines. Total visibility. Actual recommendations.

01 / TRACK

Real-time per-call tracking

Drop-in SDK wraps your Anthropic, OpenAI, or any provider client. Every call gets logged with cost, latency, model, agent name, and user ID. No data leaves your infra unless you opt in.

  • • Multi-vendor: Anthropic, OpenAI, Google, local models
  • • Hashed keys, zero PII by default
  • • <5ms overhead per call
# Python
from acr import track

client = track(Anthropic(), agent="research-bot")

# Use normally — every call is metered
response = client.messages.create(...)
02 / ATTRIBUTE

Per-agent / per-user breakdown

Slice and dice by agent, endpoint, user, customer tier, or any custom tag. Find the 3 calls eating 40% of your budget. See cost-per-customer if you're a B2B SaaS.

  • • Custom tags (env, feature, tenant)
  • • 14-day rolling history (free), 12mo (Pro)
  • • Slack/email alerts on threshold breach
# Top 5 agents this week
research-bot$142.20
writer$87.50
classifier$34.10
support-agent$22.80
summarizer$18.40
03 / OPTIMIZE

Routing recommendations that pay rent

We analyze your usage patterns and tell you exactly where Haiku will replace Sonnet without quality loss. Where prompt caching is leaking. Where Batch API gets you 50% off.

  • • Model downgrade suggestions (with quality safety net)
  • • Cache hit rate diagnostics
  • • Batch API candidate detection
→ Recommendation
Agent classifier uses Opus for 1,240 calls/day.
Output is structured JSON, <200 tokens.

Switch to Haiku → save ~$340/mo.
Confidence: high (98% similar workloads kept quality).
vs the field

How we compare.

Feature Cost Radar Helicone LangSmith
Starting price $29/mo $80/mo $39/mo (LC only)
Free tier 100K events/mo 10K events/mo 5K traces/mo
Open source SDK Yes (MIT) Partial No
Multi-provider Yes Yes LangChain only
Routing recommendations Yes No No
Self-host option Yes (Enterprise) Yes No
Setup time 3 lines ~10 lines + proxy ~30 min

Pricing accurate as of 2026-05. Compare directly — links in docs.

Pricing

Pay for what you measure. Not for the dashboard.

Free
$0
forever
  • 100K events/mo
  • 14-day history
  • 1 project, 1 user
  • Open source SDK
  • Email alerts
Start free
Most popular
Pro
$29/mo
per workspace, billed monthly
  • 1M events/mo
  • 12-month history
  • Unlimited projects
  • Routing recommendations
  • Slack alerts + webhooks
  • CSV export
30-day free trial
Team
$79/mo
per workspace, up to 10 users
  • 5M events/mo
  • 24-month history
  • Up to 10 team members
  • Custom dashboards
  • RBAC + audit log
  • Priority support
Start trial

Need self-hosted or >5M events/mo? Talk to us about Enterprise ($500+/mo). Annual plans get 20% off.

Built in the open

We use this every day on our own platform.

50
agents tracked
1,840
events / day
$487
saved last 30d

"Used in production: internal developer tooling — 50 agents, ~1,840 events/day. Caught 3 cost regressions before they hit the bill."

FAQ

Common questions.

Does this send my prompts to your servers?

No. By default we only log metadata: model, token counts, latency, cost. Your prompt and response content never leaves your infra. Opt-in content logging is available for debugging if you want it.

Which providers are supported?

Anthropic, OpenAI, Google (Gemini), Mistral, Groq, and any OpenAI-compatible endpoint (local Ollama, vLLM, etc.). Adding a provider takes ~50 lines — PRs welcome.

Can I self-host?

Yes — Enterprise plan ships a Docker compose stack (Postgres + the dashboard). Air-gapped deployments supported. Talk to us for licensing.

How do routing recommendations work?

We analyze your call patterns — task type, output structure, latency tolerance, observed quality signals. Then we suggest concrete swaps (Opus → Sonnet, Sonnet → Haiku) with a confidence score. You always approve before any change is enforced; we never change your code.

What if I cancel?

Account drops back to Free tier — your SDK keeps working, you keep 14 days of history, no data is deleted for 90 days. Full export available anytime as CSV / JSON.

Find out what your AI actually costs.

Three lines of code. Five-minute setup. Then never get surprised by an Anthropic bill again.