Mixpanel for Agent Products

Stop Drowning in Trace Logs.
Start Getting Answers.

TwoTail is an AI analyst that clusters your agent traces, diagnoses root causes, and runs optimization experiments.

Works with your stack

OpenTelemetry LangChain LlamaIndex CrewAI OpenAI Agents SDK Custom Agents Vercel AI SDK AutoGen OpenTelemetry LangChain LlamaIndex CrewAI OpenAI Agents SDK Custom Agents Vercel AI SDK AutoGen

From Agent Data to Agent Optimization

See how TwoTail turns raw agent telemetry into actionable improvements.

twotail — agent analytics
INCOMING TRACES trace-a82f gpt-4 search_api route trace-d1c3 claude-3 embed gpt-4-mini retry trace-f7e2 gpt-4 search_api plan embed trace-b4a9 claude-3 search_api ROOT CAUSE BREAKDOWN Model + Tool Success Rate Latency Avg Cost gpt-4 + search_api 20% 9.1s $0.42 gpt-4 + embed 50% 4.3s $0.28 claude-3 + search_api 80% 1.8s $0.12 ⚠ gpt-4 × search_api is dragging down success rate 💡 Experiment Add model router Route search tasks by complexity simple → gpt-4-mini · complex → claude-3 RESULT −38% cost/run −34% p50 latency +6% success ✓ SHIPPED Success Rate Eval Latency Cost Before After

What Changes with TwoTail

The tools you already use, supercharged with AI analysis.

Before Experiment
Agent Experiments

Run controlled before-and-after tests on prompts, models, and configs. TwoTail tracks the metrics so you can compare variants with confidence.

Before Evals
Eval Augmentation

Go beyond pass/fail scores. TwoTail enriches your evals with trace context, clusters similar failures, and surfaces the patterns your eval harness misses.

Before Debug Traces
Trace Inspection & Clustering

Stop scrolling through raw JSON. TwoTail groups traces by behavior, highlights anomalies, and lets you drill into any span with a single click.

The Complete Agent Analytics Platform

Everything you need to understand, debug, and improve your agents.

Easy Integration

Send OpenTelemetry traces from any framework. Works with LangChain, LlamaIndex, or your custom stack.

OpenTelemetry
LangChain
LlamaIndex
CrewAI

Understands Your Product

Teach TwoTail your terminology, user segments, and what success looks like.

response_quality model-as-judge eval for helpfulness

Chat to Chart

Ask questions in plain English. Get charts and insights in seconds.

show me failure rates by tool used
show me the daily failure rate

Analytics Playbooks

Pre-built analyses for common agent patterns: loops, failures, latency spikes.

Run Experiments

A/B test prompts, models, and configurations. Measure what actually improves outcomes.

A
B

Autonomous Analysis

TwoTail continuously monitors your traces and surfaces issues before you ask. Get alerts on regressions, anomalies, and emerging failure patterns — no queries needed.

Latency spike detected in search_tool — p95 up 3.2x since yesterday
New failure cluster: 12% of runs hitting rate limit on embed_api

Common Questions

Everything you need to know about getting started with TwoTail.

Get a Demo

TwoTail accepts traces via OpenTelemetry (OTLP). If your agent framework already emits OTel spans — LangChain, LlamaIndex, CrewAI, or custom setups — just point the exporter at your TwoTail endpoint. No SDK to install.

Those tools show you traces. TwoTail analyzes them. Ask questions in plain English and get charts, cluster similar failures, and surface patterns across thousands of runs — without writing queries or building dashboards.

No. If you already have OpenTelemetry instrumentation, TwoTail works with your existing setup. If you don't, adding a few lines of OTel config is all it takes.

Your data is stored in isolated Supabase-backed Postgres databases with row-level security. Each account's data is fully segregated, and all connections are encrypted in transit.

TwoTail has a free Starter tier for small projects. Check our pricing page for full details on plans and limits.

Your Agents Generate Thousands of Traces.
We Help You Understand Them.

TwoTail turns raw telemetry into the insights you need to ship better agents, faster. No more grep-ing through JSON.