Comparison
MCP-native, zero-code observability vs SDK-powered eval and experimentation platform. Two different philosophies for AI quality.
TL;DR
Feature Comparison
| Feature | Iris | Braintrust |
|---|---|---|
| Integration method | MCP config (zero code) | SDK imports (Python, TS, Go, Ruby, C#) |
| Self-hosting | Single SQLite file | Enterprise plan only (cloud-first) |
| Performance overhead | Zero (no SDK in hot path) | Async logging, minimal overhead |
| Eval approach | 12 built-in + 8 custom heuristic rules (<1ms) | LLM, code, and human scoring + datasets + experiments |
| Prompt playground | Not included | Full playground with side-by-side comparison |
| Datasets & experiments | Not included | Production traces to datasets, experiment tracking, CI integration |
| Cost tracking | Per-trace USD cost | Per-trace cost, per-user/feature/model breakdowns |
| MCP support | Protocol-native (IS an MCP server) | MCP server for querying Braintrust data |
| License | MIT (fully permissive) | Proprietary (proxy is MIT) |
| Pricing | Free & open-source | Free tier (1M spans) / Pro $249/mo / Enterprise custom |
| Tracing depth | MCP tool calls and agent traces | Full trace trees with token-level detail, visual timeline |
| Enterprise features | Roadmap (v0.5) | SOC 2, SSO, hybrid deployment, dedicated support |
Decision Guide
Last verified: March 2026. This comparison is based on publicly available documentation and may not reflect recent changes to Braintrust. We aim to keep this page accurate and fair.
See something outdated or incorrect? Report an inaccuracy — we review and update within 48 hours.
Add Iris to your MCP config. First trace in 60 seconds. No SDK, no signup, no infrastructure.