AI evaluation and testing platform for measuring and improving LLM output quality
Braintrust has built credibility partly through association with well-known engineering teams at Stripe and Figma who've publicly discussed their use of it. AI engineers on Hacker News and Latent Space discussions consistently recommend it for teams serious about LLM quality. Some evaluators prefer Langfuse for its open source self-hosting and tighter tracing integration. The online eval feature for production traffic scoring is frequently cited as the differentiator that justifies the paid tier.
Open-source AI pair programmer that works directly in your terminal
Open-source AI coding assistant for VS Code and JetBrains - bring your own model
The most widely used framework for building LLM-powered applications and agents
Static analysis tool that finds security bugs using customizable pattern rules
AI pair programmer that suggests code in real-time inside your editor
AI-native code editor built for fast, context-aware development
Anthropic's agentic CLI for autonomous coding directly in your terminal
AI agent that builds and deploys full apps from natural language descriptions