
Overview Metrics
Track key usage statistics:- Traces: Total number of prompt executions
- Threads: Number of conversation threads
- Users: Number of unique users
LLM Metrics
Monitor your AI model usage:- LLM Calls: Number of API calls made
- Total Cost: Cost of all API calls
- Tokens: Total tokens consumed
Version Tracking
- Track prompt behavior by version, compare different versions
- Filter messages, plot usage, cost, conversion on different prompts
Evaluations Metrics
- Run real-time evaluations on the traces to measure prompt performance
- Use real-time evaluators for classification of prompt outputs
Custom Graphs
- Create custom bar, line, pie, scatter, and more charts with any captured metrics
- Compare different prompts and versions
← Back to Prompt Management Overview