Metrics for Model Selection
Real-time latency and error rate metrics available in CEL expressions for intelligent routing decisions.
Traffic Inspector
View full request and response bodies for debugging in the ngrok dashboard.
Log Exporting
Export HTTP logs to external systems for logging and analysis.
The AI Gateway is in Early Access. We’re actively working on additional observability features including dedicated dashboards, provider/retry tracking, token usage visualization, and cost analytics.Have feedback? We’d love to hear what observability features would be most valuable to you. Email [email protected] with your ideas.
Current capabilities
| Feature | Status | Description |
|---|---|---|
| CEL Metrics | ✅ Available | Latency, error rates, token counts available in model_selection strategies |
| Traffic Inspector | ✅ Available | Standard HTTP request/response inspection |
| Log Exporting | ✅ Available | Standard ngrok HTTP traffic logs |
| AI-specific Events | 🚧 Roadmap | Provider selection, retry tracking, token usage events |
| Dedicated Dashboard | 🚧 Roadmap | AI Gateway-specific analytics and visualizations |
Quick example: Metrics-based routing
Use real-time metrics to route to the fastest, most reliable models:Understanding metric scope
For full details on available metrics and their scopes, see Metrics Reference.Coming soon
Enhanced observability features are actively being developed:- AI Gateway Dashboard - Dedicated views for provider performance, token usage, and costs
- Provider Tracking - See which providers handle each request
- Retry Visibility - Track failover decisions and retry attempts
- Token Analytics - Visualize token usage trends across models
- Cost Estimation - Automatic cost calculation based on provider pricing