The AI Gateway, currently in alpha for all users, lets you switch between ~100 AI models without needing to manage API keys, rate limits, or provider accounts.
The Observability tab in the Vercel dashboard now includes a dedicated AI section to surface metrics related to the AI Gateway. This update introduces visibility into:
You can view these metrics across all projects or drill into per-project and per-model usage to understand which models are performing well, how they compare on latency, and what each request would cost in production.
Learn more about Observability.
Read more
Continue reading...
The Observability tab in the Vercel dashboard now includes a dedicated AI section to surface metrics related to the AI Gateway. This update introduces visibility into:
Requests by model
Time to first token (TTFT)
Request duration
Input/output token count
Cost per request (free while in alpha)
You can view these metrics across all projects or drill into per-project and per-model usage to understand which models are performing well, how they compare on latency, and what each request would cost in production.
Learn more about Observability.
Read more
Continue reading...