Skip to main content
Providers are the AI services that host and serve models. The AI Gateway routes your requests to providers and handles authentication, failover, and observability.

Managed providers

These providers are available with AI Gateway API Keys—no provider accounts needed. ngrok handles authentication automatically.

OpenAI

GPT and o-series models. Managed keys available.

Anthropic

Claude models. Managed keys available. Supports both OpenAI and Anthropic SDK formats.

BYOK providers

These providers require you to bring your own key. See each provider’s page for setup instructions.

OpenRouter

Access hundreds of models from multiple providers through a single API.

Google

Gemini models from Google AI Studio.

Groq

LPU-accelerated inference for open-source models (Llama, Mixtral).

DeepSeek

High-performance reasoning and chat models.

Hyperbolic

Open-source model hosting with high-performance inference.

InceptionLabs

Diffusion-based language models for fast text generation.

Inference.net

Distributed inference network for AI models at scale.

Self-hosted providers

Run open-source models on your own infrastructure and connect them to the gateway.

Ollama

Run open-source models locally with Ollama.

vLLM

High-performance inference server.

LM Studio

Desktop app for local model inference.

Azure OpenAI

Microsoft’s OpenAI service on Azure.

How provider selection works

When a request arrives, the gateway determines which provider to use:
  1. Explicit provider prefix: if the model name includes a provider prefix (for example, openai:gpt-4o or openrouter:anthropic/claude-3.5-sonnet), that provider is used
  2. Catalog lookup: the gateway looks up the model ID in its catalog to find the default provider
  3. Selection strategy: if configured, model selection strategies can override the default

Next steps