Skip to main content

MLflow AI Gateway

warning

MLflow AI Gateway does not support Windows.

MLflow AI Gateway provides a unified interface for deploying and managing multiple LLM providers within your organization. It simplifies interactions with services like OpenAI, Anthropic, and others through a single, secure endpoint.

The gateway server excels in production environments where organizations need to manage multiple LLM providers securely while maintaining operational flexibility and developer productivity.

Unified Interface

Access multiple LLM providers through a single endpoint, eliminating the need to integrate with each provider individually.

Centralized Security

Store API keys in one secure location with request/response logging for audit trails and compliance.

Provider Abstraction

Switch between OpenAI, Anthropic, Azure OpenAI, and other providers without changing your application code.

Zero-Downtime Updates

Add, remove, or modify endpoints dynamically without restarting the server or disrupting running applications.

Cost Optimization

Monitor usage across providers and optimize costs by routing requests to the most efficient models.

Team Collaboration

Shared endpoint configurations and standardized access patterns across development teams.

Getting Started

Choose your path to get up and running with MLflow AI Gateway:

Quick Start

Get your AI Gateway running with OpenAI in under 5 minutes:

Install MLflow with gateway dependencies:

pip install 'mlflow[gateway]'

Supported Providers

MLflow AI Gateway supports a comprehensive range of LLM providers:

ProviderChatCompletionsEmbeddingsNotes
OpenAIGPT-3.5, GPT-4, text-embedding models
Azure OpenAIEnterprise OpenAI with Azure integration
AnthropicClaude models via Anthropic API
CohereCommand and embedding models
AWS BedrockClaude, Titan, and other Bedrock models
PaLMGoogle's PaLM models
MosaicMLMPT models and custom deployments
MLflow ModelsYour own deployed MLflow models

Core Concepts

Understanding these key concepts will help you effectively use the AI Gateway:

Endpoints

Endpoints are named configurations that define how to access a specific model from a provider. Each endpoint specifies the model, provider settings, and access parameters.

Providers

Providers are the underlying LLM services (OpenAI, Anthropic, etc.) that actually serve the models. The gateway abstracts away provider-specific details.

Routes

Routes define the URL structure for accessing endpoints. The gateway automatically creates routes based on your endpoint configurations.

Dynamic Updates

The gateway supports hot-reloading of configurations, allowing you to add, modify, or remove endpoints without restarting the server.

Next Steps

Ready to dive deeper? Explore these resources: