TrueFoundry AI Gateway is the proxy layer that sits between your applications and the LLM providers and MCP Servers. It is an enterprise-grade platform that enables users to access 1000+ LLMs using a unified interface while taking care of observability and governance.
Call 1000+ LLMs using a single endpoint with unified API interface
Generate and manage API keys for users/applications
Support for text, image, and audio inputs across compatible models
Fine-grained access control and permissions management
Control Models Usage with flexible rate limiting policies per user/model/application
Automatic failover to backup models when primary models are unavailable
Distribute requests across multiple model instances based on weight, latency or cost metrics.
Control spending and enforce cost limits for users, teams, and models
Content filtering and safety checks to ensure
Opentelemetry compliant metrics and logging for all requests.
Centralized prompt playground with versioning and management system
Process multiple requests efficiently with batch processing
Deploy and manage your own MCP servers with TrueFoundry AI Gateway.
Define which developers / teams / applications can access which MCP servers.
One API key to access all MCP servers and their tools.
Test Agents by adding tools and models from Playground
Connect to MCP Servers with a single API in the gateway.
Coming Soon
🔄 Zero Vendor Lock-In
Our gateway is OpenAI compatible and can be used with any OpenAI-compatible client. User doesn’t need to change their existing codebase or clients to use truefoundry’s AI gateway.
Truefoundry doesn’t provide a client side SDK for the gateway and instead recommend using OpenAI client or requests library. This ensures that your code doesn’t get vendor locked in to Truefoundry and you can switch Gateways easily.
🔒 Enterprise-Grade Security
🏅 Trusted by Fortune 500 Companies: Multiple Fortune 500 companies are using TrueFoundry. Read about our case studies here.
📊 Observability
🛡️ Built for Scale
🚀 Performance Benchmark: Learn how TrueFoundry AI Gateway delivers blazing-fast performance
Read our benchmark analysis →
We integrate with 1000+ LLMs through the following providers.
If you don’t see the provider you need, there is a high change it will just work as self hosted models or OpenAI provider. Please reach out to us at support@truefoundry.com and we will be happy to guide you.
The Truefoundry AI Gateway can either be used as a SaaS offering or deployed on-premise.
What's the performance impact of using the gateway?
The latency overhead is minimal, typically less than 5ms. Our benchmarks show enterprise-grade performance that scales with your needs.
Our SaaS offering is hosted in multiple regions across the world to ensure low latency and high availability. You can also deploy the gateway on-premise or on any cloud provider in your region which
is closer to your users.
AI Gateway on the edge, close to your applications for optimal performance
Can I deploy the gateway on-premise?
Yes, the AI Gateway supports on-premise deployments on any infrastructure or cloud provider, giving you complete control over your AI operations.
How do I integrate my self-hosted models?
You can easily integrate any OpenAI-compatible self-hosted model. Check our self-hosted models guide for detailed instructions.
Can I use the gateway without the full MLOps platform?
Yes, The AI Gateway can be used as a standalone solution. You can use the full MLOps platform if you’re using features like model deployment(traditional models and LLMs), model training, llm fine-tuning or training/data-processing workflows.
TrueFoundry AI Gateway is the proxy layer that sits between your applications and the LLM providers and MCP Servers. It is an enterprise-grade platform that enables users to access 1000+ LLMs using a unified interface while taking care of observability and governance.
Call 1000+ LLMs using a single endpoint with unified API interface
Generate and manage API keys for users/applications
Support for text, image, and audio inputs across compatible models
Fine-grained access control and permissions management
Control Models Usage with flexible rate limiting policies per user/model/application
Automatic failover to backup models when primary models are unavailable
Distribute requests across multiple model instances based on weight, latency or cost metrics.
Control spending and enforce cost limits for users, teams, and models
Content filtering and safety checks to ensure
Opentelemetry compliant metrics and logging for all requests.
Centralized prompt playground with versioning and management system
Process multiple requests efficiently with batch processing
Deploy and manage your own MCP servers with TrueFoundry AI Gateway.
Define which developers / teams / applications can access which MCP servers.
One API key to access all MCP servers and their tools.
Test Agents by adding tools and models from Playground
Connect to MCP Servers with a single API in the gateway.
Coming Soon
🔄 Zero Vendor Lock-In
Our gateway is OpenAI compatible and can be used with any OpenAI-compatible client. User doesn’t need to change their existing codebase or clients to use truefoundry’s AI gateway.
Truefoundry doesn’t provide a client side SDK for the gateway and instead recommend using OpenAI client or requests library. This ensures that your code doesn’t get vendor locked in to Truefoundry and you can switch Gateways easily.
🔒 Enterprise-Grade Security
🏅 Trusted by Fortune 500 Companies: Multiple Fortune 500 companies are using TrueFoundry. Read about our case studies here.
📊 Observability
🛡️ Built for Scale
🚀 Performance Benchmark: Learn how TrueFoundry AI Gateway delivers blazing-fast performance
Read our benchmark analysis →
We integrate with 1000+ LLMs through the following providers.
If you don’t see the provider you need, there is a high change it will just work as self hosted models or OpenAI provider. Please reach out to us at support@truefoundry.com and we will be happy to guide you.
The Truefoundry AI Gateway can either be used as a SaaS offering or deployed on-premise.
What's the performance impact of using the gateway?
The latency overhead is minimal, typically less than 5ms. Our benchmarks show enterprise-grade performance that scales with your needs.
Our SaaS offering is hosted in multiple regions across the world to ensure low latency and high availability. You can also deploy the gateway on-premise or on any cloud provider in your region which
is closer to your users.
AI Gateway on the edge, close to your applications for optimal performance
Can I deploy the gateway on-premise?
Yes, the AI Gateway supports on-premise deployments on any infrastructure or cloud provider, giving you complete control over your AI operations.
How do I integrate my self-hosted models?
You can easily integrate any OpenAI-compatible self-hosted model. Check our self-hosted models guide for detailed instructions.
Can I use the gateway without the full MLOps platform?
Yes, The AI Gateway can be used as a standalone solution. You can use the full MLOps platform if you’re using features like model deployment(traditional models and LLMs), model training, llm fine-tuning or training/data-processing workflows.