🔥 Introducing the Helicone AI Gateway (in beta) - reach 100+ models with a single integration.

Helicone AI Gateway - Now Available!

June 19, 2025

We’re thrilled to announce the launch of Helicone AI Gateway - a powerful open-source solution for routing, caching, and managing your LLM traffic at scale.

🚀 What is Helicone AI Gateway?

The AI Gateway is a high-performance proxy that sits between your application and LLM providers, offering enterprise-grade features:

  • Smart Load Balancing: Distribute requests across multiple providers
  • Intelligent Caching: Reduce costs with semantic caching
  • Automatic Failover: Seamlessly switch providers during outages
  • Rate Limiting: Protect against abuse and control costs
  • Built-in Observability: Full integration with Helicone’s analytics

💻 Get Started

The AI Gateway is available as a separate open-source project:

GitHub Repository: github.com/helicone/ai-gateway

Quick start with Docker:

docker run -p 8080:8080 helicone/ai-gateway

📚 Learn More

🔧 Key Features

  • Multi-Provider Support: OpenAI, Anthropic, Azure, and more
  • Request Routing: Route by model, cost, or custom rules
  • Security: API key management and request validation
  • Performance: Built in Rust for minimal latency overhead

Start using the AI Gateway today to take control of your LLM infrastructure!