Introducing AI Gateway

05 January 20265 min read
Introducing AI Gateway

Today we're officially launching AI Gateway on Cencori.

AI teams should not need one SDK for provider access, another layer for failover, a separate security proxy, another tool for logs, and a custom billing system on top of all of it just to ship one product.

AI Gateway is our answer to that problem.

It gives you a single control point for AI traffic: one API for routing, reliability, security, observability, cost tracking, and monetization.

What Ships Today

One API for Every Major Provider

Route traffic through one integration instead of wiring every provider separately.

  • OpenAI
  • Anthropic
  • Google
  • xAI
  • Meta
  • Mistral
  • DeepSeek
  • Groq
  • Cohere
  • Perplexity
  • OpenAI-compatible endpoints and custom provider setups

The gateway supports native Cencori endpoints for chat, embeddings, images, audio, and moderation, plus an OpenAI-compatible endpoint for teams that want the fastest migration path.

Codetext
const client = new OpenAI({
  apiKey: process.env.CENCORI_API_KEY,
  baseURL: "https://api.cencori.com/v1",
});
 
const response = await client.chat.completions.create({
  model: "gpt-4o",
  messages: [{ role: "user", content: "Hello!" }],
});

You can also use the TypeScript SDK, Python SDK, streaming responses, tool calling, and Vercel AI SDK integrations on top of the same gateway layer.

Routing, Retries, and Automatic Failover

Reliability is handled in the gateway, not bolted onto each app.

  • automatic retries for transient upstream failures
  • fallback routing when a provider returns 429, 5xx, or times out
  • model equivalence mapping across providers
  • configurable fallback policies and retry limits
  • circuit breakers that stop sending traffic into a degraded path
  • load balancing across multiple provider keys to reduce hot spots and rate-limit pressure

That means your application can send requests to Cencori once, and Cencori handles the resilience work behind the scenes.

Built-In AI Security

Every request can be inspected before it reaches a model, and every response can be evaluated before it reaches your user.

  • prompt injection and jailbreak detection
  • PII scanning for prompts and responses
  • content filtering across harmful categories
  • unified moderation controls
  • configurable actions to mask, redact, block, or allow with logging
  • structured security errors and incident records for blocked requests

Cencori's security layer is designed for production systems where safety has to happen in real time, not as an afterthought.

Full Observability

AI Gateway ships with observability built in.

  • request and response logs for every model call
  • provider, model, status, token usage, latency, and time-to-first-token
  • cost tracking per request with continuously updated provider pricing
  • analytics for spend, throughput, latency percentiles, and error rates
  • traces for multi-step workflows, agent runs, retrieval chains, and tool calls
  • web telemetry so HTTP traffic and AI traffic can be viewed together
  • export paths for enterprise teams that need raw data in their own systems

For most teams, this removes the need to stitch together a separate logging or AI observability product just to understand production traffic.

Billing and Monetization

The gateway does more than route traffic. It also helps teams control and monetize AI usage.

  • built-in cost tracking
  • rate limiting and usage governance
  • spend visibility per request and per project
  • end-user billing when you pass a user id through the gateway
  • quota enforcement, rate plans, markup pricing, and invoices
  • Stripe Connect support for payout flows
  • Usage Events API for teams that want the billing layer even if they still call providers directly

If you already route traffic through Cencori, billing becomes dramatically simpler because usage, cost, and enforcement all happen in the same place.

Developer Platform and Open Integration Surface

AI Gateway is built to fit into existing stacks without forcing a rewrite.

  • OpenAI-compatible chat endpoint
  • TypeScript and Python SDKs
  • support for streaming and tool calling
  • API key management
  • custom providers and model mappings
  • Vercel AI SDK support
  • clean migration path for teams moving from direct provider usage

The goal is simple: swap one base URL, keep shipping.

Compliance, Auditability, and Enterprise Controls

For teams moving AI into production, operational controls matter as much as model access.

AI Gateway ships with:

  • audit logs for administrative changes
  • security incident logging
  • policy enforcement controls
  • project-level governance
  • compliance-ready records for teams that need traceability

This is what turns an AI proxy into production infrastructure.

Why This Matters

Before AI Gateway, teams typically had to solve the same set of infrastructure problems over and over:

  • how to route between providers
  • how to survive provider outages
  • how to inspect and block risky prompts
  • how to understand latency and cost
  • how to rate limit and meter usage
  • how to charge end-users for AI features

That slows teams down and creates brittle systems.

AI Gateway makes Cencori the runtime control point for every model request. You integrate once. We handle the hard infrastructure layer behind it.

Get Started

This is the first major layer of the Cencori platform, and it is live now.