Introducing AI Gateway

Today we're officially launching AI Gateway on Cencori.
AI teams should not need one SDK for provider access, another layer for failover, a separate security proxy, another tool for logs, and a custom billing system on top of all of it just to ship one product.
AI Gateway is our answer to that problem.
It gives you a single control point for AI traffic: one API for routing, reliability, security, observability, cost tracking, and monetization.
What Ships Today
One API for Every Major Provider
Route traffic through one integration instead of wiring every provider separately.
- OpenAI
- Anthropic
- xAI
- Meta
- Mistral
- DeepSeek
- Groq
- Cohere
- Perplexity
- OpenAI-compatible endpoints and custom provider setups
The gateway supports native Cencori endpoints for chat, embeddings, images, audio, and moderation, plus an OpenAI-compatible endpoint for teams that want the fastest migration path.
const client = new OpenAI({
apiKey: process.env.CENCORI_API_KEY,
baseURL: "https://api.cencori.com/v1",
});
const response = await client.chat.completions.create({
model: "gpt-4o",
messages: [{ role: "user", content: "Hello!" }],
});You can also use the TypeScript SDK, Python SDK, streaming responses, tool calling, and Vercel AI SDK integrations on top of the same gateway layer.
Routing, Retries, and Automatic Failover
Reliability is handled in the gateway, not bolted onto each app.
- automatic retries for transient upstream failures
- fallback routing when a provider returns
429,5xx, or times out - model equivalence mapping across providers
- configurable fallback policies and retry limits
- circuit breakers that stop sending traffic into a degraded path
- load balancing across multiple provider keys to reduce hot spots and rate-limit pressure
That means your application can send requests to Cencori once, and Cencori handles the resilience work behind the scenes.
Built-In AI Security
Every request can be inspected before it reaches a model, and every response can be evaluated before it reaches your user.
- prompt injection and jailbreak detection
- PII scanning for prompts and responses
- content filtering across harmful categories
- unified moderation controls
- configurable actions to mask, redact, block, or allow with logging
- structured security errors and incident records for blocked requests
Cencori's security layer is designed for production systems where safety has to happen in real time, not as an afterthought.
Full Observability
AI Gateway ships with observability built in.
- request and response logs for every model call
- provider, model, status, token usage, latency, and time-to-first-token
- cost tracking per request with continuously updated provider pricing
- analytics for spend, throughput, latency percentiles, and error rates
- traces for multi-step workflows, agent runs, retrieval chains, and tool calls
- web telemetry so HTTP traffic and AI traffic can be viewed together
- export paths for enterprise teams that need raw data in their own systems
For most teams, this removes the need to stitch together a separate logging or AI observability product just to understand production traffic.
Billing and Monetization
The gateway does more than route traffic. It also helps teams control and monetize AI usage.
- built-in cost tracking
- rate limiting and usage governance
- spend visibility per request and per project
- end-user billing when you pass a
userid through the gateway - quota enforcement, rate plans, markup pricing, and invoices
- Stripe Connect support for payout flows
- Usage Events API for teams that want the billing layer even if they still call providers directly
If you already route traffic through Cencori, billing becomes dramatically simpler because usage, cost, and enforcement all happen in the same place.
Developer Platform and Open Integration Surface
AI Gateway is built to fit into existing stacks without forcing a rewrite.
- OpenAI-compatible chat endpoint
- TypeScript and Python SDKs
- support for streaming and tool calling
- API key management
- custom providers and model mappings
- Vercel AI SDK support
- clean migration path for teams moving from direct provider usage
The goal is simple: swap one base URL, keep shipping.
Compliance, Auditability, and Enterprise Controls
For teams moving AI into production, operational controls matter as much as model access.
AI Gateway ships with:
- audit logs for administrative changes
- security incident logging
- policy enforcement controls
- project-level governance
- compliance-ready records for teams that need traceability
This is what turns an AI proxy into production infrastructure.
Why This Matters
Before AI Gateway, teams typically had to solve the same set of infrastructure problems over and over:
- how to route between providers
- how to survive provider outages
- how to inspect and block risky prompts
- how to understand latency and cost
- how to rate limit and meter usage
- how to charge end-users for AI features
That slows teams down and creates brittle systems.
AI Gateway makes Cencori the runtime control point for every model request. You integrate once. We handle the hard infrastructure layer behind it.
Get Started
This is the first major layer of the Cencori platform, and it is live now.