Cencori Documentation

Cencori is the essential security and compliance layer for AI-integrated applications. It sits between your users and your AI models, ensuring safety, observability, and control.

The Problem We Solve

Modern applications are increasingly powered by AI, but this introduces unprecedented security and operational challenges:

  • Prompt Injection Attacks: Malicious users can manipulate AI responses by injecting instructions into prompts, potentially exposing sensitive data or bypassing security controls.
  • PII & Data Leakage: Without proper filtering, users can inadvertently send personally identifiable information (PII) or sensitive business data to third-party AI providers.
  • Uncontrolled Costs: AI APIs charge per token. Without rate limiting and monitoring, a single bug or bad actor can drain your budget overnight.
  • No Audit Trail: Compliance frameworks (SOC 2, GDPR, HIPAA) require immutable logs of all AI interactions. Most teams build this from scratch.
  • Model Lock-In: Switching between OpenAI, Anthropic, or Google requires rewriting integration code and losing observability continuity.
  • Blind Spots: You can't debug what you can't see. When AI responses are wrong, you need to inspect the exact prompt and model parameters.

Why Cencori?

Cencori is purpose-built infrastructure for AI-powered applications. It provides the security, observability, and control layer that traditional web infrastructure (like firewalls, load balancers, and logging services) provides for standard HTTP traffic—but for AI.

For Vibe Coders

Build rapidly with tools like Cursor, v0, and Lovable while keeping your generated apps secure by default. Cencori acts as a safety net, catching security issues that AI coding assistants might miss.

For AI Companies

Enterprise-grade observability, audit logs, and policy enforcement for your AI features. Multi-tenant rate limiting, cost attribution, and compliance reporting out of the box.

How It Works

Cencori acts as a transparent proxy layer between your application and AI providers. Instead of calling OpenAI, Anthropic, or Google directly, you route requests through Cencori.

Your
Application
Cencori
AI Models

Every AI request flows through Cencori's policy engine, which checks for:

  • Security threats (prompt injection, jailbreaks)
  • PII and sensitive data
  • Rate limits and cost thresholds
  • Compliance requirements

Core Features

Immutable Audit Logging

Every AI request and response is logged with complete context:

  • Full prompt and completion text
  • Model parameters (temperature, max tokens)
  • User identity and session metadata
  • Token usage and cost attribution
  • Timestamps and request IDs for tracing

Real-Time Threat Detection

Cencori identifies and blocks malicious activity before it reaches your AI providers:

  • Prompt injection attempts
  • Jailbreak patterns
  • PII exposure risks
  • Excessive token usage spikes

Granular Rate Limiting

Control costs and prevent abuse with multi-dimensional limits:

  • Per-user request limits
  • Per-organization token budgets
  • Model-specific throttling
  • Time-based quotas (hourly, daily, monthly)

Multi-Provider Support

Switch between AI providers without changing your application code. Cencori provides a unified API that works with all major providers.