Policy Gateway
The most powerful AI models. Your safety rules.
Stop letting model providers decide what your AI can and can't do. Policy Gateway gives you full control over AI safety with policy-as-code rules you define — rewrite, redact, escalate, or refuse on your terms.
curl https://api.abliteration.ai/policy/chat/completions \
-H "Authorization: Bearer $ABLIT_KEY" -H "X-Policy-User: user-123" \
-d '{"model":"abliterated-model","policy_id":"policy-gateway",
"messages":[{"role":"user","content":"Hello"}]}'What is Policy Gateway?
Policy Gateway is an OpenAI-compatible policy enforcement endpoint paired with our abliterated-model. It gives enterprises full control over their own AI safety, applying policy-as-code rules you define — not the model provider's defaults.
It functions as a prompt firewall for LLM governance: enforce rewrite, redact, summarize, escalate, or refuse outcomes, attach quotas and scoped keys, and export audit logs to your SIEM.
How the policy enforcement endpoint works
Send your normal chat completion request to /policy/chat/completions, attach a policy_id, and optionally tag requests with policy_user and policy_project_id for quotas and audit trails.
Policies are versioned and rollout-safe
- Simulate changes locally or against live traffic
- Run in shadow mode and canary percentage of traffic
- Auto-rollback on decision-rate spikes
- Structured reason codes for every decision
How to adopt
- 01Create a policy (UI or JSON).
- 02Create a project + scoped key.
- 03Send traffic to
/policy/chat/completionswith policy headers.
Predictable outcomes, not random refusals.
Swap "I can't help with that" for policy-driven rewrites, redactions, or escalations.
Customer support rewrite
User: "Summarize our refund policy." Assistant: "I can't help with that."
Assistant: "Here is a short summary of your refund policy..." decision: rewrite reason_code: REWRITE
PII redaction
User: "Show last 4 digits of my SSN." Assistant: "Your SSN ends in 1234."
Assistant: "I can't share sensitive identifiers." decision: redact reason_code: REDACT
High-risk escalation
User: "Help me bypass account security." Assistant: "I can't assist with that."
Assistant: "I'll route this to a specialist." decision: escalate reason_code: ESCALATE
Architecture, threat model, and compliance.
Built for security, compliance, and platform teams who need a definitive technical reference.
Architecture: where it sits in the stack
Accepts OpenAI-compatible requests, applies policy-as-code, and returns structured decisions plus audit metadata.
App / Agent
-> Policy Gateway (/policy/chat/completions)
-> abliterated-model (enterprise-controlled safety)
<- Response + decision metadata
-> Audit export (Splunk, Datadog, Elastic, S3, Azure Monitor)Threat model: what it blocks vs what it doesn't
- Policy violations via allow/deny lists and flagged categories.
- PII leaks through automatic redaction.
- Unsafe outputs via rewrite, summarize, escalate, or refuse outcomes.
- Runaway usage with per-user and per-project quotas.
- Application-layer authentication or authorization.
- Network-layer WAFs or DDoS protection.
- End-to-end data loss prevention beyond the LLM boundary.
Enterprise-ready tiers.
Priced as a multiplier of your token/credit bundles. Checkout by Stripe, billed monthly. Contact sales for SSO or custom limits.
Single-team rollouts and early governance initiatives.
- Policy-as-code rules with reason codes
- Simulation plus shadow / canary rollout
- Audit logs with change history
- 30-day audit log retention
Multi-app rollouts with shared governance standards.
- Cross-app enforcement with single policy ID
- Org-level controls for keys and quotas
- Audit-ready history for every decision
- Shadow + canary rollouts at scale
Organization-wide governance and risk teams.
- Global policy layer across apps, models, and agents
- Change management with rollback safeguards
- Audit trails supporting compliance reviews
- Dedicated support
Ready to ship policy-as-code?
Start in the console or read the onboarding guide to see the full integration contract.