Use Cases
AI for legitimate penetration testing without hacking filters
This guide is for red teaming and security firms whose legitimate penetration testing prompts trigger generic "hacking" filters on mainstream APIs.
abliteration.ai gives authorized security teams developer-controlled model access plus Policy Gateway controls, so you can keep lawful testing workflows moving without handing policy ownership to a vendor.
Quick start
{
"model": "abliterated-model",
"messages": [
{
"role": "system",
"content": "You support authorized security testing workflows. Keep outputs structured and professional."
},
{
"role": "user",
"content": "Create a JSON checklist for an authorized internal penetration test of a web app. Include reconnaissance, validation, reporting, and evidence collection."
}
],
"temperature": 0.2
}Free preview for 5 messages. Sign up to continue.
Service notes
- Pricing model: Usage-based pricing (~$5 per 1M tokens) billed on total tokens (input + output). See the API pricing page for current plans.
- Data retention: No prompt/output retention by default. Operational telemetry (token counts, timestamps, error codes) is retained for billing and reliability.
- Compatibility: OpenAI-style /v1/chat/completions request and response format with a base URL switch.
- Latency: Depends on model size, prompt length, and load. Streaming reduces time-to-first-token.
- Throughput: Team plans include priority throughput. Actual throughput varies with demand.
- Rate limits: Limits vary by plan and load. Handle 429s with backoff and respect any Retry-After header.
Why generic hacking filters break legitimate pentesting
Security firms often use the same vocabulary as malicious actors because the work itself involves exploit validation, attack paths, and defensive verification. Generic provider-side filters can flatten those distinctions and interrupt authorized work.
- Automated pentest and red-team pipelines become unreliable when prompts are refused unpredictably.
- Analysts lose time rewriting lawful prompts to satisfy vendor filters instead of finishing the engagement.
- Client trust drops when provider behavior changes mid-project and your toolchain becomes inconsistent.
What teams use it for
The goal is not unrestricted abuse. It is stable support for authorized, contracted, and well-scoped security testing workflows.
- Finding triage and evidence summarization.
- Attack-path narration and report drafting.
- Internal playbook generation for defensive validation.
- Structured extraction from scanner outputs, screenshots, and analyst notes.
How Policy Gateway helps security firms
If you need control instead of blanket refusals, Policy Gateway lets you keep a lawful security workflow while still enforcing your own boundaries.
- Create a project per client or engagement and issue scoped keys.
- Allow approved security-testing categories while escalating out-of-scope or customer-specific requests.
- Export decision metadata to your SIEM for internal review and enterprise clients.
- Use shadow mode or canaries before changing policy on live engagements.
Privacy and procurement posture
Security engagements often involve sensitive screenshots, internal hostnames, and vulnerability evidence.
- No prompt/output retention by default; requests are processed transiently.
- Payloads are never used for model training or fine-tuning.
- Operational telemetry is retained only for billing and reliability.
Common errors & fixes
- 401 Unauthorized: Check that your API key is set and sent as a Bearer token.
- 404 Not Found: Make sure the base URL ends with /v1 and you call /chat/completions.
- 400 Bad Request: Verify the model id and that messages are an array of { role, content } objects.
- 429 Rate limit: Back off and retry. Use the Retry-After header for pacing.