Integrations
SillyTavern Integration Guide
SillyTavern is a popular frontend for AI roleplay and interactive entertainment. You can use abliteration.ai as a drop-in OpenAI-compatible backend when standard providers refuse roleplay scenarios.
Uncensored models are available for teams that need them, with developer-controlled policy enforcement.
Quick start
Service notes
- Pricing model: Usage-based pricing (~$5 per 1M tokens) billed on total tokens (input + output). See the API pricing page for current plans.
- Data retention: No prompt/output retention by default. Operational telemetry (token counts, timestamps, error codes) is retained for billing and reliability.
- Compatibility: OpenAI-style /v1/chat/completions request and response format with a base URL switch.
- Latency: Depends on model size, prompt length, and load. Streaming reduces time-to-first-token.
- Throughput: Team plans include priority throughput. Actual throughput varies with demand.
- Rate limits: Limits vary by plan and load. Handle 429s with backoff and respect any Retry-After header.
Configuration Steps
1. Open SillyTavern and go to the API Connections menu (plug icon).
- Select OpenAI as the API source.
- Check Custom Endpoint (or Reverse Proxy).
- Set the API URL to:
https://api.abliteration.ai/v1 - Paste your abliteration.ai Secret Key into the API Key field.
- Click Connect and select
abliterated-model.
Why use this backend?
Standard providers often refuse roleplay scenarios. This API is designed to process your character cards and system prompts without moralizing refusals.
Common errors & fixes
- 401 Unauthorized: Check that your API key is set and sent as a Bearer token.
- 404 Not Found: Make sure the base URL ends with /v1 and you call /chat/completions.
- 400 Bad Request: Verify the model id and that messages are an array of { role, content } objects.
- 429 Rate limit: Back off and retry. Use the Retry-After header for pacing.