| title | Quick Start |
|---|---|
| description | Start compressing LLM calls in under 2 minutes. Two lines of code. |
Your API key starts with sk-occ- and works like any other LLM API key.
OpenCompress is a drop-in replacement for any OpenAI-compatible endpoint. Change two values:
| Setting | Before | After |
|---|---|---|
| Base URL | https://api.openai.com/v1 |
https://www.opencompress.ai/api/v1 |
| API Key | sk-... (OpenAI) |
sk-occ-... (OpenCompress) |
Everything else — model names, message format, streaming, tool calls — stays identical.
from openai import OpenAI
client = OpenAI(
base_url="https://www.opencompress.ai/api/v1",
api_key="sk-occ-your-key-here",
)
response = client.chat.completions.create(
model="gpt-4o-mini",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Explain quantum computing in simple terms."},
],
)
print(response.choices[0].message.content)import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://www.opencompress.ai/api/v1",
apiKey: "sk-occ-your-key-here",
});
const response = await client.chat.completions.create({
model: "gpt-4o-mini",
messages: [
{ role: "system", content: "You are a helpful assistant." },
{ role: "user", content: "Explain quantum computing in simple terms." },
],
});
console.log(response.choices[0].message.content);curl https://www.opencompress.ai/api/v1/chat/completions \
-H "Authorization: Bearer sk-occ-your-key-here" \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-4o-mini",
"messages": [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Explain quantum computing in simple terms."}
]
}'Your App OpenCompress LLM Provider
│ │ │
│─── POST /chat/completions ─│ │
│ (original prompt) │ │
│ │── compress prompt ──┐ │
│ │ │ │
│ │◄─ compressed ───────┘ │
│ │ │
│ │─── forward compressed ────────►│
│ │ │
│ │◄── response ──────────────────│
│◄── response ───────────────│ │
│ │ │
│ │── calculate savings │
│ │── deduct balance │
OpenCompress works with any model available on OpenRouter:
GPT-4o, GPT-4o-mini, GPT-4.1, GPT-4.1-mini, GPT-4.1-nano Claude Opus 4.6, Claude Sonnet 4.6, Claude Haiku 4.5 Gemini 2.5 Pro, Gemini 2.5 Flash Llama 4 Maverick, Llama 4 Scout DeepSeek V3, DeepSeek R1 Use your own API key with any provider Understand the compression pipeline Use your own LLM API key