Skip to main content

Get your API key

1

Create an account

2

Add funds

Deposit credits via Stripe. Start with $10 to test.
3

Generate an API key

Go to the API Keys section and create a new key. Copy it — it’s shown only once.
Your API key starts with sk-occ- and works like any other LLM API key.

Integration

OpenCompress is a drop-in replacement for any OpenAI-compatible endpoint. Change two values:
SettingBeforeAfter
Base URLhttps://api.openai.com/v1https://www.opencompress.ai/api/v1
API Keysk-... (OpenAI)sk-occ-... (OpenCompress)
Everything else — model names, message format, streaming, tool calls — stays identical.

Code examples

from openai import OpenAI

client = OpenAI(
    base_url="https://www.opencompress.ai/api/v1",
    api_key="sk-occ-your-key-here",
)

response = client.chat.completions.create(
    model="gpt-4o-mini",
    messages=[
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": "Explain quantum computing in simple terms."},
    ],
)

print(response.choices[0].message.content)
Your prompts are compressed transparently before reaching the model. The response you receive is identical in format to a direct OpenAI API call.

What happens behind the scenes

Your App                    OpenCompress                     LLM Provider
   │                            │                                │
   │─── POST /chat/completions ─│                                │
   │    (original prompt)       │                                │
   │                            │── compress prompt ──┐          │
   │                            │                     │          │
   │                            │◄─ compressed ───────┘          │
   │                            │                                │
   │                            │─── forward compressed ────────►│
   │                            │                                │
   │                            │◄── response ──────────────────│
   │◄── response ───────────────│                                │
   │                            │                                │
   │                            │── calculate savings            │
   │                            │── deduct balance               │

Supported models

OpenCompress works with any model available on OpenRouter:

OpenAI

GPT-4o, GPT-4o-mini, GPT-4.1, GPT-4.1-mini, GPT-4.1-nano

Anthropic

Claude Opus 4.6, Claude Sonnet 4.6, Claude Haiku 4.5

Google

Gemini 2.5 Pro, Gemini 2.5 Flash

Meta

Llama 4 Maverick, Llama 4 Scout

DeepSeek

DeepSeek V3, DeepSeek R1

BYOK

Use your own API key with any provider

Next steps

How It Works

Understand the compression pipeline

BYOK Mode

Use your own LLM API key