The Headroom proxy server is a production-ready HTTP server that applies context optimization to all requests passing through it.
# Basic usage
headroom proxy
# Custom port
headroom proxy --port 8080
# With all options
headroom proxy \
--host 0.0.0.0 \
--port 8787 \
--log-file /var/log/headroom.jsonl \
--budget 100.0| Option | Default | Description |
|---|---|---|
--host |
127.0.0.1 |
Host to bind to |
--port |
8787 |
Port to bind to |
--no-optimize |
false |
Disable optimization (passthrough mode) |
--no-cache |
false |
Disable semantic caching |
--no-rate-limit |
false |
Disable rate limiting |
--log-file |
None | Path to JSONL log file |
--budget |
None | Daily budget limit in USD |
--openai-api-url |
https://api.openai.com |
Custom OpenAI API URL endpoint |
| Option | Default | Description |
|---|---|---|
--no-intelligent-context |
false |
Disable IntelligentContextManager (fall back to RollingWindow) |
--no-intelligent-scoring |
false |
Disable multi-factor importance scoring (use position-based) |
--no-compress-first |
false |
Disable trying deeper compression before dropping messages |
By default, the proxy uses IntelligentContextManager which scores messages by multiple factors (recency, semantic similarity, TOIN-learned patterns, error indicators, forward references) and drops lowest-scored messages first. This is smarter than simple age-based truncation.
CCR Integration: When messages are dropped, they're stored in CCR so the LLM can retrieve them if needed. The inserted marker includes the CCR reference. Drops are also recorded to TOIN, so the system learns which message patterns are important across all users.
# Use legacy RollingWindow (drops oldest first)
headroom proxy --no-intelligent-context
# Disable semantic scoring (faster, but less intelligent)
headroom proxy --no-intelligent-scoring| Option | Default | Description |
|---|---|---|
--llmlingua |
false |
Enable LLMLingua-2 ML-based compression |
--llmlingua-device |
auto |
Device for model: auto, cuda, cpu, mps |
--llmlingua-rate |
0.3 |
Target compression rate (0.3 = keep 30% of tokens) |
Note: LLMLingua requires additional dependencies: pip install headroom-ai[llmlingua]
# Enable LLMLingua with GPU acceleration
headroom proxy --llmlingua --llmlingua-device cuda
# More aggressive compression (keep only 20%)
headroom proxy --llmlingua --llmlingua-rate 0.2
# Conservative compression for code (keep 50%)
headroom proxy --llmlingua --llmlingua-rate 0.5curl http://localhost:8787/healthResponse:
{
"status": "healthy",
"optimize": true,
"stats": {
"total_requests": 42,
"tokens_saved": 15000,
"savings_percent": 45.2
}
}curl http://localhost:8787/statscurl http://localhost:8787/metricsThe proxy supports both Anthropic and OpenAI API formats:
# Anthropic format
POST /v1/messages
# OpenAI format
POST /v1/chat/completions# Start proxy
headroom proxy --port 8787
# In another terminal
ANTHROPIC_BASE_URL=http://localhost:8787 claude- Start the proxy:
headroom proxy - In Cursor settings, set the base URL to
http://localhost:8787
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:8787/v1",
api_key="your-api-key", # Still needed for upstream
)When enabled, the proxy uses Microsoft's LLMLingua-2 model for ML-based token compression:
headroom proxy --llmlinguaHow it works:
- LLMLinguaCompressor is added to the transform pipeline (before RollingWindow)
- Automatically detects content type (JSON, code, text) and adjusts compression
- Stores original content in CCR for retrieval if needed
Startup feedback:
# When enabled and available:
LLMLingua: ENABLED (device=cuda, rate=0.3)
# When installed but not enabled (helpful hint):
LLMLingua: available (enable with --llmlingua for ML compression)
# When enabled but not installed:
WARNING: LLMLingua requested but not installed. Install with: pip install headroom-ai[llmlingua]
Why opt-in?
| Concern | Default Proxy | With LLMLingua |
|---|---|---|
| Dependencies | ~50MB | +2GB (torch, transformers) |
| Cold start | <1s | 10-30s (model load) |
| Memory | ~100MB | +1GB (model in RAM) |
| Overhead | <5ms | 50-200ms per request |
Enable LLMLingua when maximum compression justifies the resource cost.
The proxy caches responses for repeated queries:
- LRU eviction with configurable max entries
- TTL-based expiration
- Cache key based on message content hash
Token bucket rate limiting protects against runaway costs:
- Configurable requests per minute
- Configurable tokens per minute
- Per-API-key tracking
Track spending and enforce budgets:
- Real-time cost estimation
- Budget periods: hourly, daily, monthly
- Automatic request rejection when over budget
Export metrics for monitoring:
headroom_requests_total
headroom_tokens_saved_total
headroom_cost_usd_total
headroom_latency_ms_sum
export HEADROOM_HOST=0.0.0.0
export HEADROOM_PORT=8787
export HEADROOM_BUDGET=100.0
export OPENAI_TARGET_API_URL=https://custom.openai.endpoint.com
headroom proxyFor production deployments:
# Use a process manager
pip install gunicorn
# Run with gunicorn
gunicorn headroom.proxy.server:app \
--workers 4 \
--bind 0.0.0.0:8787 \
--worker-class uvicorn.workers.UvicornWorkerOr with Docker:
FROM python:3.11-slim
RUN pip install headroom[proxy]
EXPOSE 8787
CMD ["headroom", "proxy", "--host", "0.0.0.0"]