algorithmicsuperintelligence/optillm

Optimizing inference proxy for LLMs

62
/ 100
Established

Implements 20+ inference-time optimization techniques—including MARS, CePO, chain-of-thought reflection, and Monte Carlo tree search—that layer multiple reasoning strategies to achieve 2-10x accuracy gains on math and coding tasks. Acts as an OpenAI API-compatible proxy that intercepts requests and automatically applies selected techniques based on model prefix (e.g., `moa-gpt-4o-mini`), requiring no model retraining or client-side changes. Supports 100+ models across OpenAI, Anthropic, Google, and other providers via LiteLLM, with multi-variant Docker images for full, proxy-only, or offline deployment scenarios.

3,377 stars. Actively maintained with 6 commits in the last 30 days.

No Package No Dependents
Maintenance 17 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 19 / 25

How are scores calculated?

Stars

3,377

Forks

265

Language

Python

License

Apache-2.0

Last pushed

Jan 28, 2026

Commits (30d)

6

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/algorithmicsuperintelligence/optillm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.