algorithmicsuperintelligence/optillm
Optimizing inference proxy for LLMs
Implements 20+ inference-time optimization techniques—including MARS, CePO, chain-of-thought reflection, and Monte Carlo tree search—that layer multiple reasoning strategies to achieve 2-10x accuracy gains on math and coding tasks. Acts as an OpenAI API-compatible proxy that intercepts requests and automatically applies selected techniques based on model prefix (e.g., `moa-gpt-4o-mini`), requiring no model retraining or client-side changes. Supports 100+ models across OpenAI, Anthropic, Google, and other providers via LiteLLM, with multi-variant Docker images for full, proxy-only, or offline deployment scenarios.
3,377 stars. Actively maintained with 6 commits in the last 30 days.
Stars
3,377
Forks
265
Language
Python
License
Apache-2.0
Category
Last pushed
Jan 28, 2026
Commits (30d)
6
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/algorithmicsuperintelligence/optillm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related tools
langfuse/langfuse
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management,...
Arize-ai/phoenix
AI Observability & Evaluation
Mirascope/mirascope
The LLM Anti-Framework
Helicone/helicone
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
Agenta-AI/agenta
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM...