optillm and LLMstudio

These two tools are complements: OptiLLM focuses on optimizing inference for deployed LLMs, while LLMstudio provides a framework to bring LLM applications to production, including potentially integrating and leveraging such optimization proxies.

optillm
62
Established
LLMstudio
67
Established
Maintenance 17/25
Adoption 10/25
Maturity 16/25
Community 19/25
Maintenance 10/25
Adoption 16/25
Maturity 25/25
Community 16/25
Stars: 3,377
Forks: 265
Downloads:
Commits (30d): 6
Language: Python
License: Apache-2.0
Stars: 371
Forks: 39
Downloads: 563
Commits (30d): 0
Language: Python
License: MPL-2.0
No Package No Dependents
No risk flags

About optillm

algorithmicsuperintelligence/optillm

Optimizing inference proxy for LLMs

Implements 20+ inference-time optimization techniques—including MARS, CePO, chain-of-thought reflection, and Monte Carlo tree search—that layer multiple reasoning strategies to achieve 2-10x accuracy gains on math and coding tasks. Acts as an OpenAI API-compatible proxy that intercepts requests and automatically applies selected techniques based on model prefix (e.g., `moa-gpt-4o-mini`), requiring no model retraining or client-side changes. Supports 100+ models across OpenAI, Anthropic, Google, and other providers via LiteLLM, with multi-variant Docker images for full, proxy-only, or offline deployment scenarios.

About LLMstudio

TensorOpsAI/LLMstudio

Framework to bring LLM applications to production

Provides a unified proxy layer across OpenAI, Anthropic, and Google LLMs plus local models via Ollama, with smart routing and fallback mechanisms for reliability. Includes a web-based prompt playground UI, Python SDK, request monitoring/logging, and LangChain compatibility for seamless integration into existing projects. Supports batch calling and deploys as a server with separate proxy and tracker APIs.

Scores updated daily from GitHub, PyPI, and npm data. How scores work