prompt-optimizer and promptolution

These are competitors offering similar prompt optimization capabilities through different architectural approaches—one as a standalone optimizer tool and the other as a modular framework—where users would typically adopt one based on preference for simplicity versus customization.

prompt-optimizer
72
Verified
promptolution
46
Emerging
Maintenance 25/25
Adoption 10/25
Maturity 16/25
Community 21/25
Maintenance 10/25
Adoption 9/25
Maturity 16/25
Community 11/25
Stars: 24,228
Forks: 2,893
Downloads:
Commits (30d): 81
Language: TypeScript
License:
Stars: 114
Forks: 8
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
No Package No Dependents
No Package No Dependents

About prompt-optimizer

linshenkx/prompt-optimizer

一款提示词优化器,助力于编写高质量的提示词

Supports multi-model LLM backends (OpenAI, Gemini, DeepSeek, etc.) with dual optimization modes for system and user prompts, plus advanced testing via context variables, multi-turn sessions, and function calling. Available as web app, desktop client, Chrome extension, Docker container, and MCP server for Claude Desktop integration—with client-side data processing and optional password protection for secure deployment.

About promptolution

automl/promptolution

A unified, modular Framework for Prompt Optimization

Supports multiple state-of-the-art prompt optimization algorithms (CAPO, EvoPrompt, OPRO) with a unified LLM backend spanning API-based models, local inference via vLLM/transformers, and cluster deployments. Built-in response caching, parallelized inference, and detailed token tracking enable cost-efficient, reproducible large-scale experiments. Decomposes optimization into modular components—Task, Predictor, LLM, and Optimizer—allowing researchers to customize any stage without rigid abstractions.

Scores updated daily from GitHub, PyPI, and npm data. How scores work