CoolPrompt and promptolution

These two frameworks are competitors, as both aim to provide unified and modular solutions for automatic prompt optimization, suggesting they offer overlapping functionalities for similar use cases.

CoolPrompt
62
Established
promptolution
46
Emerging
Maintenance 13/25
Adoption 14/25
Maturity 25/25
Community 10/25
Maintenance 10/25
Adoption 9/25
Maturity 16/25
Community 11/25
Stars: 178
Forks: 9
Downloads: 84
Commits (30d): 0
Language: Python
License: Apache-2.0
Stars: 114
Forks: 8
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
No risk flags
No Package No Dependents

About CoolPrompt

CTLab-ITMO/CoolPrompt

Automatic Prompt Optimization Framework

Implements multiple optimization algorithms (HyPE, ReflectivePrompt, DistillPrompt) that iteratively refine prompts through LLM-based feedback and evaluation metrics. LLM-agnostic architecture supports any Langchain-compatible model while generating synthetic evaluation data when datasets are unavailable, and automatically detects task types for scenarios without explicit specifications.

About promptolution

automl/promptolution

A unified, modular Framework for Prompt Optimization

Supports multiple state-of-the-art prompt optimization algorithms (CAPO, EvoPrompt, OPRO) with a unified LLM backend spanning API-based models, local inference via vLLM/transformers, and cluster deployments. Built-in response caching, parallelized inference, and detailed token tracking enable cost-efficient, reproducible large-scale experiments. Decomposes optimization into modular components—Task, Predictor, LLM, and Optimizer—allowing researchers to customize any stage without rigid abstractions.

Scores updated daily from GitHub, PyPI, and npm data. How scores work