tatsu-lab/alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
Implements length-controlled win-rate scoring to mitigate output length bias, achieving 0.98 Spearman correlation with ChatBot Arena while costing under $10 and completing in under 3 minutes. Uses LLM-based pairwise comparisons (GPT-4 by default) against a reference model, validated against 20K human annotations with built-in caching and output randomization. Provides a toolkit for constructing custom evaluators with batching and multi-annotator support, plus curated evaluation datasets and leaderboards for benchmarking instruction-following models.
1,957 stars. No commits in the last 6 months.
Stars
1,957
Forks
305
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Aug 09, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tatsu-lab/alpaca_eval"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Related models
eth-sri/matharena
Evaluation of LLMs on latest math competitions
HPAI-BSC/TuRTLe
TuRTLe: A Unified Evaluation of LLMs for RTL Generation 🐢 (MLCAD 2025)
nlp-uoregon/mlmm-evaluation
Multilingual Large Language Models Evaluation Benchmark
JinjieNi/MixEval
The official evaluation suite and dynamic data release for MixEval.
princeton-nlp/LLMBar
[ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Following