onejune2018/Awesome-LLM-Eval

Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs. 一个由工具、基准/数据、演示、排行榜和大模型等组成的精选列表,主要面向基础大模型评测,旨在探求生成式AI的技术边界.

48
/ 100
Emerging

Organizes evaluation resources around an **anthropomorphic and value-oriented taxonomy** that extends beyond traditional benchmarks to assess reasoning robustness, uncertainty quantification, and long-context capabilities. Integrates references to specialized evaluation tools (OpenCompass, DeepEval, AlpacaEval), domain-specific benchmarks (RAG, agents, coding, multimodal), and LLMOps frameworks, while maintaining structured categorization of pre-trained, instruction-tuned, and aligned models. Backed by a peer-reviewed survey paper that provides the methodological foundation for its continuously updated categorization scheme.

616 stars.

No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

616

Forks

51

Language

License

MIT

Last pushed

Nov 24, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/onejune2018/Awesome-LLM-Eval"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.