embeddings-benchmark/mteb

MTEB: Massive Text Embedding Benchmark

99
/ 100
Verified

Provides standardized evaluation across 100+ tasks spanning classification, clustering, retrieval, and semantic textual similarity for both text and multimodal embeddings. Integrates with Hugging Face ecosystem (SentenceTransformers, transformers) and offers a unified Python API plus CLI for benchmarking custom or pretrained models against a curated leaderboard. Supports multilingual evaluation with automatic caching, batch processing, and reproducible result tracking across embedding model implementations.

3,159 stars and 1,555,633 monthly downloads. Used by 5 other packages. Actively maintained with 107 commits in the last 30 days. Available on PyPI.

Maintenance 25 / 25
Adoption 25 / 25
Maturity 25 / 25
Community 24 / 25

How are scores calculated?

Stars

3,159

Forks

568

Language

Python

License

Apache-2.0

Last pushed

Mar 12, 2026

Monthly downloads

1,555,633

Commits (30d)

107

Dependencies

13

Reverse dependents

5

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/embeddings-benchmark/mteb"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.