mteb and mleb

MLEB is a specialized legal-domain benchmark that extends the evaluation methodology of MTEB to a specific corpus, making them complementary tools where MLEB users would typically also use MTEB for cross-domain baseline comparisons.

mteb
99
Verified
mleb
37
Emerging
Maintenance 25/25
Adoption 25/25
Maturity 25/25
Community 24/25
Maintenance 10/25
Adoption 7/25
Maturity 9/25
Community 11/25
Stars: 3,159
Forks: 568
Downloads: 1,555,633
Commits (30d): 107
Language: Python
License: Apache-2.0
Stars: 32
Forks: 4
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No risk flags
No Package No Dependents

About mteb

embeddings-benchmark/mteb

MTEB: Massive Text Embedding Benchmark

Provides standardized evaluation across 100+ tasks spanning classification, clustering, retrieval, and semantic textual similarity for both text and multimodal embeddings. Integrates with Hugging Face ecosystem (SentenceTransformers, transformers) and offers a unified Python API plus CLI for benchmarking custom or pretrained models against a curated leaderboard. Supports multilingual evaluation with automatic caching, batch processing, and reproducible result tracking across embedding model implementations.

About mleb

isaacus-dev/mleb

The code used to evaluate embedding models on the Massive Legal Embedding Benchmark (MLEB).

Related comparisons

Scores updated daily from GitHub, PyPI, and npm data. How scores work