embeddings-benchmark/results
Data for the MTEB leaderboard
Stores standardized evaluation results from the MTEB (Massive Text Embedding Benchmark) package across diverse embedding models and tasks. Results are submitted directly to this repository rather than via Hugging Face model cards, enabling verification that scores match verified model implementations. The leaderboard aggregates these results to provide comparable benchmarks across retrieval, clustering, semantic search, and other embedding-based tasks.
Stars
47
Forks
135
Language
Python
License
—
Category
Last pushed
Mar 13, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/embeddings-benchmark/results"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Compare
Higher-rated alternatives
embeddings-benchmark/mteb
MTEB: Massive Text Embedding Benchmark
yannvgn/laserembeddings
LASER multilingual sentence embeddings as a pip package
harmonydata/harmony
The Harmony Python library: a research tool for psychologists to harmonise data and...
MilaNLProc/honest
A Python package to compute HONEST, a score to measure hurtful sentence completions in language...
fresh-stack/freshstack
This repository helps you evaluate your models on the FreshStack benchmark!