embeddings-benchmark/mteb
MTEB: Massive Text Embedding Benchmark
This tool helps machine learning engineers and researchers assess the quality and performance of different text embedding models. You provide a text embedding model and specific evaluation tasks (like text classification or retrieval). The output is a clear set of metrics showing how well the model performs on those tasks, allowing for informed comparison and selection of the best model.
3,159 stars. Used by 6 other packages. Actively maintained with 107 commits in the last 30 days. Available on PyPI.
Use this if you need to systematically compare and evaluate multiple text embedding models against standardized benchmarks to choose the most effective one for your application.
Not ideal if you are looking for a tool to train new text embedding models or to apply embeddings directly in a production system without prior evaluation.
Stars
3,159
Forks
568
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 12, 2026
Commits (30d)
107
Dependencies
13
Reverse dependents
6
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/embeddings-benchmark/mteb"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Compare
Related tools
harmonydata/harmony
The Harmony Python library: a research tool for psychologists to harmonise data and...
yannvgn/laserembeddings
LASER multilingual sentence embeddings as a pip package
embeddings-benchmark/results
Data for the MTEB leaderboard
Hironsan/awesome-embedding-models
A curated list of awesome embedding models tutorials, projects and communities.
fresh-stack/freshstack
This repository helps you evaluate your models on the FreshStack benchmark!