mteb and mleb

MLEB is a specialized legal-domain benchmark that extends the evaluation methodology of MTEB to a specific corpus, making them complementary tools where MLEB users would typically also use MTEB for cross-domain baseline comparisons.

mteb
86
Verified
mleb
43
Emerging
Maintenance 22/25
Adoption 15/25
Maturity 25/25
Community 24/25
Maintenance 10/25
Adoption 7/25
Maturity 15/25
Community 11/25
Stars: 3,159
Forks: 568
Downloads:
Commits (30d): 107
Language: Python
License: Apache-2.0
Stars: 32
Forks: 4
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No risk flags
No Package No Dependents

About mteb

embeddings-benchmark/mteb

MTEB: Massive Text Embedding Benchmark

This tool helps machine learning engineers and researchers assess the quality and performance of different text embedding models. You provide a text embedding model and specific evaluation tasks (like text classification or retrieval). The output is a clear set of metrics showing how well the model performs on those tasks, allowing for informed comparison and selection of the best model.

natural-language-processing model-evaluation text-embeddings information-retrieval machine-learning-research

About mleb

isaacus-dev/mleb

The code used to evaluate embedding models on the Massive Legal Embedding Benchmark (MLEB).

This tool helps legal tech developers, researchers, and data scientists evaluate how well their legal text embedding models understand and reason about legal documents. You input your embedding model and relevant API keys, and it outputs a performance score across diverse legal datasets. This helps you understand your model's strengths and weaknesses in legal applications.

legal-AI legal-tech natural-language-processing legal-research AI-model-evaluation

Related comparisons

Scores updated daily from GitHub, PyPI, and npm data. How scores work