

Massive Text Embedding Benchmark (MTEB) - a comprehensive benchmark for evaluating text embedding models across 8 embedding tasks and 58 datasets in 112 languages. Provides a standardized leaderboard for comparing embedding quality across classification, clustering, retrieval, reranking, semantic textual similarity, and summarization tasks.
Loading more......
MTEB (Massive Text Embedding Benchmark) is the most comprehensive benchmark for evaluating text embedding models. It covers 8 different embedding tasks across 58 datasets in 112 languages, providing a standardized way to compare embedding quality.
MTEB evaluates models across 8 core tasks:
Historically strong performers include:
Different metrics for each task:
from mteb import MTEB
# Evaluate a model
evaluation = MTEB(tasks=["Banking77Classification"])
results = evaluation.run(model)
Official leaderboard maintained on Hugging Face: https://huggingface.co/spaces/mteb/leaderboard
Free and open-source benchmark.