

Open-source embedding model from Nomic AI based on ModernBERT-base with 149M parameters. Supports 8192 token sequences and Matryoshka Representation Learning for 3x memory reduction.
Loading more......
ModernBERT Embed is an embedding model trained from ModernBERT-base, bringing the new advances of ModernBERT to embeddings. Trained on the Nomic Embed weakly-supervised and supervised datasets.
This model requires prefixes to be added to the input (similar to Nomic Embed):
search_query: prefix to queriessearch_document: prefix to documentsApache 2.0 licensed - fully commercially permissible
Requires transformers>=4.48.0
Available on Hugging Face at nomic-ai/modernbert-embed-base