Massive Text Embedding Benchmark

non-profit
Activity Feed

AI & ML interests

Massive Text Embeddings Benchmark

Recent Activity

Muennighoff  updated a dataset about 3 hours ago
mteb/arena-results
orionweller  updated a Space about 10 hours ago
mteb/leaderboard
View all activity

mteb's activity

tomaarsen 
posted an update 3 days ago
view post
Post
2282
That didn't take long! Nomic AI has finetuned the new ModernBERT-base encoder model into a strong embedding model for search, classification, clustering and more!

Details:
🤖 Based on ModernBERT-base with 149M parameters.
📊 Outperforms both nomic-embed-text-v1 and nomic-embed-text-v1.5 on MTEB!
🏎️ Immediate FA2 and unpacking support for super efficient inference.
🪆 Trained with Matryoshka support, i.e. 2 valid output dimensionalities: 768 and 256.
➡️ Maximum sequence length of 8192 tokens!
2️⃣ Trained in 2 stages: unsupervised contrastive data -> high quality labeled datasets.
➕ Integrated in Sentence Transformers, Transformers, LangChain, LlamaIndex, Haystack, etc.
🏛️ Apache 2.0 licensed: fully commercially permissible

Try it out here: nomic-ai/modernbert-embed-base

Very nice work by Zach Nussbaum and colleagues at Nomic AI.
Muennighoff 
updated a Space 3 days ago