Going multimodal: How Prezi is leveraging the Hub and the Expert Support Program to accelerate their ML roadmap Jun 19, 2024 • 11
Smarter, Better, Faster, Longer: A Modern Bidirectional Encoder for Fast, Memory Efficient, and Long Context Finetuning and Inference Paper • 2412.13663 • Published 16 days ago • 116
Qwen2-VL Collection Vision-language model series based on Qwen2 • 16 items • Updated 28 days ago • 186
PaliGemma 2 Release Collection Vision-Language Models available in multiple 3B, 10B and 28B variants. • 23 items • Updated 21 days ago • 121
PaliGemma 2: A Family of Versatile VLMs for Transfer Paper • 2412.03555 • Published 30 days ago • 119
PixMo Collection A set of vision-language datasets built by Ai2 and used to train the Molmo family of models. Read more at https://molmo.allenai.org/blog • 9 items • Updated Nov 27, 2024 • 53
Qwen2.5 Collection Qwen2.5 language models, including pretrained and instruction-tuned models of 7 sizes, including 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B. • 45 items • Updated Nov 28, 2024 • 452
SmolLM2 Collection State-of-the-art compact LLMs for on-device applications: 1.7B, 360M, 135M • 15 items • Updated 12 days ago • 196
view article Article How to optimize your data labelling project with custom interfaces By burtenshaw • Oct 16, 2024 • 18
Molmo and PixMo: Open Weights and Open Data for State-of-the-Art Multimodal Models Paper • 2409.17146 • Published Sep 25, 2024 • 104
Molmo Collection Artifacts for open multimodal language models. • 5 items • Updated Nov 27, 2024 • 290
Moshi v0.1 Release Collection MLX, Candle & PyTorch model checkpoints released as part of the Moshi release from Kyutai. Run inference via: https://github.com/kyutai-labs/moshi • 13 items • Updated Sep 18, 2024 • 225
Political DEBATE: Efficient Zero-shot and Few-shot Classifiers for Political Text Paper • 2409.02078 • Published Sep 3, 2024 • 9