OLMo 2 Collection Artifacts for the second set of OLMo models. • 26 items • Updated about 6 hours ago • 86
DeepHermes Collection Preview models of hybrid reasoner Hermes series • 6 items • Updated about 6 hours ago • 13
Jamba 1.6 Collection The AI21 Jamba family of models are hybrid SSM-Transformer foundation models, outperforming open model competitors on quality and speed. • 2 items • Updated 7 days ago • 11
C4AI Aya Vision Collection Aya Vision is a state-of-the-art family of vision models that brings multimodal capabilities to 23 languages. • 5 items • Updated 9 days ago • 63
Phi-4 Collection Phi-4 family of small language and multi-modal models. • 7 items • Updated 10 days ago • 109
Foundation Text-Generation Models Below 360M Parameters Collection Great candidates for fine-tuning targeting Wllama and Transformers.js for mobile devices, ordered by number of parameters. • 34 items • Updated 4 days ago • 28
Ovis2 Collection Our latest advancement in multi-modal large language models (MLLMs) • 8 items • Updated about 10 hours ago • 55
SuperGPQA: Scaling LLM Evaluation across 285 Graduate Disciplines Paper • 2502.14739 • Published 21 days ago • 97
Step-Audio Collection Step-Audio model family, including Audio-Tokenizer, Audio-Chat and TTS • 3 items • Updated 24 days ago • 30
Hamanasu Collection A brand new series of Models from yours truly, Designed for Intelligence, Creativity and Roleplay. • 16 items • Updated 3 days ago • 5
OLMoE (January 2025) Collection Improved OLMoE for iOS app. Read more: https://allenai.org/blog/olmoe-app • 10 items • Updated about 7 hours ago • 9
SFTvsRL Models & Data Collection This collection contains 4 initial checkpoints for https://github.com/LeslieTrue/SFTvsRL and necessary data for V-IRL training. • 7 items • Updated about 17 hours ago • 8
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training Paper • 2501.17161 • Published Jan 28 • 108