Distributed Training: Train BART/T5 for Summarization using 🤗 Transformers and Amazon SageMaker Apr 8, 2021
Post-Training Releases November 2024 Collection Includes papers with post-training sides from best open-models from November, including OpenCoder, SmolLM-v2, Orca Agent Instruct, Tülü 3 • 3 items • Updated 5 days ago
Hymba: A Hybrid-head Architecture for Small Language Models Paper • 2411.13676 • Published 8 days ago • 37