-
Larimar: Large Language Models with Episodic Memory Control
Paper • 2403.11901 • Published • 30 -
Sparse Upcycling: Training Mixture-of-Experts from Dense Checkpoints
Paper • 2212.05055 • Published • 5 -
Mixture-of-Depths: Dynamically allocating compute in transformer-based language models
Paper • 2404.02258 • Published • 102 -
Multi-Head Mixture-of-Experts
Paper • 2404.15045 • Published • 55
Collections
Discover the best community collections!
Collections including paper arxiv:2404.19756
-
Is Cosine-Similarity of Embeddings Really About Similarity?
Paper • 2403.05440 • Published • 3 -
GISTEmbed: Guided In-sample Selection of Training Negatives for Text Embedding Fine-tuning
Paper • 2402.16829 • Published -
Make Your LLM Fully Utilize the Context
Paper • 2404.16811 • Published • 52 -
KAN: Kolmogorov-Arnold Networks
Paper • 2404.19756 • Published • 102
-
VideoElevator: Elevating Video Generation Quality with Versatile Text-to-Image Diffusion Models
Paper • 2403.05438 • Published • 15 -
Chronos: Learning the Language of Time Series
Paper • 2403.07815 • Published • 43 -
Adding NVMe SSDs to Enable and Accelerate 100B Model Fine-tuning on a Single GPU
Paper • 2403.06504 • Published • 52 -
Latent-NeRF for Shape-Guided Generation of 3D Shapes and Textures
Paper • 2211.07600 • Published
-
Measuring the Effects of Data Parallelism on Neural Network Training
Paper • 1811.03600 • Published • 2 -
Adafactor: Adaptive Learning Rates with Sublinear Memory Cost
Paper • 1804.04235 • Published • 2 -
EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks
Paper • 1905.11946 • Published • 3 -
Yi: Open Foundation Models by 01.AI
Paper • 2403.04652 • Published • 59
-
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Paper • 2402.01739 • Published • 26 -
Rethinking Interpretability in the Era of Large Language Models
Paper • 2402.01761 • Published • 19 -
Self-Discover: Large Language Models Self-Compose Reasoning Structures
Paper • 2402.03620 • Published • 107 -
Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model
Paper • 2402.07827 • Published • 43
-
Blending Is All You Need: Cheaper, Better Alternative to Trillion-Parameters LLM
Paper • 2401.02994 • Published • 45 -
MambaByte: Token-free Selective State Space Model
Paper • 2401.13660 • Published • 47 -
Repeat After Me: Transformers are Better than State Space Models at Copying
Paper • 2402.01032 • Published • 22 -
BlackMamba: Mixture of Experts for State-Space Models
Paper • 2402.01771 • Published • 22
-
VideoSwap: Customized Video Subject Swapping with Interactive Semantic Point Correspondence
Paper • 2312.02087 • Published • 19 -
FaceStudio: Put Your Face Everywhere in Seconds
Paper • 2312.02663 • Published • 28 -
Orthogonal Adaptation for Modular Customization of Diffusion Models
Paper • 2312.02432 • Published • 12 -
ReconFusion: 3D Reconstruction with Diffusion Priors
Paper • 2312.02981 • Published • 8