nezubn
's Collections
Simple and Scalable Strategies to Continually Pre-train Large Language
Models
Paper
β’
2403.08763
β’
Published
β’
49
Jamba: A Hybrid Transformer-Mamba Language Model
Paper
β’
2403.19887
β’
Published
β’
104
Transformer-Lite: High-efficiency Deployment of Large Language Models on
Mobile Phone GPUs
Paper
β’
2403.20041
β’
Published
β’
34
Advancing LLM Reasoning Generalists with Preference Trees
Paper
β’
2404.02078
β’
Published
β’
44
Mixture-of-Depths: Dynamically allocating compute in transformer-based
language models
Paper
β’
2404.02258
β’
Published
β’
104
Understanding LLMs: A Comprehensive Overview from Training to Inference
Paper
β’
2401.02038
β’
Published
β’
62
Make Your LLM Fully Utilize the Context
Paper
β’
2404.16811
β’
Published
β’
52
SUTRA: Scalable Multilingual Language Model Architecture
Paper
β’
2405.06694
β’
Published
β’
37
Beyond Scaling Laws: Understanding Transformer Performance with
Associative Memory
Paper
β’
2405.08707
β’
Published
β’
27
LoRA Learns Less and Forgets Less
Paper
β’
2405.09673
β’
Published
β’
88
Layer-Condensed KV Cache for Efficient Inference of Large Language
Models
Paper
β’
2405.10637
β’
Published
β’
19
π
Llm Pricing