Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens Paper • 2411.17691 • Published 28 days ago • 9
Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens Paper • 2411.17691 • Published 28 days ago • 9 • 4
Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens Paper • 2411.17691 • Published 28 days ago • 9 • 4
Relaxed Recursive Transformers: Effective Parameter Sharing with Layer-wise LoRA Paper • 2410.20672 • Published Oct 28 • 6 • 3
Scaling Synthetic Data Creation with 1,000,000,000 Personas Paper • 2406.20094 • Published Jun 28 • 95
BERT-of-Theseus: Compressing BERT by Progressive Module Replacing Paper • 2002.02925 • Published Feb 7, 2020
Improving Sequence-to-Sequence Pre-training via Sequence Span Rewriting Paper • 2101.00416 • Published Jan 2, 2021
Unleashing Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration Paper • 2307.05300 • Published Jul 11, 2023 • 18
In-context Autoencoder for Context Compression in a Large Language Model Paper • 2307.06945 • Published Jul 13, 2023 • 27
SCALE: Synergized Collaboration of Asymmetric Language Translation Engines Paper • 2309.17061 • Published Sep 29, 2023 • 1
Unlocking Efficiency in Large Language Model Inference: A Comprehensive Survey of Speculative Decoding Paper • 2401.07851 • Published Jan 15 • 1
Pay Attention to Your Tone: Introducing a New Dataset for Polite Language Rewrite Paper • 2212.10190 • Published Dec 20, 2022
In-context Autoencoder for Context Compression in a Large Language Model Paper • 2307.06945 • Published Jul 13, 2023 • 27