Collections
Discover the best community collections!
Collections including paper arxiv:2404.05405
-
SciLitLLM: How to Adapt LLMs for Scientific Literature Understanding
Paper • 2408.15545 • Published • 34 -
Controllable Text Generation for Large Language Models: A Survey
Paper • 2408.12599 • Published • 62 -
To Code, or Not To Code? Exploring Impact of Code in Pre-training
Paper • 2408.10914 • Published • 41 -
Automated Design of Agentic Systems
Paper • 2408.08435 • Published • 38
-
Compression Represents Intelligence Linearly
Paper • 2404.09937 • Published • 27 -
MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Paper • 2404.06395 • Published • 21 -
Long-context LLMs Struggle with Long In-context Learning
Paper • 2404.02060 • Published • 35 -
Are large language models superhuman chemists?
Paper • 2404.01475 • Published • 16
-
Physics of Language Models: Part 1, Context-Free Grammar
Paper • 2305.13673 • Published • 7 -
Physics of Language Models: Part 3.2, Knowledge Manipulation
Paper • 2309.14402 • Published • 6 -
Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws
Paper • 2404.05405 • Published • 9 -
Physics of Language Models: Part 3.1, Knowledge Storage and Extraction
Paper • 2309.14316 • Published • 7
-
SELF: Language-Driven Self-Evolution for Large Language Model
Paper • 2310.00533 • Published • 2 -
GrowLength: Accelerating LLMs Pretraining by Progressively Growing Training Length
Paper • 2310.00576 • Published • 2 -
A Pretrainer's Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity
Paper • 2305.13169 • Published • 3 -
Transformers Can Achieve Length Generalization But Not Robustly
Paper • 2402.09371 • Published • 12