Efficient Continual Pre-training by Mitigating the Stability Gap Paper • 2406.14833 • Published Jun 21 • 19
Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research Paper • 2402.00159 • Published Jan 31 • 59
Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision Paper • 2312.09390 • Published Dec 14, 2023 • 32
ToolTalk: Evaluating Tool-Usage in a Conversational Setting Paper • 2311.10775 • Published Nov 15, 2023 • 7
Memory Augmented Language Models through Mixture of Word Experts Paper • 2311.10768 • Published Nov 15, 2023 • 16
PPTC Benchmark: Evaluating Large Language Models for PowerPoint Task Completion Paper • 2311.01767 • Published Nov 3, 2023 • 18