Qian Liu's picture

Qian Liu

SivilTaram

AI & ML interests

Cooking cool things

Recent Activity

updated a model about 4 hours ago
SivilTaram/tongyao_models
updated a model about 4 hours ago
SivilTaram/tongyao_models
updated a model about 4 hours ago
SivilTaram/tongyao_models
View all activity

Organizations

Microsoft's profile picture Spaces-explorers's profile picture Multimodal Art Projection's profile picture Sea AI Lab's profile picture XLang NLP Lab's profile picture Table Research Lab's profile picture BigCode's profile picture OpenLemur's profile picture Sea Language Team's profile picture LoraHub's profile picture StarPEFT's profile picture Qwen's profile picture S3Eval's profile picture GAIR-ProX's profile picture Learning UnkNown librAry's profile picture code-world-model's profile picture Social Post Explorers's profile picture Sailor2's profile picture Sea AI Lab-Sailor's profile picture OpenCoder's profile picture Sailor2 Evaluation's profile picture ml-fw-prerelease's profile picture Data Is Better Together Contributor's profile picture

Posts 4

view post
Post
2623
Still following your human intuition to mix corpora from different sources for pre-training 🧠? Everyone says that data mixture has a big impact on model performance, but how - and why🕵️? Did you know that web corpora are actually highly impactful for downstream tasks 🏆?

Check out our preprint "RegMix: Data Mixture as Regression for Language Model Pre-training" 📄

🔬 In this paper, we've proposed an automatic data mixture method RegMix that achieves a 6.3% improvement over human selection on the widely used HellaSwag benchmark - and it only needs a 2% extra training FLOPs! 📈

📄 Paper: RegMix: Data Mixture as Regression for Language Model Pre-training (2407.01492)
💻 Code: https://github.com/sail-sg/regmix
📊 Collection: sail/regmix-data-mixture-as-regression-6682b6caab37b9442877f0ce
🎮 Demo: https://huggingface.co/spaces/sail/RegMix

Articles 3

Article
10

RegMix: Data Mixture as Regression for Language Model Pre-training