Model Card for Model ID

AI ์ ๋น ๋ฐ์ดํฐ ๋ถ์ ์ ๋ฌธ ๊ธฐ์ ์ธ Linkbricks์ ๋ฐ์ดํฐ์ฌ์ด์ธํฐ์คํธ์ธ ์ง์ค์ฑ(Saxo) ์ด์ฌ๊ฐ meta-llama/Meta-Llama-3-8B๋ฅผ ๋ฒ ์ด์ค๋ชจ๋ธ๋ก GCP์์ H100-80G 8๊ฐ๋ฅผ ํตํด SFT-DPO ํ๋ จํ ํ๊ธ ๊ธฐ๋ฐ LLAMA3-8b 4๊ฐ์ MoE(Mixture of Expert)๋ชจ๋ธ. ํ ํฌ๋์ด์ ๋ ๋ผ๋ง3๋ ๋์ผํ๋ฉฐ ํ๊ธ VOCA ํ์ฅ์ ํ์ง ์์ ๋ฒ์ ์ ๋๋ค. ์ผ๋ฐ์ง์์๋ต(์ฑํ )-์๋ฃ-๊ตฐ์ฌ-์ฝ๋ฉ ํนํ LLM์ ํตํฉ
Dr. Yunsung Ji (Saxo), a data scientist at Linkbricks, a company specializing in AI and big data analytics, trained the meta-llama/Meta-Llama-3-8B base model on 8 H100-60Gs on GCP for 4 hours of instructional training (8000 Tokens). Accelerate, Deepspeed Zero-3 libraries were used.
- Downloads last month
- 7
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
Model tree for Saxo/Linkbricks-Horizon-AI-Korean-LLAMA3blend-4x8b
Base model
meta-llama/Meta-Llama-3-8B-Instruct