ENERGY-DRINK-LOVE/eeve_dpo-v3
Our Team
- Jingyeom Kim
- Youjin Chung
Model
Base Model
Hardware and Software
- Hardware: A100 * 8 for training our model
- Deepspeed library & Huggingface TRL Trainer
Dataset
- DPO_dataset
- μ체 μ μ dpo dataset(AI-hub dataset νμ©)
- OpenOrca DPO λ± μμ΄ λ°μ΄ν°μ λ²μ(ENERGY-DRINK-LOVE/translate_share_gpt_dedup_llama_SFT_1024, μ체λͺ¨λΈ νμ©)
Training Method
Benchmark
Task | 0-shot | 5-shot |
---|---|---|
kobest_boolq | 0.950142 | 0.944444 |
kobest_copa | 0.751 | 0.835 |
kobest_hellaswag | 0.474 | 0.508 |
kobest_sentineg | 0.811083 | 0.972292 |
Average | 0.74655625 | 0.81493399 |
- (240307κΈ°μ€ 7λ±)
Average Ko-ARC Ko-HellaSwag Ko-MMLU Ko-TruthfulQA Ko-CommonGen V2 57.97 57.51 67.01 56.3 54.86 54.19
- Downloads last month
- 2,049
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
Model tree for ENERGY-DRINK-LOVE/eeve_dpo-v3
Base model
upstage/SOLAR-10.7B-v1.0
Finetuned
yanolja/EEVE-Korean-10.8B-v1.0
Finetuned
yanolja/EEVE-Korean-Instruct-10.8B-v1.0