ENERGY-DRINK-LOVE/eeve_dpo-v3

Our Team

  • Jingyeom Kim
  • Youjin Chung

Model

Base Model

Hardware and Software

  • Hardware: A100 * 8 for training our model
  • Deepspeed library & Huggingface TRL Trainer

Dataset

  • DPO_dataset
    • 자체 μ œμž‘ dpo dataset(AI-hub dataset ν™œμš©)
    • OpenOrca DPO λ“± μ˜μ–΄ 데이터셋 λ²ˆμ—­(ENERGY-DRINK-LOVE/translate_share_gpt_dedup_llama_SFT_1024, 자체λͺ¨λΈ ν™œμš©)

Training Method

Benchmark

Ko LM Eval Harness

Task 0-shot 5-shot
kobest_boolq 0.950142 0.944444
kobest_copa 0.751 0.835
kobest_hellaswag 0.474 0.508
kobest_sentineg 0.811083 0.972292
Average 0.74655625 0.81493399

Ko-LLM-Leaderboard

  • (240307κΈ°μ€€ 7λ“±)
    Average Ko-ARC Ko-HellaSwag Ko-MMLU Ko-TruthfulQA Ko-CommonGen V2
    57.97 57.51 67.01 56.3 54.86 54.19
Downloads last month
2,049
Safetensors
Model size
10.8B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for ENERGY-DRINK-LOVE/eeve_dpo-v3

Finetuned
(18)
this model
Quantizations
3 models

Spaces using ENERGY-DRINK-LOVE/eeve_dpo-v3 3