Model Description

This is the meta-llama/Llama-3.2-1B base model fine tuned on the mlabonne/orpo-dpo-mix-40k dataset.

Evaluation Results

We used lm-evalutation-harness from EleutherAI to evaluate this fine-tuned version of meta-llama/Llama-3.2-1B on the 'Hellaswag' benchmark.

Hellaswag

Tasks Version Filter n-shot Metric Value Stderr
hellaswag 1 none 0 acc ↑ 0.4773 ± 0.0050
none 0 acc_norm ↑ 0.6358 ± 0.0048
Downloads last month
3
Safetensors
Model size
1.24B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for DamiFass/llama3.2-1B-finetuned-on-mlabonne

Finetuned
(249)
this model