Quick test tune overtop of meta-llama/Llama-3.2-3B-Instruct using a ~50/50 mix of instruct and completion data.

Note: Training nowhere near complete so I'm unsure how strong of an effect it had. Still refuses requests like meta-llama/Llama-3.2-3B-Instruct.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 22.41
IFEval (0-Shot) 69.31
BBH (3-Shot) 23.81
MATH Lvl 5 (4-Shot) 10.42
GPQA (0-shot) 3.24
MuSR (0-shot) 4.05
MMLU-PRO (5-shot) 23.64
Downloads last month
25
Safetensors
Model size
3.21B params
Tensor type
BF16
·
Inference API
Unable to determine this model's library. Check the docs .

Model tree for PJMixers-Dev/LLaMa-3.2-Instruct-JankMix-v0.1-SFT-3B

Finetuned
(166)
this model
Merges
7 models
Quantizations
2 models

Evaluation results