leaderboard-pr-bot's picture
Adding Evaluation Results
f7caaaa
|
raw
history blame
2.36 kB
metadata
license: apache-2.0

Buy Me A Coffee

Merge of Open-Orca/Mistral-7B-SlimOrca and Open-Orca/Mistral-7B-OpenOrca using ties merge.

Weights

Density

Quantizationed versions

Quantizationed versions of this model is available thanks to TheBloke.

GPTQ
GGUF
AWQ

Evaluation Results (Open LLM Leaderboard)

Metric Value
Avg. 66.54
ARC (25-shot) 62.97
HellaSwag (10-shot) 83.49
MMLU (5-shot) 62.3
TruthfulQA (0-shot) 57.39

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 53.43
ARC (25-shot) 62.97
HellaSwag (10-shot) 83.49
MMLU (5-shot) 62.3
TruthfulQA (0-shot) 57.39
Winogrande (5-shot) 77.43
GSM8K (5-shot) 21.46
DROP (3-shot) 9.01