Update README.md
Browse files
README.md
CHANGED
@@ -15,10 +15,15 @@ llama3-8b-spaetzle-v13 is a merge of the following models:
|
|
15 |
* [Azure99/blossom-v5-llama3-8b](https://huggingface.co/Azure99/blossom-v5-llama3-8b)
|
16 |
* [VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct)
|
17 |
|
|
|
|
|
|
|
18 |
It should work not too bad for German or English, e.g. it achieves 64.14(.10) on EQ Bench v2_de with 170(171)/171 per [q4km GGUF](https://huggingface.co/cstr/llama3-8b-spaetzle-v13-GGUF) (bpe fixed).
|
19 |
And for English EQ-Bench Score (v2): 75.59, Parseable: 171.
|
20 |
|
21 |
-
|
|
|
|
|
22 |
|
23 |
## Sample output
|
24 |
|
|
|
15 |
* [Azure99/blossom-v5-llama3-8b](https://huggingface.co/Azure99/blossom-v5-llama3-8b)
|
16 |
* [VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct)
|
17 |
|
18 |
+
No change in llama3 prompt format template.
|
19 |
+
|
20 |
+
## Benchmarks
|
21 |
It should work not too bad for German or English, e.g. it achieves 64.14(.10) on EQ Bench v2_de with 170(171)/171 per [q4km GGUF](https://huggingface.co/cstr/llama3-8b-spaetzle-v13-GGUF) (bpe fixed).
|
22 |
And for English EQ-Bench Score (v2): 75.59, Parseable: 171.
|
23 |
|
24 |
+
| Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
|
25 |
+
|------------------------------|------------|--------|-----------|--------|------------|------------|--------|
|
26 |
+
| cstr/llama3-8b-spaetzle-v13 | 71.26 | 68.69 | 85.05 | 68.06 | 59.43 | 79.24 | 67.1 |
|
27 |
|
28 |
## Sample output
|
29 |
|