cstr commited on
Commit
c29116a
1 Parent(s): c8736e2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -1
README.md CHANGED
@@ -15,10 +15,15 @@ llama3-8b-spaetzle-v13 is a merge of the following models:
15
  * [Azure99/blossom-v5-llama3-8b](https://huggingface.co/Azure99/blossom-v5-llama3-8b)
16
  * [VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct)
17
 
 
 
 
18
  It should work not too bad for German or English, e.g. it achieves 64.14(.10) on EQ Bench v2_de with 170(171)/171 per [q4km GGUF](https://huggingface.co/cstr/llama3-8b-spaetzle-v13-GGUF) (bpe fixed).
19
  And for English EQ-Bench Score (v2): 75.59, Parseable: 171.
20
 
21
- No change in llama3 prompt format template.
 
 
22
 
23
  ## Sample output
24
 
 
15
  * [Azure99/blossom-v5-llama3-8b](https://huggingface.co/Azure99/blossom-v5-llama3-8b)
16
  * [VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct)
17
 
18
+ No change in llama3 prompt format template.
19
+
20
+ ## Benchmarks
21
  It should work not too bad for German or English, e.g. it achieves 64.14(.10) on EQ Bench v2_de with 170(171)/171 per [q4km GGUF](https://huggingface.co/cstr/llama3-8b-spaetzle-v13-GGUF) (bpe fixed).
22
  And for English EQ-Bench Score (v2): 75.59, Parseable: 171.
23
 
24
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
25
+ |------------------------------|------------|--------|-----------|--------|------------|------------|--------|
26
+ | cstr/llama3-8b-spaetzle-v13 | 71.26 | 68.69 | 85.05 | 68.06 | 59.43 | 79.24 | 67.1 |
27
 
28
  ## Sample output
29