sort english base model results by descending order
Browse files
README.md
CHANGED
@@ -50,12 +50,12 @@ The evaluation was done **five-shot** with native prompts and only a sample of 1
|
|
50 |
|
51 |
| Model | ARC | BBH | HellaSwag | MMLU | GSM8k | Average |
|
52 |
| ---------------------------------------- | ----- | ----- | --------- | ----- | ----- | ------- |
|
53 |
-
|
|
54 |
| google/gemma-2-9b | 68.00 | 53.53 | 82.73 | 70.26 | 63.53 | 67.61 |
|
|
|
55 |
| meta-llama/Meta-Llama-3-8B | 57.85 | 46.09 | 81.89 | 65.10 | 45.34 | 59.25 |
|
56 |
-
| Qwen/Qwen2-7B | 61.86 | 53.10 | 80.63 | 70.45 | 78.09 | 68.83 |
|
57 |
-
| Sail/Sailor-7B | 50.34 | 35.65 | 76.11 | 52.80 | 33.81 | 49.74 |
|
58 |
| mistralai/Mistral-7B-v0.3 | 59.56 | 44.89 | 82.97 | 62.36 | 33.36 | 56.63 |
|
|
|
59 |
|
60 |
|
61 |
## Training Details
|
|
|
50 |
|
51 |
| Model | ARC | BBH | HellaSwag | MMLU | GSM8k | Average |
|
52 |
| ---------------------------------------- | ----- | ----- | --------- | ----- | ----- | ------- |
|
53 |
+
| Qwen/Qwen2-7B | 61.86 | 53.10 | 80.63 | 70.45 | 78.09 | 68.83 |
|
54 |
| google/gemma-2-9b | 68.00 | 53.53 | 82.73 | 70.26 | 63.53 | 67.61 |
|
55 |
+
| aisingapore/llama3-8b-cpt-sealionv2-base | 58.87 | 47.70 | 81.14 | 63.11 | 50.49 | 60.26 |
|
56 |
| meta-llama/Meta-Llama-3-8B | 57.85 | 46.09 | 81.89 | 65.10 | 45.34 | 59.25 |
|
|
|
|
|
57 |
| mistralai/Mistral-7B-v0.3 | 59.56 | 44.89 | 82.97 | 62.36 | 33.36 | 56.63 |
|
58 |
+
| Sail/Sailor-7B | 50.34 | 35.65 | 76.11 | 52.80 | 33.81 | 49.74 |
|
59 |
|
60 |
|
61 |
## Training Details
|