Adding Evaluation Results

#6
Files changed (1) hide show
  1. README.md +18 -4
README.md CHANGED
@@ -1,12 +1,9 @@
1
  ---
2
  license: llama3
3
- base_model: meta-llama/Meta-Llama-3-70B
4
  tags:
5
  - generated_from_trainer
6
  - axolotl
7
- model-index:
8
- - name: out
9
- results: []
10
  datasets:
11
  - cognitivecomputations/Dolphin-2.9
12
  - teknium/OpenHermes-2.5
@@ -16,6 +13,9 @@ datasets:
16
  - microsoft/orca-math-word-problems-200k
17
  - Locutusque/function-calling-chatml
18
  - internlm/Agent-FLAN
 
 
 
19
  ---
20
 
21
  # Dolphin 2.9.1 Llama 3 70b 🐬
@@ -510,3 +510,17 @@ The following hyperparameters were used during training:
510
  - Pytorch 2.2.2+cu121
511
  - Datasets 2.19.1
512
  - Tokenizers 0.19.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: llama3
 
3
  tags:
4
  - generated_from_trainer
5
  - axolotl
6
+ base_model: meta-llama/Meta-Llama-3-70B
 
 
7
  datasets:
8
  - cognitivecomputations/Dolphin-2.9
9
  - teknium/OpenHermes-2.5
 
13
  - microsoft/orca-math-word-problems-200k
14
  - Locutusque/function-calling-chatml
15
  - internlm/Agent-FLAN
16
+ model-index:
17
+ - name: out
18
+ results: []
19
  ---
20
 
21
  # Dolphin 2.9.1 Llama 3 70b 🐬
 
510
  - Pytorch 2.2.2+cu121
511
  - Datasets 2.19.1
512
  - Tokenizers 0.19.1
513
+
514
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
515
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_cognitivecomputations__dolphin-2.9.1-llama-3-70b)
516
+
517
+ | Metric |Value|
518
+ |-------------------|----:|
519
+ |Avg. |23.41|
520
+ |IFEval (0-Shot) |37.60|
521
+ |BBH (3-Shot) |31.10|
522
+ |MATH Lvl 5 (4-Shot)| 5.44|
523
+ |GPQA (0-shot) | 7.83|
524
+ |MuSR (0-shot) |23.70|
525
+ |MMLU-PRO (5-shot) |34.78|
526
+