Text Generation
Safetensors
English
llama
shining-valiant
shining-valiant-2
valiant
valiant-labs
llama-3.1
llama-3.1-instruct
llama-3.1-instruct-8b
llama-3
llama-3-instruct
llama-3-instruct-8b
8b
science
physics
biology
chemistry
compsci
computer-science
engineering
technical
conversational
chat
instruct
Eval Results
Adding Evaluation Results
#6
by
T145
- opened
README.md
CHANGED
@@ -177,7 +177,7 @@ model-index:
|
|
177 |
num_few_shot: 0
|
178 |
metrics:
|
179 |
- type: inst_level_strict_acc and prompt_level_strict_acc
|
180 |
-
value:
|
181 |
name: strict accuracy
|
182 |
source:
|
183 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
@@ -207,7 +207,7 @@ model-index:
|
|
207 |
num_few_shot: 4
|
208 |
metrics:
|
209 |
- type: exact_match
|
210 |
-
value:
|
211 |
name: exact match
|
212 |
source:
|
213 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
@@ -222,7 +222,7 @@ model-index:
|
|
222 |
num_few_shot: 0
|
223 |
metrics:
|
224 |
- type: acc_norm
|
225 |
-
value: 8.
|
226 |
name: acc_norm
|
227 |
source:
|
228 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
@@ -237,7 +237,7 @@ model-index:
|
|
237 |
num_few_shot: 0
|
238 |
metrics:
|
239 |
- type: acc_norm
|
240 |
-
value: 7.
|
241 |
name: acc_norm
|
242 |
source:
|
243 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
@@ -254,7 +254,7 @@ model-index:
|
|
254 |
num_few_shot: 5
|
255 |
metrics:
|
256 |
- type: acc
|
257 |
-
value: 26.
|
258 |
name: accuracy
|
259 |
source:
|
260 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
|
|
177 |
num_few_shot: 0
|
178 |
metrics:
|
179 |
- type: inst_level_strict_acc and prompt_level_strict_acc
|
180 |
+
value: 64.96
|
181 |
name: strict accuracy
|
182 |
source:
|
183 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
|
|
207 |
num_few_shot: 4
|
208 |
metrics:
|
209 |
- type: exact_match
|
210 |
+
value: 12.92
|
211 |
name: exact match
|
212 |
source:
|
213 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
|
|
222 |
num_few_shot: 0
|
223 |
metrics:
|
224 |
- type: acc_norm
|
225 |
+
value: 8.05
|
226 |
name: acc_norm
|
227 |
source:
|
228 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
|
|
237 |
num_few_shot: 0
|
238 |
metrics:
|
239 |
- type: acc_norm
|
240 |
+
value: 7.46
|
241 |
name: acc_norm
|
242 |
source:
|
243 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|
|
|
254 |
num_few_shot: 5
|
255 |
metrics:
|
256 |
- type: acc
|
257 |
+
value: 26.46
|
258 |
name: accuracy
|
259 |
source:
|
260 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ValiantLabs/Llama3.1-8B-ShiningValiant2
|