Bram Vanroy commited on
Commit
38b8e33
1 Parent(s): dbe0b3a

add Mistral evals

Browse files
evals/arc/arc_nl_Mistral-7B-v0.1.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "results": {
3
  "arc_nl": {
4
- "acc": 0.41916167664670656,
5
- "acc_stderr": 0.01443766613132885,
6
- "acc_norm": 0.42172797262617623,
7
- "acc_norm_stderr": 0.014449768025902365
8
  }
9
  },
10
  "versions": {
@@ -12,8 +12,8 @@
12
  },
13
  "config": {
14
  "model": "hf-auto",
15
- "model_args": "pretrained=mistralai/Mistral-7B-v0.1,use_accelerate=True,device_map_option=auto,dtype=bfloat16,load_in_8bit=True",
16
- "batch_size": "auto",
17
  "device": "cuda",
18
  "no_cache": false,
19
  "limit": null,
 
1
  {
2
  "results": {
3
  "arc_nl": {
4
+ "acc": 0.41146278870829767,
5
+ "acc_stderr": 0.014398950037131928,
6
+ "acc_norm": 0.42771599657827203,
7
+ "acc_norm_stderr": 0.014476451794175197
8
  }
9
  },
10
  "versions": {
 
12
  },
13
  "config": {
14
  "model": "hf-auto",
15
+ "model_args": "pretrained=mistralai/Mistral-7B-v0.1,use_accelerate=True,device_map_option=auto,dtype=auto,load_in_8bit=True",
16
+ "batch_size": 2,
17
  "device": "cuda",
18
  "no_cache": false,
19
  "limit": null,
evals/hellaswag/hellaswag_nl_Mistral-7B-v0.1.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hellaswag_nl": {
4
+ "acc": 0.4418780356179169,
5
+ "acc_stderr": 0.005159605430771658,
6
+ "acc_norm": 0.5829465731246627,
7
+ "acc_norm_stderr": 0.005122841872653067
8
+ }
9
+ },
10
+ "versions": {
11
+ "hellaswag_nl": 1
12
+ },
13
+ "config": {
14
+ "model": "hf-auto",
15
+ "model_args": "pretrained=mistralai/Mistral-7B-v0.1,use_accelerate=True,device_map_option=auto,dtype=auto,load_in_8bit=True",
16
+ "batch_size": 2,
17
+ "device": "cuda",
18
+ "no_cache": false,
19
+ "limit": null,
20
+ "bootstrap_iters": 100000,
21
+ "description_dict": {}
22
+ }
23
+ }
evals/truthfulqa/truthfulqa_nl_Mistral-7B-v0.1.json CHANGED
@@ -12,7 +12,7 @@
12
  },
13
  "config": {
14
  "model": "hf-auto",
15
- "model_args": "pretrained=mistralai/Mistral-7B-v0.1,use_accelerate=True,device_map_option=auto,dtype=bfloat16,load_in_8bit=True",
16
  "batch_size": 2,
17
  "device": "cuda",
18
  "no_cache": false,
 
12
  },
13
  "config": {
14
  "model": "hf-auto",
15
+ "model_args": "pretrained=mistralai/Mistral-7B-v0.1,use_accelerate=True,device_map_option=auto,dtype=auto,load_in_8bit=True",
16
  "batch_size": 2,
17
  "device": "cuda",
18
  "no_cache": false,