Update README.md
Browse files
README.md
CHANGED
@@ -59,35 +59,20 @@ Hermes 3 is competitive, if not superior, to Llama-3.1 Instruct models at genera
|
|
59 |
|
60 |
## GPT4All:
|
61 |
```
|
62 |
-
|
63 |
| Task |Version| Metric |Value | |Stderr|
|
64 |
-
|
65 |
|-------------|------:|--------|-----:|---|-----:|
|
66 |
-
|
67 |
|arc_challenge| 0|acc |0.5529|± |0.0145|
|
68 |
-
|
69 |
| | |acc_norm|0.5870|± |0.0144|
|
70 |
-
|
71 |
|arc_easy | 0|acc |0.8371|± |0.0076|
|
72 |
-
|
73 |
| | |acc_norm|0.8144|± |0.0080|
|
74 |
-
|
75 |
|boolq | 1|acc |0.8599|± |0.0061|
|
76 |
-
|
77 |
|hellaswag | 0|acc |0.6133|± |0.0049|
|
78 |
-
|
79 |
| | |acc_norm|0.7989|± |0.0040|
|
80 |
-
|
81 |
|openbookqa | 0|acc |0.3940|± |0.0219|
|
82 |
-
|
83 |
| | |acc_norm|0.4680|± |0.0223|
|
84 |
-
|
85 |
|piqa | 0|acc |0.8063|± |0.0092|
|
86 |
-
|
87 |
| | |acc_norm|0.8156|± |0.0090|
|
88 |
-
|
89 |
|winogrande | 0|acc |0.7372|± |0.0124|
|
90 |
-
|
91 |
```
|
92 |
|
93 |
Average: 72.59
|
|
|
59 |
|
60 |
## GPT4All:
|
61 |
```
|
|
|
62 |
| Task |Version| Metric |Value | |Stderr|
|
|
|
63 |
|-------------|------:|--------|-----:|---|-----:|
|
|
|
64 |
|arc_challenge| 0|acc |0.5529|± |0.0145|
|
|
|
65 |
| | |acc_norm|0.5870|± |0.0144|
|
|
|
66 |
|arc_easy | 0|acc |0.8371|± |0.0076|
|
|
|
67 |
| | |acc_norm|0.8144|± |0.0080|
|
|
|
68 |
|boolq | 1|acc |0.8599|± |0.0061|
|
|
|
69 |
|hellaswag | 0|acc |0.6133|± |0.0049|
|
|
|
70 |
| | |acc_norm|0.7989|± |0.0040|
|
|
|
71 |
|openbookqa | 0|acc |0.3940|± |0.0219|
|
|
|
72 |
| | |acc_norm|0.4680|± |0.0223|
|
|
|
73 |
|piqa | 0|acc |0.8063|± |0.0092|
|
|
|
74 |
| | |acc_norm|0.8156|± |0.0090|
|
|
|
75 |
|winogrande | 0|acc |0.7372|± |0.0124|
|
|
|
76 |
```
|
77 |
|
78 |
Average: 72.59
|