Update README.md
Browse files
README.md
CHANGED
@@ -186,8 +186,7 @@ without assuming the other birds fly away, then after
|
|
186 |
|
187 |
### Evaluate the model
|
188 |
|
189 |
-
|
190 |
-
<!-- pip3 install lm-eval==0.4.7
|
191 |
we found lm-eval is very unstable for this model. Please set `add_bos_token=True `to align with the origin model. **Please use autogptq format**
|
192 |
|
193 |
```bash
|
@@ -195,20 +194,22 @@ lm-eval --model hf --model_args pretrained=OPEA/DeepSeek-R1-Distill-Llama-70B-in
|
|
195 |
```
|
196 |
| Metric | BF16 | INT4 |
|
197 |
| :------------------------ | :---------------------- | :--------------- |
|
198 |
-
| avg | 0.
|
199 |
-
|
200 |
-
|
|
201 |
-
|
|
202 |
-
|
|
203 |
-
|
|
204 |
-
|
|
205 |
-
|
|
206 |
-
|
|
207 |
-
|
|
208 |
-
|
|
209 |
-
|
|
210 |
-
|
|
211 |
-
|
|
|
|
|
|
212 |
|
213 |
|
214 |
|
|
|
186 |
|
187 |
### Evaluate the model
|
188 |
|
189 |
+
pip3 install lm-eval==0.4.7
|
|
|
190 |
we found lm-eval is very unstable for this model. Please set `add_bos_token=True `to align with the origin model. **Please use autogptq format**
|
191 |
|
192 |
```bash
|
|
|
194 |
```
|
195 |
| Metric | BF16 | INT4 |
|
196 |
| :------------------------ | :---------------------- | :--------------- |
|
197 |
+
| avg | 0.6636 | 0.6678 |
|
198 |
+
|----------------------|--------|--------|
|
199 |
+
| leaderboard_mmlu_pro | 0.4913 | 0.4780 |
|
200 |
+
| mmlu | 0.7752 | 0.7791 |
|
201 |
+
| lambada_openai | 0.6977 | 0.6996 |
|
202 |
+
| hellaswag | 0.6408 | 0.6438 |
|
203 |
+
| winogrande | 0.7530 | 0.7782 |
|
204 |
+
| piqa | 0.8112 | 0.8194 |
|
205 |
+
| truthfulqa_mc1 | 0.3709 | 0.3721 |
|
206 |
+
| openbookqa | 0.3380 | 0.3600 |
|
207 |
+
| boolq | 0.8847 | 0.8917 |
|
208 |
+
| arc_easy | 0.8131 | 0.8106 |
|
209 |
+
| arc_challenge | 0.5512 | 0.5239 |
|
210 |
+
| leaderboard_ifeval | 0.4421 | 0.4208 |
|
211 |
+
| gsm8k | 0.9295 | 0.9265 |
|
212 |
+
|
213 |
|
214 |
|
215 |
|