Update README.md
Browse files
README.md
CHANGED
@@ -17,15 +17,12 @@ note that this is a zero-shot setting as opposite to open llm leaderboard's few-
|
|
17 |
Model ARC_C HellaSwag PIQA Winogrande Average
|
18 |
palmer-001 | 0.2807 | 0.5524 | 0.7106 | 0.5896 | 0.5333 |
|
19 |
palmer-003-turbo | 0.3106 | 0.5806 | 0.7247 | 0.5951 | 0.5527 |
|
20 |
-
p-003-turbo-2401 |
|
21 |
palmer-002 | 0.3242 | 0.5956 | 0.7345 | 0.5888 | 0.5607 |
|
22 |
```
|
23 |
|
24 |
This model is as good as tinyllama base while being half the size.
|
25 |
|
26 |
-
### training π¦Ύ
|
27 |
-
Training took 1.5 rtx 2060 gpu hours. It was trained on 15,000 gpt-4 shuffled samples. palmer was fine-tuned using lower learning rates ensuring it keeps as much general knowledge as possible.
|
28 |
-
|
29 |
### prompt π
|
30 |
```
|
31 |
no prompt π
|
|
|
17 |
Model ARC_C HellaSwag PIQA Winogrande Average
|
18 |
palmer-001 | 0.2807 | 0.5524 | 0.7106 | 0.5896 | 0.5333 |
|
19 |
palmer-003-turbo | 0.3106 | 0.5806 | 0.7247 | 0.5951 | 0.5527 |
|
20 |
+
p-003-turbo-2401 | 0.3114 | ~~~~~~ | 0.7258 | 0.5959 | ~~~~~~ | (this)
|
21 |
palmer-002 | 0.3242 | 0.5956 | 0.7345 | 0.5888 | 0.5607 |
|
22 |
```
|
23 |
|
24 |
This model is as good as tinyllama base while being half the size.
|
25 |
|
|
|
|
|
|
|
26 |
### prompt π
|
27 |
```
|
28 |
no prompt π
|