Update README.md
Browse files
README.md
CHANGED
@@ -28,7 +28,9 @@ widget:
|
|
28 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/5f0b462819cb630495b814d7/DVA4MnFUs3UHBnTrX9jG6.png" style="max-width: 550px; height: auto;">
|
29 |
</div>
|
30 |
|
31 |
-
## Please note:
|
|
|
|
|
32 |
|
33 |
# Model Card for Minerva-7B-base-v1.0
|
34 |
|
@@ -133,9 +135,11 @@ We assessed our model using the [LM-Evaluation-Harness](https://github.com/Eleut
|
|
133 |
|
134 |
All the reported benchmark data was already present in the LM-Evaluation-Harness suite.
|
135 |
|
136 |
-
|
137 |
-
|
138 |
-
|
|
|
|
|
139 |
<!-- | [xcopa](https://huggingface.co/datasets/xcopa) (0-shot) | 0.694 |
|
140 |
| [Hellaswag](https://huggingface.co/datasets/alexandrainst/m_hellaswag) (5-shot) | 0.5293 |
|
141 |
| [Belebele](https://huggingface.co/datasets/facebook/belebele) (5-shot) | 0.2333 |
|
@@ -144,9 +148,9 @@ All the reported benchmark data was already present in the LM-Evaluation-Harness
|
|
144 |
| [M MMLU](https://huggingface.co/datasets/alexandrainst/m_mmlu) (5-shot) | 0.2612 |
|
145 |
| [arc challenge](https://huggingface.co/datasets/alexandrainst/m_arc) (5-shot) | 0.3268 | -->
|
146 |
|
147 |
-
**English** Data:
|
148 |
-
| Task | Accuracy |
|
149 |
-
| --- | --- |
|
150 |
<!-- | [Hellaswag](https://huggingface.co/datasets/Rowan/hellaswag) (5-shot) | 0.6168 |
|
151 |
| [piqa](https://huggingface.co/datasets/piqa) (5-shot) | 0.7535 |
|
152 |
| [sciq](https://huggingface.co/datasets/sciq) (5-shot) | 0.925 |
|
|
|
28 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/5f0b462819cb630495b814d7/DVA4MnFUs3UHBnTrX9jG6.png" style="max-width: 550px; height: auto;">
|
29 |
</div>
|
30 |
|
31 |
+
## Please note:
|
32 |
+
|
33 |
+
This model is currently training, the weights available at the moment are taken from the checkpoint at 10K steps out of 590K.
|
34 |
|
35 |
# Model Card for Minerva-7B-base-v1.0
|
36 |
|
|
|
135 |
|
136 |
All the reported benchmark data was already present in the LM-Evaluation-Harness suite.
|
137 |
|
138 |
+
_Scores will be available at later stage._
|
139 |
+
|
140 |
+
<!-- **Italian** Data: -->
|
141 |
+
<!-- | Task | Accuracy |
|
142 |
+
| --- | --- | -->
|
143 |
<!-- | [xcopa](https://huggingface.co/datasets/xcopa) (0-shot) | 0.694 |
|
144 |
| [Hellaswag](https://huggingface.co/datasets/alexandrainst/m_hellaswag) (5-shot) | 0.5293 |
|
145 |
| [Belebele](https://huggingface.co/datasets/facebook/belebele) (5-shot) | 0.2333 |
|
|
|
148 |
| [M MMLU](https://huggingface.co/datasets/alexandrainst/m_mmlu) (5-shot) | 0.2612 |
|
149 |
| [arc challenge](https://huggingface.co/datasets/alexandrainst/m_arc) (5-shot) | 0.3268 | -->
|
150 |
|
151 |
+
<!-- **English** Data: -->
|
152 |
+
<!-- | Task | Accuracy |
|
153 |
+
| --- | --- | -->
|
154 |
<!-- | [Hellaswag](https://huggingface.co/datasets/Rowan/hellaswag) (5-shot) | 0.6168 |
|
155 |
| [piqa](https://huggingface.co/datasets/piqa) (5-shot) | 0.7535 |
|
156 |
| [sciq](https://huggingface.co/datasets/sciq) (5-shot) | 0.925 |
|