Update README.md (#8)
Browse files- Update README.md (ceefe6791c7907f137f4119367b654ba936f74e9)
README.md
CHANGED
@@ -198,8 +198,9 @@ Data used for model training and how the data was processed.
|
|
198 |
### Training Dataset
|
199 |
|
200 |
These models were trained on a dataset of text data that includes a wide variety
|
201 |
-
of sources. The 27B model was trained with 13 trillion tokens
|
202 |
-
|
|
|
203 |
|
204 |
* Web Documents: A diverse collection of web text ensures the model is exposed
|
205 |
to a broad range of linguistic styles, topics, and vocabulary. Primarily
|
@@ -382,7 +383,7 @@ and in brief in the
|
|
382 |
<tr>
|
383 |
<th>Evaluation</th>
|
384 |
<th>Capability</th>
|
385 |
-
<th>Gemma 2 27B</th>
|
386 |
</tr>
|
387 |
</thead>
|
388 |
<tbody>
|
|
|
198 |
### Training Dataset
|
199 |
|
200 |
These models were trained on a dataset of text data that includes a wide variety
|
201 |
+
of sources. The 27B model was trained with 13 trillion tokens, the 9B model was
|
202 |
+
trained with 8 trillion tokens, and 2B model was trained with 2 trillion tokens.
|
203 |
+
Here are the key components:
|
204 |
|
205 |
* Web Documents: A diverse collection of web text ensures the model is exposed
|
206 |
to a broad range of linguistic styles, topics, and vocabulary. Primarily
|
|
|
383 |
<tr>
|
384 |
<th>Evaluation</th>
|
385 |
<th>Capability</th>
|
386 |
+
<th>Gemma 2 IT 27B</th>
|
387 |
</tr>
|
388 |
</thead>
|
389 |
<tbody>
|