Update README.md (#8)
Browse files- Update README.md (ceefe6791c7907f137f4119367b654ba936f74e9)
README.md
CHANGED
|
@@ -198,8 +198,9 @@ Data used for model training and how the data was processed.
|
|
| 198 |
### Training Dataset
|
| 199 |
|
| 200 |
These models were trained on a dataset of text data that includes a wide variety
|
| 201 |
-
of sources. The 27B model was trained with 13 trillion tokens
|
| 202 |
-
|
|
|
|
| 203 |
|
| 204 |
* Web Documents: A diverse collection of web text ensures the model is exposed
|
| 205 |
to a broad range of linguistic styles, topics, and vocabulary. Primarily
|
|
@@ -382,7 +383,7 @@ and in brief in the
|
|
| 382 |
<tr>
|
| 383 |
<th>Evaluation</th>
|
| 384 |
<th>Capability</th>
|
| 385 |
-
<th>Gemma 2 27B</th>
|
| 386 |
</tr>
|
| 387 |
</thead>
|
| 388 |
<tbody>
|
|
|
|
| 198 |
### Training Dataset
|
| 199 |
|
| 200 |
These models were trained on a dataset of text data that includes a wide variety
|
| 201 |
+
of sources. The 27B model was trained with 13 trillion tokens, the 9B model was
|
| 202 |
+
trained with 8 trillion tokens, and 2B model was trained with 2 trillion tokens.
|
| 203 |
+
Here are the key components:
|
| 204 |
|
| 205 |
* Web Documents: A diverse collection of web text ensures the model is exposed
|
| 206 |
to a broad range of linguistic styles, topics, and vocabulary. Primarily
|
|
|
|
| 383 |
<tr>
|
| 384 |
<th>Evaluation</th>
|
| 385 |
<th>Capability</th>
|
| 386 |
+
<th>Gemma 2 IT 27B</th>
|
| 387 |
</tr>
|
| 388 |
</thead>
|
| 389 |
<tbody>
|