Fixing typo
Browse files
README.md
CHANGED
|
@@ -69,7 +69,7 @@ Model Details:
|
|
| 69 |
- 420B tokens of pre-processed, tokenized and shuffled text extraced from the **[DOLMA](https://allenai.org/dolma)** corpus
|
| 70 |
- We use this dataset to train our model suite
|
| 71 |
|
| 72 |
-
2. **[pretokenized-dolma-
|
| 73 |
- A smaller version of the **pretokenized-dolma** corpus for quick experiments
|
| 74 |
|
| 75 |
3. **[pretokenized-paloma](https://huggingface.co/datasets/pico-lm/pretokenized-paloma)**
|
|
|
|
| 69 |
- 420B tokens of pre-processed, tokenized and shuffled text extraced from the **[DOLMA](https://allenai.org/dolma)** corpus
|
| 70 |
- We use this dataset to train our model suite
|
| 71 |
|
| 72 |
+
2. **[pretokenized-dolma-tinsy](https://huggingface.co/datasets/pico-lm/pretokenized-dolma-tinsy)**
|
| 73 |
- A smaller version of the **pretokenized-dolma** corpus for quick experiments
|
| 74 |
|
| 75 |
3. **[pretokenized-paloma](https://huggingface.co/datasets/pico-lm/pretokenized-paloma)**
|