| license: apache-2.0 | |
| datasets: | |
| - ThomasTheMaker/pretokenized-dolma-10M | |
| - allenai/dolma | |
| language: | |
| - en | |
| An 11M model, pre-trained on 10M rows of dataset from Dolma |
| license: apache-2.0 | |
| datasets: | |
| - ThomasTheMaker/pretokenized-dolma-10M | |
| - allenai/dolma | |
| language: | |
| - en | |
| An 11M model, pre-trained on 10M rows of dataset from Dolma |