--- license: apache-2.0 datasets: - ThomasTheMaker/pretokenized-dolma-10M - allenai/dolma language: - en --- An 11M model, pre-trained on 10M rows of dataset from Dolma