| _target_: src.datamodules.language_modeling_hf.LMDataModule | |
| dataset_name: the_pile | |
| dataset_config_name: null | |
| tokenizer_name: gpt2 | |
| cache_dir: ${oc.env:DATA_DIR,${data_dir}}/the_pile/cache | |
| max_length: 2048 | |
| add_eos: True | |
| batch_size: 4 # per GPU | |
| batch_size_eval: ${eval:${.batch_size} * 2} | |
| num_workers: 64 # For preprocessing only | |
| use_shmem: False | |
| shuffle: True | |
| pin_memory: True | |
| __train_len: ${div_up:374337375694, ${.max_length}} | |