ssuresh commited on
Commit
58b4fa2
·
verified ·
1 Parent(s): 59680ab

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +109 -15
README.md CHANGED
@@ -1,25 +1,119 @@
1
- # Pre-training dataset
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
 
3
- This is a Next Token Prediction dataset processed by combining multiple text datasets.
4
- This dataset is suitable for foundational training of smaller LLM models.
5
 
6
- It consists of samples from
7
- - HuggingFaceFW/fineweb
8
- - HuggingFaceTB/finemath
9
- - mlfoundations/dclm-baseline-1.0
10
- - HuggingFaceTB/cosmopedia
11
- - karpathy/fineweb-edu-100b-shuffle
 
 
 
 
 
12
 
13
  ## Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ```
15
- from dataset import load_dataset
16
 
17
- ds = load_dataset("ssuresh/nemo-stage1-50M-samples", split="train")
18
- print(ds)
 
 
 
 
 
19
 
 
 
 
 
 
 
 
20
  ```
21
 
 
22
 
23
- ## TODO:
24
- 1. Update the dataset to add coding samples
25
- 2. Update the datatset to add more foundational training dataset
 
1
+ ---
2
+ license: apache-2.0
3
+ task_categories:
4
+ - text-generation
5
+ - text-classification
6
+ language:
7
+ - en
8
+ tags:
9
+ - nemo
10
+ - pretraining
11
+ - text-generation
12
+ - large-language-model
13
+ - stage1
14
+ size_categories:
15
+ - 10M<n<100M
16
+ ---
17
 
18
+ # NeMo Stage1 Pretraining Dataset - 50M Samples
 
19
 
20
+ This dataset contains 50 million text samples for NeMo model pretraining (Stage 1). The dataset is organized in chunks for efficient loading and processing.
21
+
22
+ ## Dataset Details
23
+
24
+ - **Total Samples**: ~50,000,000
25
+ - **Format**: JSONL (JSON Lines)
26
+ - **Structure**: Each sample contains `{"id": number, "text": "content"}`
27
+ - **Chunks**: 47 files (chunk_000.jsonl to chunk_046.jsonl)
28
+ - **Samples per chunk**: ~1,000,000
29
+ - **Language**: English
30
+ - **Task**: Text generation pretraining
31
 
32
  ## Usage
33
+
34
+ ### Load the entire dataset:
35
+ ```python
36
+ from datasets import load_dataset
37
+
38
+ # Load all chunks
39
+ dataset = load_dataset('ssuresh/nemo-stage1-50M-samples', split='train')
40
+ print(f"Loaded {len(dataset)} samples")
41
+ ```
42
+
43
+ ### Load specific chunks:
44
+ ```python
45
+ # Load a specific chunk
46
+ chunk_dataset = load_dataset('ssuresh/nemo-stage1-50M-samples',
47
+ data_files='chunks/chunk_000.jsonl',
48
+ split='train')
49
+ print(f"Loaded {len(chunk_dataset)} samples from chunk_000")
50
+ ```
51
+
52
+ ### Load multiple chunks:
53
+ ```python
54
+ # Load multiple chunks
55
+ chunk_files = [f'chunks/chunk_{i:03d}.jsonl' for i in range(5)] # First 5 chunks
56
+ multi_chunk_dataset = load_dataset('ssuresh/nemo-stage1-50M-samples',
57
+ data_files=chunk_files,
58
+ split='train')
59
+ print(f"Loaded {len(multi_chunk_dataset)} samples from 5 chunks")
60
+ ```
61
+
62
+ ## Dataset Structure
63
+
64
+ Each sample in the dataset has the following structure:
65
+ ```json
66
+ {
67
+ "id": 0,
68
+ "text": "Your text content here..."
69
+ }
70
+ ```
71
+
72
+ ## Chunk Information
73
+
74
+ - **chunk_000.jsonl** to **chunk_046.jsonl**: Individual chunk files
75
+ - Each chunk contains approximately 1,000,000 samples
76
+ - Chunks are numbered sequentially from 000 to 046
77
+ - Total: 47 chunks
78
+
79
+ ## Sample Data
80
+
81
+ The first 100 samples are available in `sample_data.json` for inspection and understanding the data format.
82
+
83
+ ## Memory-Efficient Loading
84
+
85
+ This dataset is designed for memory-efficient loading:
86
+ - Load individual chunks to avoid memory issues
87
+ - Use streaming for very large datasets
88
+ - Chunk-based processing for distributed training
89
+
90
+ ## Training Usage
91
+
92
+ For NeMo training, you can use this dataset with the `use_processed_datasets: true` setting in your configuration:
93
+
94
+ ```yaml
95
+ data:
96
+ use_processed_datasets: true
97
+ dataset_name: "ssuresh/nemo-stage1-50M-samples"
98
  ```
 
99
 
100
+ ## License
101
+
102
+ This dataset is released under the Apache 2.0 License.
103
+
104
+ ## Citation
105
+
106
+ If you use this dataset in your research, please cite:
107
 
108
+ ```bibtex
109
+ @dataset{nemo_stage1_50m,
110
+ title={NeMo Stage1 Pretraining Dataset - 50M Samples},
111
+ author={Suresh},
112
+ year={2024},
113
+ url={https://huggingface.co/datasets/ssuresh/nemo-stage1-50M-samples}
114
+ }
115
  ```
116
 
117
+ ## Contact
118
 
119
+ For questions or issues with this dataset, please open an issue on the Hugging Face Hub or contact the dataset maintainer.