adityasasidhar commited on
Commit
8181090
·
verified ·
1 Parent(s): 044d232

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +115 -18
README.md CHANGED
@@ -1,18 +1,115 @@
1
- ---
2
- license: mit
3
- dataset_info:
4
- features:
5
- - name: text
6
- dtype: string
7
- splits:
8
- - name: train
9
- num_bytes: 141038347
10
- num_examples: 826715
11
- download_size: 82296094
12
- dataset_size: 141038347
13
- configs:
14
- - config_name: default
15
- data_files:
16
- - split: train
17
- path: data/train-*
18
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: cc-by-sa-4.0
5
+ size_categories:
6
+ - 10M<n<100M
7
+ task_categories:
8
+ - text-generation
9
+ - fill-mask
10
+ pretty_name: Wikipedia English Cleaned
11
+ tags:
12
+ - wikipedia
13
+ - english
14
+ - language-modeling
15
+ ---
16
+
17
+ # Wikipedia English Cleaned
18
+
19
+ ## Dataset Description
20
+
21
+ A cleaned English Wikipedia text corpus suitable for language model training. This dataset contains English Wikipedia articles processed and cleaned for use in training small language models.
22
+
23
+ ### Dataset Summary
24
+
25
+ - **Language**: English
26
+ - **Size**: ~133 MB (plain text)
27
+ - **Format**: Plain text (.txt)
28
+ - **License**: CC-BY-SA 4.0 (Wikipedia content license)
29
+
30
+ ### Source Data
31
+
32
+ The dataset is derived from English Wikipedia articles, cleaned and formatted for language model training.
33
+
34
+ ## Dataset Structure
35
+
36
+ ### Data Fields
37
+
38
+ The dataset consists of plain text files containing Wikipedia articles, with one article or paragraph per line.
39
+
40
+ ### Data Splits
41
+
42
+ This dataset is provided as a single text file without predefined splits. Users can create their own train/validation/test splits as needed.
43
+
44
+ ## Usage
45
+
46
+ ### Loading the Dataset
47
+
48
+ ```python
49
+ from datasets import load_dataset
50
+
51
+ # Load the dataset
52
+ dataset = load_dataset("adityasasidhar/Wikipedia_Cleaned")
53
+
54
+ # Or load directly as text
55
+ from datasets import load_dataset
56
+ dataset = load_dataset("adityasasidhar/Wikipedia_Cleaned", split="train")
57
+ ```
58
+
59
+ ### Example Use Case
60
+
61
+ This dataset was used to train a Small Language Model (SLM) with the following characteristics:
62
+ - **Model Size**: 15.58M parameters
63
+ - **Architecture**: Decoder-only Transformer
64
+ - **Training**: Combined with TinyStories dataset for ~100M tokens total
65
+
66
+ ## Dataset Creation
67
+
68
+ ### Curation Rationale
69
+
70
+ This dataset was created to provide clean, factual English text for training small language models. Wikipedia provides high-quality, encyclopedic content that helps models learn proper grammar, factual knowledge, and formal writing style.
71
+
72
+ ### Source Data
73
+
74
+ - **Source**: English Wikipedia
75
+ - **Processing**: Cleaned and formatted for language model training
76
+ - **Quality**: High-quality encyclopedic content
77
+
78
+ ## Considerations for Using the Data
79
+
80
+ ### Social Impact
81
+
82
+ This dataset contains factual, encyclopedic content from Wikipedia. Users should be aware that:
83
+ - Wikipedia content reflects the biases and perspectives of its editors
84
+ - The dataset is suitable for general language modeling tasks
85
+ - For specific domains, additional fine-tuning may be necessary
86
+
87
+ ### Limitations
88
+
89
+ - The dataset represents a snapshot of Wikipedia at a specific point in time
90
+ - May not include the most recent information
91
+ - Content is limited to English language articles
92
+
93
+ ## Additional Information
94
+
95
+ ### Licensing Information
96
+
97
+ The dataset is released under the **CC-BY-SA 4.0** license, consistent with Wikipedia's content license.
98
+
99
+ ### Citation Information
100
+
101
+ If you use this dataset, please cite:
102
+
103
+ ```bibtex
104
+ @misc{wikipedia_cleaned_2026,
105
+ author = {Aditya Sasidhar},
106
+ title = {Wikipedia English Cleaned},
107
+ year = {2026},
108
+ publisher = {Hugging Face},
109
+ howpublished = {\url{https://huggingface.co/datasets/adityasasidhar/Wikipedia_Cleaned}}
110
+ }
111
+ ```
112
+
113
+ ### Contributions
114
+
115
+ Dataset curated and uploaded by [Aditya Sasidhar](https://github.com/adityasasidhar).