|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: cc-by-sa-4.0 |
|
|
size_categories: |
|
|
- 10M<n<100M |
|
|
task_categories: |
|
|
- text-generation |
|
|
- fill-mask |
|
|
pretty_name: Wikipedia English Cleaned |
|
|
tags: |
|
|
- wikipedia |
|
|
- english |
|
|
- language-modeling |
|
|
--- |
|
|
|
|
|
# Wikipedia English Cleaned |
|
|
|
|
|
## Dataset Description |
|
|
|
|
|
A cleaned English Wikipedia text corpus suitable for language model training. This dataset contains English Wikipedia articles processed and cleaned for use in training small language models. |
|
|
|
|
|
### Dataset Summary |
|
|
|
|
|
- **Language**: English |
|
|
- **Size**: ~133 MB (plain text) |
|
|
- **Format**: Plain text (.txt) |
|
|
- **License**: CC-BY-SA 4.0 (Wikipedia content license) |
|
|
|
|
|
### Source Data |
|
|
|
|
|
The dataset is derived from English Wikipedia articles, cleaned and formatted for language model training. |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
### Data Fields |
|
|
|
|
|
The dataset consists of plain text files containing Wikipedia articles, with one article or paragraph per line. |
|
|
|
|
|
### Data Splits |
|
|
|
|
|
This dataset is provided as a single text file without predefined splits. Users can create their own train/validation/test splits as needed. |
|
|
|
|
|
## Usage |
|
|
|
|
|
### Loading the Dataset |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
# Load the dataset |
|
|
dataset = load_dataset("adityasasidhar/Wikipedia_Cleaned") |
|
|
|
|
|
# Or load directly as text |
|
|
from datasets import load_dataset |
|
|
dataset = load_dataset("adityasasidhar/Wikipedia_Cleaned", split="train") |
|
|
``` |
|
|
|
|
|
### Example Use Case |
|
|
|
|
|
This dataset was used to train a Small Language Model (SLM) with the following characteristics: |
|
|
- **Model Size**: 15.58M parameters |
|
|
- **Architecture**: Decoder-only Transformer |
|
|
- **Training**: Combined with TinyStories dataset for ~100M tokens total |
|
|
|
|
|
## Dataset Creation |
|
|
|
|
|
### Curation Rationale |
|
|
|
|
|
This dataset was created to provide clean, factual English text for training small language models. Wikipedia provides high-quality, encyclopedic content that helps models learn proper grammar, factual knowledge, and formal writing style. |
|
|
|
|
|
### Source Data |
|
|
|
|
|
- **Source**: English Wikipedia |
|
|
- **Processing**: Cleaned and formatted for language model training |
|
|
- **Quality**: High-quality encyclopedic content |
|
|
|
|
|
## Considerations for Using the Data |
|
|
|
|
|
### Social Impact |
|
|
|
|
|
This dataset contains factual, encyclopedic content from Wikipedia. Users should be aware that: |
|
|
- Wikipedia content reflects the biases and perspectives of its editors |
|
|
- The dataset is suitable for general language modeling tasks |
|
|
- For specific domains, additional fine-tuning may be necessary |
|
|
|
|
|
### Limitations |
|
|
|
|
|
- The dataset represents a snapshot of Wikipedia at a specific point in time |
|
|
- May not include the most recent information |
|
|
- Content is limited to English language articles |
|
|
|
|
|
## Additional Information |
|
|
|
|
|
### Licensing Information |
|
|
|
|
|
The dataset is released under the **CC-BY-SA 4.0** license, consistent with Wikipedia's content license. |
|
|
|
|
|
### Citation Information |
|
|
|
|
|
If you use this dataset, please cite: |
|
|
|
|
|
```bibtex |
|
|
@misc{wikipedia_cleaned_2026, |
|
|
author = {Aditya Sasidhar}, |
|
|
title = {Wikipedia English Cleaned}, |
|
|
year = {2026}, |
|
|
publisher = {Hugging Face}, |
|
|
howpublished = {\url{https://huggingface.co/datasets/adityasasidhar/Wikipedia_Cleaned}} |
|
|
} |
|
|
``` |
|
|
|
|
|
### Contributions |
|
|
|
|
|
Dataset curated and uploaded by [Aditya Sasidhar](https://github.com/adityasasidhar). |
|
|
|