| --- |
| license: mit |
| --- |
| |
| # Dataset Card for Cleaned & Chunked WikiText Corpus |
|
|
| ## π Dataset Description |
|
|
| This dataset is a cleaned and chunked version of WikiText-style corpus, designed for language model pretraining and evaluation. |
|
|
| The preprocessing pipeline follows a lightweight data curation paradigm: |
|
|
| > **cleaning β deduplication β normalization β token-aware chunking** |
|
|
| Specifically, the dataset includes: |
|
|
| - Removal of special and noisy characters |
| - Cleaning of formatting artifacts (e.g., HTML tags, irregular symbols) |
| - Deduplication to reduce redundant or highly similar text samples |
| - Text normalization (whitespace, encoding, etc.) |
| - Chunking of long documents into smaller segments suitable for model training |
|
|
| Each sample corresponds to a **cleaned text chunk**, rather than a full original document. |
|
|
| --- |
|
|
| ## π Data Structure |
|
|
| Each sample in the dataset follows this structure: |
|
|
| ```json |
| { |
| "uid": "string", |
| "content": "string", |
| "meta_data": { |
| "index": "int", |
| "total": "int", |
| "length": "int" |
| } |
| } |