| | --- |
| | language: |
| | - ko |
| | dataset_info: |
| | features: |
| | - name: text |
| | dtype: string |
| | - name: source |
| | dtype: string |
| | - name: token_count |
| | dtype: int64 |
| | - name: __index_level_0__ |
| | dtype: int64 |
| | splits: |
| | - name: train |
| | num_bytes: 8555372905 |
| | num_examples: 1284879 |
| | download_size: 4472792071 |
| | dataset_size: 8555372905 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: train |
| | path: data/train-* |
| | --- |
| | |
| | # KOREAN-WEBTEXT |
| |
|
| | **KOREAN-WEBTEXT** is a high-quality Korean language corpus consisting of 2.2 billion tokens. The data has been collected from the following sources: |
| |
|
| | - **cc100** |
| | - **oscar-corpus/OSCAR-2201** |
| | - **oscar-corpus/OSCAR-2109** |
| | - **oscar-corpus/OSCAR-2301** |
| | - **ontocord/CulturaY** |
| | - **Additional credible internet sources collected by out team** |
| |
|
| | (We are working to add more sources) |
| |
|
| | The dataset undergoes rigorous filtering at both the sentence and document levels to ensure quality of text data. Additionally, simple deduplication processes are applied to further refine the dataset. |
| |
|
| | ## Dataset Structure |
| |
|
| | ### Sentence-Level Filters |
| |
|
| | The following filters are applied at the sentence level: |
| |
|
| | 1. **Repetition Check**: The ratio of repetition for any word in a line should not exceed 0.2. |
| | 2. **Punctuation Check**: Lines must end with one of these punctuation marks: `.`, `?`, `]`, or `"`. |
| | 3. **Token Count Check**: The line must contain more than 16 tokens. |
| | 4. **Character Count Check**: The line must contain more than 32 characters. |
| |
|
| | ### Document-Level Filters |
| |
|
| | The following filters are applied at the document level: |
| |
|
| | 1. **Token Count Check**: Documents must contain more than 512 tokens. |
| | 2. **Stopwords Removal**: Documents containing any of the following stopwords are removed: |
| | ```python |
| | stopwords = [ |
| | 'www', 'http', '...', 'ㅋㅋㅋ', '약관', 'is', '카지노', '토토', '\u3000', |
| | '■', '▲', '010', '.kr', '@', '마사지', '스웨디시', '대선' |
| | ] |
| | ``` |
| |
|
| | ### Deduplication Processes |
| |
|
| | To ensure data uniqueness, the following deduplication steps are applied: |
| |
|
| | 1. **Exact Deduplication**: Removal of exact duplicate lines. |
| | 2. **First 15 Tokens Deduplication**: Removal of lines with identical first 15 tokens. |
| | 3. **Last 15 Tokens Deduplication**: Removal of lines with identical last 15 tokens. |
| |
|
| | ## Usage |
| |
|
| | While the dataset may be small for pretraining models due to its size, we expect it to be better suited for ablation studies. |
| |
|
| | ### Examples |
| |
|
| | #### Loading the Dataset |
| |
|
| | To load and use the dataset, you can use the following example code: |
| |
|
| | ```python |
| | import datasets |
| | |
| | dataset = datasets.load_dataset('HAERAE-HUB/KOREAN-WEBTEXT-1B') |
| | ``` |
| |
|
| | ## Citation |
| |
|
| | If you use this dataset in your research, please cite it as follows: |
| |
|
| | ``` |
| | @dataset{KOREAN-WEBTEXT, |
| | title={KOREAN-WEBTEXT: A High-Quality Korean Language Corpus}, |
| | author={HAERAE-Team}, |
| | year={2024}, |
| | howpublished={\url{https://huggingface.co/datasets/HAERAE-HUB/KOREAN-WEBTEXT}}, |
| | } |
| | ``` |
| |
|
| | ## Contact |
| |
|
| | For more information or questions about the dataset, please contact the maintainers at [spthsrbwls123@yonsei.ac.kr]. |
| |
|
| | --- |
| |
|