Datasets:

Modalities:
Text
Formats:
json
ArXiv:
Libraries:
Datasets
Dask
License:
CASIA-LM commited on
Commit
5aad68f
·
verified ·
1 Parent(s): db70cca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -39,7 +39,7 @@ We have released the latest and largest Chinese dataset, ChineseWebText 2.0, whi
39
 
40
  We introduce a new toolchain, MDFG-tool (see Figure 1). We begin with the coarse-grained filtering module, which applies rule-based methods to clean the data, focusing on criteria such as text length and sensitive words to ensure data quality. After cleaning, we evaluate the text quality using a BERT-based model. This process generates a quality score, and by selecting an appropriate threshold, we can extract high-quality text data that meets our needs. Next, we use FastText for both single-label and multi-label classification of the cleaned data. Meanwhile, we conduct toxicity assessment. The FastText model is used to filter out toxic content and assign toxicity scores to each text. This scoring system allows researchers to set thresholds for identifying and selecting harmful texts for further training.
41
 
42
- <div align=center><img src="./structure.png" style="zoom:67%;" /></div>
43
 
44
  ### Data Analysis
45
 
@@ -50,7 +50,7 @@ In order to provide a high-level overview of the preparation and preprocessing s
50
  After collecting raw data from various sources, we initially obtain a original Chinese dataset totaling 6.6 TB. However, due to a significant amount of irrelevant and noisy content in some sources, a manual sampling analysis is performed in preparation stage. If irrelevant text accounted for more than 50\% of a source, the data from that source will be discarded entirely. As a result, a substantial portion of the data is removed during the preparation stage, retaining only 67.68\% of the original dataset. In preprocessing stage, four rule-based steps are implemented to filter the remained data. First, the Data Length step remove overly short texts to ensure that each text contains sufficient informational content. Next, the Character Proportion step eliminate texts with a high percentage of noisy characters, such as English, Traditional Chinese characters, or other irrelevant symbols. Finally, the Sensitive Words step and the Deduplication step are employed to remove toxic content and duplicate texts from the dataset. After the preprocessing stage, we produce a high-quality Chinese text dataset totaling 3.8 TB. In the next stage, each text in this high-quality dataset will be enriched with fine-grained annotations, including a quality score, domain lablels, a toxicity score and a toxicity label.
51
 
52
  <div align="center">
53
- <img src="/picture/data_statistics.png" width="67%" />
54
  <br>
55
  <em>Figure 1: Description of Image 1</em>
56
  </div>
@@ -62,12 +62,12 @@ After collecting raw data from various sources, we initially obtain a original C
62
  <table>
63
  <tr>
64
  <td>
65
- <img src="/picture/quality_distribution.pdf" alt="Image 1" style="width:100%;">
66
  <br>
67
  <em>Figure 1: Description of Image 1</em>
68
  </td>
69
  <td>
70
- <img src="/picture/human_acceptance.pdf" alt="Image 2" style="width:100%;">
71
  <br>
72
  <em>Figure 1: Description of Image 1</em>
73
  </td>
 
39
 
40
  We introduce a new toolchain, MDFG-tool (see Figure 1). We begin with the coarse-grained filtering module, which applies rule-based methods to clean the data, focusing on criteria such as text length and sensitive words to ensure data quality. After cleaning, we evaluate the text quality using a BERT-based model. This process generates a quality score, and by selecting an appropriate threshold, we can extract high-quality text data that meets our needs. Next, we use FastText for both single-label and multi-label classification of the cleaned data. Meanwhile, we conduct toxicity assessment. The FastText model is used to filter out toxic content and assign toxicity scores to each text. This scoring system allows researchers to set thresholds for identifying and selecting harmful texts for further training.
41
 
42
+ <div align=center><img src="./picture/structure.png" style="zoom:67%;" /></div>
43
 
44
  ### Data Analysis
45
 
 
50
  After collecting raw data from various sources, we initially obtain a original Chinese dataset totaling 6.6 TB. However, due to a significant amount of irrelevant and noisy content in some sources, a manual sampling analysis is performed in preparation stage. If irrelevant text accounted for more than 50\% of a source, the data from that source will be discarded entirely. As a result, a substantial portion of the data is removed during the preparation stage, retaining only 67.68\% of the original dataset. In preprocessing stage, four rule-based steps are implemented to filter the remained data. First, the Data Length step remove overly short texts to ensure that each text contains sufficient informational content. Next, the Character Proportion step eliminate texts with a high percentage of noisy characters, such as English, Traditional Chinese characters, or other irrelevant symbols. Finally, the Sensitive Words step and the Deduplication step are employed to remove toxic content and duplicate texts from the dataset. After the preprocessing stage, we produce a high-quality Chinese text dataset totaling 3.8 TB. In the next stage, each text in this high-quality dataset will be enriched with fine-grained annotations, including a quality score, domain lablels, a toxicity score and a toxicity label.
51
 
52
  <div align="center">
53
+ <img src="./picture/data_statistics.png" width="67%" />
54
  <br>
55
  <em>Figure 1: Description of Image 1</em>
56
  </div>
 
62
  <table>
63
  <tr>
64
  <td>
65
+ <img src="./picture/quality_distribution.pdf" alt="Image 1" style="width:100%;">
66
  <br>
67
  <em>Figure 1: Description of Image 1</em>
68
  </td>
69
  <td>
70
+ <img src="./picture/human_acceptance.pdf" alt="Image 2" style="width:100%;">
71
  <br>
72
  <em>Figure 1: Description of Image 1</em>
73
  </td>