|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: correct_text |
|
|
dtype: string |
|
|
- name: error_text |
|
|
dtype: string |
|
|
- name: __index_level_0__ |
|
|
dtype: int64 |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 895314409 |
|
|
num_examples: 3175684 |
|
|
- name: test |
|
|
num_bytes: 111815675 |
|
|
num_examples: 396961 |
|
|
- name: validation |
|
|
num_bytes: 112054245 |
|
|
num_examples: 396961 |
|
|
download_size: 742034799 |
|
|
dataset_size: 1119184329 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
- split: test |
|
|
path: data/test-* |
|
|
- split: validation |
|
|
path: data/validation-* |
|
|
license: mit |
|
|
task_categories: |
|
|
- fill-mask |
|
|
- text-generation |
|
|
- token-classification |
|
|
language: |
|
|
- vi |
|
|
--- |
|
|
|
|
|
## Dataset Card for Vietnamese Text Correction Dataset |
|
|
|
|
|
## Dataset Description |
|
|
|
|
|
This dataset contains Vietnamese text pairs for training and evaluating text correction models. Each example consists of an erroneous text and its corrected version, making it ideal for: |
|
|
|
|
|
- **Grammar correction** |
|
|
- **Spelling correction** |
|
|
- **Text normalization** |
|
|
- **Language model fine-tuning** |
|
|
|
|
|
### Dataset Summary |
|
|
|
|
|
- **Language**: Vietnamese (vi) |
|
|
- **Format**: Text correction pairs |
|
|
- **Size**: ~4.0M examples across train/validation/test splits |
|
|
- **Domain**: General Vietnamese text from various sources |
|
|
- **Collection Method**: Automated collection and human verification |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
### Data Instances |
|
|
|
|
|
A typical data point looks like this: |
|
|
|
|
|
```json |
|
|
{ |
|
|
"correct_text": "Đây là một câu tiếng Việt chuẩn xác.", |
|
|
"error_text": "Đây là một câu tieengs Việt chuẩn xác.", |
|
|
"__index_level_0__": 42 |
|
|
} |
|
|
``` |
|
|
|
|
|
### Data Fields |
|
|
|
|
|
- **correct_text**: The corrected Vietnamese text (string) |
|
|
- **error_text**: The original text with errors (string) |
|
|
|
|
|
### Data Splits |
|
|
|
|
|
| Split | Examples | Size | |
|
|
|-------|----------|------| |
|
|
| **train** | 3,175,684 | 895 MB | |
|
|
| **test** | 396,961 | 112 MB | |
|
|
| **validation** | 396,961 | 112 MB | |
|
|
|
|
|
## Dataset Creation |
|
|
|
|
|
### Source Data |
|
|
|
|
|
The dataset was created by collecting Vietnamese text from multiple sources including: |
|
|
- Web crawling of Vietnamese websites |
|
|
- Social media posts |
|
|
- News articles |
|
|
- User-generated content |
|
|
|
|
|
### Processing Steps |
|
|
|
|
|
1. **Text Collection**: Gathered raw Vietnamese text from diverse sources |
|
|
2. **Error Injection**: Applied various error types (spelling, grammar, typos) |
|
|
3. **Human Annotation**: Native speakers corrected the erroneous texts |
|
|
4. **Quality Filtering**: Removed low-quality or nonsensical examples |
|
|
5. **Deduplication**: Ensured no duplicate text pairs |
|
|
|
|
|
## Usage |
|
|
|
|
|
### Loading the Dataset |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
# Load the dataset |
|
|
dataset = load_dataset("8Opt/vn-text-correction-0001") |
|
|
|
|
|
# Access different splits |
|
|
train_data = dataset['train'] |
|
|
test_data = dataset['test'] |
|
|
val_data = dataset['validation'] |
|
|
|
|
|
# Example usage |
|
|
example = train_data[0] |
|
|
print(f"Original: {example['error_text']}") |
|
|
print(f"Corrected: {example['correct_text']}") |
|
|
``` |
|
|
|
|
|
### For Training |
|
|
|
|
|
```python |
|
|
# Prepare for training a correction model |
|
|
def preprocess_function(examples): |
|
|
inputs = [f"Correct this text: {err}" for err in examples['error_text']] |
|
|
targets = examples['correct_text'] |
|
|
return {'input_text': inputs, 'target_text': targets} |
|
|
|
|
|
# Apply preprocessing |
|
|
tokenized_datasets = dataset.map(preprocess_function, batched=True) |
|
|
``` |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
### Metrics |
|
|
|
|
|
Common evaluation metrics for this dataset include: |
|
|
- **BLEU Score**: Measures n-gram overlap with reference corrections |
|
|
- **ROUGE Score**: Evaluates summary-quality corrections |
|
|
- **Character/Word Error Rate**: Traditional correction metrics |
|
|
- **Human Evaluation**: Manual assessment of correction quality |
|
|
|
|
|
## Limitations and Bias |
|
|
|
|
|
- **Domain Coverage**: Dataset may not cover all Vietnamese domains equally |
|
|
- **Error Types**: Focuses on common errors; rare error types may be underrepresented |
|
|
- **Regional Variations**: Northern/Central/Southern Vietnamese differences may exist |
|
|
- **Contemporary Usage**: May not capture very recent slang or terminology |
|
|
|
|
|
## Ethical Considerations |
|
|
|
|
|
- All text data is publicly available Vietnamese content |
|
|
- No personally identifiable information included |
|
|
- Dataset intended for research and educational purposes only |
|
|
|
|
|
|
|
|
## Contributing |
|
|
|
|
|
We welcome contributions! If you find issues or want to improve the dataset: |
|
|
|
|
|
1. Open an issue on the https://huggingface.co/datasets/8Opt/vn-text-correction-0001. |
|
|
2. Submit a pull request with improvements |
|
|
3. Report any data quality issues or annotation errors |
|
|
|
|
|
## Contact |
|
|
|
|
|
For questions or feedback about this dataset, please contact: [minh.leduc.0210@gmail.com] |