ViLexNorm / README.md
AnnyNguyen's picture
Create README.md
bde58f9 verified
---
task_categories:
- token-classification
language:
- vi
---
## Dataset Card for ViLexNorm
### 1. Dataset Summary
**ViLexNorm** is a Vietnamese lexical normalization corpus of **10,467** comment pairs, each comprising an *original* noisy social‑media comment and its *normalized* counterpart. In this unified version, all pairs are merged into one CSV with a `type` column indicating `train` / `dev` / `test`, and two additional columns `input`/`output` containing the tokenized forms.
### 2. Supported Tasks and Metrics
* **Primary Task**: Sequence‑to‑sequence lexical normalization
* **Metric**:
* **Error Reduction Rate (ERR)** (van der Goot 2019)
* **Token‑level accuracy**
### 3. Languages
* Vietnamese
### 4. Dataset Structure
| Column | Type | Description |
| ------------ | ------ | -------------------------------------------- |
| `original` | string | The raw, unnormalized comment. |
| `normalized` | string | The corrected, normalized comment. |
| `input` | list | Tokenized original text (list of strings). |
| `output` | list | Tokenized normalized text (list of strings). |
| `type` | string | Split: `train` / `validation` / `test`. |
| `dataset` | string | Always `ViLexNorm` for provenance. |
### 5. Data Fields
* **original** (`str`): Noisy input sentence.
* **normalized** (`str`): Human‑annotated normalized sentence.
* **input** (`List[str]`): Token list of `original`.
* **output** (`List[str]`): Token list of `normalized`.
* **type** (`str`): Which split the example belongs to.
* **dataset** (`str`): Always `ViLexNorm`.
### 6. Usage
```python
from datasets import load_dataset
ds = load_dataset("visolex/ViLexNorm")
train = ds.filter(lambda ex: ex["type"] == "train")
val = ds.filter(lambda ex: ex["type"] == "dev")
test = ds.filter(lambda ex: ex["type"] == "test")
print(train[0])
```
### 7. Source & Links
* **Paper**: Nguyen et al. (2024), “ViLexNorm: A Lexical Normalization Corpus for Vietnamese Social Media Text”
[https://aclanthology.org/2024.eacl-long.85](https://aclanthology.org/2024.eacl-long.85)
* **Hugging Face** (this unified version):
[https://huggingface.co/datasets/visolex/ViLexNorm](https://huggingface.co/datasets/visolex/ViLexNorm)
* **Original GitHub** (if available):
[https://github.com/ngxtnhi/ViLexNorm](https://github.com/ngxtnhi/ViLexNorm)
### 8. Contact Information
* **Ms. Thanh‑Nhi Nguyen**: [21521232@gm.uit.edu.vn](mailto:21521232@gm.uit.edu.vn)
* **Mr. Thanh‑Phong Le**: [21520395@gm.uit.edu.vn](mailto:21520395@gm.uit.edu.vn)
### 9. Licensing and Citation
#### License
Released under **CC BY‑NC‑SA 4.0** (Creative Commons Attribution‑NonCommercial‑ShareAlike 4.0 International).
#### How to Cite
```bibtex
@inproceedings{nguyen-etal-2024-vilexnorm,
title = {ViLexNorm: A Lexical Normalization Corpus for Vietnamese Social Media Text},
author = {Nguyen, Thanh-Nhi and Le, Thanh-Phong and Nguyen, Kiet},
booktitle = {Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)},
month = mar,
year = {2024},
address = {St. Julian's, Malta},
publisher = {Association for Computational Linguistics},
url = {https://aclanthology.org/2024.eacl-long.85},
pages = {1421--1437}
}
```