| import glob |
| import os |
| from pathlib import Path |
|
|
| import datasets |
|
|
|
|
| _CITATION = """\ |
| @article{boonkwan2020annotation, |
| title={The Annotation Guideline of LST20 Corpus}, |
| author={Boonkwan, Prachya and Luantangsrisuk, Vorapon and Phaholphinyo, Sitthaa and Kriengket, Kanyanat and Leenoi, Dhanon and Phrombut, Charun and Boriboon, Monthika and Kosawat, Krit and Supnithi, Thepchai}, |
| journal={arXiv preprint arXiv:2008.05055}, |
| year={2020} |
| } |
| """ |
|
|
| _DESCRIPTION = """\ |
| LST20 Corpus is a dataset for Thai language processing developed by National Electronics and Computer Technology Center (NECTEC), Thailand. |
| It offers five layers of linguistic annotation: word boundaries, POS tagging, named entities, clause boundaries, and sentence boundaries. |
| At a large scale, it consists of 3,164,002 words, 288,020 named entities, 248,181 clauses, and 74,180 sentences, while it is annotated with |
| 16 distinct POS tags. All 3,745 documents are also annotated with one of 15 news genres. Regarding its sheer size, this dataset is |
| considered large enough for developing joint neural models for NLP. |
| Manually download at https://aiforthai.in.th/corpus.php |
| """ |
|
|
|
|
| class Lst20Config(datasets.BuilderConfig): |
| """BuilderConfig for Lst20""" |
|
|
| def __init__(self, **kwargs): |
| """BuilderConfig for Lst20. |
| |
| Args: |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(Lst20Config, self).__init__(**kwargs) |
|
|
|
|
| class Lst20(datasets.GeneratorBasedBuilder): |
| """Lst20 dataset.""" |
|
|
| _SENTENCE_SPLITTERS = ["", " ", "\n"] |
| _TRAINING_FOLDER = "train" |
| _VALID_FOLDER = "eval" |
| _TEST_FOLDER = "test" |
| _POS_TAGS = ["NN", "VV", "PU", "CC", "PS", "AX", "AV", "FX", "NU", "AJ", "CL", "PR", "NG", "PA", "XX", "IJ"] |
| _NER_TAGS = [ |
| "O", |
| "B_BRN", |
| "B_DES", |
| "B_DTM", |
| "B_LOC", |
| "B_MEA", |
| "B_NUM", |
| "B_ORG", |
| "B_PER", |
| "B_TRM", |
| "B_TTL", |
| "I_BRN", |
| "I_DES", |
| "I_DTM", |
| "I_LOC", |
| "I_MEA", |
| "I_NUM", |
| "I_ORG", |
| "I_PER", |
| "I_TRM", |
| "I_TTL", |
| "E_BRN", |
| "E_DES", |
| "E_DTM", |
| "E_LOC", |
| "E_MEA", |
| "E_NUM", |
| "E_ORG", |
| "E_PER", |
| "E_TRM", |
| "E_TTL", |
| ] |
| _CLAUSE_TAGS = ["O", "B_CLS", "I_CLS", "E_CLS"] |
|
|
| BUILDER_CONFIGS = [ |
| Lst20Config(name="lst20", version=datasets.Version("1.0.0"), description="LST20 dataset"), |
| ] |
|
|
| @property |
| def manual_download_instructions(self): |
| return """\ |
| You need to |
| 1. Manually download `AIFORTHAI-LST20Corpus.tar.gz` from https://aiforthai.in.th/corpus.php (login required; website mostly in Thai) |
| 2. Extract the .tar.gz; this will result in folder `LST20Corpus` |
| The <path/to/folder> can e.g. be `~/Downloads/LST20Corpus`. |
| lst20 can then be loaded using the following command `datasets.load_dataset("lst20", data_dir="<path/to/folder>")`. |
| """ |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "id": datasets.Value("string"), |
| "fname": datasets.Value("string"), |
| "tokens": datasets.Sequence(datasets.Value("string")), |
| "pos_tags": datasets.Sequence(datasets.features.ClassLabel(names=self._POS_TAGS)), |
| "ner_tags": datasets.Sequence(datasets.features.ClassLabel(names=self._NER_TAGS)), |
| "clause_tags": datasets.Sequence(datasets.features.ClassLabel(names=self._CLAUSE_TAGS)), |
| } |
| ), |
| supervised_keys=None, |
| homepage="https://aiforthai.in.th/", |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| """Returns SplitGenerators.""" |
|
|
| data_dir = os.path.abspath(os.path.expanduser(dl_manager.manual_dir)) |
|
|
| |
| if not os.path.exists(data_dir): |
| raise FileNotFoundError( |
| f"{data_dir} does not exist. Make sure you insert a manual dir via `datasetts.load_dataset('lst20', data_dir=...)`. Manual download instructions: {self.manual_download_instructions})" |
| ) |
|
|
| |
| nb_train = len(glob.glob(os.path.join(data_dir, "train", "*.txt"))) |
| nb_valid = len(glob.glob(os.path.join(data_dir, "eval", "*.txt"))) |
| nb_test = len(glob.glob(os.path.join(data_dir, "test", "*.txt"))) |
| assert ( |
| nb_train > 0 |
| ), f"No files found in train/*.txt.\nManual download instructions:{self.manual_download_instructions})" |
| assert ( |
| nb_valid > 0 |
| ), f"No files found in eval/*.txt.\nManual download instructions:{self.manual_download_instructions})" |
| assert ( |
| nb_test > 0 |
| ), f"No files found in test/*.txt.\nManual download instructions:{self.manual_download_instructions})" |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={"filepath": os.path.join(data_dir, self._TRAINING_FOLDER)}, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.VALIDATION, |
| gen_kwargs={"filepath": os.path.join(data_dir, self._VALID_FOLDER)}, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| gen_kwargs={"filepath": os.path.join(data_dir, self._TEST_FOLDER)}, |
| ), |
| ] |
|
|
| def _generate_examples(self, filepath): |
| for file_idx, fname in enumerate(sorted(glob.glob(os.path.join(filepath, "*.txt")))): |
| with open(fname, encoding="utf-8") as f: |
| guid = 0 |
| tokens = [] |
| pos_tags = [] |
| ner_tags = [] |
| clause_tags = [] |
|
|
| for line in f: |
| if line in self._SENTENCE_SPLITTERS: |
| if tokens: |
| yield f"{file_idx}_{guid}", { |
| "id": str(guid), |
| "fname": Path(fname).name, |
| "tokens": tokens, |
| "pos_tags": pos_tags, |
| "ner_tags": ner_tags, |
| "clause_tags": clause_tags, |
| } |
| guid += 1 |
| tokens = [] |
| pos_tags = [] |
| ner_tags = [] |
| clause_tags = [] |
| else: |
| |
| splits = line.split("\t") |
| |
| ner_tag = splits[2] if splits[2] in self._NER_TAGS else "O" |
| tokens.append(splits[0]) |
| pos_tags.append(splits[1]) |
| ner_tags.append(ner_tag) |
| clause_tags.append(splits[3].rstrip()) |
| |
| if tokens: |
| yield f"{file_idx}_{guid}", { |
| "id": str(guid), |
| "fname": Path(fname).name, |
| "tokens": tokens, |
| "pos_tags": pos_tags, |
| "ner_tags": ner_tags, |
| "clause_tags": clause_tags, |
| } |
|
|