| |
| |
| |
| |
| """ |
| Thank https://dev.to/ton_ami/text-data-augmentation-synonym-replacement-4h8l |
| """ |
| __all__ = [ |
| "WordNetAug", |
| "postype2wordnet", |
| ] |
|
|
| import itertools |
| from collections import OrderedDict |
| from typing import List |
|
|
| from nltk.corpus import wordnet as wn |
|
|
| from pythainlp.corpus import wordnet |
| from pythainlp.tag import pos_tag |
| from pythainlp.tokenize import word_tokenize |
|
|
| orchid = { |
| "": "", |
| |
| "NOUN": wn.NOUN, |
| "NCMN": wn.NOUN, |
| "NTTL": wn.NOUN, |
| "CNIT": wn.NOUN, |
| "CLTV": wn.NOUN, |
| "CMTR": wn.NOUN, |
| "CFQC": wn.NOUN, |
| "CVBL": wn.NOUN, |
| |
| "VACT": wn.VERB, |
| "VSTA": wn.VERB, |
| |
| "PROPN": "", |
| "NPRP": "", |
| |
| "ADJ": wn.ADJ, |
| "NONM": wn.ADJ, |
| "VATT": wn.ADJ, |
| "DONM": wn.ADJ, |
| |
| "ADV": wn.ADV, |
| "ADVN": wn.ADV, |
| "ADVI": wn.ADV, |
| "ADVP": wn.ADV, |
| "ADVS": wn.ADV, |
| |
| "INT": "", |
| |
| "PRON": "", |
| "PPRS": "", |
| "PDMN": "", |
| "PNTR": "", |
| |
| "DET": "", |
| "DDAN": "", |
| "DDAC": "", |
| "DDBQ": "", |
| "DDAQ": "", |
| "DIAC": "", |
| "DIBQ": "", |
| "DIAQ": "", |
| |
| "NUM": "", |
| "NCNM": "", |
| "NLBL": "", |
| "DCNM": "", |
| |
| "AUX": "", |
| "XVBM": "", |
| "XVAM": "", |
| "XVMM": "", |
| "XVBB": "", |
| "XVAE": "", |
| |
| "ADP": "", |
| "RPRE": "", |
| |
| "CCONJ": "", |
| "JCRG": "", |
| |
| "SCONJ": "", |
| "PREL": "", |
| "JSBR": "", |
| "JCMP": "", |
| |
| "PART": "", |
| "FIXN": "", |
| "FIXV": "", |
| "EAFF": "", |
| "EITT": "", |
| "AITT": "", |
| "NEG": "", |
| |
| "PUNCT": "", |
| "PUNC": "", |
| } |
|
|
|
|
| def postype2wordnet(pos: str, corpus: str): |
| """ |
| Convert part-of-speech type to wordnet type |
| |
| :param str pos: POS type |
| :param str corpus: part-of-speech corpus |
| |
| **Options for corpus** |
| * *orchid* - Orchid Corpus |
| """ |
| if corpus not in ["orchid"]: |
| return None |
| return orchid[pos] |
|
|
|
|
| class WordNetAug: |
| """ |
| Text Augment using wordnet |
| """ |
|
|
| def __init__(self): |
| pass |
|
|
| def find_synonyms( |
| self, word: str, pos: str = None, postag_corpus: str = "orchid" |
| ) -> List[str]: |
| """ |
| Find synonyms using wordnet |
| |
| :param str word: word |
| :param str pos: part-of-speech type |
| :param str postag_corpus: name of POS tag corpus |
| :return: list of synonyms |
| :rtype: List[str] |
| """ |
| self.synonyms = [] |
| if pos is None: |
| self.list_synsets = wordnet.synsets(word) |
| else: |
| self.p2w_pos = postype2wordnet(pos, postag_corpus) |
| if self.p2w_pos != "": |
| self.list_synsets = wordnet.synsets(word, pos=self.p2w_pos) |
| else: |
| self.list_synsets = wordnet.synsets(word) |
|
|
| for self.synset in wordnet.synsets(word): |
| for self.syn in self.synset.lemma_names(lang="tha"): |
| self.synonyms.append(self.syn) |
|
|
| self.synonyms_without_duplicates = list( |
| OrderedDict.fromkeys(self.synonyms) |
| ) |
| return self.synonyms_without_duplicates |
|
|
| def augment( |
| self, |
| sentence: str, |
| tokenize: object = word_tokenize, |
| max_syn_sent: int = 6, |
| postag: bool = True, |
| postag_corpus: str = "orchid", |
| ) -> List[List[str]]: |
| """ |
| Text Augment using wordnet |
| |
| :param str sentence: Thai sentence |
| :param object tokenize: function for tokenizing words |
| :param int max_syn_sent: maximum number of synonymous sentences |
| :param bool postag: use part-of-speech |
| :param str postag_corpus: name of POS tag corpus |
| |
| :return: list of synonyms |
| :rtype: List[Tuple[str]] |
| |
| :Example: |
| :: |
| |
| from pythainlp.augment import WordNetAug |
| |
| aug = WordNetAug() |
| aug.augment("เราชอบไปโรงเรียน") |
| # output: [('เรา', 'ชอบ', 'ไป', 'ร.ร.'), |
| ('เรา', 'ชอบ', 'ไป', 'รร.'), |
| ('เรา', 'ชอบ', 'ไป', 'โรงเรียน'), |
| ('เรา', 'ชอบ', 'ไป', 'อาคารเรียน'), |
| ('เรา', 'ชอบ', 'ไปยัง', 'ร.ร.'), |
| ('เรา', 'ชอบ', 'ไปยัง', 'รร.')] |
| """ |
| new_sentences = [] |
| self.list_words = tokenize(sentence) |
| self.list_synonym = [] |
| self.p_all = 1 |
| if postag: |
| self.list_pos = pos_tag(self.list_words, corpus=postag_corpus) |
| for word, pos in self.list_pos: |
| self.temp = self.find_synonyms(word, pos, postag_corpus) |
| if not self.temp: |
| self.list_synonym.append([word]) |
| else: |
| self.list_synonym.append(self.temp) |
| self.p_all *= len(self.temp) |
| else: |
| for word in self.list_words: |
| self.temp = self.find_synonyms(word) |
| if not self.temp: |
| self.list_synonym.append([word]) |
| else: |
| self.list_synonym.append(self.temp) |
| self.p_all *= len(self.temp) |
| if max_syn_sent > self.p_all: |
| max_syn_sent = self.p_all |
| for x in list(itertools.product(*self.list_synonym))[0:max_syn_sent]: |
| new_sentences.append(x) |
| return new_sentences |
|
|