Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
| import re | |
| def japanese_cleaners(text): | |
| from text.japanese import japanese_to_romaji_with_accent | |
| text = japanese_to_romaji_with_accent(text) | |
| if len(text) == 0 or re.match('[A-Za-z]', text[-1]): | |
| text += '.' | |
| return text | |
| def japanese_cleaners2(text): | |
| return japanese_cleaners(text).replace('ts', 'ʦ').replace('...', '…') | |
| def korean_cleaners(text): | |
| '''Pipeline for Korean text''' | |
| from text.korean import latin_to_hangul, number_to_hangul, divide_hangul | |
| text = latin_to_hangul(text) | |
| text = number_to_hangul(text) | |
| text = divide_hangul(text) | |
| if len(text) == 0 or re.match('[\u3131-\u3163]', text[-1]): | |
| text += '.' | |
| return text | |
| def chinese_cleaners(text): | |
| '''Pipeline for Chinese text''' | |
| from text.mandarin import number_to_chinese, chinese_to_bopomofo, latin_to_bopomofo | |
| text = number_to_chinese(text) | |
| text = chinese_to_bopomofo(text) | |
| text = latin_to_bopomofo(text) | |
| if len(text) == 0 or re.match('[ˉˊˇˋ˙]', text[-1]): | |
| text += '。' | |
| return text | |
| def zh_ja_mixture_cleaners(text): | |
| from text.mandarin import chinese_to_romaji | |
| from text.japanese import japanese_to_romaji_with_accent | |
| chinese_texts = re.findall(r'\[ZH\].*?\[ZH\]', text) | |
| japanese_texts = re.findall(r'\[JA\].*?\[JA\]', text) | |
| for chinese_text in chinese_texts: | |
| cleaned_text = chinese_to_romaji(chinese_text[4:-4]) | |
| text = text.replace(chinese_text, cleaned_text + ' ', 1) | |
| for japanese_text in japanese_texts: | |
| cleaned_text = japanese_to_romaji_with_accent( | |
| japanese_text[4:-4]).replace('ts', 'ʦ').replace('u', 'ɯ').replace('...', '…') | |
| text = text.replace(japanese_text, cleaned_text + ' ', 1) | |
| text = text[:-1] | |
| if len(text) == 0 or re.match('[A-Za-zɯɹəɥ→↓↑]', text[-1]): | |
| text += '.' | |
| return text | |
| def sanskrit_cleaners(text): | |
| text = text.replace('॥', '।').replace('ॐ', 'ओम्') | |
| if len(text) == 0 or text[-1] != '।': | |
| text += ' ।' | |
| return text | |
| def cjks_cleaners(text): | |
| from text.mandarin import chinese_to_lazy_ipa | |
| from text.japanese import japanese_to_ipa | |
| from text.korean import korean_to_lazy_ipa | |
| from text.sanskrit import devanagari_to_ipa | |
| chinese_texts = re.findall(r'\[ZH\].*?\[ZH\]', text) | |
| japanese_texts = re.findall(r'\[JA\].*?\[JA\]', text) | |
| korean_texts = re.findall(r'\[KO\].*?\[KO\]', text) | |
| sanskrit_texts = re.findall(r'\[SA\].*?\[SA\]', text) | |
| for chinese_text in chinese_texts: | |
| cleaned_text = chinese_to_lazy_ipa(chinese_text[4:-4]) | |
| text = text.replace(chinese_text, cleaned_text + ' ', 1) | |
| for japanese_text in japanese_texts: | |
| cleaned_text = japanese_to_ipa(japanese_text[4:-4]) | |
| text = text.replace(japanese_text, cleaned_text + ' ', 1) | |
| for korean_text in korean_texts: | |
| cleaned_text = korean_to_lazy_ipa(korean_text[4:-4]) | |
| text = text.replace(korean_text, cleaned_text + ' ', 1) | |
| for sanskrit_text in sanskrit_texts: | |
| cleaned_text = devanagari_to_ipa(sanskrit_text[4:-4]) | |
| text = text.replace(sanskrit_text, cleaned_text + ' ', 1) | |
| text = text[:-1] | |
| if len(text) == 0 or re.match(r'[^\.,!\?\-…~]', text[-1]): | |
| text += '.' | |
| return text | |