|
|
from typing import List, Tuple |
|
|
import json |
|
|
import os.path as osp |
|
|
import os |
|
|
|
|
|
HALF2FULL = {i: i + 0xFEE0 for i in range(0x21, 0x7F)} |
|
|
HALF2FULL[0x20] = 0x3000 |
|
|
|
|
|
FULL2HALF = dict((i + 0xFEE0, i) for i in range(0x21, 0x7F)) |
|
|
FULL2HALF[0x3000] = 0x20 |
|
|
FULL2HALF[0x3002] = 0x2E |
|
|
|
|
|
LANGSET_CJK = {'简体中文', '繁體中文', '日本語'} |
|
|
LANGSET_CH = {'简体中文', '繁體中文'} |
|
|
|
|
|
PUNSET_RIGHT_ENG = {'.', '?', '!', ':', ';', ')', '}', "\""} |
|
|
PUNCTUATION_L = {'「', '『', '【', '《', '〈', '〔', '[', '{', '(', '(', '[', '{', '“', '‘'} |
|
|
|
|
|
PKUSEG_PUNCSET = {' ', '.', ' '} |
|
|
PKUSEGPATH = r'data/pkusegscores.json' |
|
|
PKUSEGSCORES = None |
|
|
CHSEG = None |
|
|
|
|
|
def full_len(s: str): |
|
|
""" |
|
|
Convert all ASCII characters to their full-width counterpart. |
|
|
https://stackoverflow.com/questions/2422177/python-how-can-i-replace-full-width-characters-with-half-width-characters |
|
|
""" |
|
|
return s.translate(HALF2FULL) |
|
|
|
|
|
def half_len(s): |
|
|
''' |
|
|
Convert full-width characters to ASCII counterpart |
|
|
''' |
|
|
return s.translate(FULL2HALF) |
|
|
|
|
|
def seg_to_chars(text: str) -> List[str]: |
|
|
text = text.replace('\n', '') |
|
|
return [c for c in text] |
|
|
|
|
|
def seg_eng(text: str) -> List[str]: |
|
|
text = text.replace(' ', ' ').replace(' .', '.').replace('\n', ' ') |
|
|
processed_text = '' |
|
|
|
|
|
|
|
|
text_len = len(text) |
|
|
for ii, c in enumerate(text): |
|
|
if c in PUNSET_RIGHT_ENG and ii < text_len - 1: |
|
|
next_c = text[ii + 1] |
|
|
if next_c.isalpha() or next_c.isnumeric(): |
|
|
processed_text += c + ' ' |
|
|
else: |
|
|
processed_text += c |
|
|
else: |
|
|
processed_text += c |
|
|
|
|
|
word_list = processed_text.split(' ') |
|
|
word_num = len(word_list) |
|
|
if word_num <= 1: |
|
|
return word_list |
|
|
|
|
|
words = [] |
|
|
skip_next = False |
|
|
for ii, word in enumerate(word_list): |
|
|
if skip_next: |
|
|
skip_next = False |
|
|
continue |
|
|
if len(word) < 3: |
|
|
append_left, append_right = False, False |
|
|
len_word, len_next, len_prev = len(word), -1, -1 |
|
|
if ii < word_num - 1: |
|
|
len_next = len(word_list[ii + 1]) |
|
|
if ii > 0: |
|
|
len_prev = len(words[-1]) |
|
|
cond_next = (len_word == 2 and len_next <= 4) or len_word == 1 |
|
|
cond_prev = (len_word == 2 and len_prev <= 4) or len_word == 1 |
|
|
if len_next > 0 and len_prev > 0: |
|
|
if len_next < len_prev: |
|
|
append_right = cond_next |
|
|
else: |
|
|
append_left = cond_prev |
|
|
elif len_next > 0: |
|
|
append_right = cond_next |
|
|
elif len_prev > 0: |
|
|
append_left = cond_prev |
|
|
|
|
|
if append_left: |
|
|
words[-1] = words[-1] + ' ' + word |
|
|
elif append_right: |
|
|
words.append(word + ' ' + word_list[ii + 1]) |
|
|
skip_next = True |
|
|
else: |
|
|
words.append(word) |
|
|
continue |
|
|
words.append(word) |
|
|
return words |
|
|
|
|
|
def _seg_ch_pkg(text: str) -> List[str]: |
|
|
|
|
|
if text == ' ': |
|
|
return [' '] |
|
|
elif text == '': |
|
|
return [] |
|
|
|
|
|
segments = CHSEG.cut(text) |
|
|
num_segments = len(segments) |
|
|
if num_segments == 0: |
|
|
return [] |
|
|
if num_segments == 1: |
|
|
return [segments[0][0]] |
|
|
|
|
|
words = [] |
|
|
tags = [] |
|
|
max_concat_len = 4 |
|
|
skip_next = False |
|
|
try: |
|
|
for ii, (word, tag) in enumerate(segments): |
|
|
if skip_next: |
|
|
skip_next = False |
|
|
continue |
|
|
|
|
|
len_word, len_next, len_prev = len(word), -1, -1 |
|
|
next_valid, prev_valid = False, False |
|
|
word_next, tag_next = '', '' |
|
|
word_prev, tag_prev = '', '' |
|
|
score_next, score_prev = 0, 0 |
|
|
if ii < num_segments - 1: |
|
|
word_next, tag_next = segments[ii + 1] |
|
|
len_next = len(word_next) |
|
|
next_valid = True |
|
|
if tag_next != 'w' and not word_next in PKUSEG_PUNCSET: |
|
|
score_next = PKUSEGSCORES[tag][tag_next] |
|
|
|
|
|
if ii > 0: |
|
|
word_prev, tag_prev = words[-1], segments[ii - 1][1] |
|
|
len_prev = len(word_prev) |
|
|
prev_valid = True |
|
|
if tag_prev != 'w' and not word_prev[-1] in PKUSEG_PUNCSET: |
|
|
score_prev = PKUSEGSCORES[tag_prev][tag] |
|
|
|
|
|
append_prev, append_next = False, False |
|
|
|
|
|
if tag == 'w' or word in PKUSEG_PUNCSET: |
|
|
if word in PUNCTUATION_L: |
|
|
append_next = next_valid |
|
|
elif len_word <= 1: |
|
|
append_prev = prev_valid |
|
|
else: |
|
|
next_valid = score_next > 0 and len_next < max_concat_len |
|
|
prev_valid = score_prev > 0 and len_prev < max_concat_len |
|
|
need_concat = len_word < max_concat_len |
|
|
append_prev = score_prev == 1 |
|
|
append_next = score_next == 1 |
|
|
if score_prev != 1 and score_next != 1 and need_concat: |
|
|
append_prev = prev_valid |
|
|
append_next = next_valid |
|
|
if append_next and append_prev: |
|
|
if len_prev == len_next: |
|
|
if score_prev >= score_next: |
|
|
append_next = False |
|
|
else: |
|
|
append_prev = False |
|
|
elif len_prev < len_next: |
|
|
append_next = False |
|
|
else: |
|
|
append_prev = False |
|
|
|
|
|
if append_next and append_prev: |
|
|
words[-1] = word_prev + word + word_next |
|
|
tags[-1] = tags[-1] + [tag, tag_next] |
|
|
skip_next = True |
|
|
elif append_prev: |
|
|
words[-1] = words[-1] + word |
|
|
tags[-1].append(tag) |
|
|
elif append_next: |
|
|
words.append(word + word_next) |
|
|
tags.append([tag, tag_next]) |
|
|
skip_next = True |
|
|
else: |
|
|
words.append(word) |
|
|
tags.append([tag]) |
|
|
except Exception as e: |
|
|
print('exp at line: ', text) |
|
|
raise e |
|
|
return words |
|
|
|
|
|
def seg_ch_pkg(text: str): |
|
|
|
|
|
global CHSEG |
|
|
if CHSEG is None: |
|
|
try: |
|
|
import pkuseg |
|
|
except: |
|
|
import spacy_pkuseg as pkuseg |
|
|
CHSEG = pkuseg.pkuseg(postag=True) |
|
|
|
|
|
|
|
|
fullen_text = full_len(text).replace(' ', ' ') |
|
|
cvt_back = False |
|
|
if fullen_text != text: |
|
|
cvt_back = True |
|
|
text = fullen_text |
|
|
|
|
|
global PKUSEGSCORES |
|
|
if PKUSEGSCORES is None: |
|
|
with open(PKUSEGPATH, 'r', encoding='utf8') as f: |
|
|
PKUSEGSCORES = json.loads(f.read()) |
|
|
|
|
|
text_list = text.replace('\n', '').replace(' ', ' ').split(' ') |
|
|
result_list = [] |
|
|
for ii, text in enumerate(text_list): |
|
|
words = None |
|
|
if text: |
|
|
words = _seg_ch_pkg(text) |
|
|
if words is not None: |
|
|
if ii > 0: |
|
|
words[0] = ' ' + words[0] |
|
|
result_list.extend(words) |
|
|
|
|
|
if cvt_back: |
|
|
|
|
|
result_list = [half_len(word) for word in result_list] |
|
|
return result_list |
|
|
|
|
|
def seg_text(text: str, lang: str) -> Tuple[List, str]: |
|
|
delimiter = '' |
|
|
if lang in LANGSET_CH: |
|
|
words = seg_ch_pkg(text) |
|
|
elif lang in LANGSET_CJK: |
|
|
words = seg_to_chars(text) |
|
|
else: |
|
|
words = seg_eng(text) |
|
|
delimiter = ' ' |
|
|
return words, delimiter |
|
|
|
|
|
def is_cjk(lang: str) -> bool: |
|
|
return lang in LANGSET_CJK |