# -*- coding: utf-8 -*- # SPDX-FileCopyrightText: 2016-2025 PyThaiNLP Project # SPDX-FileType: SOURCE # SPDX-License-Identifier: Apache-2.0 """ Tokenizers at different levels of linguistic analysis. """ __all__ = [ "THAI2FIT_TOKENIZER", "Tokenizer", "Trie", "paragraph_tokenize", "sent_tokenize", "subword_tokenize", "syllable_tokenize", "word_detokenize", "word_tokenize", ] from pythainlp.corpus import thai_syllables, thai_words from pythainlp.util.trie import Trie DEFAULT_WORD_TOKENIZE_ENGINE = "newmm" DEFAULT_SENT_TOKENIZE_ENGINE = "crfcut" DEFAULT_SUBWORD_TOKENIZE_ENGINE = "tcc" DEFAULT_SYLLABLE_TOKENIZE_ENGINE = "han_solo" DEFAULT_WORD_DICT_TRIE = Trie(thai_words()) DEFAULT_SYLLABLE_DICT_TRIE = Trie(thai_syllables()) DEFAULT_DICT_TRIE = DEFAULT_WORD_DICT_TRIE from pythainlp.tokenize.core import ( Tokenizer, paragraph_tokenize, sent_tokenize, subword_tokenize, syllable_tokenize, word_detokenize, word_tokenize, ) from pythainlp.corpus import get_corpus as _get_corpus THAI2FIT_TOKENIZER = Tokenizer( custom_dict=_get_corpus("words_th_thai2fit_201810.txt"), engine="mm" )