"""Vietnamese sentence tokenization using trained NLTK Punkt model. Usage: from sent_tokenize import sent_tokenize sentences = sent_tokenize("Xin chào. Tôi là Claude.") """ import json from os.path import dirname, join from nltk.tokenize.punkt import PunktParameters, PunktSentenceTokenizer _tokenizer = None def _load_tokenizer(): """Load PunktSentenceTokenizer with trained parameters.""" global _tokenizer if _tokenizer is not None: return params_path = join(dirname(__file__), "punkt_params_trained.json") with open(params_path, encoding="utf-8") as f: data = json.load(f) params = PunktParameters() params.abbrev_types = set(data.get("abbrev_types", [])) params.sent_starters = set(data.get("sent_starters", [])) collocations = set() for colloc_str in data.get("collocations", []): parts = colloc_str.split(" ", 1) if len(parts) == 2: collocations.add(tuple(parts)) params.collocations = collocations _tokenizer = PunktSentenceTokenizer(params) def sent_tokenize(text: str) -> list[str]: """Tokenize Vietnamese text into sentences using trained Punkt model. Args: text: Input Vietnamese text string. Returns: List of sentence strings. """ if not text or not text.strip(): return [] _load_tokenizer() return _tokenizer.tokenize(text)