from transformers import AutoTokenizer from hanuman_loader import HanumanModel def run(prompt: str = "สวัสดีครับ ช่วยสรุปประเทศไทยสั้น ๆ"): repo_id = "ZombitX64/GPT4All-Model" tokenizer = AutoTokenizer.from_pretrained(repo_id) model_wrapper = HanumanModel.from_pretrained(repo_id, map_location='cpu') model = model_wrapper.model inputs = tokenizer(prompt, return_tensors='pt') out = model.generate(inputs['input_ids'], max_new_tokens=50, temperature=1.2, top_k=50, top_p=0.95) print(tokenizer.decode(out[0], skip_special_tokens=True)) if __name__ == '__main__': run()