from transformers import GPT2LMHeadModel, GPT2TokenizerFast MODEL_DIR = "./68h" tokenizer = GPT2TokenizerFast.from_pretrained(MODEL_DIR) model = GPT2LMHeadModel.from_pretrained(MODEL_DIR) while True: prompt = input(">>> ") if prompt == "exit": break inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate( **inputs, max_new_tokens=150, do_sample=True, temperature=0.8 ) print(tokenizer.decode(outputs[0], skip_special_tokens=True))