from transformers import AutoModelForCausalLM, AutoTokenizer import torch model = AutoModelForCausalLM.from_pretrained( "Girinath11/recursive-language-model-48m", trust_remote_code=True ) tokenizer = AutoTokenizer.from_pretrained("gpt2") tokenizer.pad_token = tokenizer.eos_token device = "cuda" if torch.cuda.is_available() else "cpu" model = model.to(device) model.eval() print(f"Model loaded on {device}\n") prompts = [ "The future of artificial intelligence", "Once upon a time", "The key to success is" ] for prompt in prompts: print(f"Prompt: {prompt}") inputs = tokenizer.encode(prompt, return_tensors="pt").to(device) with torch.no_grad(): outputs = model.generate( input_ids, max_new_tokens=60, temperature=0.7, top_p=0.9, top_k=50, repetition_penalty=1.2, no_repeat_ngram_size=3, do_sample=True, pad_token_id=tokenizer.eos_token_id ) text = tokenizer.decode(outputs[0], skip_special_tokens=True) print(f"{text}\n")