| from transformers import AutoTokenizer | |
| from retnet.modeling_retnet import RetNetForCausalLM | |
| model = RetNetForCausalLM.from_pretrained("./") | |
| tokenizer = AutoTokenizer.from_pretrained('gpt2') | |
| tokenizer.model_max_length = 16384 | |
| tokenizer.pad_token = tokenizer.eos_token | |
| tokenizer.unk_token = tokenizer.eos_token | |
| tokenizer.bos_token = tokenizer.eos_token | |
| inputs = tokenizer("Hello, my dog is cute and ", return_tensors="pt") | |
| # Generate output with max_length parameter | |
| generation_output = model.generate(**inputs, max_length=50) # Adjust max_length as needed | |
| output = tokenizer.decode(generation_output[0], skip_special_tokens=True) | |
| print(output) |