Spaces:
No application file
No application file
| from transformers import GPT2Tokenizer, GPT2Model | |
| # Load the tokenizer and model | |
| tokenizer = GPT2Tokenizer.from_pretrained('gpt2') | |
| model = GPT2Model.from_pretrained('gpt2') | |
| # Create a prompt | |
| prompt = "I need a hug." | |
| # Tokenize the prompt | |
| inputs = tokenizer.encode(prompt, return_tensors='pt') | |
| # Generate text using the model | |
| outputs = model.generate(inputs, max_length=50, num_return_sequences=1) | |
| # Decode the generated text | |
| generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True) | |
| # Print the generated text | |
| print(generated_text) |