|
|
import gradio as gr |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from transformers import pipeline |
|
|
|
|
|
pipe = pipeline("text-generation", model="gpt2") |
|
|
|
|
|
|
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained("gpt2") |
|
|
model = AutoModelForCausalLM.from_pretrained("gpt2") |
|
|
|
|
|
|
|
|
text_input = "One upon a time there was a tree" |
|
|
max_length = 100 |
|
|
temperature = 0.8 |
|
|
top_k = 100 |
|
|
|
|
|
input_ids = tokenizer.encode(text_input,return_tensors='pt') |
|
|
|
|
|
output = model.generate(input_ids, max_length=max_length, temperature=temperature, top_k=top_k, do_sample = True) |
|
|
|
|
|
|
|
|
response = tokenizer.decode(output[0], skip_special_token=True) |
|
|
print(response) |