Spaces:
Runtime error
Runtime error
| from transformers import AutoTokenizer, AutoModelForCausalLM | |
| model_name = 'armandnlp/gpt2-TOD_finetuned_SGD' | |
| tokenizer_TOD = AutoTokenizer.from_pretrained(model_name) | |
| model_TOD = AutoModelForCausalLM.from_pretrained(model_name) | |
| def generate_response(prompt): | |
| input_ids = tokenizer_TOD(prompt, return_tensors="pt").input_ids | |
| outputs = model_TOD.generate(input_ids, | |
| do_sample=False, | |
| max_length=1024, | |
| eos_token_id=50262) | |
| return tokenizer_TOD.batch_decode(outputs)[0] | |
| #<|context|> <|user|> I want to go to the restaurant.<|endofcontext|> | |
| iface = gr.Interface(fn=generate_response, | |
| inputs="text", | |
| outputs="text", | |
| title="gpt2-TOD", | |
| examples=[["<|context|> <|user|> I'm super hungry ! I want to go to the restaurant.<|endofcontext|>"], | |
| "<|context|> <|user|> I want to go to the restaurant.\ | |
| <|system|> What food would you like to eat ? <|user|> Italian sounds good. <|endofcontext|>"], | |
| description="Passing in a task-oriented dialogue context generates a belief state, actions to take and a response based on those actions", | |
| ) | |
| iface.launch() | |