Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -19,6 +19,12 @@ def generate2(prompt, max_new_tokens):
|
|
| 19 |
generated_ids = model2.generate(batch["input_ids"], max_new_tokens=150)
|
| 20 |
output = tokenizer2.batch_decode(generated_ids, skip_special_tokens=True)
|
| 21 |
return output[0]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
def generate_prompt(type, prompt, max_new_tokens):
|
| 23 |
if type==1:
|
| 24 |
return generate(prompt)
|
|
@@ -29,4 +35,4 @@ input_component = gr.Textbox(label = "Input a persona, e.g. photographer", value
|
|
| 29 |
output_component = gr.Textbox(label = "Prompt")
|
| 30 |
examples = [["photographer"], ["developer"]]
|
| 31 |
description = ""
|
| 32 |
-
gr.Interface(
|
|
|
|
| 19 |
generated_ids = model2.generate(batch["input_ids"], max_new_tokens=150)
|
| 20 |
output = tokenizer2.batch_decode(generated_ids, skip_special_tokens=True)
|
| 21 |
return output[0]
|
| 22 |
+
def generate2_test(prompt):
|
| 23 |
+
batch = tokenizer2(prompt, return_tensors="pt")
|
| 24 |
+
generated_ids = model2.generate(batch["input_ids"], max_new_tokens=150)
|
| 25 |
+
output = tokenizer2.batch_decode(generated_ids, skip_special_tokens=True)
|
| 26 |
+
return output[0]
|
| 27 |
+
|
| 28 |
def generate_prompt(type, prompt, max_new_tokens):
|
| 29 |
if type==1:
|
| 30 |
return generate(prompt)
|
|
|
|
| 35 |
output_component = gr.Textbox(label = "Prompt")
|
| 36 |
examples = [["photographer"], ["developer"]]
|
| 37 |
description = ""
|
| 38 |
+
gr.Interface(generate2_test, inputs = input_component, outputs=output_component, examples=examples, title = "π¨π»βπ€ ChatGPT Prompt Generator v12 π¨π»βπ€", description=description).launch()
|