Spaces:
Sleeping
Sleeping
Commit
·
ff8bdc6
1
Parent(s):
d28e427
feat: add api endpoint
Browse files
app.py
CHANGED
|
@@ -14,8 +14,8 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
| 14 |
|
| 15 |
classifier = pipeline("text-classification", model="saiteki-kai/QA-DeBERTa-v3-large")
|
| 16 |
|
| 17 |
-
@spaces.GPU(
|
| 18 |
-
def generate(prompts):
|
| 19 |
messages = [[{"role": "user", "content": message}] for message in prompts]
|
| 20 |
|
| 21 |
texts = tokenizer.apply_chat_template(
|
|
@@ -37,6 +37,8 @@ def generate(prompts):
|
|
| 37 |
|
| 38 |
return responses, classifier([text + "[SEP]" + response for text, response in zip(texts, responses)])
|
| 39 |
|
|
|
|
|
|
|
|
|
|
| 40 |
|
| 41 |
-
demo = gr.Interface(fn=generate, inputs=gr.Text(), outputs=gr.Text())
|
| 42 |
demo.launch()
|
|
|
|
| 14 |
|
| 15 |
classifier = pipeline("text-classification", model="saiteki-kai/QA-DeBERTa-v3-large")
|
| 16 |
|
| 17 |
+
@spaces.GPU()
|
| 18 |
+
def generate(prompts: list[str]) -> tuple[list[str], list[dict[str, float]]]:
|
| 19 |
messages = [[{"role": "user", "content": message}] for message in prompts]
|
| 20 |
|
| 21 |
texts = tokenizer.apply_chat_template(
|
|
|
|
| 37 |
|
| 38 |
return responses, classifier([text + "[SEP]" + response for text, response in zip(texts, responses)])
|
| 39 |
|
| 40 |
+
with gr.Blocks() as demo:
|
| 41 |
+
gr.Markdown("Welcome")
|
| 42 |
+
gr.api(generate, api_name="predict")
|
| 43 |
|
|
|
|
| 44 |
demo.launch()
|