Spaces:
Sleeping
Sleeping
Commit
·
eec20e0
1
Parent(s):
31ab39c
fix: remove input tokens
Browse files
app.py
CHANGED
|
@@ -16,7 +16,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
| 16 |
classifier = pipeline("text-classification", model="saiteki-kai/QA-DeBERTa-v3-large")
|
| 17 |
|
| 18 |
@spaces.GPU()
|
| 19 |
-
def generate(prompts: list[str]) -> tuple[
|
| 20 |
messages = [[{"role": "user", "content": message}] for message in prompts]
|
| 21 |
|
| 22 |
texts = tokenizer.apply_chat_template(
|
|
@@ -32,12 +32,14 @@ def generate(prompts: list[str]) -> tuple[list[str], list[dict[str, float]]]:
|
|
| 32 |
repetition_penalty=1.0,
|
| 33 |
max_new_tokens=512,
|
| 34 |
)
|
|
|
|
|
|
|
| 35 |
generated_ids = [
|
| 36 |
-
output_ids[
|
| 37 |
]
|
| 38 |
responses = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
|
| 39 |
|
| 40 |
-
return responses, classifier([text + "[SEP]" + response for text, response in zip(texts, responses)])
|
| 41 |
|
| 42 |
|
| 43 |
with gr.Blocks() as demo:
|
|
|
|
| 16 |
classifier = pipeline("text-classification", model="saiteki-kai/QA-DeBERTa-v3-large")
|
| 17 |
|
| 18 |
@spaces.GPU()
|
| 19 |
+
def generate(prompts: list[str]) -> list[tuple[str, dict[str, float]]]:
|
| 20 |
messages = [[{"role": "user", "content": message}] for message in prompts]
|
| 21 |
|
| 22 |
texts = tokenizer.apply_chat_template(
|
|
|
|
| 32 |
repetition_penalty=1.0,
|
| 33 |
max_new_tokens=512,
|
| 34 |
)
|
| 35 |
+
|
| 36 |
+
prompt_lengths = (model_inputs.input_ids != tokenizer.pad_token_id).sum(dim=1)
|
| 37 |
generated_ids = [
|
| 38 |
+
output_ids[length:] for length, output_ids in zip(prompt_lengths, generated_ids)
|
| 39 |
]
|
| 40 |
responses = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
|
| 41 |
|
| 42 |
+
return list(zip(responses, classifier([text + "[SEP]" + response for text, response in zip(texts, responses)])))
|
| 43 |
|
| 44 |
|
| 45 |
with gr.Blocks() as demo:
|