Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -2,6 +2,7 @@ import gradio as gr
|
|
| 2 |
from sentence_transformers import SentenceTransformer, util
|
| 3 |
from transformers import GPT2LMHeadModel, GPT2Tokenizer
|
| 4 |
import os
|
|
|
|
| 5 |
|
| 6 |
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
| 7 |
|
|
@@ -60,7 +61,7 @@ def generate_response(user_query, relevant_segment):
|
|
| 60 |
input_ids = tokenizer.encode(user_message, return_tensors='pt')
|
| 61 |
|
| 62 |
# Create attention mask (1 for real tokens, 0 for padding tokens)
|
| 63 |
-
attention_mask = (input_ids != tokenizer.pad_token_id).long
|
| 64 |
|
| 65 |
# Generate the response using the model
|
| 66 |
output = model.generate(
|
|
|
|
| 2 |
from sentence_transformers import SentenceTransformer, util
|
| 3 |
from transformers import GPT2LMHeadModel, GPT2Tokenizer
|
| 4 |
import os
|
| 5 |
+
import torch
|
| 6 |
|
| 7 |
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
| 8 |
|
|
|
|
| 61 |
input_ids = tokenizer.encode(user_message, return_tensors='pt')
|
| 62 |
|
| 63 |
# Create attention mask (1 for real tokens, 0 for padding tokens)
|
| 64 |
+
attention_mask = (input_ids != tokenizer.pad_token_id).to(torch.long)
|
| 65 |
|
| 66 |
# Generate the response using the model
|
| 67 |
output = model.generate(
|