Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -58,7 +58,19 @@ def generate_response(user_query, relevant_segment):
|
|
| 58 |
|
| 59 |
# Encode the input and generate a response
|
| 60 |
input_ids = tokenizer.encode(user_message, return_tensors='pt')
|
| 61 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 62 |
output_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
| 63 |
|
| 64 |
# Append assistant's message to messages list for context
|
|
|
|
| 58 |
|
| 59 |
# Encode the input and generate a response
|
| 60 |
input_ids = tokenizer.encode(user_message, return_tensors='pt')
|
| 61 |
+
|
| 62 |
+
# Create attention mask (1 for real tokens, 0 for padding tokens)
|
| 63 |
+
attention_mask = (input_ids != tokenizer.pad_token_id).long()
|
| 64 |
+
|
| 65 |
+
# Generate the response using the model
|
| 66 |
+
output = model.generate(
|
| 67 |
+
input_ids,
|
| 68 |
+
attention_mask=attention_mask,
|
| 69 |
+
max_length=150,
|
| 70 |
+
num_return_sequences=1,
|
| 71 |
+
pad_token_id=tokenizer.eos_token_id # Set pad_token_id to eos_token_id
|
| 72 |
+
)
|
| 73 |
+
|
| 74 |
output_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
| 75 |
|
| 76 |
# Append assistant's message to messages list for context
|