Update project_model.py
Browse files- project_model.py +2 -1
project_model.py
CHANGED
|
@@ -72,6 +72,7 @@ class VisualQAState:
|
|
| 72 |
"role": "system", # System prompt
|
| 73 |
"content": (
|
| 74 |
"You are a helpful visual assistant designed for visually impaired users. "
|
|
|
|
| 75 |
"You must provide detailed, descriptive, and spatially-aware answers based on the given image and conversation history. "
|
| 76 |
"Always describe what you see clearly and help the user understand the scene. "
|
| 77 |
'If unsure, say "I am not certain."'
|
|
@@ -202,7 +203,7 @@ def process_inputs(
|
|
| 202 |
session.add_question(question)
|
| 203 |
|
| 204 |
# Send prompt to Gemma 3
|
| 205 |
-
gemma_output = gemma_pipe(text=session.message_history, max_new_tokens=
|
| 206 |
|
| 207 |
# Debugging: Check the output structure
|
| 208 |
print("Gemma Output:", gemma_output)
|
|
|
|
| 72 |
"role": "system", # System prompt
|
| 73 |
"content": (
|
| 74 |
"You are a helpful visual assistant designed for visually impaired users. "
|
| 75 |
+
"You must understand the question being asked and answer accordinly."
|
| 76 |
"You must provide detailed, descriptive, and spatially-aware answers based on the given image and conversation history. "
|
| 77 |
"Always describe what you see clearly and help the user understand the scene. "
|
| 78 |
'If unsure, say "I am not certain."'
|
|
|
|
| 203 |
session.add_question(question)
|
| 204 |
|
| 205 |
# Send prompt to Gemma 3
|
| 206 |
+
gemma_output = gemma_pipe(text=session.message_history, max_new_tokens=500)
|
| 207 |
|
| 208 |
# Debugging: Check the output structure
|
| 209 |
print("Gemma Output:", gemma_output)
|