Spaces:
Running on Zero
Running on Zero
Commit ·
89d297d
1
Parent(s): 5a4a0a6
fixed prompt format!
Browse files- __pycache__/model.cpython-310.pyc +0 -0
- __pycache__/vlm_inference.cpython-310.pyc +0 -0
- app.py +4 -4
__pycache__/model.cpython-310.pyc
CHANGED
|
Binary files a/__pycache__/model.cpython-310.pyc and b/__pycache__/model.cpython-310.pyc differ
|
|
|
__pycache__/vlm_inference.cpython-310.pyc
CHANGED
|
Binary files a/__pycache__/vlm_inference.cpython-310.pyc and b/__pycache__/vlm_inference.cpython-310.pyc differ
|
|
|
app.py
CHANGED
|
@@ -68,10 +68,10 @@ def chat_fn(message, history, temperature, top_p, top_k):
|
|
| 68 |
images=image.convert("RGB"),
|
| 69 |
return_tensors="pt"
|
| 70 |
)["pixel_values"].to(device)
|
| 71 |
-
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
|
| 75 |
|
| 76 |
# ★ ここが重要:累積して yield
|
| 77 |
output = ""
|
|
|
|
| 68 |
images=image.convert("RGB"),
|
| 69 |
return_tensors="pt"
|
| 70 |
)["pixel_values"].to(device)
|
| 71 |
+
|
| 72 |
+
### NEW ###
|
| 73 |
+
prompt = f"<user>\n{text}<assistant>\n"
|
| 74 |
+
### NEW ###
|
| 75 |
|
| 76 |
# ★ ここが重要:累積して yield
|
| 77 |
output = ""
|