Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,8 +5,7 @@ import os
|
|
| 5 |
from huggingface_hub import hf_hub_download
|
| 6 |
import base64
|
| 7 |
from llama_cpp import Llama
|
| 8 |
-
from llama_cpp.
|
| 9 |
-
|
| 10 |
|
| 11 |
os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
|
| 12 |
MODEL_LIST = ["openbmb/MiniCPM-Llama3-V-2_5","openbmb/MiniCPM-Llama3-V-2_5-int4"]
|
|
@@ -27,7 +26,7 @@ CSS = """
|
|
| 27 |
}
|
| 28 |
"""
|
| 29 |
|
| 30 |
-
|
| 31 |
repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
|
| 32 |
filename="*mmproj*",
|
| 33 |
)
|
|
@@ -35,7 +34,7 @@ chat_handler = LlamaChatCompletionHandler.from_pretrained(
|
|
| 35 |
llm = Llama.from_pretrained(
|
| 36 |
repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
|
| 37 |
filename="ggml-model-Q5_K_M.gguf",
|
| 38 |
-
|
| 39 |
n_ctx=4096,
|
| 40 |
verbose=True
|
| 41 |
)
|
|
|
|
| 5 |
from huggingface_hub import hf_hub_download
|
| 6 |
import base64
|
| 7 |
from llama_cpp import Llama
|
| 8 |
+
from llama_cpp.llama_tokenizer import LlamaHFTokenizer
|
|
|
|
| 9 |
|
| 10 |
os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
|
| 11 |
MODEL_LIST = ["openbmb/MiniCPM-Llama3-V-2_5","openbmb/MiniCPM-Llama3-V-2_5-int4"]
|
|
|
|
| 26 |
}
|
| 27 |
"""
|
| 28 |
|
| 29 |
+
tokenizer = LlamaHFTokenizer.from_pretrained(
|
| 30 |
repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
|
| 31 |
filename="*mmproj*",
|
| 32 |
)
|
|
|
|
| 34 |
llm = Llama.from_pretrained(
|
| 35 |
repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
|
| 36 |
filename="ggml-model-Q5_K_M.gguf",
|
| 37 |
+
tokenizer=tokenizer,
|
| 38 |
n_ctx=4096,
|
| 39 |
verbose=True
|
| 40 |
)
|