Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -4,12 +4,14 @@ from llama_cpp import Llama
|
|
| 4 |
try:
|
| 5 |
llm = Llama.from_pretrained(
|
| 6 |
repo_id="operablepattern/gemma-2b-it-Q",
|
| 7 |
-
filename="*
|
|
|
|
| 8 |
verbose=True
|
| 9 |
)
|
| 10 |
except:
|
| 11 |
llm = Llama(
|
| 12 |
model_path="./gemma-2b-it-Q5_K_M.gguf",
|
|
|
|
| 13 |
verbose=True
|
| 14 |
)
|
| 15 |
|
|
@@ -21,5 +23,4 @@ def response(message, history):
|
|
| 21 |
gr.ChatInterface(
|
| 22 |
fn=response,
|
| 23 |
title="Chat with Gemma",
|
| 24 |
-
theme='sky'
|
| 25 |
).queue().launch()
|
|
|
|
| 4 |
try:
|
| 5 |
llm = Llama.from_pretrained(
|
| 6 |
repo_id="operablepattern/gemma-2b-it-Q",
|
| 7 |
+
filename="*Q5_K_M.gguf",
|
| 8 |
+
max_tokens=32,
|
| 9 |
verbose=True
|
| 10 |
)
|
| 11 |
except:
|
| 12 |
llm = Llama(
|
| 13 |
model_path="./gemma-2b-it-Q5_K_M.gguf",
|
| 14 |
+
max_tokens=32,
|
| 15 |
verbose=True
|
| 16 |
)
|
| 17 |
|
|
|
|
| 23 |
gr.ChatInterface(
|
| 24 |
fn=response,
|
| 25 |
title="Chat with Gemma",
|
|
|
|
| 26 |
).queue().launch()
|