Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -8,7 +8,7 @@ from llama_cpp import Llama
|
|
| 8 |
from huggingface_hub import hf_hub_download
|
| 9 |
|
| 10 |
hf_hub_download(repo_id="Pi3141/alpaca-lora-7B-ggml", filename="ggml-model-q4_1.bin", local_dir=".")
|
| 11 |
-
llm = Llama(model_path="./ggml-model-q4_1.bin"
|
| 12 |
|
| 13 |
|
| 14 |
ins = '''Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
|
@@ -132,7 +132,7 @@ with gr.Blocks(theme=seafoam, analytics_enabled=False, css=css) as demo:
|
|
| 132 |
gr.Examples(
|
| 133 |
examples=examples,
|
| 134 |
inputs=[instruction],
|
| 135 |
-
cache_examples=
|
| 136 |
fn=process_example,
|
| 137 |
outputs=[output],
|
| 138 |
)
|
|
|
|
| 8 |
from huggingface_hub import hf_hub_download
|
| 9 |
|
| 10 |
hf_hub_download(repo_id="Pi3141/alpaca-lora-7B-ggml", filename="ggml-model-q4_1.bin", local_dir=".")
|
| 11 |
+
llm = Llama(model_path="./ggml-model-q4_1.bin")
|
| 12 |
|
| 13 |
|
| 14 |
ins = '''Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
|
|
|
| 132 |
gr.Examples(
|
| 133 |
examples=examples,
|
| 134 |
inputs=[instruction],
|
| 135 |
+
cache_examples=False,
|
| 136 |
fn=process_example,
|
| 137 |
outputs=[output],
|
| 138 |
)
|