Update app.py
Browse files
app.py
CHANGED
|
@@ -14,12 +14,13 @@ logging.basicConfig(stream=sys.stdout, level=logging.INFO)
|
|
| 14 |
logging.getLogger().addHandler(logging.StreamHandler(stream=sys.stdout))
|
| 15 |
|
| 16 |
def configure_llama_model():
|
| 17 |
-
|
|
|
|
| 18 |
llm = LlamaCPP(
|
| 19 |
model_url=model_url,
|
| 20 |
temperature=0.3,
|
| 21 |
-
max_new_tokens=
|
| 22 |
-
context_window=
|
| 23 |
model_kwargs={"n_gpu_layers": 1},
|
| 24 |
messages_to_prompt=messages_to_prompt,
|
| 25 |
completion_to_prompt=completion_to_prompt,
|
|
|
|
| 14 |
logging.getLogger().addHandler(logging.StreamHandler(stream=sys.stdout))
|
| 15 |
|
| 16 |
def configure_llama_model():
|
| 17 |
+
|
| 18 |
+
model_url = 'https://huggingface.co/TheBloke/Mistral-7B-v0.1-GGUF/resolve/main/mistral-7b-v0.1.Q4_K_M.gguf'
|
| 19 |
llm = LlamaCPP(
|
| 20 |
model_url=model_url,
|
| 21 |
temperature=0.3,
|
| 22 |
+
max_new_tokens=256,
|
| 23 |
+
context_window=3900,
|
| 24 |
model_kwargs={"n_gpu_layers": 1},
|
| 25 |
messages_to_prompt=messages_to_prompt,
|
| 26 |
completion_to_prompt=completion_to_prompt,
|