Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -12,7 +12,7 @@ model_path = hf_hub_download(repo_id = repo_id, filename=model_file)
|
|
| 12 |
# n_threads
|
| 13 |
llm = Llama(model_path=model_path,n_gpu_layers=30,n_ctx=512,temperature=0.2,repeat_penalty=1.1,top_k_sampling=40,top_p_sampling=0.95,min_p_sampling=0.05)
|
| 14 |
def generate_llm_response(prompt):
|
| 15 |
-
output = llm(prompt, max_tokens=
|
| 16 |
return output["choices"][0]["text"]
|
| 17 |
|
| 18 |
import streamlit as st
|
|
|
|
| 12 |
# n_threads
|
| 13 |
llm = Llama(model_path=model_path,n_gpu_layers=30,n_ctx=512,temperature=0.2,repeat_penalty=1.1,top_k_sampling=40,top_p_sampling=0.95,min_p_sampling=0.05)
|
| 14 |
def generate_llm_response(prompt):
|
| 15 |
+
output = llm(prompt, max_tokens=256)
|
| 16 |
return output["choices"][0]["text"]
|
| 17 |
|
| 18 |
import streamlit as st
|