Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,13 +5,13 @@ import argparse
|
|
| 5 |
from huggingface_hub import snapshot_download
|
| 6 |
from llama_cpp import Llama
|
| 7 |
|
| 8 |
-
repo_name = '
|
| 9 |
-
model_file = "
|
| 10 |
|
| 11 |
snapshot_download(repo_id=repo_name, local_dir=".", allow_patterns=model_file)
|
| 12 |
|
| 13 |
DEFAULT_MODEL_PATH = model_file
|
| 14 |
-
llm = Llama(model_path=model_file, model_type="
|
| 15 |
|
| 16 |
def predict(input, chatbot, max_length, top_p, temperature, history):
|
| 17 |
chatbot.append((input, ""))
|
|
@@ -36,8 +36,8 @@ def reset_state():
|
|
| 36 |
return [], []
|
| 37 |
|
| 38 |
with gr.Blocks() as demo:
|
| 39 |
-
gr.HTML("""<h1 align="center">
|
| 40 |
-
<h3 align="center">This is unofficial demo of `
|
| 41 |
<h4 align="center">Hit the like button if you liked! 🤗</h4>""")
|
| 42 |
|
| 43 |
chatbot = gr.Chatbot()
|
|
|
|
| 5 |
from huggingface_hub import snapshot_download
|
| 6 |
from llama_cpp import Llama
|
| 7 |
|
| 8 |
+
repo_name = 'mradermacher/Turkcell-LLM-7b-v1-GGUF'
|
| 9 |
+
model_file = "Turkcell-LLM-7b-v1.Q5_K_M.gguf"
|
| 10 |
|
| 11 |
snapshot_download(repo_id=repo_name, local_dir=".", allow_patterns=model_file)
|
| 12 |
|
| 13 |
DEFAULT_MODEL_PATH = model_file
|
| 14 |
+
llm = Llama(model_path=model_file, model_type="mistral")
|
| 15 |
|
| 16 |
def predict(input, chatbot, max_length, top_p, temperature, history):
|
| 17 |
chatbot.append((input, ""))
|
|
|
|
| 36 |
return [], []
|
| 37 |
|
| 38 |
with gr.Blocks() as demo:
|
| 39 |
+
gr.HTML("""<h1 align="center">TurkcellLLM Chatbot Demo</h1>
|
| 40 |
+
<h3 align="center">This is unofficial demo of `mradermacher/Turkcell-LLM-7b-v1-GGUF` model based on Mistral architecture.</h3>
|
| 41 |
<h4 align="center">Hit the like button if you liked! 🤗</h4>""")
|
| 42 |
|
| 43 |
chatbot = gr.Chatbot()
|