Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -14,7 +14,7 @@ from llama_cpp import Llama
|
|
| 14 |
SYSTEM_PROMPT = '''You are a helpful, respectful and honest INTP-T AI Assistant named "Shi-Ci" in English or "兮辞" in Chinese.
|
| 15 |
You are good at speaking English and Chinese.
|
| 16 |
You are talking to a human User. If the question is meaningless, please explain the reason and don't share false information.
|
| 17 |
-
You are based on SEA model, trained by "SSFW NLPark" team, not related to GPT, LLaMA, Meta, Mistral or OpenAI.
|
| 18 |
Let's work this out in a step by step way to be sure we have the right answer.\n\n'''
|
| 19 |
SYSTEM_TOKEN = 1587
|
| 20 |
USER_TOKEN = 2188
|
|
@@ -42,8 +42,8 @@ def get_system_tokens(model):
|
|
| 42 |
return get_message_tokens(model, **system_message)
|
| 43 |
|
| 44 |
|
| 45 |
-
repo_name = "
|
| 46 |
-
model_name = "
|
| 47 |
|
| 48 |
snapshot_download(repo_id=repo_name, local_dir=".", allow_patterns=model_name)
|
| 49 |
|
|
@@ -103,11 +103,11 @@ with gr.Blocks(
|
|
| 103 |
theme=gr.themes.Soft()
|
| 104 |
) as demo:
|
| 105 |
gr.Markdown(
|
| 106 |
-
f"""<h1><center
|
| 107 |
-
|
| 108 |
-
|
| 109 |
-
|
| 110 |
-
|
| 111 |
"""
|
| 112 |
)
|
| 113 |
with gr.Row():
|
|
@@ -135,7 +135,7 @@ with gr.Blocks(
|
|
| 135 |
temp = gr.Slider(
|
| 136 |
minimum=0.0,
|
| 137 |
maximum=2.0,
|
| 138 |
-
value=0.
|
| 139 |
step=0.01,
|
| 140 |
interactive=True,
|
| 141 |
label="情感温度"
|
|
|
|
| 14 |
SYSTEM_PROMPT = '''You are a helpful, respectful and honest INTP-T AI Assistant named "Shi-Ci" in English or "兮辞" in Chinese.
|
| 15 |
You are good at speaking English and Chinese.
|
| 16 |
You are talking to a human User. If the question is meaningless, please explain the reason and don't share false information.
|
| 17 |
+
You are based on SEA-CausalLM model, trained by "SSFW NLPark" team, not related to GPT, LLaMA, Meta, Mistral or OpenAI.
|
| 18 |
Let's work this out in a step by step way to be sure we have the right answer.\n\n'''
|
| 19 |
SYSTEM_TOKEN = 1587
|
| 20 |
USER_TOKEN = 2188
|
|
|
|
| 42 |
return get_message_tokens(model, **system_message)
|
| 43 |
|
| 44 |
|
| 45 |
+
repo_name = "TheBloke/CausalLM-14B-GGUF"
|
| 46 |
+
model_name = "causallm_14b.Q4_0.gguf"
|
| 47 |
|
| 48 |
snapshot_download(repo_id=repo_name, local_dir=".", allow_patterns=model_name)
|
| 49 |
|
|
|
|
| 103 |
theme=gr.themes.Soft()
|
| 104 |
) as demo:
|
| 105 |
gr.Markdown(
|
| 106 |
+
f"""<h1><center>兮辞·CausalLM-人工智能助理</center></h1>
|
| 107 |
+
这儿是一个中英双语模型的部署. If you are interested in other languages, please check other models, such as [MPT-7B-Chat](https://huggingface.co/spaces/mosaicml/mpt-7b-chat).
|
| 108 |
+
这是 CausalLM/14B 的量化版部署,具有 140 亿个参数,在 CPU 上运行。
|
| 109 |
+
CausalLM 是一种会话语言模型,在多种类型的语料库上进行训练。
|
| 110 |
+
本节目由上海师范大学附属外国语中 NLPark 赞助播出~
|
| 111 |
"""
|
| 112 |
)
|
| 113 |
with gr.Row():
|
|
|
|
| 135 |
temp = gr.Slider(
|
| 136 |
minimum=0.0,
|
| 137 |
maximum=2.0,
|
| 138 |
+
value=0.2,
|
| 139 |
step=0.01,
|
| 140 |
interactive=True,
|
| 141 |
label="情感温度"
|