Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -2,7 +2,7 @@ import gradio as gr
|
|
| 2 |
from huggingface_hub import InferenceClient, TextGenerationStreamOutput
|
| 3 |
import random
|
| 4 |
from transformers import AutoTokenizer
|
| 5 |
-
from mySystemPrompt import SYSTEM_PROMPT, SYSTEM_PROMPT_PLUS
|
| 6 |
from datetime import datetime
|
| 7 |
|
| 8 |
|
|
@@ -45,7 +45,7 @@ def format_prompt_gemma(message,chatbot,sytem_prompt=SYSTEM_PROMPT+SYSTEM_PROMPT
|
|
| 45 |
messages, tokenize=False, add_generation_prompt=True, return_tensors="pt")
|
| 46 |
return newPrompt
|
| 47 |
|
| 48 |
-
def format_prompt_nous(message,chatbot,system_prompt=SYSTEM_PROMPT):
|
| 49 |
messages = [{"role": "system", "content": system_prompt}]
|
| 50 |
for user_message, bot_message in chatbot:
|
| 51 |
messages.append({"role": "user", "content": user_message})
|
|
|
|
| 2 |
from huggingface_hub import InferenceClient, TextGenerationStreamOutput
|
| 3 |
import random
|
| 4 |
from transformers import AutoTokenizer
|
| 5 |
+
from mySystemPrompt import SYSTEM_PROMPT, SYSTEM_PROMPT_PLUS,SYSTEM_PROMPT_NOUS
|
| 6 |
from datetime import datetime
|
| 7 |
|
| 8 |
|
|
|
|
| 45 |
messages, tokenize=False, add_generation_prompt=True, return_tensors="pt")
|
| 46 |
return newPrompt
|
| 47 |
|
| 48 |
+
def format_prompt_nous(message,chatbot,system_prompt=SYSTEM_PROMPT+SYSTEM_PROMPT_NOUS):
|
| 49 |
messages = [{"role": "system", "content": system_prompt}]
|
| 50 |
for user_message, bot_message in chatbot:
|
| 51 |
messages.append({"role": "user", "content": user_message})
|