Update app.py
Browse files
app.py
CHANGED
|
@@ -31,19 +31,6 @@ tokens = model.generate(
|
|
| 31 |
do_sample=True,
|
| 32 |
stopping_criteria=StoppingCriteriaList([StopOnTokens()])
|
| 33 |
)
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
gr.Markdown("## StableLM-Tuned-Alpha-7b Chat")
|
| 37 |
-
gr.HTML('''<center><a href="https://huggingface.co/spaces/stabilityai/stablelm-tuned-alpha-chat?duplicate=true"><img src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a>Duplicate the Space to skip the queue and run in a private space</center>''')
|
| 38 |
-
chatbot = gr.Chatbot().style(height=500)
|
| 39 |
-
with gr.Row():
|
| 40 |
-
with gr.Column(scale=0.70):
|
| 41 |
-
msg = gr.Textbox(label="Chat Message Box", placeholder="Chat Message Box", show_label=False).style(container=False)
|
| 42 |
-
with gr.Column(scale=0.30):
|
| 43 |
-
with gr.Row():
|
| 44 |
-
submit = gr.Button("Submit")
|
| 45 |
-
clear = gr.Button("Clear")
|
| 46 |
-
system_msg = gr.Textbox(
|
| 47 |
-
start_message, label="System Message", interactive=False, visible=False)
|
| 48 |
|
| 49 |
print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|
|
|
|
| 31 |
do_sample=True,
|
| 32 |
stopping_criteria=StoppingCriteriaList([StopOnTokens()])
|
| 33 |
)
|
| 34 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
|
| 36 |
print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|