Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -138,8 +138,9 @@ def chat_llama3_8b(message: str,
|
|
| 138 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
|
| 139 |
|
| 140 |
streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
|
| 141 |
-
print(max_new_tokens)
|
| 142 |
max_new_tokens=4096
|
|
|
|
| 143 |
generate_kwargs = dict(
|
| 144 |
input_ids= input_ids,
|
| 145 |
streamer=streamer,
|
|
@@ -180,14 +181,16 @@ with gr.Blocks(fill_height=True, css=css) as demo:
|
|
| 180 |
gr.Slider(minimum=0,
|
| 181 |
maximum=1,
|
| 182 |
step=0.1,
|
| 183 |
-
value=0.
|
| 184 |
label="Temperature",
|
|
|
|
| 185 |
render=False),
|
| 186 |
gr.Slider(minimum=128,
|
| 187 |
maximum=4096,
|
| 188 |
step=1,
|
| 189 |
value=4096,
|
| 190 |
label="Max new tokens",
|
|
|
|
| 191 |
render=False),
|
| 192 |
],
|
| 193 |
examples=[
|
|
|
|
| 138 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
|
| 139 |
|
| 140 |
streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
|
| 141 |
+
# print(max_new_tokens)
|
| 142 |
max_new_tokens=4096
|
| 143 |
+
temperature=0.9
|
| 144 |
generate_kwargs = dict(
|
| 145 |
input_ids= input_ids,
|
| 146 |
streamer=streamer,
|
|
|
|
| 181 |
gr.Slider(minimum=0,
|
| 182 |
maximum=1,
|
| 183 |
step=0.1,
|
| 184 |
+
value=0.9,
|
| 185 |
label="Temperature",
|
| 186 |
+
interactive = False,
|
| 187 |
render=False),
|
| 188 |
gr.Slider(minimum=128,
|
| 189 |
maximum=4096,
|
| 190 |
step=1,
|
| 191 |
value=4096,
|
| 192 |
label="Max new tokens",
|
| 193 |
+
interactive = False,
|
| 194 |
render=False),
|
| 195 |
],
|
| 196 |
examples=[
|