Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -46,7 +46,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 46 |
model.eval()
|
| 47 |
|
| 48 |
|
| 49 |
-
@spaces.GPU(duration=
|
| 50 |
def generate(
|
| 51 |
message: str,
|
| 52 |
chat_history: list[tuple[str, str]],
|
|
@@ -84,8 +84,8 @@ def generate(
|
|
| 84 |
outputs.append(text)
|
| 85 |
yield "".join(outputs)
|
| 86 |
|
| 87 |
-
outputs.append("\n\nI have now attempted to solve the optimization modeling task! Please try executing the code in your environment, making sure it is equipped with `coptpy`.")
|
| 88 |
-
yield "".join(outputs)
|
| 89 |
|
| 90 |
|
| 91 |
chat_interface = gr.ChatInterface(
|
|
|
|
| 46 |
model.eval()
|
| 47 |
|
| 48 |
|
| 49 |
+
@spaces.GPU(duration=100)
|
| 50 |
def generate(
|
| 51 |
message: str,
|
| 52 |
chat_history: list[tuple[str, str]],
|
|
|
|
| 84 |
outputs.append(text)
|
| 85 |
yield "".join(outputs)
|
| 86 |
|
| 87 |
+
# outputs.append("\n\nI have now attempted to solve the optimization modeling task! Please try executing the code in your environment, making sure it is equipped with `coptpy`.")
|
| 88 |
+
# yield "".join(outputs)
|
| 89 |
|
| 90 |
|
| 91 |
chat_interface = gr.ChatInterface(
|