Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -4,7 +4,7 @@ import ollama
|
|
| 4 |
public_ip = os.environ['PUBLIC_IP']
|
| 5 |
port = os.environ['PORT']
|
| 6 |
|
| 7 |
-
model = 'llama3.
|
| 8 |
from ollama import Client
|
| 9 |
client = Client(host=f'http://{public_ip}:{port}')
|
| 10 |
|
|
@@ -42,7 +42,7 @@ chatbot = gr.ChatInterface(
|
|
| 42 |
gr.Slider(0.0,2.0, label="temperature", value=0.4, info="The temperature of the model. Increasing the temperature will make the model answer more creatively. (Default: 0.8)"),
|
| 43 |
],
|
| 44 |
title="Trashcan AI",
|
| 45 |
-
description="LLama3.
|
| 46 |
theme="finlaymacklon/smooth_slate",
|
| 47 |
submit_btn="Send",
|
| 48 |
retry_btn="🔄 Regenerate Response",
|
|
|
|
| 4 |
public_ip = os.environ['PUBLIC_IP']
|
| 5 |
port = os.environ['PORT']
|
| 6 |
|
| 7 |
+
model = 'llama3.2'
|
| 8 |
from ollama import Client
|
| 9 |
client = Client(host=f'http://{public_ip}:{port}')
|
| 10 |
|
|
|
|
| 42 |
gr.Slider(0.0,2.0, label="temperature", value=0.4, info="The temperature of the model. Increasing the temperature will make the model answer more creatively. (Default: 0.8)"),
|
| 43 |
],
|
| 44 |
title="Trashcan AI",
|
| 45 |
+
description="LLama3.2 hosted on a 2013 \"Trashcan\" Mac Pro with ollama",
|
| 46 |
theme="finlaymacklon/smooth_slate",
|
| 47 |
submit_btn="Send",
|
| 48 |
retry_btn="🔄 Regenerate Response",
|