Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -24,9 +24,9 @@ import traceback
|
|
| 24 |
|
| 25 |
try:
|
| 26 |
response = client.text_generation("Hello, how are you?", max_new_tokens=50)
|
| 27 |
-
print("
|
| 28 |
except Exception as e:
|
| 29 |
-
print("
|
| 30 |
traceback.print_exc()'''
|
| 31 |
|
| 32 |
|
|
@@ -54,7 +54,7 @@ class BasicAgent:
|
|
| 54 |
|
| 55 |
# Load model
|
| 56 |
llm = InferenceClientModel(
|
| 57 |
-
model_id="
|
| 58 |
token=os.getenv("test"),
|
| 59 |
max_tokens=256,
|
| 60 |
temperature=0.2,
|
|
|
|
| 24 |
|
| 25 |
try:
|
| 26 |
response = client.text_generation("Hello, how are you?", max_new_tokens=50)
|
| 27 |
+
print("Model response:", response)
|
| 28 |
except Exception as e:
|
| 29 |
+
print("Direct call error:", repr(e))
|
| 30 |
traceback.print_exc()'''
|
| 31 |
|
| 32 |
|
|
|
|
| 54 |
|
| 55 |
# Load model
|
| 56 |
llm = InferenceClientModel(
|
| 57 |
+
model_id="mistralai/Mistral-7B-Instruct-v0.3",
|
| 58 |
token=os.getenv("test"),
|
| 59 |
max_tokens=256,
|
| 60 |
temperature=0.2,
|