Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -165,7 +165,7 @@ def llm_chat_response(text, image_base64=None):
|
|
| 165 |
|
| 166 |
logger.info("Initializing InferenceClient...")
|
| 167 |
client = InferenceClient(
|
| 168 |
-
provider="
|
| 169 |
api_key=HF_TOKEN
|
| 170 |
)
|
| 171 |
|
|
@@ -199,7 +199,7 @@ def llm_chat_response(text, image_base64=None):
|
|
| 199 |
logger.info("Sending request to model...")
|
| 200 |
try:
|
| 201 |
completion = client.chat.completions.create(
|
| 202 |
-
model="
|
| 203 |
messages=messages,
|
| 204 |
max_tokens=500
|
| 205 |
)
|
|
|
|
| 165 |
|
| 166 |
logger.info("Initializing InferenceClient...")
|
| 167 |
client = InferenceClient(
|
| 168 |
+
provider="cerebras", # Using cerebras as in this example
|
| 169 |
api_key=HF_TOKEN
|
| 170 |
)
|
| 171 |
|
|
|
|
| 199 |
logger.info("Sending request to model...")
|
| 200 |
try:
|
| 201 |
completion = client.chat.completions.create(
|
| 202 |
+
model="meta-llama/Llama-4-Scout-17B-16E-Instruct",
|
| 203 |
messages=messages,
|
| 204 |
max_tokens=500
|
| 205 |
)
|