try token
Browse files
app.py
CHANGED
|
@@ -58,9 +58,10 @@ def get_top_chunks(query):
|
|
| 58 |
return top_chunks
|
| 59 |
|
| 60 |
client = InferenceClient(
|
| 61 |
-
model=
|
| 62 |
-
token=
|
| 63 |
-
)
|
|
|
|
| 64 |
#client is where you can change the LLM model!
|
| 65 |
def respond(message,history):
|
| 66 |
if not message.strip():
|
|
|
|
| 58 |
return top_chunks
|
| 59 |
|
| 60 |
client = InferenceClient(
|
| 61 |
+
model='Qwen/Qwen2.5-72B-Instruct',
|
| 62 |
+
token='HF_TOKEN'
|
| 63 |
+
)
|
| 64 |
+
|
| 65 |
#client is where you can change the LLM model!
|
| 66 |
def respond(message,history):
|
| 67 |
if not message.strip():
|