updating Deep_Research models from Qwen3-235B to GLM-4.6
Browse files- Modules/Deep_Research.py +2 -2
Modules/Deep_Research.py
CHANGED
|
@@ -418,7 +418,7 @@ def Deep_Research(
|
|
| 418 |
def _invoke_chat(messages, provider: str, max_tokens: int, temp: float, top_p: float):
|
| 419 |
client = InferenceClient(provider=provider, api_key=HF_TEXTGEN_TOKEN)
|
| 420 |
return client.chat.completions.create(
|
| 421 |
-
model="
|
| 422 |
messages=messages,
|
| 423 |
max_tokens=max_tokens,
|
| 424 |
temperature=temp,
|
|
@@ -492,7 +492,7 @@ def Deep_Research(
|
|
| 492 |
except Exception:
|
| 493 |
prompt_chars = -1
|
| 494 |
print(f"[PIPELINE] Fetch complete: pages={len(pages)}, unique_urls={len(pages.keys())}, prompt_chars={prompt_chars}", flush=True)
|
| 495 |
-
print("[PIPELINE] Starting inference (provider=cerebras, model=
|
| 496 |
|
| 497 |
try:
|
| 498 |
print("[LLM] Attempt 1: provider=cerebras, max_tokens=32768", flush=True)
|
|
|
|
| 418 |
def _invoke_chat(messages, provider: str, max_tokens: int, temp: float, top_p: float):
|
| 419 |
client = InferenceClient(provider=provider, api_key=HF_TEXTGEN_TOKEN)
|
| 420 |
return client.chat.completions.create(
|
| 421 |
+
model="zai-org/GLM-4.6",
|
| 422 |
messages=messages,
|
| 423 |
max_tokens=max_tokens,
|
| 424 |
temperature=temp,
|
|
|
|
| 492 |
except Exception:
|
| 493 |
prompt_chars = -1
|
| 494 |
print(f"[PIPELINE] Fetch complete: pages={len(pages)}, unique_urls={len(pages.keys())}, prompt_chars={prompt_chars}", flush=True)
|
| 495 |
+
print("[PIPELINE] Starting inference (provider=cerebras, model=zai-org/GLM-4.6)", flush=True)
|
| 496 |
|
| 497 |
try:
|
| 498 |
print("[LLM] Attempt 1: provider=cerebras, max_tokens=32768", flush=True)
|