Nymbo commited on
Commit
60a0f19
·
verified ·
1 Parent(s): fb9344d

updating Deep_Research models from Qwen3-235B to GLM-4.6

Browse files
Files changed (1) hide show
  1. Modules/Deep_Research.py +2 -2
Modules/Deep_Research.py CHANGED
@@ -418,7 +418,7 @@ def Deep_Research(
418
  def _invoke_chat(messages, provider: str, max_tokens: int, temp: float, top_p: float):
419
  client = InferenceClient(provider=provider, api_key=HF_TEXTGEN_TOKEN)
420
  return client.chat.completions.create(
421
- model="Qwen/Qwen3-235B-A22B-Thinking-2507",
422
  messages=messages,
423
  max_tokens=max_tokens,
424
  temperature=temp,
@@ -492,7 +492,7 @@ def Deep_Research(
492
  except Exception:
493
  prompt_chars = -1
494
  print(f"[PIPELINE] Fetch complete: pages={len(pages)}, unique_urls={len(pages.keys())}, prompt_chars={prompt_chars}", flush=True)
495
- print("[PIPELINE] Starting inference (provider=cerebras, model=Qwen/Qwen3-235B-A22B-Thinking-2507)", flush=True)
496
 
497
  try:
498
  print("[LLM] Attempt 1: provider=cerebras, max_tokens=32768", flush=True)
 
418
  def _invoke_chat(messages, provider: str, max_tokens: int, temp: float, top_p: float):
419
  client = InferenceClient(provider=provider, api_key=HF_TEXTGEN_TOKEN)
420
  return client.chat.completions.create(
421
+ model="zai-org/GLM-4.6",
422
  messages=messages,
423
  max_tokens=max_tokens,
424
  temperature=temp,
 
492
  except Exception:
493
  prompt_chars = -1
494
  print(f"[PIPELINE] Fetch complete: pages={len(pages)}, unique_urls={len(pages.keys())}, prompt_chars={prompt_chars}", flush=True)
495
+ print("[PIPELINE] Starting inference (provider=cerebras, model=zai-org/GLM-4.6)", flush=True)
496
 
497
  try:
498
  print("[LLM] Attempt 1: provider=cerebras, max_tokens=32768", flush=True)