LiamKhoaLe commited on
Commit
140713c
·
1 Parent(s): ac4ae39

Upd user_id

Browse files
helpers/namer.py CHANGED
@@ -79,7 +79,9 @@ Return only the session name, nothing else."""
79
  system_prompt=sys_prompt,
80
  user_prompt=user_prompt,
81
  gemini_rotator=None,
82
- nvidia_rotator=nvidia_rotator
 
 
83
  )
84
 
85
  logger.info(f"[NAMER] Raw API response: {response}")
 
79
  system_prompt=sys_prompt,
80
  user_prompt=user_prompt,
81
  gemini_rotator=None,
82
+ nvidia_rotator=nvidia_rotator,
83
+ user_id=user_id,
84
+ context="session_naming"
85
  )
86
 
87
  logger.info(f"[NAMER] Raw API response: {response}")
legacy.py CHANGED
@@ -673,7 +673,7 @@ async def generate_report(
673
 
674
  try:
675
  selection_filter = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
676
- filter_response = await generate_answer_with_model(selection_filter, filter_sys, filter_user, gemini_rotator, nvidia_rotator)
677
  logger.info(f"[REPORT] Raw filter response: {filter_response}")
678
  # Try to parse the filter response to get relevant chunks
679
  import json
@@ -709,7 +709,7 @@ async def generate_report(
709
  try:
710
  # Step 1: Outline with Flash/Med
711
  selection_outline = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
712
- outline_md = await generate_answer_with_model(selection_outline, sys_outline, user_outline, gemini_rotator, nvidia_rotator)
713
  except Exception as e:
714
  logger.warning(f"Report outline failed: {e}")
715
  outline_md = "# Report Outline\n\n- Introduction\n- Key Topics\n- Conclusion"
@@ -731,7 +731,7 @@ async def generate_report(
731
 
732
  try:
733
  selection_report = {"provider": "gemini", "model": os.getenv("GEMINI_PRO", "gemini-2.5-pro")}
734
- report_md = await generate_answer_with_model(selection_report, sys_report, user_report, gemini_rotator, nvidia_rotator)
735
  except Exception as e:
736
  logger.error(f"Report generation failed: {e}")
737
  report_md = outline_md + "\n\n" + file_summary
 
673
 
674
  try:
675
  selection_filter = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
676
+ filter_response = await generate_answer_with_model(selection_filter, filter_sys, filter_user, gemini_rotator, nvidia_rotator, user_id="system", context="legacy_filter")
677
  logger.info(f"[REPORT] Raw filter response: {filter_response}")
678
  # Try to parse the filter response to get relevant chunks
679
  import json
 
709
  try:
710
  # Step 1: Outline with Flash/Med
711
  selection_outline = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
712
+ outline_md = await generate_answer_with_model(selection_outline, sys_outline, user_outline, gemini_rotator, nvidia_rotator, user_id="system", context="legacy_outline")
713
  except Exception as e:
714
  logger.warning(f"Report outline failed: {e}")
715
  outline_md = "# Report Outline\n\n- Introduction\n- Key Topics\n- Conclusion"
 
731
 
732
  try:
733
  selection_report = {"provider": "gemini", "model": os.getenv("GEMINI_PRO", "gemini-2.5-pro")}
734
+ report_md = await generate_answer_with_model(selection_report, sys_report, user_report, gemini_rotator, nvidia_rotator, user_id="system", context="legacy_report")
735
  except Exception as e:
736
  logger.error(f"Report generation failed: {e}")
737
  report_md = outline_md + "\n\n" + file_summary
memo/sessions.py CHANGED
@@ -174,7 +174,9 @@ Is this a context switch?"""
174
  system_prompt=sys_prompt,
175
  user_prompt=user_prompt,
176
  gemini_rotator=None,
177
- nvidia_rotator=nvidia_rotator
 
 
178
  )
179
 
180
  # Parse JSON response
 
174
  system_prompt=sys_prompt,
175
  user_prompt=user_prompt,
176
  gemini_rotator=None,
177
+ nvidia_rotator=nvidia_rotator,
178
+ user_id="system",
179
+ context="context_switch_detection"
180
  )
181
 
182
  # Parse JSON response
utils/api/router.py CHANGED
@@ -135,11 +135,11 @@ async def generate_answer_with_model(selection: Dict[str, Any], system_prompt: s
135
  if model in [GEMINI_PRO, GEMINI_MED]:
136
  logger.info(f"Falling back from {model} to NVIDIA_LARGE")
137
  fallback_selection = {"provider": "nvidia_large", "model": NVIDIA_LARGE}
138
- return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator)
139
  elif model == GEMINI_SMALL:
140
  logger.info(f"Falling back from {model} to NVIDIA_SMALL")
141
  fallback_selection = {"provider": "nvidia", "model": NVIDIA_SMALL}
142
- return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator)
143
  else:
144
  logger.error(f"No fallback defined for Gemini model: {model}")
145
  return "I couldn't parse the model response."
 
135
  if model in [GEMINI_PRO, GEMINI_MED]:
136
  logger.info(f"Falling back from {model} to NVIDIA_LARGE")
137
  fallback_selection = {"provider": "nvidia_large", "model": NVIDIA_LARGE}
138
+ return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator, user_id, context)
139
  elif model == GEMINI_SMALL:
140
  logger.info(f"Falling back from {model} to NVIDIA_SMALL")
141
  fallback_selection = {"provider": "nvidia", "model": NVIDIA_SMALL}
142
+ return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator, user_id, context)
143
  else:
144
  logger.error(f"No fallback defined for Gemini model: {model}")
145
  return "I couldn't parse the model response."
utils/service/pdf.py CHANGED
@@ -691,7 +691,7 @@ Return only the formatted references, one per line, numbered sequentially."""
691
  user_prompt = f"Format these sources in IEEE style:\n\n{source_data}"
692
 
693
  selection = {"provider": "nvidia", "model": "meta/llama-3.1-8b-instruct"}
694
- response = await generate_answer_with_model(selection, sys_prompt, user_prompt, None, nvidia_rotator)
695
 
696
  # Parse the response into individual references
697
  references = [line.strip() for line in response.split('\n') if line.strip() and line.strip().startswith('[')]
 
691
  user_prompt = f"Format these sources in IEEE style:\n\n{source_data}"
692
 
693
  selection = {"provider": "nvidia", "model": "meta/llama-3.1-8b-instruct"}
694
+ response = await generate_answer_with_model(selection, sys_prompt, user_prompt, None, nvidia_rotator, user_id="system", context="pdf_citation")
695
 
696
  # Parse the response into individual references
697
  references = [line.strip() for line in response.split('\n') if line.strip() and line.strip().startswith('[')]