Spaces:
Sleeping
Sleeping
Commit ·
140713c
1
Parent(s): ac4ae39
Upd user_id
Browse files- helpers/namer.py +3 -1
- legacy.py +3 -3
- memo/sessions.py +3 -1
- utils/api/router.py +2 -2
- utils/service/pdf.py +1 -1
helpers/namer.py
CHANGED
|
@@ -79,7 +79,9 @@ Return only the session name, nothing else."""
|
|
| 79 |
system_prompt=sys_prompt,
|
| 80 |
user_prompt=user_prompt,
|
| 81 |
gemini_rotator=None,
|
| 82 |
-
nvidia_rotator=nvidia_rotator
|
|
|
|
|
|
|
| 83 |
)
|
| 84 |
|
| 85 |
logger.info(f"[NAMER] Raw API response: {response}")
|
|
|
|
| 79 |
system_prompt=sys_prompt,
|
| 80 |
user_prompt=user_prompt,
|
| 81 |
gemini_rotator=None,
|
| 82 |
+
nvidia_rotator=nvidia_rotator,
|
| 83 |
+
user_id=user_id,
|
| 84 |
+
context="session_naming"
|
| 85 |
)
|
| 86 |
|
| 87 |
logger.info(f"[NAMER] Raw API response: {response}")
|
legacy.py
CHANGED
|
@@ -673,7 +673,7 @@ async def generate_report(
|
|
| 673 |
|
| 674 |
try:
|
| 675 |
selection_filter = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
|
| 676 |
-
filter_response = await generate_answer_with_model(selection_filter, filter_sys, filter_user, gemini_rotator, nvidia_rotator)
|
| 677 |
logger.info(f"[REPORT] Raw filter response: {filter_response}")
|
| 678 |
# Try to parse the filter response to get relevant chunks
|
| 679 |
import json
|
|
@@ -709,7 +709,7 @@ async def generate_report(
|
|
| 709 |
try:
|
| 710 |
# Step 1: Outline with Flash/Med
|
| 711 |
selection_outline = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
|
| 712 |
-
outline_md = await generate_answer_with_model(selection_outline, sys_outline, user_outline, gemini_rotator, nvidia_rotator)
|
| 713 |
except Exception as e:
|
| 714 |
logger.warning(f"Report outline failed: {e}")
|
| 715 |
outline_md = "# Report Outline\n\n- Introduction\n- Key Topics\n- Conclusion"
|
|
@@ -731,7 +731,7 @@ async def generate_report(
|
|
| 731 |
|
| 732 |
try:
|
| 733 |
selection_report = {"provider": "gemini", "model": os.getenv("GEMINI_PRO", "gemini-2.5-pro")}
|
| 734 |
-
report_md = await generate_answer_with_model(selection_report, sys_report, user_report, gemini_rotator, nvidia_rotator)
|
| 735 |
except Exception as e:
|
| 736 |
logger.error(f"Report generation failed: {e}")
|
| 737 |
report_md = outline_md + "\n\n" + file_summary
|
|
|
|
| 673 |
|
| 674 |
try:
|
| 675 |
selection_filter = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
|
| 676 |
+
filter_response = await generate_answer_with_model(selection_filter, filter_sys, filter_user, gemini_rotator, nvidia_rotator, user_id="system", context="legacy_filter")
|
| 677 |
logger.info(f"[REPORT] Raw filter response: {filter_response}")
|
| 678 |
# Try to parse the filter response to get relevant chunks
|
| 679 |
import json
|
|
|
|
| 709 |
try:
|
| 710 |
# Step 1: Outline with Flash/Med
|
| 711 |
selection_outline = {"provider": "gemini", "model": os.getenv("GEMINI_MED", "gemini-2.5-flash")}
|
| 712 |
+
outline_md = await generate_answer_with_model(selection_outline, sys_outline, user_outline, gemini_rotator, nvidia_rotator, user_id="system", context="legacy_outline")
|
| 713 |
except Exception as e:
|
| 714 |
logger.warning(f"Report outline failed: {e}")
|
| 715 |
outline_md = "# Report Outline\n\n- Introduction\n- Key Topics\n- Conclusion"
|
|
|
|
| 731 |
|
| 732 |
try:
|
| 733 |
selection_report = {"provider": "gemini", "model": os.getenv("GEMINI_PRO", "gemini-2.5-pro")}
|
| 734 |
+
report_md = await generate_answer_with_model(selection_report, sys_report, user_report, gemini_rotator, nvidia_rotator, user_id="system", context="legacy_report")
|
| 735 |
except Exception as e:
|
| 736 |
logger.error(f"Report generation failed: {e}")
|
| 737 |
report_md = outline_md + "\n\n" + file_summary
|
memo/sessions.py
CHANGED
|
@@ -174,7 +174,9 @@ Is this a context switch?"""
|
|
| 174 |
system_prompt=sys_prompt,
|
| 175 |
user_prompt=user_prompt,
|
| 176 |
gemini_rotator=None,
|
| 177 |
-
nvidia_rotator=nvidia_rotator
|
|
|
|
|
|
|
| 178 |
)
|
| 179 |
|
| 180 |
# Parse JSON response
|
|
|
|
| 174 |
system_prompt=sys_prompt,
|
| 175 |
user_prompt=user_prompt,
|
| 176 |
gemini_rotator=None,
|
| 177 |
+
nvidia_rotator=nvidia_rotator,
|
| 178 |
+
user_id="system",
|
| 179 |
+
context="context_switch_detection"
|
| 180 |
)
|
| 181 |
|
| 182 |
# Parse JSON response
|
utils/api/router.py
CHANGED
|
@@ -135,11 +135,11 @@ async def generate_answer_with_model(selection: Dict[str, Any], system_prompt: s
|
|
| 135 |
if model in [GEMINI_PRO, GEMINI_MED]:
|
| 136 |
logger.info(f"Falling back from {model} to NVIDIA_LARGE")
|
| 137 |
fallback_selection = {"provider": "nvidia_large", "model": NVIDIA_LARGE}
|
| 138 |
-
return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator)
|
| 139 |
elif model == GEMINI_SMALL:
|
| 140 |
logger.info(f"Falling back from {model} to NVIDIA_SMALL")
|
| 141 |
fallback_selection = {"provider": "nvidia", "model": NVIDIA_SMALL}
|
| 142 |
-
return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator)
|
| 143 |
else:
|
| 144 |
logger.error(f"No fallback defined for Gemini model: {model}")
|
| 145 |
return "I couldn't parse the model response."
|
|
|
|
| 135 |
if model in [GEMINI_PRO, GEMINI_MED]:
|
| 136 |
logger.info(f"Falling back from {model} to NVIDIA_LARGE")
|
| 137 |
fallback_selection = {"provider": "nvidia_large", "model": NVIDIA_LARGE}
|
| 138 |
+
return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator, user_id, context)
|
| 139 |
elif model == GEMINI_SMALL:
|
| 140 |
logger.info(f"Falling back from {model} to NVIDIA_SMALL")
|
| 141 |
fallback_selection = {"provider": "nvidia", "model": NVIDIA_SMALL}
|
| 142 |
+
return await generate_answer_with_model(fallback_selection, system_prompt, user_prompt, gemini_rotator, nvidia_rotator, user_id, context)
|
| 143 |
else:
|
| 144 |
logger.error(f"No fallback defined for Gemini model: {model}")
|
| 145 |
return "I couldn't parse the model response."
|
utils/service/pdf.py
CHANGED
|
@@ -691,7 +691,7 @@ Return only the formatted references, one per line, numbered sequentially."""
|
|
| 691 |
user_prompt = f"Format these sources in IEEE style:\n\n{source_data}"
|
| 692 |
|
| 693 |
selection = {"provider": "nvidia", "model": "meta/llama-3.1-8b-instruct"}
|
| 694 |
-
response = await generate_answer_with_model(selection, sys_prompt, user_prompt, None, nvidia_rotator)
|
| 695 |
|
| 696 |
# Parse the response into individual references
|
| 697 |
references = [line.strip() for line in response.split('\n') if line.strip() and line.strip().startswith('[')]
|
|
|
|
| 691 |
user_prompt = f"Format these sources in IEEE style:\n\n{source_data}"
|
| 692 |
|
| 693 |
selection = {"provider": "nvidia", "model": "meta/llama-3.1-8b-instruct"}
|
| 694 |
+
response = await generate_answer_with_model(selection, sys_prompt, user_prompt, None, nvidia_rotator, user_id="system", context="pdf_citation")
|
| 695 |
|
| 696 |
# Parse the response into individual references
|
| 697 |
references = [line.strip() for line in response.split('\n') if line.strip() and line.strip().startswith('[')]
|