sample_id
stringlengths
21
196
text
stringlengths
105
936k
metadata
dict
category
stringclasses
6 values
browser-use/browser-use:tests/scripts/debug_iframe_scrolling.py
""" Debug test for iframe scrolling issue where DOM tree only shows top elements after scrolling. This test verifies that after scrolling inside an iframe, the selector_map correctly contains lower input elements like City, State, Zip Code, etc. """ import asyncio import sys from pathlib import Path # Add parent directory to path to import browser_use modules sys.path.insert(0, str(Path(__file__).parent.parent.parent)) from browser_use.agent.service import Agent from browser_use.agent.views import ActionModel from browser_use.browser import BrowserProfile, BrowserSession from browser_use.browser.events import BrowserStateRequestEvent # Import the mock LLM helper from conftest from tests.ci.conftest import create_mock_llm async def debug_iframe_scrolling(): """Debug iframe scrolling and DOM visibility issue.""" print('Starting iframe scrolling debug test...') # Create the sequence of actions for the mock LLM # We need to format these as the LLM would return them actions = [ # First action: Navigate to the test URL """ { "thinking": "Navigating to the iframe test page", "evaluation_previous_goal": null, "memory": "Starting test", "next_goal": "Navigate to the iframe test page", "action": [ { "navigate": { "url": "https://browser-use.github.io/stress-tests/challenges/iframe-inception-level1.html", "new_tab": false } } ] } """, # Second action: Input text in the first name field (to verify we can interact) """ { "thinking": "Inputting text in the first name field to test interaction", "evaluation_previous_goal": "Successfully navigated to the page", "memory": "Page loaded with nested iframes", "next_goal": "Type text in the first name field", "action": [ { "input_text": { "index": 1, "text": "TestName" } } ] } """, # Third action: Scroll the iframe (element_index=2 should be the iframe) """ { "thinking": "Scrolling inside the iframe to reveal lower form elements", "evaluation_previous_goal": "Successfully typed in first name field", "memory": "Typed TestName in first field", "next_goal": "Scroll inside the innermost iframe to see more form fields", "action": [ { "scroll": { "down": true, "num_pages": 1.0, "index": 2 } } ] } """, # Fourth action: Done """ { "thinking": "Completed scrolling, ready to inspect DOM", "evaluation_previous_goal": "Successfully scrolled inside iframe", "memory": "Scrolled to reveal lower form fields", "next_goal": "Task completed", "action": [ { "done": { "text": "Scrolling completed", "success": true } } ] } """, ] # Create mock LLM with our action sequence mock_llm = create_mock_llm(actions=actions) # Create browser session with headless=False so we can see what's happening browser_session = BrowserSession( browser_profile=BrowserProfile( headless=False, # Set to False to see the browser user_data_dir=None, # Use temporary directory keep_alive=True, enable_default_extensions=True, cross_origin_iframes=True, # Enable cross-origin iframe support ) ) try: # Start the browser session await browser_session.start() print('Browser session started') # Create an agent with the mock LLM agent = Agent( task='Navigate to the iframe test page and scroll inside the iframe', llm=mock_llm, browser_session=browser_session, ) # Helper function to capture and analyze DOM state async def capture_dom_state(label: str) -> dict: """Capture DOM state and return analysis""" print(f'\nπŸ“Έ Capturing DOM state: {label}') state_event = browser_session.event_bus.dispatch( BrowserStateRequestEvent(include_dom=True, include_screenshot=False, include_recent_events=False) ) browser_state = await state_event.event_result() if browser_state and browser_state.dom_state and browser_state.dom_state.selector_map: selector_map = browser_state.dom_state.selector_map element_count = len(selector_map) # Check for specific elements found_elements = {} expected_checks = [ ('First Name', ['firstName', 'first name']), ('Last Name', ['lastName', 'last name']), ('Email', ['email']), ('City', ['city']), ('State', ['state']), ('Zip', ['zip', 'zipCode']), ] for name, keywords in expected_checks: for index, element in selector_map.items(): element_str = str(element).lower() if any(kw.lower() in element_str for kw in keywords): found_elements[name] = True break return { 'label': label, 'total_elements': element_count, 'found_elements': found_elements, 'selector_map': selector_map, } return {'label': label, 'error': 'No DOM state available'} # Capture initial state before any actions print('\n' + '=' * 80) print('PHASE 1: INITIAL PAGE LOAD') print('=' * 80) # Navigate to the page first from browser_use.tools.service import Tools tools = Tools() # Create the action model for navigation goto_action = ActionModel.model_validate_json(actions[0]) await tools.act(goto_action, browser_session) await asyncio.sleep(2) # Wait for page to fully load initial_state = await capture_dom_state('INITIAL (after page load)') # Now run the rest of the actions via the agent print('\n' + '=' * 80) print('PHASE 2: EXECUTING ACTIONS') print('=' * 80) # Create new agent with remaining actions remaining_actions = actions[1:] # Skip the navigation we already did mock_llm_remaining = create_mock_llm(actions=remaining_actions) agent = Agent( task='Input text and scroll inside the iframe', llm=mock_llm_remaining, browser_session=browser_session, ) # Hook into agent actions to capture state after each one states = [] original_act = tools.act async def wrapped_act(action, session): result = await original_act(action, session) # Capture state after each action action_type = 'unknown' if hasattr(action, 'input_text') and action.input_text: action_type = 'input_text' await asyncio.sleep(1) # Give time for DOM to update state = await capture_dom_state('AFTER INPUT_TEXT') states.append(state) elif hasattr(action, 'scroll') and action.scroll: action_type = 'scroll' await asyncio.sleep(2) # Give more time after scroll state = await capture_dom_state('AFTER SCROLL') states.append(state) return result tools.act = wrapped_act # Run the agent with remaining actions result = await agent.run() print(f'\nAgent completed with result: {result}') # Analyze all captured states print('\n' + '=' * 80) print('PHASE 3: ANALYSIS OF DOM STATES') print('=' * 80) all_states = [initial_state] + states for state in all_states: if 'error' in state: print(f'\n❌ {state["label"]}: {state["error"]}') else: print(f'\nπŸ“Š {state["label"]}:') print(f' Total elements: {state["total_elements"]}') print(' Found elements:') for elem_name, found in state['found_elements'].items(): status = 'βœ“' if found else 'βœ—' print(f' {status} {elem_name}') # Compare states print('\n' + '=' * 80) print('COMPARISON SUMMARY') print('=' * 80) if len(all_states) >= 3: initial = all_states[0] after_input = all_states[1] if len(all_states) > 1 else None after_scroll = all_states[2] if len(all_states) > 2 else None print('\nElement count changes:') print(f' Initial: {initial.get("total_elements", 0)} elements') if after_input: print(f' After input_text: {after_input.get("total_elements", 0)} elements') if after_scroll: print(f' After scroll: {after_scroll.get("total_elements", 0)} elements') # Check if lower form fields appear after scroll if after_scroll and 'found_elements' in after_scroll: lower_fields = ['City', 'State', 'Zip'] missing_fields = [f for f in lower_fields if not after_scroll['found_elements'].get(f, False)] if missing_fields: print('\n⚠️ BUG CONFIRMED: Lower form fields missing after scroll:') for field in missing_fields: print(f' βœ— {field}') print('\nThis confirms that scrolling inside iframes does not update the DOM tree properly.') else: print('\nβœ… SUCCESS: All lower form fields are visible after scrolling!') # Show first few elements from final state for debugging if states and 'selector_map' in states[-1]: print('\n' + '=' * 80) print('DEBUG: First 5 elements in final selector_map') print('=' * 80) final_map = states[-1]['selector_map'] for i, (index, element) in enumerate(list(final_map.items())[:5]): elem_preview = str(element)[:150] print(f'\n [{index}]: {elem_preview}...') # Keep browser open for manual inspection if needed print('\n' + '=' * 80) print('Test complete. Browser will remain open for 10 seconds for inspection...') print('=' * 80) await asyncio.sleep(10) finally: # Clean up print('\nCleaning up...') await browser_session.kill() await browser_session.event_bus.stop(clear=True, timeout=5) print('Browser session closed') if __name__ == '__main__': # Run the debug test asyncio.run(debug_iframe_scrolling())
{ "repo_id": "browser-use/browser-use", "file_path": "tests/scripts/debug_iframe_scrolling.py", "license": "MIT License", "lines": 257, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
browser-use/browser-use:tests/scripts/test_frame_hierarchy.py
#!/usr/bin/env python3 """Test frame hierarchy for any URL passed as argument.""" import asyncio import sys from browser_use.browser import BrowserSession from browser_use.browser.events import BrowserStartEvent from browser_use.browser.profile import BrowserProfile async def analyze_frame_hierarchy(url): """Analyze and display complete frame hierarchy for a URL.""" profile = BrowserProfile(headless=True, user_data_dir=None) session = BrowserSession(browser_profile=profile) try: print('πŸš€ Starting browser...') await session.on_BrowserStartEvent(BrowserStartEvent()) print(f'πŸ“ Navigating to: {url}') await session._cdp_navigate(url) await asyncio.sleep(3) print('\n' + '=' * 80) print('FRAME HIERARCHY ANALYSIS') print('=' * 80) # Get all targets from SessionManager all_targets = session.session_manager.get_all_targets() # Separate by type page_targets = [target for target in all_targets.values() if target.target_type == 'page'] iframe_targets = [target for target in all_targets.values() if target.target_type == 'iframe'] print('\nπŸ“Š Target Summary:') print(f' Total targets: {len(all_targets)}') print(f' Page targets: {len(page_targets)}') print(f' Iframe targets (OOPIFs): {len(iframe_targets)}') # Show all targets print('\nπŸ“‹ All Targets:') for i, (target_id, target) in enumerate(all_targets.items()): if target.target_type in ['page', 'iframe']: print(f'\n [{i + 1}] Type: {target.target_type}') print(f' URL: {target.url}') print(f' Target ID: {target.target_id[:30]}...') # Check if target has active sessions using the public API try: cdp_session = await session.get_or_create_cdp_session(target.target_id, focus=False) has_session = cdp_session is not None except Exception: has_session = False print(f' Has Session: {has_session}') # Get main page frame tree main_target = next((t for t in page_targets if url in t.url), page_targets[0] if page_targets else None) if main_target: print('\nπŸ“ Main Page Frame Tree:') print(f' Target: {main_target.url}') print(f' Target ID: {main_target.target_id[:30]}...') s = await session.cdp_client.send.Target.attachToTarget(params={'targetId': main_target.target_id, 'flatten': True}) sid = s['sessionId'] try: await session.cdp_client.send.Page.enable(session_id=sid) tree = await session.cdp_client.send.Page.getFrameTree(session_id=sid) print('\n Frame Tree Structure:') def print_tree(node, indent=0, parent_id=None): frame = node['frame'] frame_id = frame.get('id', 'unknown') frame_url = frame.get('url', 'none') prefix = ' ' * indent + ('└─ ' if indent > 0 else '') print(f'{prefix}Frame: {frame_url}') print(f'{" " * (indent + 1)}ID: {frame_id[:30]}...') if parent_id: print(f'{" " * (indent + 1)}Parent: {parent_id[:30]}...') # Check cross-origin status cross_origin = frame.get('crossOriginIsolatedContextType', 'unknown') if cross_origin != 'NotIsolated': print(f'{" " * (indent + 1)}⚠️ Cross-Origin: {cross_origin}') # Process children for child in node.get('childFrames', []): print_tree(child, indent + 1, frame_id) print_tree(tree['frameTree']) finally: await session.cdp_client.send.Target.detachFromTarget(params={'sessionId': sid}) # Show iframe target trees if iframe_targets: print('\nπŸ”Έ OOPIF Target Frame Trees:') for iframe_target in iframe_targets: print(f'\n OOPIF Target: {iframe_target.url}') print(f' Target ID: {iframe_target.target_id[:30]}...') s = await session.cdp_client.send.Target.attachToTarget( params={'targetId': iframe_target.target_id, 'flatten': True} ) sid = s['sessionId'] try: await session.cdp_client.send.Page.enable(session_id=sid) tree = await session.cdp_client.send.Page.getFrameTree(session_id=sid) frame = tree['frameTree']['frame'] print(f' Frame ID: {frame.get("id", "unknown")[:30]}...') print(f' Frame URL: {frame.get("url", "none")}') print(' ⚠️ This frame runs in a separate process (OOPIF)') except Exception as e: print(f' Error: {e}') finally: await session.cdp_client.send.Target.detachFromTarget(params={'sessionId': sid}) # Now show unified view from get_all_frames print('\n' + '=' * 80) print('UNIFIED FRAME HIERARCHY (get_all_frames method)') print('=' * 80) all_frames, target_sessions = await session.get_all_frames() # Clean up sessions for tid, sess_id in target_sessions.items(): try: await session.cdp_client.send.Target.detachFromTarget(params={'sessionId': sess_id}) except Exception: pass print('\nπŸ“Š Frame Statistics:') print(f' Total frames discovered: {len(all_frames)}') # Separate root and child frames root_frames = [] child_frames = [] for frame_id, frame_info in all_frames.items(): if not frame_info.get('parentFrameId'): root_frames.append((frame_id, frame_info)) else: child_frames.append((frame_id, frame_info)) print(f' Root frames: {len(root_frames)}') print(f' Child frames: {len(child_frames)}') # Display all frames with details print('\nπŸ“‹ All Frames:') for i, (frame_id, frame_info) in enumerate(all_frames.items()): url = frame_info.get('url', 'none') parent = frame_info.get('parentFrameId') target_id = frame_info.get('frameTargetId', 'unknown') is_cross = frame_info.get('isCrossOrigin', False) print(f'\n [{i + 1}] Frame URL: {url}') print(f' Frame ID: {frame_id[:30]}...') print(f' Parent Frame ID: {parent[:30] + "..." if parent else "None (ROOT)"}') print(f' Target ID: {target_id[:30]}...') print(f' Cross-Origin: {is_cross}') # Highlight problems if not parent and 'v0-simple-landing' in url: print(' ❌ PROBLEM: Cross-origin frame incorrectly marked as root!') elif not parent and url != 'about:blank' and url not in ['chrome://newtab/', 'about:blank']: # Check if this should be the main frame if any(url in t.url for t in page_targets): print(' βœ… Correctly identified as root frame') if is_cross: print(' πŸ”Έ This is a cross-origin frame (OOPIF)') # Show parent-child relationships print('\n🌳 Frame Relationships:') # Build a tree structure def print_frame_tree(frame_id, frame_info, indent=0, visited=None): if visited is None: visited = set() if frame_id in visited: return visited.add(frame_id) url = frame_info.get('url', 'none') prefix = ' ' * indent + ('└─ ' if indent > 0 else '') print(f'{prefix}{url[:60]}...') print(f'{" " * (indent + 1)}[{frame_id[:20]}...]') # Find children for child_id, child_info in all_frames.items(): if child_info.get('parentFrameId') == frame_id: print_frame_tree(child_id, child_info, indent + 1, visited) # Print trees starting from roots for frame_id, frame_info in root_frames: print('\n Tree starting from root:') print_frame_tree(frame_id, frame_info) print('\n' + '=' * 80) print('βœ… Analysis complete!') print('=' * 80) except Exception as e: print(f'❌ Error: {e}') import traceback traceback.print_exc() finally: # Stop the CDP client first before killing the browser print('\nπŸ›‘ Shutting down...') # Close CDP connection first while browser is still alive if session._cdp_client_root: try: await session._cdp_client_root.stop() except Exception: pass # Ignore errors if already disconnected # Then stop the browser process from browser_use.browser.events import BrowserStopEvent stop_event = session.event_bus.dispatch(BrowserStopEvent()) try: await asyncio.wait_for(stop_event, timeout=2.0) except TimeoutError: print('⚠️ Browser stop timed out') def main(): if len(sys.argv) != 2: print('Usage: python test_frame_hierarchy.py <URL>') print('\nExample URLs to test:') print(' https://v0-website-with-clickable-elements.vercel.app/nested-iframe') print(' https://v0-website-with-clickable-elements.vercel.app/cross-origin') print(' https://v0-website-with-clickable-elements.vercel.app/shadow-dom') sys.exit(1) url = sys.argv[1] asyncio.run(analyze_frame_hierarchy(url)) # Ensure clean exit print('βœ… Script completed') sys.exit(0) if __name__ == '__main__': main()
{ "repo_id": "browser-use/browser-use", "file_path": "tests/scripts/test_frame_hierarchy.py", "license": "MIT License", "lines": 198, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
browser-use/browser-use:browser_use/cli.py
# pyright: reportMissingImports=false # Check for MCP mode early to prevent logging initialization import sys if '--mcp' in sys.argv: import logging import os os.environ['BROWSER_USE_LOGGING_LEVEL'] = 'critical' os.environ['BROWSER_USE_SETUP_LOGGING'] = 'false' logging.disable(logging.CRITICAL) # Special case: install command doesn't need CLI dependencies if len(sys.argv) > 1 and sys.argv[1] == 'install': import platform import subprocess print('πŸ“¦ Installing Chromium browser + system dependencies...') print('⏳ This may take a few minutes...\n') # Build command - only use --with-deps on Linux (it fails on Windows/macOS) cmd = ['uvx', 'playwright', 'install', 'chromium'] if platform.system() == 'Linux': cmd.append('--with-deps') cmd.append('--no-shell') result = subprocess.run(cmd) if result.returncode == 0: print('\nβœ… Installation complete!') print('πŸš€ Ready to use! Run: uvx browser-use') else: print('\n❌ Installation failed') sys.exit(1) sys.exit(0) # Check for init subcommand early to avoid loading TUI dependencies if 'init' in sys.argv: from browser_use.init_cmd import INIT_TEMPLATES from browser_use.init_cmd import main as init_main # Check if --template or -t flag is present without a value # If so, just remove it and let init_main handle interactive mode if '--template' in sys.argv or '-t' in sys.argv: try: template_idx = sys.argv.index('--template') if '--template' in sys.argv else sys.argv.index('-t') template = sys.argv[template_idx + 1] if template_idx + 1 < len(sys.argv) else None # If template is not provided or is another flag, remove the flag and use interactive mode if not template or template.startswith('-'): if '--template' in sys.argv: sys.argv.remove('--template') else: sys.argv.remove('-t') except (ValueError, IndexError): pass # Remove 'init' from sys.argv so click doesn't see it as an unexpected argument sys.argv.remove('init') init_main() sys.exit(0) # Check for --template flag early to avoid loading TUI dependencies if '--template' in sys.argv: from pathlib import Path import click from browser_use.init_cmd import INIT_TEMPLATES # Parse template and output from sys.argv try: template_idx = sys.argv.index('--template') template = sys.argv[template_idx + 1] if template_idx + 1 < len(sys.argv) else None except (ValueError, IndexError): template = None # If template is not provided or is another flag, use interactive mode if not template or template.startswith('-'): # Redirect to init command with interactive template selection from browser_use.init_cmd import main as init_main # Remove --template from sys.argv sys.argv.remove('--template') init_main() sys.exit(0) # Validate template name if template not in INIT_TEMPLATES: click.echo(f'❌ Invalid template. Choose from: {", ".join(INIT_TEMPLATES.keys())}', err=True) sys.exit(1) # Check for --output flag output = None if '--output' in sys.argv or '-o' in sys.argv: try: output_idx = sys.argv.index('--output') if '--output' in sys.argv else sys.argv.index('-o') output = sys.argv[output_idx + 1] if output_idx + 1 < len(sys.argv) else None except (ValueError, IndexError): pass # Check for --force flag force = '--force' in sys.argv or '-f' in sys.argv # Determine output path output_path = Path(output) if output else Path.cwd() / f'browser_use_{template}.py' # Read and write template try: templates_dir = Path(__file__).parent / 'cli_templates' template_file = INIT_TEMPLATES[template]['file'] template_path = templates_dir / template_file content = template_path.read_text(encoding='utf-8') # Write file with safety checks if output_path.exists() and not force: click.echo(f'⚠️ File already exists: {output_path}') if not click.confirm('Overwrite?', default=False): click.echo('❌ Cancelled') sys.exit(1) output_path.parent.mkdir(parents=True, exist_ok=True) output_path.write_text(content, encoding='utf-8') click.echo(f'βœ… Created {output_path}') click.echo('\nNext steps:') click.echo(' 1. Install browser-use:') click.echo(' uv pip install browser-use') click.echo(' 2. Set up your API key in .env file or environment:') click.echo(' BROWSER_USE_API_KEY=your-key') click.echo(' (Get your key at https://cloud.browser-use.com/new-api-key)') click.echo(' 3. Run your script:') click.echo(f' python {output_path.name}') except Exception as e: click.echo(f'❌ Error: {e}', err=True) sys.exit(1) sys.exit(0) import asyncio import json import logging import os import time from pathlib import Path from typing import Any from dotenv import load_dotenv from browser_use.llm.anthropic.chat import ChatAnthropic from browser_use.llm.google.chat import ChatGoogle from browser_use.llm.openai.chat import ChatOpenAI load_dotenv() from browser_use import Agent, Controller from browser_use.agent.views import AgentSettings from browser_use.browser import BrowserProfile, BrowserSession from browser_use.logging_config import addLoggingLevel from browser_use.telemetry import CLITelemetryEvent, ProductTelemetry from browser_use.utils import get_browser_use_version try: import click from textual import events from textual.app import App, ComposeResult from textual.binding import Binding from textual.containers import Container, HorizontalGroup, VerticalScroll from textual.widgets import Footer, Header, Input, Label, Link, RichLog, Static except ImportError: print( '⚠️ CLI addon is not installed. Please install it with: `pip install "browser-use[cli]"` and try again.', file=sys.stderr ) sys.exit(1) try: import readline READLINE_AVAILABLE = True except ImportError: # readline not available on Windows by default READLINE_AVAILABLE = False os.environ['BROWSER_USE_LOGGING_LEVEL'] = 'result' from browser_use.config import CONFIG # Set USER_DATA_DIR now that CONFIG is imported USER_DATA_DIR = CONFIG.BROWSER_USE_PROFILES_DIR / 'cli' # Ensure directories exist CONFIG.BROWSER_USE_CONFIG_FILE.parent.mkdir(parents=True, exist_ok=True) USER_DATA_DIR.mkdir(parents=True, exist_ok=True) # Default User settings MAX_HISTORY_LENGTH = 100 # Directory setup will happen in functions that need CONFIG # Logo components with styling for rich panels BROWSER_LOGO = """ [white] ++++++ +++++++++ [/] [white] +++ +++++ +++ [/] [white] ++ ++++ ++ ++ [/] [white] ++ +++ +++ ++ [/] [white] ++++ +++ [/] [white] +++ +++ [/] [white] +++ +++ [/] [white] ++ +++ +++ ++ [/] [white] ++ ++++ ++ ++ [/] [white] +++ ++++++ +++ [/] [white] ++++++ +++++++ [/] [white]β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•— β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•— β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•— β–ˆβ–ˆβ•— β–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—[/] [darkorange]β–ˆβ–ˆβ•— β–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—[/] [white]β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•”β•β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•”β•β•β•β•β•β–ˆβ–ˆβ•”β•β•β•β•β•β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—[/] [darkorange]β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•”β•β•β•β•β•β–ˆβ–ˆβ•”β•β•β•β•β•[/] [white]β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•‘ β–ˆβ•— β–ˆβ–ˆβ•‘β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•— β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•[/] [darkorange]β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—[/] [white]β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•‘β–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘β•šβ•β•β•β•β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•”β•β•β• β–ˆβ–ˆβ•”β•β•β–ˆβ–ˆβ•—[/] [darkorange]β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β•šβ•β•β•β•β–ˆβ–ˆβ•‘β–ˆβ–ˆβ•”β•β•β•[/] [white]β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘β•šβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•β•šβ–ˆβ–ˆβ–ˆβ•”β–ˆβ–ˆβ–ˆβ•”β•β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•‘β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—β–ˆβ–ˆβ•‘ β–ˆβ–ˆβ•‘[/] [darkorange]β•šβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•”β•β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•‘β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ•—[/] [white]β•šβ•β•β•β•β•β• β•šβ•β• β•šβ•β• β•šβ•β•β•β•β•β• β•šβ•β•β•β•šβ•β•β• β•šβ•β•β•β•β•β•β•β•šβ•β•β•β•β•β•β•β•šβ•β• β•šβ•β•[/] [darkorange]β•šβ•β•β•β•β•β• β•šβ•β•β•β•β•β•β•β•šβ•β•β•β•β•β•β•[/] """ # Common UI constants TEXTUAL_BORDER_STYLES = {'logo': 'blue', 'info': 'blue', 'input': 'orange3', 'working': 'yellow', 'completion': 'green'} def get_default_config() -> dict[str, Any]: """Return default configuration dictionary using the new config system.""" # Load config from the new config system config_data = CONFIG.load_config() # Extract browser profile, llm, and agent configs browser_profile = config_data.get('browser_profile', {}) llm_config = config_data.get('llm', {}) agent_config = config_data.get('agent', {}) return { 'model': { 'name': llm_config.get('model'), 'temperature': llm_config.get('temperature', 0.0), 'api_keys': { 'OPENAI_API_KEY': llm_config.get('api_key', CONFIG.OPENAI_API_KEY), 'ANTHROPIC_API_KEY': CONFIG.ANTHROPIC_API_KEY, 'GOOGLE_API_KEY': CONFIG.GOOGLE_API_KEY, 'DEEPSEEK_API_KEY': CONFIG.DEEPSEEK_API_KEY, 'GROK_API_KEY': CONFIG.GROK_API_KEY, }, }, 'agent': agent_config, 'browser': { 'headless': browser_profile.get('headless', True), 'keep_alive': browser_profile.get('keep_alive', True), 'ignore_https_errors': browser_profile.get('ignore_https_errors', False), 'user_data_dir': browser_profile.get('user_data_dir'), 'allowed_domains': browser_profile.get('allowed_domains'), 'wait_between_actions': browser_profile.get('wait_between_actions'), 'is_mobile': browser_profile.get('is_mobile'), 'device_scale_factor': browser_profile.get('device_scale_factor'), 'disable_security': browser_profile.get('disable_security'), }, 'command_history': [], } def load_user_config() -> dict[str, Any]: """Load user configuration using the new config system.""" # Just get the default config which already loads from the new system config = get_default_config() # Load command history from a separate file if it exists history_file = CONFIG.BROWSER_USE_CONFIG_DIR / 'command_history.json' if history_file.exists(): try: with open(history_file) as f: config['command_history'] = json.load(f) except (FileNotFoundError, json.JSONDecodeError): config['command_history'] = [] return config def save_user_config(config: dict[str, Any]) -> None: """Save command history only (config is saved via the new system).""" # Only save command history to a separate file if 'command_history' in config and isinstance(config['command_history'], list): # Ensure command history doesn't exceed maximum length history = config['command_history'] if len(history) > MAX_HISTORY_LENGTH: history = history[-MAX_HISTORY_LENGTH:] # Save to separate history file history_file = CONFIG.BROWSER_USE_CONFIG_DIR / 'command_history.json' with open(history_file, 'w', encoding='utf-8') as f: json.dump(history, f, indent=2, ensure_ascii=False) def update_config_with_click_args(config: dict[str, Any], ctx: click.Context) -> dict[str, Any]: """Update configuration with command-line arguments.""" # Ensure required sections exist if 'model' not in config: config['model'] = {} if 'browser' not in config: config['browser'] = {} # Update configuration with command-line args if provided if ctx.params.get('model'): config['model']['name'] = ctx.params['model'] if ctx.params.get('headless') is not None: config['browser']['headless'] = ctx.params['headless'] if ctx.params.get('window_width'): config['browser']['window_width'] = ctx.params['window_width'] if ctx.params.get('window_height'): config['browser']['window_height'] = ctx.params['window_height'] if ctx.params.get('user_data_dir'): config['browser']['user_data_dir'] = ctx.params['user_data_dir'] if ctx.params.get('profile_directory'): config['browser']['profile_directory'] = ctx.params['profile_directory'] if ctx.params.get('cdp_url'): config['browser']['cdp_url'] = ctx.params['cdp_url'] # Consolidated proxy dict proxy: dict[str, str] = {} if ctx.params.get('proxy_url'): proxy['server'] = ctx.params['proxy_url'] if ctx.params.get('no_proxy'): # Store as comma-separated list string to match Chrome flag proxy['bypass'] = ','.join([p.strip() for p in ctx.params['no_proxy'].split(',') if p.strip()]) if ctx.params.get('proxy_username'): proxy['username'] = ctx.params['proxy_username'] if ctx.params.get('proxy_password'): proxy['password'] = ctx.params['proxy_password'] if proxy: config['browser']['proxy'] = proxy return config def setup_readline_history(history: list[str]) -> None: """Set up readline with command history.""" if not READLINE_AVAILABLE: return # Add history items to readline for item in history: readline.add_history(item) def get_llm(config: dict[str, Any]): """Get the language model based on config and available API keys.""" model_config = config.get('model', {}) model_name = model_config.get('name') temperature = model_config.get('temperature', 0.0) # Get API key from config or environment api_key = model_config.get('api_keys', {}).get('OPENAI_API_KEY') or CONFIG.OPENAI_API_KEY if model_name: if model_name.startswith('gpt'): if not api_key and not CONFIG.OPENAI_API_KEY: print('⚠️ OpenAI API key not found. Please update your config or set OPENAI_API_KEY environment variable.') sys.exit(1) return ChatOpenAI(model=model_name, temperature=temperature, api_key=api_key or CONFIG.OPENAI_API_KEY) elif model_name.startswith('claude'): if not CONFIG.ANTHROPIC_API_KEY: print('⚠️ Anthropic API key not found. Please update your config or set ANTHROPIC_API_KEY environment variable.') sys.exit(1) return ChatAnthropic(model=model_name, temperature=temperature) elif model_name.startswith('gemini'): if not CONFIG.GOOGLE_API_KEY: print('⚠️ Google API key not found. Please update your config or set GOOGLE_API_KEY environment variable.') sys.exit(1) return ChatGoogle(model=model_name, temperature=temperature) elif model_name.startswith('oci'): # OCI models require additional configuration print( '⚠️ OCI models require manual configuration. Please use the ChatOCIRaw class directly with your OCI credentials.' ) sys.exit(1) # Auto-detect based on available API keys if api_key or CONFIG.OPENAI_API_KEY: return ChatOpenAI(model='gpt-5-mini', temperature=temperature, api_key=api_key or CONFIG.OPENAI_API_KEY) elif CONFIG.ANTHROPIC_API_KEY: return ChatAnthropic(model='claude-4-sonnet', temperature=temperature) elif CONFIG.GOOGLE_API_KEY: return ChatGoogle(model='gemini-2.5-pro', temperature=temperature) else: print( '⚠️ No API keys found. Please update your config or set one of: OPENAI_API_KEY, ANTHROPIC_API_KEY, or GOOGLE_API_KEY.' ) sys.exit(1) class RichLogHandler(logging.Handler): """Custom logging handler that redirects logs to a RichLog widget.""" def __init__(self, rich_log: RichLog): super().__init__() self.rich_log = rich_log def emit(self, record): try: msg = self.format(record) self.rich_log.write(msg) except Exception: self.handleError(record) class BrowserUseApp(App): """Browser-use TUI application.""" # Make it an inline app instead of fullscreen # MODES = {"light"} # Ensure app is inline, not fullscreen CSS = """ #main-container { height: 100%; layout: vertical; } #logo-panel, #links-panel, #paths-panel, #info-panels { border: solid $primary; margin: 0 0 0 0; padding: 0; } #info-panels { display: none; layout: vertical; height: auto; min-height: 5; margin: 0 0 1 0; } #top-panels { layout: horizontal; height: auto; width: 100%; } #browser-panel, #model-panel { width: 1fr; height: 100%; padding: 1; border-right: solid $primary; } #model-panel { border-right: none; } #tasks-panel { height: auto; max-height: 10; overflow-y: scroll; padding: 1; border-top: solid $primary; } #browser-info, #model-info, #tasks-info { height: auto; margin: 0; padding: 0; background: transparent; overflow-y: auto; min-height: 3; } #three-column-container { height: 1fr; layout: horizontal; width: 100%; display: none; } #main-output-column { width: 1fr; height: 100%; border: solid $primary; padding: 0; margin: 0 1 0 0; } #events-column { width: 1fr; height: 100%; border: solid $warning; padding: 0; margin: 0 1 0 0; } #cdp-column { width: 1fr; height: 100%; border: solid $accent; padding: 0; margin: 0; } #main-output-log, #events-log, #cdp-log { height: 100%; overflow-y: scroll; background: $surface; color: $text; width: 100%; padding: 1; } #events-log { color: $warning; } #cdp-log { color: $accent-lighten-2; } #logo-panel { width: 100%; height: auto; content-align: center middle; text-align: center; } #links-panel { width: 100%; padding: 1; border: solid $primary; height: auto; } .link-white { color: white; } .link-purple { color: purple; } .link-magenta { color: magenta; } .link-green { color: green; } HorizontalGroup { height: auto; } .link-label { width: auto; } .link-url { width: auto; } .link-row { width: 100%; height: auto; } #paths-panel { color: $text-muted; } #task-input-container { border: solid $accent; padding: 1; margin-bottom: 1; height: auto; dock: bottom; } #task-label { color: $accent; padding-bottom: 1; } #task-input { width: 100%; } """ BINDINGS = [ Binding('ctrl+c', 'quit', 'Quit', priority=True, show=True), Binding('ctrl+q', 'quit', 'Quit', priority=True), Binding('ctrl+d', 'quit', 'Quit', priority=True), Binding('up', 'input_history_prev', 'Previous command', show=False), Binding('down', 'input_history_next', 'Next command', show=False), ] def __init__(self, config: dict[str, Any], *args, **kwargs): super().__init__(*args, **kwargs) self.config = config self.browser_session: BrowserSession | None = None # Will be set before app.run_async() self.controller: Controller | None = None # Will be set before app.run_async() self.agent: Agent | None = None self.llm: Any | None = None # Will be set before app.run_async() self.task_history = config.get('command_history', []) # Track current position in history for up/down navigation self.history_index = len(self.task_history) # Initialize telemetry self._telemetry = ProductTelemetry() # Store for event bus handler self._event_bus_handler_id = None self._event_bus_handler_func = None # Timer for info panel updates self._info_panel_timer = None def setup_richlog_logging(self) -> None: """Set up logging to redirect to RichLog widget instead of stdout.""" # Try to add RESULT level if it doesn't exist try: addLoggingLevel('RESULT', 35) except AttributeError: pass # Level already exists, which is fine # Get the main output RichLog widget rich_log = self.query_one('#main-output-log', RichLog) # Create and set up the custom handler log_handler = RichLogHandler(rich_log) log_type = os.getenv('BROWSER_USE_LOGGING_LEVEL', 'result').lower() class BrowserUseFormatter(logging.Formatter): def format(self, record): # if isinstance(record.name, str) and record.name.startswith('browser_use.'): # record.name = record.name.split('.')[-2] return super().format(record) # Set up the formatter based on log type if log_type == 'result': log_handler.setLevel('RESULT') log_handler.setFormatter(BrowserUseFormatter('%(message)s')) else: log_handler.setFormatter(BrowserUseFormatter('%(levelname)-8s [%(name)s] %(message)s')) # Configure root logger - Replace ALL handlers, not just stdout handlers root = logging.getLogger() # Clear all existing handlers to prevent output to stdout/stderr root.handlers = [] root.addHandler(log_handler) # Set log level based on environment variable if log_type == 'result': root.setLevel('RESULT') elif log_type == 'debug': root.setLevel(logging.DEBUG) else: root.setLevel(logging.INFO) # Configure browser_use logger and all its sub-loggers browser_use_logger = logging.getLogger('browser_use') browser_use_logger.propagate = False # Don't propagate to root logger browser_use_logger.handlers = [log_handler] # Replace any existing handlers browser_use_logger.setLevel(root.level) # Also ensure agent loggers go to the main output # Use a wildcard pattern to catch all agent-related loggers for logger_name in ['browser_use.Agent', 'browser_use.controller', 'browser_use.agent', 'browser_use.agent.service']: agent_logger = logging.getLogger(logger_name) agent_logger.propagate = False agent_logger.handlers = [log_handler] agent_logger.setLevel(root.level) # Also catch any dynamically created agent loggers with task IDs for name, logger in logging.Logger.manager.loggerDict.items(): if isinstance(name, str) and 'browser_use.Agent' in name: if isinstance(logger, logging.Logger): logger.propagate = False logger.handlers = [log_handler] logger.setLevel(root.level) # Silence third-party loggers but keep them using our handler for logger_name in [ 'WDM', 'httpx', 'selenium', 'playwright', 'urllib3', 'asyncio', 'openai', 'httpcore', 'charset_normalizer', 'anthropic._base_client', 'PIL.PngImagePlugin', 'trafilatura.htmlprocessing', 'trafilatura', 'groq', 'portalocker', 'portalocker.utils', ]: third_party = logging.getLogger(logger_name) third_party.setLevel(logging.ERROR) third_party.propagate = False third_party.handlers = [log_handler] # Use our handler to prevent stdout/stderr leakage def on_mount(self) -> None: """Set up components when app is mounted.""" # We'll use a file logger since stdout is now controlled by Textual logger = logging.getLogger('browser_use.on_mount') logger.debug('on_mount() method started') # Step 1: Set up custom logging to RichLog logger.debug('Setting up RichLog logging...') try: self.setup_richlog_logging() logger.debug('RichLog logging set up successfully') except Exception as e: logger.error(f'Error setting up RichLog logging: {str(e)}', exc_info=True) raise RuntimeError(f'Failed to set up RichLog logging: {str(e)}') # Step 2: Set up input history logger.debug('Setting up readline history...') try: if READLINE_AVAILABLE and self.task_history: for item in self.task_history: readline.add_history(item) logger.debug(f'Added {len(self.task_history)} items to readline history') else: logger.debug('No readline history to set up') except Exception as e: logger.error(f'Error setting up readline history: {str(e)}', exc_info=False) # Non-critical, continue # Step 3: Focus the input field logger.debug('Focusing input field...') try: input_field = self.query_one('#task-input', Input) input_field.focus() logger.debug('Input field focused') except Exception as e: logger.error(f'Error focusing input field: {str(e)}', exc_info=True) # Non-critical, continue # Step 5: Setup CDP logger and event bus listener if browser session is available logger.debug('Setting up CDP logging and event bus listener...') try: self.setup_cdp_logger() if self.browser_session: self.setup_event_bus_listener() logger.debug('CDP logging and event bus setup complete') except Exception as e: logger.error(f'Error setting up CDP logging/event bus: {str(e)}', exc_info=True) # Non-critical, continue # Capture telemetry for CLI start self._telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='start', mode='interactive', model=self.llm.model if self.llm and hasattr(self.llm, 'model') else None, model_provider=self.llm.provider if self.llm and hasattr(self.llm, 'provider') else None, ) ) logger.debug('on_mount() completed successfully') def on_input_key_up(self, event: events.Key) -> None: """Handle up arrow key in the input field.""" # For textual key events, we need to check focus manually input_field = self.query_one('#task-input', Input) if not input_field.has_focus: return # Only process if we have history if not self.task_history: return # Move back in history if possible if self.history_index > 0: self.history_index -= 1 task_input = self.query_one('#task-input', Input) task_input.value = self.task_history[self.history_index] # Move cursor to end of text task_input.cursor_position = len(task_input.value) # Prevent default behavior (cursor movement) event.prevent_default() event.stop() def on_input_key_down(self, event: events.Key) -> None: """Handle down arrow key in the input field.""" # For textual key events, we need to check focus manually input_field = self.query_one('#task-input', Input) if not input_field.has_focus: return # Only process if we have history if not self.task_history: return # Move forward in history or clear input if at the end if self.history_index < len(self.task_history) - 1: self.history_index += 1 task_input = self.query_one('#task-input', Input) task_input.value = self.task_history[self.history_index] # Move cursor to end of text task_input.cursor_position = len(task_input.value) elif self.history_index == len(self.task_history) - 1: # At the end of history, go to "new line" state self.history_index += 1 self.query_one('#task-input', Input).value = '' # Prevent default behavior (cursor movement) event.prevent_default() event.stop() async def on_key(self, event: events.Key) -> None: """Handle key events at the app level to ensure graceful exit.""" # Handle Ctrl+C, Ctrl+D, and Ctrl+Q for app exit if event.key == 'ctrl+c' or event.key == 'ctrl+d' or event.key == 'ctrl+q': await self.action_quit() event.stop() event.prevent_default() def on_input_submitted(self, event: Input.Submitted) -> None: """Handle task input submission.""" if event.input.id == 'task-input': task = event.input.value if not task.strip(): return # Add to history if it's new if task.strip() and (not self.task_history or task != self.task_history[-1]): self.task_history.append(task) self.config['command_history'] = self.task_history save_user_config(self.config) # Reset history index to point past the end of history self.history_index = len(self.task_history) # Hide logo, links, and paths panels self.hide_intro_panels() # Process the task self.run_task(task) # Clear the input event.input.value = '' def hide_intro_panels(self) -> None: """Hide the intro panels, show info panels and the three-column view.""" try: # Get the panels logo_panel = self.query_one('#logo-panel') links_panel = self.query_one('#links-panel') paths_panel = self.query_one('#paths-panel') info_panels = self.query_one('#info-panels') three_column = self.query_one('#three-column-container') # Hide intro panels if they're visible and show info panels + three-column view if logo_panel.display: logging.debug('Hiding intro panels and showing info panels + three-column view') logo_panel.display = False links_panel.display = False paths_panel.display = False # Show info panels and three-column container info_panels.display = True three_column.display = True # Start updating info panels self.update_info_panels() logging.debug('Info panels and three-column view should now be visible') except Exception as e: logging.error(f'Error in hide_intro_panels: {str(e)}') def setup_event_bus_listener(self) -> None: """Setup listener for browser session event bus.""" if not self.browser_session or not self.browser_session.event_bus: return # Clean up any existing handler before registering a new one if self._event_bus_handler_func is not None: try: # Remove handler from the event bus's internal handlers dict if hasattr(self.browser_session.event_bus, 'handlers'): # Find and remove our handler function from all event patterns for event_type, handler_list in list(self.browser_session.event_bus.handlers.items()): # Remove our specific handler function object if self._event_bus_handler_func in handler_list: handler_list.remove(self._event_bus_handler_func) logging.debug(f'Removed old handler from event type: {event_type}') except Exception as e: logging.debug(f'Error cleaning up event bus handler: {e}') self._event_bus_handler_func = None self._event_bus_handler_id = None try: # Get the events log widget events_log = self.query_one('#events-log', RichLog) except Exception: # Widget not ready yet return # Create handler to log all events def log_event(event): event_name = event.__class__.__name__ # Format event data nicely try: if hasattr(event, 'model_dump'): event_data = event.model_dump(exclude_unset=True) # Remove large fields if 'screenshot' in event_data: event_data['screenshot'] = '<bytes>' if 'dom_state' in event_data: event_data['dom_state'] = '<truncated>' event_str = str(event_data) if event_data else '' else: event_str = str(event) # Truncate long strings if len(event_str) > 200: event_str = event_str[:200] + '...' events_log.write(f'[yellow]β†’ {event_name}[/] {event_str}') except Exception as e: events_log.write(f'[red]β†’ {event_name}[/] (error formatting: {e})') # Store the handler function before registering it self._event_bus_handler_func = log_event self._event_bus_handler_id = id(log_event) # Register wildcard handler for all events self.browser_session.event_bus.on('*', log_event) logging.debug(f'Registered new event bus handler with id: {self._event_bus_handler_id}') def setup_cdp_logger(self) -> None: """Setup CDP message logger to capture already-transformed CDP logs.""" # No need to configure levels - setup_logging() already handles that # We just need to capture the transformed logs and route them to the CDP pane # Get the CDP log widget cdp_log = self.query_one('#cdp-log', RichLog) # Create custom handler for CDP logging class CDPLogHandler(logging.Handler): def __init__(self, rich_log: RichLog): super().__init__() self.rich_log = rich_log def emit(self, record): try: msg = self.format(record) # Truncate very long messages if len(msg) > 300: msg = msg[:300] + '...' # Color code by level if record.levelno >= logging.ERROR: self.rich_log.write(f'[red]{msg}[/]') elif record.levelno >= logging.WARNING: self.rich_log.write(f'[yellow]{msg}[/]') else: self.rich_log.write(f'[cyan]{msg}[/]') except Exception: self.handleError(record) # Setup handler for cdp_use loggers cdp_handler = CDPLogHandler(cdp_log) cdp_handler.setFormatter(logging.Formatter('%(message)s')) cdp_handler.setLevel(logging.DEBUG) # Route CDP logs to the CDP pane # These are already transformed by cdp_use and at the right level from setup_logging for logger_name in ['websockets.client', 'cdp_use', 'cdp_use.client', 'cdp_use.cdp', 'cdp_use.cdp.registry']: logger = logging.getLogger(logger_name) # Add our handler (don't replace - keep existing console handler too) if cdp_handler not in logger.handlers: logger.addHandler(cdp_handler) def scroll_to_input(self) -> None: """Scroll to the input field to ensure it's visible.""" input_container = self.query_one('#task-input-container') input_container.scroll_visible() def run_task(self, task: str) -> None: """Launch the task in a background worker.""" # Create or update the agent agent_settings = AgentSettings.model_validate(self.config.get('agent', {})) # Get the logger logger = logging.getLogger('browser_use.app') # Make sure intro is hidden and log is ready self.hide_intro_panels() # Clear the main output log to start fresh rich_log = self.query_one('#main-output-log', RichLog) rich_log.clear() if self.agent is None: if not self.llm: raise RuntimeError('LLM not initialized') self.agent = Agent( task=task, llm=self.llm, controller=self.controller if self.controller else Controller(), browser_session=self.browser_session, source='cli', **agent_settings.model_dump(), ) # Update our browser_session reference to point to the agent's if hasattr(self.agent, 'browser_session'): self.browser_session = self.agent.browser_session # Set up event bus listener (will clean up any old handler first) self.setup_event_bus_listener() else: self.agent.add_new_task(task) # Let the agent run in the background async def agent_task_worker() -> None: logger.debug('\nπŸš€ Working on task: %s', task) # Set flags to indicate the agent is running if self.agent: self.agent.running = True # type: ignore self.agent.last_response_time = 0 # type: ignore # Panel updates are already happening via the timer in update_info_panels task_start_time = time.time() error_msg = None try: # Capture telemetry for message sent self._telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='message_sent', mode='interactive', model=self.llm.model if self.llm and hasattr(self.llm, 'model') else None, model_provider=self.llm.provider if self.llm and hasattr(self.llm, 'provider') else None, ) ) # Run the agent task, redirecting output to RichLog through our handler if self.agent: await self.agent.run() except Exception as e: error_msg = str(e) logger.error('\nError running agent: %s', str(e)) finally: # Clear the running flag if self.agent: self.agent.running = False # type: ignore # Capture telemetry for task completion duration = time.time() - task_start_time self._telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='task_completed' if error_msg is None else 'error', mode='interactive', model=self.llm.model if self.llm and hasattr(self.llm, 'model') else None, model_provider=self.llm.provider if self.llm and hasattr(self.llm, 'provider') else None, duration_seconds=duration, error_message=error_msg, ) ) logger.debug('\nβœ… Task completed!') # Make sure the task input container is visible task_input_container = self.query_one('#task-input-container') task_input_container.display = True # Refocus the input field input_field = self.query_one('#task-input', Input) input_field.focus() # Ensure the input is visible by scrolling to it self.call_after_refresh(self.scroll_to_input) # Run the worker self.run_worker(agent_task_worker, name='agent_task') def action_input_history_prev(self) -> None: """Navigate to the previous item in command history.""" # Only process if we have history and input is focused input_field = self.query_one('#task-input', Input) if not input_field.has_focus or not self.task_history: return # Move back in history if possible if self.history_index > 0: self.history_index -= 1 input_field.value = self.task_history[self.history_index] # Move cursor to end of text input_field.cursor_position = len(input_field.value) def action_input_history_next(self) -> None: """Navigate to the next item in command history or clear input.""" # Only process if we have history and input is focused input_field = self.query_one('#task-input', Input) if not input_field.has_focus or not self.task_history: return # Move forward in history or clear input if at the end if self.history_index < len(self.task_history) - 1: self.history_index += 1 input_field.value = self.task_history[self.history_index] # Move cursor to end of text input_field.cursor_position = len(input_field.value) elif self.history_index == len(self.task_history) - 1: # At the end of history, go to "new line" state self.history_index += 1 input_field.value = '' async def action_quit(self) -> None: """Quit the application and clean up resources.""" # Note: We don't need to close the browser session here because: # 1. If an agent exists, it already called browser_session.stop() in its run() method # 2. If keep_alive=True (default), we want to leave the browser running anyway # This prevents the duplicate "stop() called" messages in the logs # Flush telemetry before exiting self._telemetry.flush() # Exit the application self.exit() print('\nTry running tasks on our cloud: https://browser-use.com') def compose(self) -> ComposeResult: """Create the UI layout.""" yield Header() # Main container for app content with Container(id='main-container'): # Logo panel yield Static(BROWSER_LOGO, id='logo-panel', markup=True) # Links panel with URLs with Container(id='links-panel'): with HorizontalGroup(classes='link-row'): yield Static('Run at scale on cloud: [blink]☁️[/] ', markup=True, classes='link-label') yield Link('https://browser-use.com', url='https://browser-use.com', classes='link-white link-url') yield Static('') # Empty line with HorizontalGroup(classes='link-row'): yield Static('Chat & share on Discord: πŸš€ ', markup=True, classes='link-label') yield Link( 'https://discord.gg/ESAUZAdxXY', url='https://discord.gg/ESAUZAdxXY', classes='link-purple link-url' ) with HorizontalGroup(classes='link-row'): yield Static('Get prompt inspiration: 🦸 ', markup=True, classes='link-label') yield Link( 'https://github.com/browser-use/awesome-prompts', url='https://github.com/browser-use/awesome-prompts', classes='link-magenta link-url', ) with HorizontalGroup(classes='link-row'): yield Static('[dim]Report any issues:[/] πŸ› ', markup=True, classes='link-label') yield Link( 'https://github.com/browser-use/browser-use/issues', url='https://github.com/browser-use/browser-use/issues', classes='link-green link-url', ) # Paths panel yield Static( f' βš™οΈ Settings saved to: {str(CONFIG.BROWSER_USE_CONFIG_FILE.resolve()).replace(str(Path.home()), "~")}\n' f' πŸ“ Outputs & recordings saved to: {str(Path(".").resolve()).replace(str(Path.home()), "~")}', id='paths-panel', markup=True, ) # Info panels (hidden by default, shown when task starts) with Container(id='info-panels'): # Top row with browser and model panels side by side with Container(id='top-panels'): # Browser panel with Container(id='browser-panel'): yield RichLog(id='browser-info', markup=True, highlight=True, wrap=True) # Model panel with Container(id='model-panel'): yield RichLog(id='model-info', markup=True, highlight=True, wrap=True) # Tasks panel (full width, below browser and model) with VerticalScroll(id='tasks-panel'): yield RichLog(id='tasks-info', markup=True, highlight=True, wrap=True, auto_scroll=True) # Three-column container (hidden by default) with Container(id='three-column-container'): # Column 1: Main output with VerticalScroll(id='main-output-column'): yield RichLog(highlight=True, markup=True, id='main-output-log', wrap=True, auto_scroll=True) # Column 2: Event bus events with VerticalScroll(id='events-column'): yield RichLog(highlight=True, markup=True, id='events-log', wrap=True, auto_scroll=True) # Column 3: CDP messages with VerticalScroll(id='cdp-column'): yield RichLog(highlight=True, markup=True, id='cdp-log', wrap=True, auto_scroll=True) # Task input container (now at the bottom) with Container(id='task-input-container'): yield Label('πŸ” What would you like me to do on the web?', id='task-label') yield Input(placeholder='Enter your task...', id='task-input') yield Footer() def update_info_panels(self) -> None: """Update all information panels with current state.""" try: # Update actual content self.update_browser_panel() self.update_model_panel() self.update_tasks_panel() except Exception as e: logging.error(f'Error in update_info_panels: {str(e)}') finally: # Always schedule the next update - will update at 1-second intervals # This ensures continuous updates even if agent state changes self.set_timer(1.0, self.update_info_panels) def update_browser_panel(self) -> None: """Update browser information panel with details about the browser.""" browser_info = self.query_one('#browser-info', RichLog) browser_info.clear() # Try to use the agent's browser session if available browser_session = self.browser_session if hasattr(self, 'agent') and self.agent and hasattr(self.agent, 'browser_session'): browser_session = self.agent.browser_session if browser_session: try: # Check if browser session has a CDP client if not hasattr(browser_session, 'cdp_client') or browser_session.cdp_client is None: browser_info.write('[yellow]Browser session created, waiting for browser to launch...[/]') return # Update our reference if we're using the agent's session if browser_session != self.browser_session: self.browser_session = browser_session # Get basic browser info from browser_profile browser_type = 'Chromium' headless = browser_session.browser_profile.headless # Determine connection type based on config connection_type = 'playwright' # Default if browser_session.cdp_url: connection_type = 'CDP' elif browser_session.browser_profile.executable_path: connection_type = 'user-provided' # Get window size details from browser_profile window_width = None window_height = None if browser_session.browser_profile.viewport: window_width = browser_session.browser_profile.viewport.width window_height = browser_session.browser_profile.viewport.height # Try to get browser PID browser_pid = 'Unknown' connected = False browser_status = '[red]Disconnected[/]' try: # Check if browser PID is available # Check if we have a CDP client if browser_session.cdp_client is not None: connected = True browser_status = '[green]Connected[/]' browser_pid = 'N/A' except Exception as e: browser_pid = f'Error: {str(e)}' # Display browser information browser_info.write(f'[bold cyan]Chromium[/] Browser ({browser_status})') browser_info.write( f'Type: [yellow]{connection_type}[/] [{"green" if not headless else "red"}]{" (headless)" if headless else ""}[/]' ) browser_info.write(f'PID: [dim]{browser_pid}[/]') browser_info.write(f'CDP Port: {browser_session.cdp_url}') if window_width and window_height: browser_info.write(f'Window: [blue]{window_width}[/] Γ— [blue]{window_height}[/]') # Include additional information about the browser if needed if connected and hasattr(self, 'agent') and self.agent: try: # Show when the browser was connected timestamp = int(time.time()) current_time = time.strftime('%H:%M:%S', time.localtime(timestamp)) browser_info.write(f'Last updated: [dim]{current_time}[/]') except Exception: pass # Show the agent's current page URL if available if browser_session.agent_focus_target_id: target = browser_session.session_manager.get_focused_target() target_url = target.url if target else 'about:blank' current_url = target_url.replace('https://', '').replace('http://', '').replace('www.', '')[:36] + '…' browser_info.write(f'πŸ‘οΈ [green]{current_url}[/]') except Exception as e: browser_info.write(f'[red]Error updating browser info: {str(e)}[/]') else: browser_info.write('[red]Browser not initialized[/]') def update_model_panel(self) -> None: """Update model information panel with details about the LLM.""" model_info = self.query_one('#model-info', RichLog) model_info.clear() if self.llm: # Get model details model_name = 'Unknown' if hasattr(self.llm, 'model_name'): model_name = self.llm.model_name elif hasattr(self.llm, 'model'): model_name = self.llm.model # Show model name if self.agent: temp_str = f'{self.llm.temperature}ΒΊC ' if self.llm.temperature else '' vision_str = '+ vision ' if self.agent.settings.use_vision else '' model_info.write( f'[white]LLM:[/] [blue]{self.llm.__class__.__name__} [yellow]{model_name}[/] {temp_str}{vision_str}' ) else: model_info.write(f'[white]LLM:[/] [blue]{self.llm.__class__.__name__} [yellow]{model_name}[/]') # Show token usage statistics if agent exists and has history if self.agent and hasattr(self.agent, 'state') and hasattr(self.agent.state, 'history'): # Calculate tokens per step num_steps = len(self.agent.history.history) # Get the last step metadata to show the most recent LLM response time if num_steps > 0 and self.agent.history.history[-1].metadata: last_step = self.agent.history.history[-1] if last_step.metadata: step_duration = last_step.metadata.duration_seconds else: step_duration = 0 # Show total duration total_duration = self.agent.history.total_duration_seconds() if total_duration > 0: model_info.write(f'[white]Total Duration:[/] [magenta]{total_duration:.2f}s[/]') # Calculate response time metrics model_info.write(f'[white]Last Step Duration:[/] [magenta]{step_duration:.2f}s[/]') # Add current state information if hasattr(self.agent, 'running'): if getattr(self.agent, 'running', False): model_info.write('[yellow]LLM is thinking[blink]...[/][/]') elif hasattr(self.agent, 'state') and hasattr(self.agent.state, 'paused') and self.agent.state.paused: model_info.write('[orange]LLM paused[/]') else: model_info.write('[red]Model not initialized[/]') def update_tasks_panel(self) -> None: """Update tasks information panel with details about the tasks and steps hierarchy.""" tasks_info = self.query_one('#tasks-info', RichLog) tasks_info.clear() if self.agent: # Check if agent has tasks task_history = [] message_history = [] # Try to extract tasks by looking at message history if hasattr(self.agent, '_message_manager') and self.agent._message_manager: message_history = self.agent._message_manager.state.history.get_messages() # Extract original task(s) original_tasks = [] for msg in message_history: if hasattr(msg, 'content'): content = msg.content if isinstance(content, str) and 'Your ultimate task is:' in content: task_text = content.split('"""')[1].strip() original_tasks.append(task_text) if original_tasks: tasks_info.write('[bold green]TASK:[/]') for i, task in enumerate(original_tasks, 1): # Only show latest task if multiple task changes occurred if i == len(original_tasks): tasks_info.write(f'[white]{task}[/]') tasks_info.write('') # Get current state information current_step = self.agent.state.n_steps if hasattr(self.agent, 'state') else 0 # Get all agent history items history_items = [] if hasattr(self.agent, 'state') and hasattr(self.agent.state, 'history'): history_items = self.agent.history.history if history_items: tasks_info.write('[bold yellow]STEPS:[/]') for idx, item in enumerate(history_items, 1): # Determine step status step_style = '[green]βœ“[/]' # For the current step, show it as in progress if idx == current_step: step_style = '[yellow]⟳[/]' # Check if this step had an error if item.result and any(result.error for result in item.result): step_style = '[red]βœ—[/]' # Show step number tasks_info.write(f'{step_style} Step {idx}/{current_step}') # Show goal if available if item.model_output and hasattr(item.model_output, 'current_state'): # Show goal for this step goal = item.model_output.current_state.next_goal if goal: # Take just the first line for display goal_lines = goal.strip().split('\n') goal_summary = goal_lines[0] tasks_info.write(f' [cyan]Goal:[/] {goal_summary}') # Show evaluation of previous goal (feedback) eval_prev = item.model_output.current_state.evaluation_previous_goal if eval_prev and idx > 1: # Only show for steps after the first eval_lines = eval_prev.strip().split('\n') eval_summary = eval_lines[0] eval_summary = eval_summary.replace('Success', 'βœ… ').replace('Failed', '❌ ').strip() tasks_info.write(f' [tan]Evaluation:[/] {eval_summary}') # Show actions taken in this step if item.model_output and item.model_output.action: tasks_info.write(' [purple]Actions:[/]') for action_idx, action in enumerate(item.model_output.action, 1): action_type = action.__class__.__name__ if hasattr(action, 'model_dump'): # For proper actions, show the action type action_dict = action.model_dump(exclude_unset=True) if action_dict: action_name = list(action_dict.keys())[0] tasks_info.write(f' {action_idx}. [blue]{action_name}[/]') # Show results or errors from this step if item.result: for result in item.result: if result.error: error_text = result.error tasks_info.write(f' [red]Error:[/] {error_text}') elif result.extracted_content: content = result.extracted_content tasks_info.write(f' [green]Result:[/] {content}') # Add a space between steps for readability tasks_info.write('') # If agent is actively running, show a status indicator if hasattr(self.agent, 'running') and getattr(self.agent, 'running', False): tasks_info.write('[yellow]Agent is actively working[blink]...[/][/]') elif hasattr(self.agent, 'state') and hasattr(self.agent.state, 'paused') and self.agent.state.paused: tasks_info.write('[orange]Agent is paused (press Enter to resume)[/]') else: tasks_info.write('[dim]Agent not initialized[/]') # Force scroll to bottom tasks_panel = self.query_one('#tasks-panel') tasks_panel.scroll_end(animate=False) async def run_prompt_mode(prompt: str, ctx: click.Context, debug: bool = False): """Run browser-use in non-interactive mode with a single prompt.""" # Import and call setup_logging to ensure proper initialization from browser_use.logging_config import setup_logging # Set up logging to only show results by default os.environ['BROWSER_USE_LOGGING_LEVEL'] = 'result' # Re-run setup_logging to apply the new log level setup_logging() # The logging is now properly configured by setup_logging() # No need to manually configure handlers since setup_logging() handles it # Initialize telemetry telemetry = ProductTelemetry() start_time = time.time() error_msg = None try: # Load config config = load_user_config() config = update_config_with_click_args(config, ctx) # Get LLM llm = get_llm(config) # Capture telemetry for CLI start in oneshot mode telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='start', mode='oneshot', model=llm.model if hasattr(llm, 'model') else None, model_provider=llm.__class__.__name__ if llm else None, ) ) # Get agent settings from config agent_settings = AgentSettings.model_validate(config.get('agent', {})) # Create browser session with config parameters browser_config = config.get('browser', {}) # Remove None values from browser_config browser_config = {k: v for k, v in browser_config.items() if v is not None} # Create BrowserProfile with user_data_dir profile = BrowserProfile(user_data_dir=str(USER_DATA_DIR), **browser_config) browser_session = BrowserSession( browser_profile=profile, ) # Create and run agent agent = Agent( task=prompt, llm=llm, browser_session=browser_session, source='cli', **agent_settings.model_dump(), ) await agent.run() # Ensure the browser session is fully stopped # The agent's close() method only kills the browser if keep_alive=False, # but we need to ensure all background tasks are stopped regardless if browser_session: try: # Kill the browser session to stop all background tasks await browser_session.kill() except Exception: # Ignore errors during cleanup pass # Capture telemetry for successful completion telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='task_completed', mode='oneshot', model=llm.model if hasattr(llm, 'model') else None, model_provider=llm.__class__.__name__ if llm else None, duration_seconds=time.time() - start_time, ) ) except Exception as e: error_msg = str(e) # Capture telemetry for error telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='error', mode='oneshot', model=llm.model if hasattr(llm, 'model') else None, model_provider=llm.__class__.__name__ if llm and 'llm' in locals() else None, duration_seconds=time.time() - start_time, error_message=error_msg, ) ) if debug: import traceback traceback.print_exc() else: print(f'Error: {str(e)}', file=sys.stderr) sys.exit(1) finally: # Ensure telemetry is flushed telemetry.flush() # Give a brief moment for cleanup to complete await asyncio.sleep(0.1) # Cancel any remaining tasks to ensure clean exit tasks = [t for t in asyncio.all_tasks() if t != asyncio.current_task()] for task in tasks: task.cancel() # Wait for all tasks to be cancelled if tasks: await asyncio.gather(*tasks, return_exceptions=True) async def textual_interface(config: dict[str, Any]): """Run the Textual interface.""" # Prevent browser_use from setting up logging at import time os.environ['BROWSER_USE_SETUP_LOGGING'] = 'false' logger = logging.getLogger('browser_use.startup') # Set up logging for Textual UI - prevent any logging to stdout def setup_textual_logging(): # Replace all handlers with null handler root_logger = logging.getLogger() for handler in root_logger.handlers: root_logger.removeHandler(handler) # Add null handler to ensure no output to stdout/stderr null_handler = logging.NullHandler() root_logger.addHandler(null_handler) logger.debug('Logging configured for Textual UI') logger.debug('Setting up Browser, Controller, and LLM...') # Step 1: Initialize BrowserSession with config logger.debug('Initializing BrowserSession...') try: # Get browser config from the config dict browser_config = config.get('browser', {}) logger.info('Browser type: chromium') # BrowserSession only supports chromium if browser_config.get('executable_path'): logger.info(f'Browser binary: {browser_config["executable_path"]}') if browser_config.get('headless'): logger.info('Browser mode: headless') else: logger.info('Browser mode: visible') # Create BrowserSession directly with config parameters # Remove None values from browser_config browser_config = {k: v for k, v in browser_config.items() if v is not None} # Create BrowserProfile with user_data_dir profile = BrowserProfile(user_data_dir=str(USER_DATA_DIR), **browser_config) browser_session = BrowserSession( browser_profile=profile, ) logger.debug('BrowserSession initialized successfully') # Set up FIFO logging pipes for streaming logs to UI try: from browser_use.logging_config import setup_log_pipes setup_log_pipes(session_id=browser_session.id) logger.debug(f'FIFO logging pipes set up for session {browser_session.id[-4:]}') except Exception as e: logger.debug(f'Could not set up FIFO logging pipes: {e}') # Browser version logging not available with CDP implementation except Exception as e: logger.error(f'Error initializing BrowserSession: {str(e)}', exc_info=True) raise RuntimeError(f'Failed to initialize BrowserSession: {str(e)}') # Step 3: Initialize Controller logger.debug('Initializing Controller...') try: controller = Controller() logger.debug('Controller initialized successfully') except Exception as e: logger.error(f'Error initializing Controller: {str(e)}', exc_info=True) raise RuntimeError(f'Failed to initialize Controller: {str(e)}') # Step 4: Get LLM logger.debug('Getting LLM...') try: # Ensure setup_logging is not called when importing modules os.environ['BROWSER_USE_SETUP_LOGGING'] = 'false' llm = get_llm(config) # Log LLM details model_name = getattr(llm, 'model_name', None) or getattr(llm, 'model', 'Unknown model') provider = llm.__class__.__name__ temperature = getattr(llm, 'temperature', 0.0) logger.info(f'LLM: {provider} ({model_name}), temperature: {temperature}') logger.debug(f'LLM initialized successfully: {provider}') except Exception as e: logger.error(f'Error getting LLM: {str(e)}', exc_info=True) raise RuntimeError(f'Failed to initialize LLM: {str(e)}') logger.debug('Initializing BrowserUseApp instance...') try: app = BrowserUseApp(config) # Pass the initialized components to the app app.browser_session = browser_session app.controller = controller app.llm = llm # Set up event bus listener now that browser session is available # Note: This needs to be called before run_async() but after browser_session is set # We'll defer this to on_mount() since it needs the widgets to be available # Configure logging for Textual UI before going fullscreen setup_textual_logging() # Log browser and model configuration that will be used browser_type = 'Chromium' # BrowserSession only supports Chromium model_name = config.get('model', {}).get('name', 'auto-detected') headless = config.get('browser', {}).get('headless', False) headless_str = 'headless' if headless else 'visible' logger.info(f'Preparing {browser_type} browser ({headless_str}) with {model_name} LLM') logger.debug('Starting Textual app with run_async()...') # No more logging after this point as we're in fullscreen mode await app.run_async() except Exception as e: logger.error(f'Error in textual_interface: {str(e)}', exc_info=True) # Note: We don't close the browser session here to avoid duplicate stop() calls # The browser session will be cleaned up by its __del__ method if needed raise async def run_auth_command(): """Run the authentication command with dummy task in UI.""" import asyncio import os from browser_use.sync.auth import DeviceAuthClient print('πŸ” Browser Use Cloud Authentication') print('=' * 40) # Ensure cloud sync is enabled (should be default, but make sure) os.environ['BROWSER_USE_CLOUD_SYNC'] = 'true' auth_client = DeviceAuthClient() print('πŸ” Debug: Checking authentication status...') print(f' API Token: {"βœ… Present" if auth_client.api_token else "❌ Missing"}') print(f' User ID: {auth_client.user_id}') print(f' Is Authenticated: {auth_client.is_authenticated}') if auth_client.auth_config.authorized_at: print(f' Authorized at: {auth_client.auth_config.authorized_at}') print() # Check if already authenticated if auth_client.is_authenticated: print('βœ… Already authenticated!') print(f' User ID: {auth_client.user_id}') print(f' Authenticated at: {auth_client.auth_config.authorized_at}') # Show cloud URL if possible frontend_url = CONFIG.BROWSER_USE_CLOUD_UI_URL or auth_client.base_url.replace('//api.', '//cloud.') print(f'\n🌐 View your runs at: {frontend_url}') return print('πŸš€ Starting authentication flow...') print(' This will open a browser window for you to sign in.') print() # Initialize variables for exception handling task_id = None sync_service = None try: # Create authentication flow with dummy task from uuid_extensions import uuid7str from browser_use.agent.cloud_events import ( CreateAgentSessionEvent, CreateAgentStepEvent, CreateAgentTaskEvent, UpdateAgentTaskEvent, ) from browser_use.sync.service import CloudSync # IDs for our session and task session_id = uuid7str() task_id = uuid7str() # Create special sync service that allows auth events sync_service = CloudSync(allow_session_events_for_auth=True) sync_service.set_auth_flow_active() # Explicitly enable auth flow sync_service.session_id = session_id # Set session ID for auth context sync_service.auth_client = auth_client # Use the same auth client instance! # 1. Create session (like main branch does at start) session_event = CreateAgentSessionEvent( id=session_id, user_id=auth_client.temp_user_id, browser_session_id=uuid7str(), browser_session_live_url='', browser_session_cdp_url='', device_id=auth_client.device_id, browser_state={ 'viewport': {'width': 1280, 'height': 720}, 'user_agent': None, 'headless': True, 'initial_url': None, 'final_url': None, 'total_pages_visited': 0, 'session_duration_seconds': 0, }, browser_session_data={ 'cookies': [], 'secrets': {}, 'allowed_domains': [], }, ) await sync_service.handle_event(session_event) # Brief delay to ensure session is created in backend before sending task await asyncio.sleep(0.5) # 2. Create task (like main branch does at start) task_event = CreateAgentTaskEvent( id=task_id, agent_session_id=session_id, llm_model='auth-flow', task='πŸ” Complete authentication and join the browser-use community', user_id=auth_client.temp_user_id, device_id=auth_client.device_id, done_output=None, user_feedback_type=None, user_comment=None, gif_url=None, ) await sync_service.handle_event(task_event) # Longer delay to ensure task is created in backend before sending step event await asyncio.sleep(1.0) # 3. Run authentication with timeout print('⏳ Waiting for authentication... (this may take up to 2 minutes for testing)') print(' Complete the authentication in your browser, then this will continue automatically.') print() try: print('πŸ”§ Debug: Starting authentication process...') print(f' Original auth client authenticated: {auth_client.is_authenticated}') print(f' Sync service auth client authenticated: {sync_service.auth_client.is_authenticated}') print(f' Same auth client? {auth_client is sync_service.auth_client}') print(f' Session ID: {sync_service.session_id}') # Create a task to show periodic status updates async def show_auth_progress(): for i in range(1, 25): # Show updates every 5 seconds for 2 minutes await asyncio.sleep(5) fresh_check = DeviceAuthClient() print(f'⏱️ Waiting for authentication... ({i * 5}s elapsed)') print(f' Status: {"βœ… Authenticated" if fresh_check.is_authenticated else "⏳ Still waiting"}') if fresh_check.is_authenticated: print('πŸŽ‰ Authentication detected! Completing...') break # Run authentication and progress updates concurrently auth_start_time = asyncio.get_event_loop().time() from browser_use.utils import create_task_with_error_handling auth_task = create_task_with_error_handling( sync_service.authenticate(show_instructions=True), name='sync_authenticate' ) progress_task = create_task_with_error_handling( show_auth_progress(), name='show_auth_progress', suppress_exceptions=True ) # Wait for authentication to complete, with timeout success = await asyncio.wait_for(auth_task, timeout=120.0) # 2 minutes for initial testing progress_task.cancel() # Stop the progress updates auth_duration = asyncio.get_event_loop().time() - auth_start_time print(f'πŸ”§ Debug: Authentication returned: {success} (took {auth_duration:.1f}s)') except TimeoutError: print('⏱️ Authentication timed out after 2 minutes.') print(' Checking if authentication completed in background...') # Create a fresh auth client to check current status fresh_auth_client = DeviceAuthClient() print('πŸ”§ Debug: Fresh auth client check:') print(f' API Token: {"βœ… Present" if fresh_auth_client.api_token else "❌ Missing"}') print(f' Is Authenticated: {fresh_auth_client.is_authenticated}') if fresh_auth_client.is_authenticated: print('βœ… Authentication was successful!') success = True # Update the sync service's auth client sync_service.auth_client = fresh_auth_client else: print('❌ Authentication not completed. Please try again.') success = False except Exception as e: print(f'❌ Authentication error: {type(e).__name__}: {e}') import traceback print(f'πŸ“„ Full traceback: {traceback.format_exc()}') success = False if success: # 4. Send step event to show progress (like main branch during execution) # Use the sync service's auth client which has the updated user_id step_event = CreateAgentStepEvent( # Remove explicit ID - let it auto-generate to avoid backend validation issues user_id=auth_client.temp_user_id, # Use same temp user_id as task for consistency device_id=auth_client.device_id, # Use consistent device_id agent_task_id=task_id, step=1, actions=[ { 'click': { 'coordinate': [800, 400], 'description': 'Click on Star button', 'success': True, }, 'done': { 'success': True, 'text': '⭐ Starred browser-use/browser-use repository! Welcome to the community!', }, } ], next_goal='⭐ Star browser-use GitHub repository to join the community', evaluation_previous_goal='Authentication completed successfully', memory='User authenticated with Browser Use Cloud and is now part of the community', screenshot_url=None, url='https://github.com/browser-use/browser-use', ) print('πŸ“€ Sending dummy step event...') await sync_service.handle_event(step_event) # Small delay to ensure step is processed before completion await asyncio.sleep(0.5) # 5. Complete task (like main branch does at end) completion_event = UpdateAgentTaskEvent( id=task_id, user_id=auth_client.temp_user_id, # Use same temp user_id as task for consistency device_id=auth_client.device_id, # Use consistent device_id done_output="πŸŽ‰ Welcome to Browser Use! You're now authenticated and part of our community. ⭐ Your future tasks will sync to the cloud automatically.", user_feedback_type=None, user_comment=None, gif_url=None, ) await sync_service.handle_event(completion_event) print('πŸŽ‰ Authentication successful!') print(' Future browser-use runs will now sync to the cloud.') else: # Failed - still complete the task with failure message completion_event = UpdateAgentTaskEvent( id=task_id, user_id=auth_client.temp_user_id, # Still temp user since auth failed device_id=auth_client.device_id, done_output='❌ Authentication failed. Please try again.', user_feedback_type=None, user_comment=None, gif_url=None, ) await sync_service.handle_event(completion_event) print('❌ Authentication failed.') print(' Please try again or check your internet connection.') except Exception as e: print(f'❌ Authentication error: {e}') # Still try to complete the task in UI with error message if task_id and sync_service: try: from browser_use.agent.cloud_events import UpdateAgentTaskEvent completion_event = UpdateAgentTaskEvent( id=task_id, user_id=auth_client.temp_user_id, device_id=auth_client.device_id, done_output=f'❌ Authentication error: {e}', user_feedback_type=None, user_comment=None, gif_url=None, ) await sync_service.handle_event(completion_event) except Exception: pass # Don't fail if we can't send the error event sys.exit(1) @click.group(invoke_without_command=True) @click.option('--version', is_flag=True, help='Print version and exit') @click.option( '--template', type=click.Choice(['default', 'advanced', 'tools'], case_sensitive=False), help='Generate a template file (default, advanced, or tools)', ) @click.option('--output', '-o', type=click.Path(), help='Output file path for template (default: browser_use_<template>.py)') @click.option('--force', '-f', is_flag=True, help='Overwrite existing files without asking') @click.option('--model', type=str, help='Model to use (e.g., gpt-5-mini, claude-4-sonnet, gemini-2.5-flash)') @click.option('--debug', is_flag=True, help='Enable verbose startup logging') @click.option('--headless', is_flag=True, help='Run browser in headless mode', default=None) @click.option('--window-width', type=int, help='Browser window width') @click.option('--window-height', type=int, help='Browser window height') @click.option( '--user-data-dir', type=str, help='Path to Chrome user data directory (e.g. ~/Library/Application Support/Google/Chrome)' ) @click.option('--profile-directory', type=str, help='Chrome profile directory name (e.g. "Default", "Profile 1")') @click.option('--cdp-url', type=str, help='Connect to existing Chrome via CDP URL (e.g. http://localhost:9222)') @click.option('--proxy-url', type=str, help='Proxy server for Chromium traffic (e.g. http://host:8080 or socks5://host:1080)') @click.option('--no-proxy', type=str, help='Comma-separated hosts to bypass proxy (e.g. localhost,127.0.0.1,*.internal)') @click.option('--proxy-username', type=str, help='Proxy auth username') @click.option('--proxy-password', type=str, help='Proxy auth password') @click.option('-p', '--prompt', type=str, help='Run a single task without the TUI (headless mode)') @click.option('--mcp', is_flag=True, help='Run as MCP server (exposes JSON RPC via stdin/stdout)') @click.pass_context def main(ctx: click.Context, debug: bool = False, **kwargs): """Browser Use - AI Agent for Web Automation Run without arguments to start the interactive TUI. Examples: uvx browser-use --template default uvx browser-use --template advanced --output my_script.py """ # Handle template generation if kwargs.get('template'): _run_template_generation(kwargs['template'], kwargs.get('output'), kwargs.get('force', False)) return if ctx.invoked_subcommand is None: # No subcommand, run the main interface run_main_interface(ctx, debug, **kwargs) def run_main_interface(ctx: click.Context, debug: bool = False, **kwargs): """Run the main browser-use interface""" if kwargs['version']: from importlib.metadata import version print(version('browser-use')) sys.exit(0) # Check if MCP server mode is activated if kwargs.get('mcp'): # Capture telemetry for MCP server mode via CLI (suppress any logging from this) try: telemetry = ProductTelemetry() telemetry.capture( CLITelemetryEvent( version=get_browser_use_version(), action='start', mode='mcp_server', ) ) except Exception: # Ignore telemetry errors in MCP mode to prevent any stdout contamination pass # Run as MCP server from browser_use.mcp.server import main as mcp_main asyncio.run(mcp_main()) return # Check if prompt mode is activated if kwargs.get('prompt'): # Set environment variable for prompt mode before running os.environ['BROWSER_USE_LOGGING_LEVEL'] = 'result' # Run in non-interactive mode asyncio.run(run_prompt_mode(kwargs['prompt'], ctx, debug)) return # Configure console logging console_handler = logging.StreamHandler(sys.stdout) console_handler.setFormatter(logging.Formatter('%(asctime)s - %(levelname)s - %(message)s', '%H:%M:%S')) # Configure root logger root_logger = logging.getLogger() root_logger.setLevel(logging.INFO if not debug else logging.DEBUG) root_logger.addHandler(console_handler) logger = logging.getLogger('browser_use.startup') logger.info('Starting Browser-Use initialization') if debug: logger.debug(f'System info: Python {sys.version.split()[0]}, Platform: {sys.platform}') logger.debug('Loading environment variables from .env file...') load_dotenv() logger.debug('Environment variables loaded') # Load user configuration logger.debug('Loading user configuration...') try: config = load_user_config() logger.debug(f'User configuration loaded from {CONFIG.BROWSER_USE_CONFIG_FILE}') except Exception as e: logger.error(f'Error loading user configuration: {str(e)}', exc_info=True) print(f'Error loading configuration: {str(e)}') sys.exit(1) # Update config with command-line arguments logger.debug('Updating configuration with command line arguments...') try: config = update_config_with_click_args(config, ctx) logger.debug('Configuration updated') except Exception as e: logger.error(f'Error updating config with command line args: {str(e)}', exc_info=True) print(f'Error updating configuration: {str(e)}') sys.exit(1) # Save updated config logger.debug('Saving user configuration...') try: save_user_config(config) logger.debug('Configuration saved') except Exception as e: logger.error(f'Error saving user configuration: {str(e)}', exc_info=True) print(f'Error saving configuration: {str(e)}') sys.exit(1) # Setup handlers for console output before entering Textual UI logger.debug('Setting up handlers for Textual UI...') # Log browser and model configuration that will be used browser_type = 'Chromium' # BrowserSession only supports Chromium model_name = config.get('model', {}).get('name', 'auto-detected') headless = config.get('browser', {}).get('headless', False) headless_str = 'headless' if headless else 'visible' logger.info(f'Preparing {browser_type} browser ({headless_str}) with {model_name} LLM') try: # Run the Textual UI interface - now all the initialization happens before we go fullscreen logger.debug('Starting Textual UI interface...') asyncio.run(textual_interface(config)) except Exception as e: # Restore console logging for error reporting root_logger.setLevel(logging.INFO) for handler in root_logger.handlers: root_logger.removeHandler(handler) root_logger.addHandler(console_handler) logger.error(f'Error initializing Browser-Use: {str(e)}', exc_info=debug) print(f'\nError launching Browser-Use: {str(e)}') if debug: import traceback traceback.print_exc() sys.exit(1) @main.command() def auth(): """Authenticate with Browser Use Cloud to sync your runs""" asyncio.run(run_auth_command()) @main.command() def install(): """Install Chromium browser with system dependencies""" import platform import subprocess print('πŸ“¦ Installing Chromium browser + system dependencies...') print('⏳ This may take a few minutes...\n') # Build command - only use --with-deps on Linux (it fails on Windows/macOS) cmd = ['uvx', 'playwright', 'install', 'chromium'] if platform.system() == 'Linux': cmd.append('--with-deps') cmd.append('--no-shell') result = subprocess.run(cmd) if result.returncode == 0: print('\nβœ… Installation complete!') print('πŸš€ Ready to use! Run: uvx browser-use') else: print('\n❌ Installation failed') sys.exit(1) # ============================================================================ # Template Generation - Generate template files # ============================================================================ # Template metadata INIT_TEMPLATES = { 'default': { 'file': 'default_template.py', 'description': 'Simplest setup - capable of any web task with minimal configuration', }, 'advanced': { 'file': 'advanced_template.py', 'description': 'All configuration options shown with defaults', }, 'tools': { 'file': 'tools_template.py', 'description': 'Custom action examples - extend the agent with your own functions', }, } def _run_template_generation(template: str, output: str | None, force: bool): """Generate a template file (called from main CLI).""" # Determine output path if output: output_path = Path(output) else: output_path = Path.cwd() / f'browser_use_{template}.py' # Read template file try: templates_dir = Path(__file__).parent / 'cli_templates' template_file = INIT_TEMPLATES[template]['file'] template_path = templates_dir / template_file content = template_path.read_text(encoding='utf-8') except Exception as e: click.echo(f'❌ Error reading template: {e}', err=True) sys.exit(1) # Write file if _write_init_file(output_path, content, force): click.echo(f'βœ… Created {output_path}') click.echo('\nNext steps:') click.echo(' 1. Install browser-use:') click.echo(' uv pip install browser-use') click.echo(' 2. Set up your API key in .env file or environment:') click.echo(' BROWSER_USE_API_KEY=your-key') click.echo(' (Get your key at https://cloud.browser-use.com/new-api-key)') click.echo(' 3. Run your script:') click.echo(f' python {output_path.name}') else: sys.exit(1) def _write_init_file(output_path: Path, content: str, force: bool = False) -> bool: """Write content to a file, with safety checks.""" # Check if file already exists if output_path.exists() and not force: click.echo(f'⚠️ File already exists: {output_path}') if not click.confirm('Overwrite?', default=False): click.echo('❌ Cancelled') return False # Ensure parent directory exists output_path.parent.mkdir(parents=True, exist_ok=True) # Write file try: output_path.write_text(content, encoding='utf-8') return True except Exception as e: click.echo(f'❌ Error writing file: {e}', err=True) return False @main.command('init') @click.option( '--template', '-t', type=click.Choice(['default', 'advanced', 'tools'], case_sensitive=False), help='Template to use', ) @click.option( '--output', '-o', type=click.Path(), help='Output file path (default: browser_use_<template>.py)', ) @click.option( '--force', '-f', is_flag=True, help='Overwrite existing files without asking', ) @click.option( '--list', '-l', 'list_templates', is_flag=True, help='List available templates', ) def init( template: str | None, output: str | None, force: bool, list_templates: bool, ): """ Generate a browser-use template file to get started quickly. Examples: \b # Interactive mode - prompts for template selection uvx browser-use init \b # Generate default template uvx browser-use init --template default \b # Generate advanced template with custom filename uvx browser-use init --template advanced --output my_script.py \b # List available templates uvx browser-use init --list """ # Handle --list flag if list_templates: click.echo('Available templates:\n') for name, info in INIT_TEMPLATES.items(): click.echo(f' {name:12} - {info["description"]}') return # Interactive template selection if not provided if not template: click.echo('Available templates:\n') for name, info in INIT_TEMPLATES.items(): click.echo(f' {name:12} - {info["description"]}') click.echo() template = click.prompt( 'Which template would you like to use?', type=click.Choice(['default', 'advanced', 'tools'], case_sensitive=False), default='default', ) # Template is guaranteed to be set at this point (either from option or prompt) assert template is not None # Determine output path if output: output_path = Path(output) else: output_path = Path.cwd() / f'browser_use_{template}.py' # Read template file try: templates_dir = Path(__file__).parent / 'cli_templates' template_file = INIT_TEMPLATES[template]['file'] template_path = templates_dir / template_file content = template_path.read_text(encoding='utf-8') except Exception as e: click.echo(f'❌ Error reading template: {e}', err=True) sys.exit(1) # Write file if _write_init_file(output_path, content, force): click.echo(f'βœ… Created {output_path}') click.echo('\nNext steps:') click.echo(' 1. Install browser-use:') click.echo(' uv pip install browser-use') click.echo(' 2. Set up your API key in .env file or environment:') click.echo(' BROWSER_USE_API_KEY=your-key') click.echo(' (Get your key at https://cloud.browser-use.com/new-api-key)') click.echo(' 3. Run your script:') click.echo(f' python {output_path.name}') else: sys.exit(1) if __name__ == '__main__': main()
{ "repo_id": "browser-use/browser-use", "file_path": "browser_use/cli.py", "license": "MIT License", "lines": 1980, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/tests/test_present_file_tool_core_logic.py
"""Core behavior tests for present_files path normalization.""" import importlib from types import SimpleNamespace present_file_tool_module = importlib.import_module( "src.tools.builtins.present_file_tool" ) def _make_runtime(outputs_path: str) -> SimpleNamespace: return SimpleNamespace( state={"thread_data": {"outputs_path": outputs_path}}, context={"thread_id": "thread-1"}, ) def test_present_files_normalizes_host_outputs_path(tmp_path): outputs_dir = tmp_path / "threads" / "thread-1" / "user-data" / "outputs" outputs_dir.mkdir(parents=True) artifact_path = outputs_dir / "report.md" artifact_path.write_text("ok") result = present_file_tool_module.present_file_tool.func( runtime=_make_runtime(str(outputs_dir)), filepaths=[str(artifact_path)], tool_call_id="tc-1", ) assert result.update["artifacts"] == ["/mnt/user-data/outputs/report.md"] assert result.update["messages"][0].content == "Successfully presented files" def test_present_files_keeps_virtual_outputs_path(tmp_path, monkeypatch): outputs_dir = tmp_path / "threads" / "thread-1" / "user-data" / "outputs" outputs_dir.mkdir(parents=True) artifact_path = outputs_dir / "summary.json" artifact_path.write_text("{}") monkeypatch.setattr( present_file_tool_module, "get_paths", lambda: SimpleNamespace( resolve_virtual_path=lambda thread_id, path: artifact_path ), ) result = present_file_tool_module.present_file_tool.func( runtime=_make_runtime(str(outputs_dir)), filepaths=["/mnt/user-data/outputs/summary.json"], tool_call_id="tc-2", ) assert result.update["artifacts"] == ["/mnt/user-data/outputs/summary.json"] def test_present_files_rejects_paths_outside_outputs(tmp_path): outputs_dir = tmp_path / "threads" / "thread-1" / "user-data" / "outputs" workspace_dir = tmp_path / "threads" / "thread-1" / "user-data" / "workspace" outputs_dir.mkdir(parents=True) workspace_dir.mkdir(parents=True) leaked_path = workspace_dir / "notes.txt" leaked_path.write_text("leak") result = present_file_tool_module.present_file_tool.func( runtime=_make_runtime(str(outputs_dir)), filepaths=[str(leaked_path)], tool_call_id="tc-3", ) assert "artifacts" not in result.update assert ( result.update["messages"][0].content == f"Error: Only files in /mnt/user-data/outputs can be presented: {leaked_path}" )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_present_file_tool_core_logic.py", "license": "MIT License", "lines": 58, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/gateway/routers/suggestions.py
import json import logging from fastapi import APIRouter from pydantic import BaseModel, Field from src.models import create_chat_model logger = logging.getLogger(__name__) router = APIRouter(prefix="/api", tags=["suggestions"]) class SuggestionMessage(BaseModel): role: str = Field(..., description="Message role: user|assistant") content: str = Field(..., description="Message content as plain text") class SuggestionsRequest(BaseModel): messages: list[SuggestionMessage] = Field(..., description="Recent conversation messages") n: int = Field(default=3, ge=1, le=5, description="Number of suggestions to generate") model_name: str | None = Field(default=None, description="Optional model override") class SuggestionsResponse(BaseModel): suggestions: list[str] = Field(default_factory=list, description="Suggested follow-up questions") def _strip_markdown_code_fence(text: str) -> str: stripped = text.strip() if not stripped.startswith("```"): return stripped lines = stripped.splitlines() if len(lines) >= 3 and lines[0].startswith("```") and lines[-1].startswith("```"): return "\n".join(lines[1:-1]).strip() return stripped def _parse_json_string_list(text: str) -> list[str] | None: candidate = _strip_markdown_code_fence(text) start = candidate.find("[") end = candidate.rfind("]") if start == -1 or end == -1 or end <= start: return None candidate = candidate[start : end + 1] try: data = json.loads(candidate) except Exception: return None if not isinstance(data, list): return None out: list[str] = [] for item in data: if not isinstance(item, str): continue s = item.strip() if not s: continue out.append(s) return out def _format_conversation(messages: list[SuggestionMessage]) -> str: parts: list[str] = [] for m in messages: role = m.role.strip().lower() if role in ("user", "human"): parts.append(f"User: {m.content.strip()}") elif role in ("assistant", "ai"): parts.append(f"Assistant: {m.content.strip()}") else: parts.append(f"{m.role}: {m.content.strip()}") return "\n".join(parts).strip() @router.post( "/threads/{thread_id}/suggestions", response_model=SuggestionsResponse, summary="Generate Follow-up Questions", description="Generate short follow-up questions a user might ask next, based on recent conversation context.", ) async def generate_suggestions(thread_id: str, request: SuggestionsRequest) -> SuggestionsResponse: if not request.messages: return SuggestionsResponse(suggestions=[]) n = request.n conversation = _format_conversation(request.messages) if not conversation: return SuggestionsResponse(suggestions=[]) prompt = ( "You are generating follow-up questions to help the user continue the conversation.\n" f"Based on the conversation below, produce EXACTLY {n} short questions the user might ask next.\n" "Requirements:\n" "- Questions must be relevant to the conversation.\n" "- Questions must be written in the same language as the user.\n" "- Keep each question concise (ideally <= 20 words / <= 40 Chinese characters).\n" "- Do NOT include numbering, markdown, or any extra text.\n" "- Output MUST be a JSON array of strings only.\n\n" "Conversation:\n" f"{conversation}\n" ).format(n=n, conversation=conversation) try: model = create_chat_model(name=request.model_name, thinking_enabled=False) response = model.invoke(prompt) raw = str(response.content or "") suggestions = _parse_json_string_list(raw) or [] cleaned = [s.replace("\n", " ").strip() for s in suggestions if s.strip()] cleaned = cleaned[:n] return SuggestionsResponse(suggestions=cleaned) except Exception as exc: logger.exception("Failed to generate suggestions: thread_id=%s err=%s", thread_id, exc) return SuggestionsResponse(suggestions=[])
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/suggestions.py", "license": "MIT License", "lines": 93, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/tests/test_suggestions_router.py
import asyncio from unittest.mock import MagicMock from src.gateway.routers import suggestions def test_strip_markdown_code_fence_removes_wrapping(): text = "```json\n[\"a\"]\n```" assert suggestions._strip_markdown_code_fence(text) == "[\"a\"]" def test_strip_markdown_code_fence_no_fence_keeps_content(): text = " [\"a\"] " assert suggestions._strip_markdown_code_fence(text) == "[\"a\"]" def test_parse_json_string_list_filters_invalid_items(): text = "```json\n[\"a\", \" \", 1, \"b\"]\n```" assert suggestions._parse_json_string_list(text) == ["a", "b"] def test_parse_json_string_list_rejects_non_list(): text = "{\"a\": 1}" assert suggestions._parse_json_string_list(text) is None def test_format_conversation_formats_roles(): messages = [ suggestions.SuggestionMessage(role="User", content="Hi"), suggestions.SuggestionMessage(role="assistant", content="Hello"), suggestions.SuggestionMessage(role="system", content="note"), ] assert suggestions._format_conversation(messages) == "User: Hi\nAssistant: Hello\nsystem: note" def test_generate_suggestions_parses_and_limits(monkeypatch): req = suggestions.SuggestionsRequest( messages=[ suggestions.SuggestionMessage(role="user", content="Hi"), suggestions.SuggestionMessage(role="assistant", content="Hello"), ], n=3, model_name=None, ) fake_model = MagicMock() fake_model.invoke.return_value = MagicMock(content="```json\n[\"Q1\", \"Q2\", \"Q3\", \"Q4\"]\n```") monkeypatch.setattr(suggestions, "create_chat_model", lambda **kwargs: fake_model) result = asyncio.run(suggestions.generate_suggestions("t1", req)) assert result.suggestions == ["Q1", "Q2", "Q3"] def test_generate_suggestions_returns_empty_on_model_error(monkeypatch): req = suggestions.SuggestionsRequest( messages=[suggestions.SuggestionMessage(role="user", content="Hi")], n=2, model_name=None, ) fake_model = MagicMock() fake_model.invoke.side_effect = RuntimeError("boom") monkeypatch.setattr(suggestions, "create_chat_model", lambda **kwargs: fake_model) result = asyncio.run(suggestions.generate_suggestions("t1", req)) assert result.suggestions == []
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_suggestions_router.py", "license": "MIT License", "lines": 47, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/community/infoquest/infoquest_client.py
"""Util that calls InfoQuest Search And Fetch API. In order to set this up, follow instructions at: https://docs.byteplus.com/en/docs/InfoQuest/What_is_Info_Quest """ import json import logging import os from typing import Any import requests logger = logging.getLogger(__name__) class InfoQuestClient: """Client for interacting with the InfoQuest web search and fetch API.""" def __init__(self, fetch_time: int = -1, fetch_timeout: int = -1, fetch_navigation_timeout: int = -1, search_time_range: int = -1): logger.info("\n============================================\nπŸš€ BytePlus InfoQuest Client Initialization πŸš€\n============================================") self.fetch_time = fetch_time self.fetch_timeout = fetch_timeout self.fetch_navigation_timeout = fetch_navigation_timeout self.search_time_range = search_time_range self.api_key_set = bool(os.getenv("INFOQUEST_API_KEY")) if logger.isEnabledFor(logging.DEBUG): config_details = ( f"\nπŸ“‹ Configuration Details:\n" f"β”œβ”€β”€ Fetch time: {fetch_time} {'(Default: No fetch time)' if fetch_time == -1 else '(Custom)'}\n" f"β”œβ”€β”€ Fetch Timeout: {fetch_timeout} {'(Default: No fetch timeout)' if fetch_timeout == -1 else '(Custom)'}\n" f"β”œβ”€β”€ Navigation Timeout: {fetch_navigation_timeout} {'(Default: No Navigation Timeout)' if fetch_navigation_timeout == -1 else '(Custom)'}\n" f"β”œβ”€β”€ Search Time Range: {search_time_range} {'(Default: No Search Time Range)' if search_time_range == -1 else '(Custom)'}\n" f"└── API Key: {'βœ… Configured' if self.api_key_set else '❌ Not set'}" ) logger.debug(config_details) logger.debug("\n" + "*" * 70 + "\n") def fetch(self, url: str, return_format: str = "html") -> str: if logger.isEnabledFor(logging.DEBUG): url_truncated = url[:50] + "..." if len(url) > 50 else url logger.debug( f"InfoQuest - Fetch API request initiated | " f"operation=crawl url | " f"url_truncated={url_truncated} | " f"has_timeout_filter={self.fetch_timeout > 0} | timeout_filter={self.fetch_timeout} | " f"has_fetch_time_filter={self.fetch_time > 0} | fetch_time_filter={self.fetch_time} | " f"has_navigation_timeout_filter={self.fetch_navigation_timeout > 0} | navi_timeout_filter={self.fetch_navigation_timeout} | " f"request_type=sync" ) # Prepare headers headers = self._prepare_headers() # Prepare request data data = self._prepare_crawl_request_data(url, return_format) logger.debug("Sending crawl request to InfoQuest API") try: response = requests.post("https://reader.infoquest.bytepluses.com", headers=headers, json=data) # Check if status code is not 200 if response.status_code != 200: error_message = f"fetch API returned status {response.status_code}: {response.text}" logger.debug("InfoQuest Crawler fetch API return status %d: %s for URL: %s", response.status_code, response.text, url) return f"Error: {error_message}" # Check for empty response if not response.text or not response.text.strip(): error_message = "no result found" logger.debug("InfoQuest Crawler returned empty response for URL: %s", url) return f"Error: {error_message}" # Try to parse response as JSON and extract reader_result try: response_data = json.loads(response.text) # Extract reader_result if it exists if "reader_result" in response_data: logger.debug("Successfully extracted reader_result from JSON response") return response_data["reader_result"] elif "content" in response_data: # Fallback to content field if reader_result is not available logger.debug("reader_result missing in JSON response, falling back to content field: %s", response_data["content"]) return response_data["content"] else: # If neither field exists, return the original response logger.warning("Neither reader_result nor content field found in JSON response") except json.JSONDecodeError: # If response is not JSON, return the original text logger.debug("Response is not in JSON format, returning as-is") return response.text # Print partial response for debugging if logger.isEnabledFor(logging.DEBUG): response_sample = response.text[:200] + ("..." if len(response.text) > 200 else "") logger.debug("Successfully received response, content length: %d bytes, first 200 chars: %s", len(response.text), response_sample) return response.text except Exception as e: error_message = f"fetch API failed: {str(e)}" logger.error(error_message) return f"Error: {error_message}" @staticmethod def _prepare_headers() -> dict[str, str]: """Prepare request headers.""" headers = { "Content-Type": "application/json", } # Add API key if available if os.getenv("INFOQUEST_API_KEY"): headers["Authorization"] = f"Bearer {os.getenv('INFOQUEST_API_KEY')}" logger.debug("API key added to request headers") else: logger.warning("InfoQuest API key is not set. Provide your own key for authentication.") return headers def _prepare_crawl_request_data(self, url: str, return_format: str) -> dict[str, Any]: """Prepare request data with formatted parameters.""" # Normalize return_format if return_format and return_format.lower() == "html": normalized_format = "HTML" else: normalized_format = return_format data = {"url": url, "format": normalized_format} # Add timeout parameters if set to positive values timeout_params = {} if self.fetch_time > 0: timeout_params["fetch_time"] = self.fetch_time if self.fetch_timeout > 0: timeout_params["timeout"] = self.fetch_timeout if self.fetch_navigation_timeout > 0: timeout_params["navi_timeout"] = self.fetch_navigation_timeout # Log applied timeout parameters if timeout_params: logger.debug("Applying timeout parameters: %s", timeout_params) data.update(timeout_params) return data def web_search_raw_results( self, query: str, site: str, output_format: str = "JSON", ) -> dict: """Get results from the InfoQuest Web-Search API synchronously.""" headers = self._prepare_headers() params = {"format": output_format, "query": query} if self.search_time_range > 0: params["time_range"] = self.search_time_range if site != "": params["site"] = site response = requests.post("https://search.infoquest.bytepluses.com", headers=headers, json=params) response.raise_for_status() # Print partial response for debugging response_json = response.json() if logger.isEnabledFor(logging.DEBUG): response_sample = json.dumps(response_json)[:200] + ("..." if len(json.dumps(response_json)) > 200 else "") logger.debug(f"Search API request completed successfully | service=InfoQuest | status=success | response_sample={response_sample}") return response_json @staticmethod def clean_results(raw_results: list[dict[str, dict[str, dict[str, Any]]]]) -> list[dict]: """Clean results from InfoQuest Web-Search API.""" logger.debug("Processing web-search results") seen_urls = set() clean_results = [] counts = {"pages": 0, "news": 0} for content_list in raw_results: content = content_list["content"] results = content["results"] if results.get("organic"): organic_results = results["organic"] for result in organic_results: clean_result = { "type": "page", } if "title" in result: clean_result["title"] = result["title"] if "desc" in result: clean_result["desc"] = result["desc"] clean_result["snippet"] = result["desc"] if "url" in result: clean_result["url"] = result["url"] url = clean_result["url"] if isinstance(url, str) and url and url not in seen_urls: seen_urls.add(url) clean_results.append(clean_result) counts["pages"] += 1 if results.get("top_stories"): news = results["top_stories"] for obj in news["items"]: clean_result = { "type": "news", } if "time_frame" in obj: clean_result["time_frame"] = obj["time_frame"] if "source" in obj: clean_result["source"] = obj["source"] title = obj.get("title") url = obj.get("url") if title: clean_result["title"] = title if url: clean_result["url"] = url if title and isinstance(url, str) and url and url not in seen_urls: seen_urls.add(url) clean_results.append(clean_result) counts["news"] += 1 logger.debug(f"Results processing completed | total_results={len(clean_results)} | pages={counts['pages']} | news_items={counts['news']} | unique_urls={len(seen_urls)}") return clean_results def web_search( self, query: str, site: str = "", output_format: str = "JSON", ) -> str: if logger.isEnabledFor(logging.DEBUG): query_truncated = query[:50] + "..." if len(query) > 50 else query logger.debug( f"InfoQuest - Search API request initiated | " f"operation=search webs | " f"query_truncated={query_truncated} | " f"has_time_filter={self.search_time_range > 0} | time_filter={self.search_time_range} | " f"has_site_filter={bool(site)} | site={site} | " f"request_type=sync" ) try: logger.debug("InfoQuest Web-Search - Executing search with parameters") raw_results = self.web_search_raw_results( query, site, output_format, ) if "search_result" in raw_results: logger.debug("InfoQuest Web-Search - Successfully extracted search_result from JSON response") results = raw_results["search_result"] logger.debug("InfoQuest Web-Search - Processing raw search results") cleaned_results = self.clean_results(results["results"]) result_json = json.dumps(cleaned_results, indent=2, ensure_ascii=False) logger.debug(f"InfoQuest Web-Search - Search tool execution completed | mode=synchronous | results_count={len(cleaned_results)}") return result_json elif "content" in raw_results: # Fallback to content field if search_result is not available error_message = "web search API return wrong format" logger.error("web search API return wrong format, no search_result nor content field found in JSON response, content: %s", raw_results["content"]) return f"Error: {error_message}" else: # If neither field exists, return the original response logger.warning("InfoQuest Web-Search - Neither search_result nor content field found in JSON response") return json.dumps(raw_results, indent=2, ensure_ascii=False) except Exception as e: error_message = f"InfoQuest Web-Search - Search tool execution failed | mode=synchronous | error={str(e)}" logger.error(error_message) return f"Error: {error_message}" @staticmethod def clean_results_with_image_search(raw_results: list[dict[str, dict[str, dict[str, Any]]]]) -> list[dict]: """Clean results from InfoQuest Web-Search API.""" logger.debug("Processing web-search results") seen_urls = set() clean_results = [] counts = {"images": 0} for content_list in raw_results: content = content_list["content"] results = content["results"] if results.get("images_results"): images_results = results["images_results"] for result in images_results: clean_result = {} if "image_url" in result: clean_result["image_url"] = result["image_url"] url = clean_result["image_url"] if isinstance(url, str) and url and url not in seen_urls: seen_urls.add(url) clean_results.append(clean_result) counts["images"] += 1 if "thumbnail_url" in result: clean_result["thumbnail_url"] = result["thumbnail_url"] if "url" in result: clean_result["url"] = result["url"] logger.debug(f"Results processing completed | total_results={len(clean_results)} | images={counts['images']} | unique_urls={len(seen_urls)}") return clean_results
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/infoquest/infoquest_client.py", "license": "MIT License", "lines": 263, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/infoquest/tools.py
from langchain.tools import tool from src.config import get_app_config from src.utils.readability import ReadabilityExtractor from .infoquest_client import InfoQuestClient readability_extractor = ReadabilityExtractor() def _get_infoquest_client() -> InfoQuestClient: search_config = get_app_config().get_tool_config("web_search") search_time_range = -1 if search_config is not None and "search_time_range" in search_config.model_extra: search_time_range = search_config.model_extra.get("search_time_range") fetch_config = get_app_config().get_tool_config("web_fetch") fetch_time = -1 if fetch_config is not None and "fetch_time" in fetch_config.model_extra: fetch_time = fetch_config.model_extra.get("fetch_time") fetch_timeout = -1 if fetch_config is not None and "timeout" in fetch_config.model_extra: fetch_timeout = fetch_config.model_extra.get("timeout") navigation_timeout = -1 if fetch_config is not None and "navigation_timeout" in fetch_config.model_extra: navigation_timeout = fetch_config.model_extra.get("navigation_timeout") return InfoQuestClient( search_time_range=search_time_range, fetch_timeout=fetch_timeout, fetch_navigation_timeout=navigation_timeout, fetch_time=fetch_time, ) @tool("web_search", parse_docstring=True) def web_search_tool(query: str) -> str: """Search the web. Args: query: The query to search for. """ client = _get_infoquest_client() return client.web_search(query) @tool("web_fetch", parse_docstring=True) def web_fetch_tool(url: str) -> str: """Fetch the contents of a web page at a given URL. Only fetch EXACT URLs that have been provided directly by the user or have been returned in results from the web_search and web_fetch tools. This tool can NOT access content that requires authentication, such as private Google Docs or pages behind login walls. Do NOT add www. to URLs that do NOT have them. URLs must include the schema: https://example.com is a valid URL while example.com is an invalid URL. Args: url: The URL to fetch the contents of. """ client = _get_infoquest_client() result = client.fetch(url) if result.startswith("Error: "): return result article = readability_extractor.extract_article(result) return article.to_markdown()[:4096]
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/infoquest/tools.py", "license": "MIT License", "lines": 50, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/tests/test_infoquest_client.py
"""Tests for InfoQuest client and tools.""" import json from unittest.mock import MagicMock, patch from src.community.infoquest import tools from src.community.infoquest.infoquest_client import InfoQuestClient class TestInfoQuestClient: def test_infoquest_client_initialization(self): """Test InfoQuestClient initialization with different parameters.""" # Test with default parameters client = InfoQuestClient() assert client.fetch_time == -1 assert client.fetch_timeout == -1 assert client.fetch_navigation_timeout == -1 assert client.search_time_range == -1 # Test with custom parameters client = InfoQuestClient(fetch_time=10, fetch_timeout=30, fetch_navigation_timeout=60, search_time_range=24) assert client.fetch_time == 10 assert client.fetch_timeout == 30 assert client.fetch_navigation_timeout == 60 assert client.search_time_range == 24 @patch("src.community.infoquest.infoquest_client.requests.post") def test_fetch_success(self, mock_post): """Test successful fetch operation.""" mock_response = MagicMock() mock_response.status_code = 200 mock_response.text = json.dumps({"reader_result": "<html><body>Test content</body></html>"}) mock_post.return_value = mock_response client = InfoQuestClient() result = client.fetch("https://example.com") assert result == "<html><body>Test content</body></html>" mock_post.assert_called_once() args, kwargs = mock_post.call_args assert args[0] == "https://reader.infoquest.bytepluses.com" assert kwargs["json"]["url"] == "https://example.com" assert kwargs["json"]["format"] == "HTML" @patch("src.community.infoquest.infoquest_client.requests.post") def test_fetch_non_200_status(self, mock_post): """Test fetch operation with non-200 status code.""" mock_response = MagicMock() mock_response.status_code = 404 mock_response.text = "Not Found" mock_post.return_value = mock_response client = InfoQuestClient() result = client.fetch("https://example.com") assert result == "Error: fetch API returned status 404: Not Found" @patch("src.community.infoquest.infoquest_client.requests.post") def test_fetch_empty_response(self, mock_post): """Test fetch operation with empty response.""" mock_response = MagicMock() mock_response.status_code = 200 mock_response.text = "" mock_post.return_value = mock_response client = InfoQuestClient() result = client.fetch("https://example.com") assert result == "Error: no result found" @patch("src.community.infoquest.infoquest_client.requests.post") def test_web_search_raw_results_success(self, mock_post): """Test successful web_search_raw_results operation.""" mock_response = MagicMock() mock_response.status_code = 200 mock_response.json.return_value = {"search_result": {"results": [{"content": {"results": {"organic": [{"title": "Test Result", "desc": "Test description", "url": "https://example.com"}]}}}], "images_results": []}} mock_post.return_value = mock_response client = InfoQuestClient() result = client.web_search_raw_results("test query", "") assert "search_result" in result mock_post.assert_called_once() args, kwargs = mock_post.call_args assert args[0] == "https://search.infoquest.bytepluses.com" assert kwargs["json"]["query"] == "test query" @patch("src.community.infoquest.infoquest_client.requests.post") def test_web_search_success(self, mock_post): """Test successful web_search operation.""" mock_response = MagicMock() mock_response.status_code = 200 mock_response.json.return_value = {"search_result": {"results": [{"content": {"results": {"organic": [{"title": "Test Result", "desc": "Test description", "url": "https://example.com"}]}}}], "images_results": []}} mock_post.return_value = mock_response client = InfoQuestClient() result = client.web_search("test query") # Check if result is a valid JSON string with expected content result_data = json.loads(result) assert len(result_data) == 1 assert result_data[0]["title"] == "Test Result" assert result_data[0]["url"] == "https://example.com" def test_clean_results(self): """Test clean_results method with sample raw results.""" raw_results = [ { "content": { "results": { "organic": [{"title": "Test Page", "desc": "Page description", "url": "https://example.com/page1"}], "top_stories": {"items": [{"title": "Test News", "source": "Test Source", "time_frame": "2 hours ago", "url": "https://example.com/news1"}]}, } } } ] cleaned = InfoQuestClient.clean_results(raw_results) assert len(cleaned) == 2 assert cleaned[0]["type"] == "page" assert cleaned[0]["title"] == "Test Page" assert cleaned[1]["type"] == "news" assert cleaned[1]["title"] == "Test News" def test_clean_results_with_image_search(self): """Test clean_results_with_image_search method with sample raw results.""" raw_results = [{"content": {"results": {"images_results": [{"image_url": "https://example.com/image1.jpg", "thumbnail_url": "https://example.com/thumb1.jpg","url": "https://example.com/page1"}]}}}] cleaned = InfoQuestClient.clean_results_with_image_search(raw_results) assert len(cleaned) == 1 assert cleaned[0]["image_url"] == "https://example.com/image1.jpg" assert cleaned[0]["thumbnail_url"] == "https://example.com/thumb1.jpg" assert cleaned[0]["url"] == "https://example.com/page1" @patch("src.community.infoquest.tools._get_infoquest_client") def test_web_search_tool(self, mock_get_client): """Test web_search_tool function.""" mock_client = MagicMock() mock_client.web_search.return_value = json.dumps([]) mock_get_client.return_value = mock_client result = tools.web_search_tool.run("test query") assert result == json.dumps([]) mock_get_client.assert_called_once() mock_client.web_search.assert_called_once_with("test query") @patch("src.community.infoquest.tools._get_infoquest_client") def test_web_fetch_tool(self, mock_get_client): """Test web_fetch_tool function.""" mock_client = MagicMock() mock_client.fetch.return_value = "<html><body>Test content</body></html>" mock_get_client.return_value = mock_client result = tools.web_fetch_tool.run("https://example.com") assert result == "# Untitled\n\nTest content" mock_get_client.assert_called_once() mock_client.fetch.assert_called_once_with("https://example.com") @patch("src.community.infoquest.tools.get_app_config") def test_get_infoquest_client(self, mock_get_app_config): """Test _get_infoquest_client function with config.""" mock_config = MagicMock() mock_config.get_tool_config.side_effect = [MagicMock(model_extra={"search_time_range": 24}), MagicMock(model_extra={"fetch_time": 10, "timeout": 30, "navigation_timeout": 60})] mock_get_app_config.return_value = mock_config client = tools._get_infoquest_client() assert client.search_time_range == 24 assert client.fetch_time == 10 assert client.fetch_timeout == 30 assert client.fetch_navigation_timeout == 60 @patch("src.community.infoquest.infoquest_client.requests.post") def test_web_search_api_error(self, mock_post): """Test web_search operation with API error.""" mock_post.side_effect = Exception("Connection error") client = InfoQuestClient() result = client.web_search("test query") assert "Error" in result
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_infoquest_client.py", "license": "MIT License", "lines": 147, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_subagent_executor.py
"""Tests for subagent executor async/sync execution paths. Covers: - SubagentExecutor.execute() synchronous execution path - SubagentExecutor._aexecute() asynchronous execution path - asyncio.run() properly executes async workflow within thread pool context - Error handling in both sync and async paths - Async tool support (MCP tools) Note: Due to circular import issues in the main codebase, conftest.py mocks src.subagents.executor. This test file uses delayed import via fixture to test the real implementation in isolation. """ import asyncio import sys from datetime import datetime from unittest.mock import MagicMock, patch import pytest # Module names that need to be mocked to break circular imports _MOCKED_MODULE_NAMES = [ "src.agents", "src.agents.thread_state", "src.agents.middlewares", "src.agents.middlewares.thread_data_middleware", "src.sandbox", "src.sandbox.middleware", "src.models", ] @pytest.fixture(scope="session", autouse=True) def _setup_executor_classes(): """Set up mocked modules and import real executor classes. This fixture runs once per session and yields the executor classes. It handles module cleanup to avoid affecting other test files. """ # Save original modules original_modules = {name: sys.modules.get(name) for name in _MOCKED_MODULE_NAMES} original_executor = sys.modules.get("src.subagents.executor") # Remove mocked executor if exists (from conftest.py) if "src.subagents.executor" in sys.modules: del sys.modules["src.subagents.executor"] # Set up mocks for name in _MOCKED_MODULE_NAMES: sys.modules[name] = MagicMock() # Import real classes inside fixture from langchain_core.messages import AIMessage, HumanMessage from src.subagents.config import SubagentConfig from src.subagents.executor import ( SubagentExecutor, SubagentResult, SubagentStatus, ) # Store classes in a dict to yield classes = { "AIMessage": AIMessage, "HumanMessage": HumanMessage, "SubagentConfig": SubagentConfig, "SubagentExecutor": SubagentExecutor, "SubagentResult": SubagentResult, "SubagentStatus": SubagentStatus, } yield classes # Cleanup: Restore original modules for name in _MOCKED_MODULE_NAMES: if original_modules[name] is not None: sys.modules[name] = original_modules[name] elif name in sys.modules: del sys.modules[name] # Restore executor module (conftest.py mock) if original_executor is not None: sys.modules["src.subagents.executor"] = original_executor elif "src.subagents.executor" in sys.modules: del sys.modules["src.subagents.executor"] # Helper classes that wrap real classes for testing class MockHumanMessage: """Mock HumanMessage for testing - wraps real class from fixture.""" def __init__(self, content, _classes=None): self._content = content self._classes = _classes def _get_real(self): return self._classes["HumanMessage"](content=self._content) class MockAIMessage: """Mock AIMessage for testing - wraps real class from fixture.""" def __init__(self, content, msg_id=None, _classes=None): self._content = content self._msg_id = msg_id self._classes = _classes def _get_real(self): msg = self._classes["AIMessage"](content=self._content) if self._msg_id: msg.id = self._msg_id return msg async def async_iterator(items): """Helper to create an async iterator from a list.""" for item in items: yield item # ----------------------------------------------------------------------------- # Fixtures # ----------------------------------------------------------------------------- @pytest.fixture def classes(_setup_executor_classes): """Provide access to executor classes.""" return _setup_executor_classes @pytest.fixture def base_config(classes): """Return a basic subagent config for testing.""" return classes["SubagentConfig"]( name="test-agent", description="Test agent", system_prompt="You are a test agent.", max_turns=10, timeout_seconds=60, ) @pytest.fixture def mock_agent(): """Return a properly configured mock agent with async stream.""" agent = MagicMock() agent.astream = MagicMock() return agent # Helper to create real message objects class _MsgHelper: """Helper to create real message objects from fixture classes.""" def __init__(self, classes): self.classes = classes def human(self, content): return self.classes["HumanMessage"](content=content) def ai(self, content, msg_id=None): msg = self.classes["AIMessage"](content=content) if msg_id: msg.id = msg_id return msg @pytest.fixture def msg(classes): """Provide message factory.""" return _MsgHelper(classes) # ----------------------------------------------------------------------------- # Async Execution Path Tests # ----------------------------------------------------------------------------- class TestAsyncExecutionPath: """Test _aexecute() async execution path.""" @pytest.mark.anyio async def test_aexecute_success(self, classes, base_config, mock_agent, msg): """Test successful async execution returns completed result.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] final_message = msg.ai("Task completed successfully", "msg-1") final_state = { "messages": [ msg.human("Do something"), final_message, ] } mock_agent.astream = lambda *args, **kwargs: async_iterator([final_state]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", trace_id="test-trace", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Do something") assert result.status == SubagentStatus.COMPLETED assert result.result == "Task completed successfully" assert result.error is None assert result.started_at is not None assert result.completed_at is not None @pytest.mark.anyio async def test_aexecute_collects_ai_messages(self, classes, base_config, mock_agent, msg): """Test that AI messages are collected during streaming.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] msg1 = msg.ai("First response", "msg-1") msg2 = msg.ai("Second response", "msg-2") chunk1 = {"messages": [msg.human("Task"), msg1]} chunk2 = {"messages": [msg.human("Task"), msg1, msg2]} mock_agent.astream = lambda *args, **kwargs: async_iterator([chunk1, chunk2]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") assert result.status == SubagentStatus.COMPLETED assert len(result.ai_messages) == 2 assert result.ai_messages[0]["id"] == "msg-1" assert result.ai_messages[1]["id"] == "msg-2" @pytest.mark.anyio async def test_aexecute_handles_duplicate_messages(self, classes, base_config, mock_agent, msg): """Test that duplicate AI messages are not added.""" SubagentExecutor = classes["SubagentExecutor"] msg1 = msg.ai("Response", "msg-1") # Same message appears in multiple chunks chunk1 = {"messages": [msg.human("Task"), msg1]} chunk2 = {"messages": [msg.human("Task"), msg1]} mock_agent.astream = lambda *args, **kwargs: async_iterator([chunk1, chunk2]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") assert len(result.ai_messages) == 1 @pytest.mark.anyio async def test_aexecute_handles_list_content(self, classes, base_config, mock_agent, msg): """Test handling of list-type content in AIMessage.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] final_message = msg.ai([{"text": "Part 1"}, {"text": "Part 2"}]) final_state = { "messages": [ msg.human("Task"), final_message, ] } mock_agent.astream = lambda *args, **kwargs: async_iterator([final_state]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") assert result.status == SubagentStatus.COMPLETED assert "Part 1" in result.result assert "Part 2" in result.result @pytest.mark.anyio async def test_aexecute_handles_agent_exception(self, classes, base_config, mock_agent): """Test that exceptions during execution are caught and returned as FAILED.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] mock_agent.astream.side_effect = Exception("Agent error") executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") assert result.status == SubagentStatus.FAILED assert "Agent error" in result.error assert result.completed_at is not None @pytest.mark.anyio async def test_aexecute_no_final_state(self, classes, base_config, mock_agent): """Test handling when no final state is returned.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] mock_agent.astream = lambda *args, **kwargs: async_iterator([]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") assert result.status == SubagentStatus.COMPLETED assert result.result == "No response generated" @pytest.mark.anyio async def test_aexecute_no_ai_message_in_state(self, classes, base_config, mock_agent, msg): """Test fallback when no AIMessage found in final state.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] final_state = {"messages": [msg.human("Task")]} mock_agent.astream = lambda *args, **kwargs: async_iterator([final_state]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") # Should fallback to string representation of last message assert result.status == SubagentStatus.COMPLETED assert "Task" in result.result # ----------------------------------------------------------------------------- # Sync Execution Path Tests # ----------------------------------------------------------------------------- class TestSyncExecutionPath: """Test execute() synchronous execution path with asyncio.run().""" def test_execute_runs_async_in_event_loop(self, classes, base_config, mock_agent, msg): """Test that execute() runs _aexecute() in a new event loop via asyncio.run().""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] final_message = msg.ai("Sync result", "msg-1") final_state = { "messages": [ msg.human("Task"), final_message, ] } mock_agent.astream = lambda *args, **kwargs: async_iterator([final_state]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = executor.execute("Task") assert result.status == SubagentStatus.COMPLETED assert result.result == "Sync result" def test_execute_in_thread_pool_context(self, classes, base_config, msg): """Test that execute() works correctly when called from a thread pool. This simulates the real-world usage where execute() is called from _execution_pool in execute_async(). """ from concurrent.futures import ThreadPoolExecutor SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] final_message = msg.ai("Thread pool result", "msg-1") final_state = { "messages": [ msg.human("Task"), final_message, ] } def run_in_thread(): mock_agent = MagicMock() mock_agent.astream = lambda *args, **kwargs: async_iterator([final_state]) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): return executor.execute("Task") # Execute in thread pool (simulating _execution_pool usage) with ThreadPoolExecutor(max_workers=1) as pool: future = pool.submit(run_in_thread) result = future.result(timeout=5) assert result.status == SubagentStatus.COMPLETED assert result.result == "Thread pool result" def test_execute_handles_asyncio_run_failure(self, classes, base_config): """Test handling when asyncio.run() itself fails.""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_aexecute") as mock_aexecute: mock_aexecute.side_effect = Exception("Asyncio run error") result = executor.execute("Task") assert result.status == SubagentStatus.FAILED assert "Asyncio run error" in result.error assert result.completed_at is not None def test_execute_with_result_holder(self, classes, base_config, mock_agent, msg): """Test execute() updates provided result_holder in real-time.""" SubagentExecutor = classes["SubagentExecutor"] SubagentResult = classes["SubagentResult"] SubagentStatus = classes["SubagentStatus"] msg1 = msg.ai("Step 1", "msg-1") chunk1 = {"messages": [msg.human("Task"), msg1]} mock_agent.astream = lambda *args, **kwargs: async_iterator([chunk1]) # Pre-create result holder (as done in execute_async) result_holder = SubagentResult( task_id="predefined-id", trace_id="test-trace", status=SubagentStatus.RUNNING, started_at=datetime.now(), ) executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = executor.execute("Task", result_holder=result_holder) # Should be the same object assert result is result_holder assert result.task_id == "predefined-id" assert result.status == SubagentStatus.COMPLETED # ----------------------------------------------------------------------------- # Async Tool Support Tests (MCP Tools) # ----------------------------------------------------------------------------- class TestAsyncToolSupport: """Test that async-only tools (like MCP tools) work correctly.""" @pytest.mark.anyio async def test_async_tool_called_in_astream(self, classes, base_config, msg): """Test that async tools are properly awaited in astream. This verifies the fix for: async MCP tools not being executed properly because they were being called synchronously. """ SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] async_tool_calls = [] async def mock_async_tool(*args, **kwargs): async_tool_calls.append("called") await asyncio.sleep(0.01) # Simulate async work return {"result": "async tool result"} mock_agent = MagicMock() # Simulate agent that calls async tools during streaming async def mock_astream(*args, **kwargs): await mock_async_tool() yield { "messages": [ msg.human("Task"), msg.ai("Done", "msg-1"), ] } mock_agent.astream = mock_astream executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = await executor._aexecute("Task") assert len(async_tool_calls) == 1 assert result.status == SubagentStatus.COMPLETED def test_sync_execute_with_async_tools(self, classes, base_config, msg): """Test that sync execute() properly runs async tools via asyncio.run().""" SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] async_tool_calls = [] async def mock_async_tool(): async_tool_calls.append("called") await asyncio.sleep(0.01) return {"result": "async result"} mock_agent = MagicMock() async def mock_astream(*args, **kwargs): await mock_async_tool() yield { "messages": [ msg.human("Task"), msg.ai("Done", "msg-1"), ] } mock_agent.astream = mock_astream executor = SubagentExecutor( config=base_config, tools=[], thread_id="test-thread", ) with patch.object(executor, "_create_agent", return_value=mock_agent): result = executor.execute("Task") assert len(async_tool_calls) == 1 assert result.status == SubagentStatus.COMPLETED # ----------------------------------------------------------------------------- # Thread Safety Tests # ----------------------------------------------------------------------------- class TestThreadSafety: """Test thread safety of executor operations.""" def test_multiple_executors_in_parallel(self, classes, base_config, msg): """Test multiple executors running in parallel via thread pool.""" from concurrent.futures import ThreadPoolExecutor, as_completed SubagentExecutor = classes["SubagentExecutor"] SubagentStatus = classes["SubagentStatus"] results = [] def execute_task(task_id: int): def make_astream(*args, **kwargs): return async_iterator( [ { "messages": [ msg.human(f"Task {task_id}"), msg.ai(f"Result {task_id}", f"msg-{task_id}"), ] } ] ) mock_agent = MagicMock() mock_agent.astream = make_astream executor = SubagentExecutor( config=base_config, tools=[], thread_id=f"thread-{task_id}", ) with patch.object(executor, "_create_agent", return_value=mock_agent): return executor.execute(f"Task {task_id}") # Execute multiple tasks in parallel with ThreadPoolExecutor(max_workers=3) as pool: futures = [pool.submit(execute_task, i) for i in range(5)] for future in as_completed(futures): results.append(future.result()) assert len(results) == 5 for result in results: assert result.status == SubagentStatus.COMPLETED assert "Result" in result.result
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_subagent_executor.py", "license": "MIT License", "lines": 477, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_uploads_middleware_core_logic.py
"""Core behaviour tests for UploadsMiddleware. Covers: - _files_from_kwargs: parsing, validation, existence check, virtual-path construction - _create_files_message: output format with new-only and new+historical files - before_agent: full injection pipeline (string & list content, preserved additional_kwargs, historical files from uploads dir, edge-cases) """ from pathlib import Path from unittest.mock import MagicMock from langchain_core.messages import AIMessage, HumanMessage from src.agents.middlewares.uploads_middleware import UploadsMiddleware from src.config.paths import Paths THREAD_ID = "thread-abc123" # --------------------------------------------------------------------------- # Helpers # --------------------------------------------------------------------------- def _middleware(tmp_path: Path) -> UploadsMiddleware: return UploadsMiddleware(base_dir=str(tmp_path)) def _runtime(thread_id: str | None = THREAD_ID) -> MagicMock: rt = MagicMock() rt.context = {"thread_id": thread_id} return rt def _uploads_dir(tmp_path: Path, thread_id: str = THREAD_ID) -> Path: d = Paths(str(tmp_path)).sandbox_uploads_dir(thread_id) d.mkdir(parents=True, exist_ok=True) return d def _human(content, files=None, **extra_kwargs): additional_kwargs = dict(extra_kwargs) if files is not None: additional_kwargs["files"] = files return HumanMessage(content=content, additional_kwargs=additional_kwargs) # --------------------------------------------------------------------------- # _files_from_kwargs # --------------------------------------------------------------------------- class TestFilesFromKwargs: def test_returns_none_when_files_field_absent(self, tmp_path): mw = _middleware(tmp_path) msg = HumanMessage(content="hello") assert mw._files_from_kwargs(msg) is None def test_returns_none_for_empty_files_list(self, tmp_path): mw = _middleware(tmp_path) msg = _human("hello", files=[]) assert mw._files_from_kwargs(msg) is None def test_returns_none_for_non_list_files(self, tmp_path): mw = _middleware(tmp_path) msg = _human("hello", files="not-a-list") assert mw._files_from_kwargs(msg) is None def test_skips_non_dict_entries(self, tmp_path): mw = _middleware(tmp_path) msg = _human("hi", files=["bad", 42, None]) assert mw._files_from_kwargs(msg) is None def test_skips_entries_with_empty_filename(self, tmp_path): mw = _middleware(tmp_path) msg = _human("hi", files=[{"filename": "", "size": 100, "path": "/mnt/user-data/uploads/x"}]) assert mw._files_from_kwargs(msg) is None def test_always_uses_virtual_path(self, tmp_path): """path field must be /mnt/user-data/uploads/<filename> regardless of what the frontend sent.""" mw = _middleware(tmp_path) msg = _human( "hi", files=[{"filename": "report.pdf", "size": 1024, "path": "/some/arbitrary/path/report.pdf"}], ) result = mw._files_from_kwargs(msg) assert result is not None assert result[0]["path"] == "/mnt/user-data/uploads/report.pdf" def test_skips_file_that_does_not_exist_on_disk(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) # file is NOT written to disk msg = _human("hi", files=[{"filename": "missing.txt", "size": 50, "path": "/mnt/user-data/uploads/missing.txt"}]) assert mw._files_from_kwargs(msg, uploads_dir) is None def test_accepts_file_that_exists_on_disk(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "data.csv").write_text("a,b,c") msg = _human("hi", files=[{"filename": "data.csv", "size": 5, "path": "/mnt/user-data/uploads/data.csv"}]) result = mw._files_from_kwargs(msg, uploads_dir) assert result is not None assert len(result) == 1 assert result[0]["filename"] == "data.csv" assert result[0]["path"] == "/mnt/user-data/uploads/data.csv" def test_skips_nonexistent_but_accepts_existing_in_mixed_list(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "present.txt").write_text("here") msg = _human( "hi", files=[ {"filename": "present.txt", "size": 4, "path": "/mnt/user-data/uploads/present.txt"}, {"filename": "gone.txt", "size": 4, "path": "/mnt/user-data/uploads/gone.txt"}, ], ) result = mw._files_from_kwargs(msg, uploads_dir) assert result is not None assert [f["filename"] for f in result] == ["present.txt"] def test_no_existence_check_when_uploads_dir_is_none(self, tmp_path): """Without an uploads_dir argument the existence check is skipped entirely.""" mw = _middleware(tmp_path) msg = _human("hi", files=[{"filename": "phantom.txt", "size": 10, "path": "/mnt/user-data/uploads/phantom.txt"}]) result = mw._files_from_kwargs(msg, uploads_dir=None) assert result is not None assert result[0]["filename"] == "phantom.txt" def test_size_is_coerced_to_int(self, tmp_path): mw = _middleware(tmp_path) msg = _human("hi", files=[{"filename": "f.txt", "size": "2048", "path": "/mnt/user-data/uploads/f.txt"}]) result = mw._files_from_kwargs(msg) assert result is not None assert result[0]["size"] == 2048 def test_missing_size_defaults_to_zero(self, tmp_path): mw = _middleware(tmp_path) msg = _human("hi", files=[{"filename": "f.txt", "path": "/mnt/user-data/uploads/f.txt"}]) result = mw._files_from_kwargs(msg) assert result is not None assert result[0]["size"] == 0 # --------------------------------------------------------------------------- # _create_files_message # --------------------------------------------------------------------------- class TestCreateFilesMessage: def _new_file(self, filename="notes.txt", size=1024): return {"filename": filename, "size": size, "path": f"/mnt/user-data/uploads/{filename}"} def test_new_files_section_always_present(self, tmp_path): mw = _middleware(tmp_path) msg = mw._create_files_message([self._new_file()], []) assert "<uploaded_files>" in msg assert "</uploaded_files>" in msg assert "uploaded in this message" in msg assert "notes.txt" in msg assert "/mnt/user-data/uploads/notes.txt" in msg def test_historical_section_present_only_when_non_empty(self, tmp_path): mw = _middleware(tmp_path) msg_no_hist = mw._create_files_message([self._new_file()], []) assert "previous messages" not in msg_no_hist hist = self._new_file("old.txt") msg_with_hist = mw._create_files_message([self._new_file()], [hist]) assert "previous messages" in msg_with_hist assert "old.txt" in msg_with_hist def test_size_formatting_kb(self, tmp_path): mw = _middleware(tmp_path) msg = mw._create_files_message([self._new_file(size=2048)], []) assert "2.0 KB" in msg def test_size_formatting_mb(self, tmp_path): mw = _middleware(tmp_path) msg = mw._create_files_message([self._new_file(size=2 * 1024 * 1024)], []) assert "2.0 MB" in msg def test_read_file_instruction_included(self, tmp_path): mw = _middleware(tmp_path) msg = mw._create_files_message([self._new_file()], []) assert "read_file" in msg def test_empty_new_files_produces_empty_marker(self, tmp_path): mw = _middleware(tmp_path) msg = mw._create_files_message([], []) assert "(empty)" in msg assert "<uploaded_files>" in msg assert "</uploaded_files>" in msg # --------------------------------------------------------------------------- # before_agent # --------------------------------------------------------------------------- class TestBeforeAgent: def _state(self, *messages): return {"messages": list(messages)} def test_returns_none_when_messages_empty(self, tmp_path): mw = _middleware(tmp_path) assert mw.before_agent({"messages": []}, _runtime()) is None def test_returns_none_when_last_message_is_not_human(self, tmp_path): mw = _middleware(tmp_path) state = self._state(HumanMessage(content="q"), AIMessage(content="a")) assert mw.before_agent(state, _runtime()) is None def test_returns_none_when_no_files_in_kwargs(self, tmp_path): mw = _middleware(tmp_path) state = self._state(_human("plain message")) assert mw.before_agent(state, _runtime()) is None def test_returns_none_when_all_files_missing_from_disk(self, tmp_path): mw = _middleware(tmp_path) _uploads_dir(tmp_path) # directory exists but is empty msg = _human("hi", files=[{"filename": "ghost.txt", "size": 10, "path": "/mnt/user-data/uploads/ghost.txt"}]) state = self._state(msg) assert mw.before_agent(state, _runtime()) is None def test_injects_uploaded_files_tag_into_string_content(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "report.pdf").write_bytes(b"pdf") msg = _human("please analyse", files=[{"filename": "report.pdf", "size": 3, "path": "/mnt/user-data/uploads/report.pdf"}]) state = self._state(msg) result = mw.before_agent(state, _runtime()) assert result is not None updated_msg = result["messages"][-1] assert isinstance(updated_msg.content, str) assert "<uploaded_files>" in updated_msg.content assert "report.pdf" in updated_msg.content assert "please analyse" in updated_msg.content def test_injects_uploaded_files_tag_into_list_content(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "data.csv").write_bytes(b"a,b") msg = _human( [{"type": "text", "text": "analyse this"}], files=[{"filename": "data.csv", "size": 3, "path": "/mnt/user-data/uploads/data.csv"}], ) state = self._state(msg) result = mw.before_agent(state, _runtime()) assert result is not None updated_msg = result["messages"][-1] assert "<uploaded_files>" in updated_msg.content assert "analyse this" in updated_msg.content def test_preserves_additional_kwargs_on_updated_message(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "img.png").write_bytes(b"png") files_meta = [{"filename": "img.png", "size": 3, "path": "/mnt/user-data/uploads/img.png", "status": "uploaded"}] msg = _human("check image", files=files_meta, element="task") state = self._state(msg) result = mw.before_agent(state, _runtime()) assert result is not None updated_kwargs = result["messages"][-1].additional_kwargs assert updated_kwargs.get("files") == files_meta assert updated_kwargs.get("element") == "task" def test_uploaded_files_returned_in_state_update(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "notes.txt").write_bytes(b"hello") msg = _human("review", files=[{"filename": "notes.txt", "size": 5, "path": "/mnt/user-data/uploads/notes.txt"}]) result = mw.before_agent(self._state(msg), _runtime()) assert result is not None assert result["uploaded_files"] == [ { "filename": "notes.txt", "size": 5, "path": "/mnt/user-data/uploads/notes.txt", "extension": ".txt", } ] def test_historical_files_from_uploads_dir_excluding_new(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "old.txt").write_bytes(b"old") (uploads_dir / "new.txt").write_bytes(b"new") msg = _human("go", files=[{"filename": "new.txt", "size": 3, "path": "/mnt/user-data/uploads/new.txt"}]) result = mw.before_agent(self._state(msg), _runtime()) assert result is not None content = result["messages"][-1].content assert "uploaded in this message" in content assert "new.txt" in content assert "previous messages" in content assert "old.txt" in content def test_no_historical_section_when_upload_dir_is_empty(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "only.txt").write_bytes(b"x") msg = _human("go", files=[{"filename": "only.txt", "size": 1, "path": "/mnt/user-data/uploads/only.txt"}]) result = mw.before_agent(self._state(msg), _runtime()) content = result["messages"][-1].content assert "previous messages" not in content def test_no_historical_scan_when_thread_id_is_none(self, tmp_path): mw = _middleware(tmp_path) msg = _human("go", files=[{"filename": "f.txt", "size": 1, "path": "/mnt/user-data/uploads/f.txt"}]) # thread_id=None β†’ _files_from_kwargs skips existence check, no dir scan result = mw.before_agent(self._state(msg), _runtime(thread_id=None)) # With no existence check, the file passes through and injection happens assert result is not None content = result["messages"][-1].content assert "previous messages" not in content def test_message_id_preserved_on_updated_message(self, tmp_path): mw = _middleware(tmp_path) uploads_dir = _uploads_dir(tmp_path) (uploads_dir / "f.txt").write_bytes(b"x") msg = _human("go", files=[{"filename": "f.txt", "size": 1, "path": "/mnt/user-data/uploads/f.txt"}]) msg.id = "original-id-42" result = mw.before_agent(self._state(msg), _runtime()) assert result["messages"][-1].id == "original-id-42"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_uploads_middleware_core_logic.py", "license": "MIT License", "lines": 269, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_memory_upload_filtering.py
"""Tests for upload-event filtering in the memory pipeline. Covers two functions introduced to prevent ephemeral file-upload context from persisting in long-term memory: - _filter_messages_for_memory (memory_middleware) - _strip_upload_mentions_from_memory (updater) """ from langchain_core.messages import AIMessage, HumanMessage, ToolMessage from src.agents.memory.updater import _strip_upload_mentions_from_memory from src.agents.middlewares.memory_middleware import _filter_messages_for_memory # --------------------------------------------------------------------------- # Helpers # --------------------------------------------------------------------------- _UPLOAD_BLOCK = ( "<uploaded_files>\n" "The following files have been uploaded and are available for use:\n\n" "- filename: secret.txt\n" " path: /mnt/user-data/uploads/abc123/secret.txt\n" " size: 42 bytes\n" "</uploaded_files>" ) def _human(text: str) -> HumanMessage: return HumanMessage(content=text) def _ai(text: str, tool_calls=None) -> AIMessage: msg = AIMessage(content=text) if tool_calls: msg.tool_calls = tool_calls return msg # =========================================================================== # _filter_messages_for_memory # =========================================================================== class TestFilterMessagesForMemory: # --- upload-only turns are excluded --- def test_upload_only_turn_is_excluded(self): """A human turn containing only <uploaded_files> (no real question) and its paired AI response must both be dropped.""" msgs = [ _human(_UPLOAD_BLOCK), _ai("I have read the file. It says: Hello."), ] result = _filter_messages_for_memory(msgs) assert result == [] def test_upload_with_real_question_preserves_question(self): """When the user asks a question alongside an upload, the question text must reach the memory queue (upload block stripped, AI response kept).""" combined = _UPLOAD_BLOCK + "\n\nWhat does this file contain?" msgs = [ _human(combined), _ai("The file contains: Hello DeerFlow."), ] result = _filter_messages_for_memory(msgs) assert len(result) == 2 human_result = result[0] assert "<uploaded_files>" not in human_result.content assert "What does this file contain?" in human_result.content assert result[1].content == "The file contains: Hello DeerFlow." # --- non-upload turns pass through unchanged --- def test_plain_conversation_passes_through(self): msgs = [ _human("What is the capital of France?"), _ai("The capital of France is Paris."), ] result = _filter_messages_for_memory(msgs) assert len(result) == 2 assert result[0].content == "What is the capital of France?" assert result[1].content == "The capital of France is Paris." def test_tool_messages_are_excluded(self): """Intermediate tool messages must never reach memory.""" msgs = [ _human("Search for something"), _ai("Calling search tool", tool_calls=[{"name": "search", "id": "1", "args": {}}]), ToolMessage(content="Search results", tool_call_id="1"), _ai("Here are the results."), ] result = _filter_messages_for_memory(msgs) human_msgs = [m for m in result if m.type == "human"] ai_msgs = [m for m in result if m.type == "ai"] assert len(human_msgs) == 1 assert len(ai_msgs) == 1 assert ai_msgs[0].content == "Here are the results." def test_multi_turn_with_upload_in_middle(self): """Only the upload turn is dropped; surrounding non-upload turns survive.""" msgs = [ _human("Hello, how are you?"), _ai("I'm doing well, thank you!"), _human(_UPLOAD_BLOCK), # upload-only β†’ dropped _ai("I read the uploaded file."), # paired AI β†’ dropped _human("What is 2 + 2?"), _ai("4"), ] result = _filter_messages_for_memory(msgs) human_contents = [m.content for m in result if m.type == "human"] ai_contents = [m.content for m in result if m.type == "ai"] assert "Hello, how are you?" in human_contents assert "What is 2 + 2?" in human_contents assert _UPLOAD_BLOCK not in human_contents assert "I'm doing well, thank you!" in ai_contents assert "4" in ai_contents # The upload-paired AI response must NOT appear assert "I read the uploaded file." not in ai_contents def test_multimodal_content_list_handled(self): """Human messages with list-style content (multimodal) are handled.""" msg = HumanMessage(content=[ {"type": "text", "text": _UPLOAD_BLOCK}, ]) msgs = [msg, _ai("Done.")] result = _filter_messages_for_memory(msgs) assert result == [] def test_file_path_not_in_filtered_content(self): """After filtering, no upload file path should appear in any message.""" combined = _UPLOAD_BLOCK + "\n\nSummarise the file please." msgs = [_human(combined), _ai("It says hello.")] result = _filter_messages_for_memory(msgs) all_content = " ".join( m.content for m in result if isinstance(m.content, str) ) assert "/mnt/user-data/uploads/" not in all_content assert "<uploaded_files>" not in all_content # =========================================================================== # _strip_upload_mentions_from_memory # =========================================================================== class TestStripUploadMentionsFromMemory: def _make_memory(self, summary: str, facts: list[dict] | None = None) -> dict: return { "user": {"topOfMind": {"summary": summary}}, "history": {"recentMonths": {"summary": ""}}, "facts": facts or [], } # --- summaries --- def test_upload_event_sentence_removed_from_summary(self): mem = self._make_memory( "User is interested in AI. " "User uploaded a test file for verification purposes. " "User prefers concise answers." ) result = _strip_upload_mentions_from_memory(mem) summary = result["user"]["topOfMind"]["summary"] assert "uploaded a test file" not in summary assert "User is interested in AI" in summary assert "User prefers concise answers" in summary def test_upload_path_sentence_removed_from_summary(self): mem = self._make_memory( "User uses Python. " "User uploaded file to /mnt/user-data/uploads/tid/data.csv. " "User likes clean code." ) result = _strip_upload_mentions_from_memory(mem) summary = result["user"]["topOfMind"]["summary"] assert "/mnt/user-data/uploads/" not in summary assert "User uses Python" in summary def test_legitimate_csv_mention_is_preserved(self): """'User works with CSV files' must NOT be deleted β€” it's not an upload event.""" mem = self._make_memory("User regularly works with CSV files for data analysis.") result = _strip_upload_mentions_from_memory(mem) assert "CSV files" in result["user"]["topOfMind"]["summary"] def test_pdf_export_preference_preserved(self): """'Prefers PDF export' is a legitimate preference, not an upload event.""" mem = self._make_memory("User prefers PDF export for reports.") result = _strip_upload_mentions_from_memory(mem) assert "PDF export" in result["user"]["topOfMind"]["summary"] def test_uploading_a_test_file_removed(self): """'uploading a test file' (with intervening words) must be caught.""" mem = self._make_memory( "User conducted a hands-on test by uploading a test file titled " "'test_deerflow_memory_bug.txt'. User is also learning Python." ) result = _strip_upload_mentions_from_memory(mem) summary = result["user"]["topOfMind"]["summary"] assert "test_deerflow_memory_bug.txt" not in summary assert "uploading a test file" not in summary # --- facts --- def test_upload_fact_removed_from_facts(self): facts = [ {"content": "User uploaded a file titled secret.txt", "category": "behavior"}, {"content": "User prefers dark mode", "category": "preference"}, {"content": "User is uploading document attachments regularly", "category": "behavior"}, ] mem = self._make_memory("summary", facts=facts) result = _strip_upload_mentions_from_memory(mem) remaining = [f["content"] for f in result["facts"]] assert "User prefers dark mode" in remaining assert not any("uploaded a file" in c for c in remaining) assert not any("uploading document" in c for c in remaining) def test_non_upload_facts_preserved(self): facts = [ {"content": "User graduated from Peking University", "category": "context"}, {"content": "User prefers Python over JavaScript", "category": "preference"}, ] mem = self._make_memory("", facts=facts) result = _strip_upload_mentions_from_memory(mem) assert len(result["facts"]) == 2 def test_empty_memory_handled_gracefully(self): mem = {"user": {}, "history": {}, "facts": []} result = _strip_upload_mentions_from_memory(mem) assert result == {"user": {}, "history": {}, "facts": []}
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_memory_upload_filtering.py", "license": "MIT License", "lines": 194, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/config/agents_config.py
"""Configuration and loaders for custom agents.""" import logging import re from typing import Any import yaml from pydantic import BaseModel from src.config.paths import get_paths logger = logging.getLogger(__name__) SOUL_FILENAME = "SOUL.md" AGENT_NAME_PATTERN = re.compile(r"^[A-Za-z0-9-]+$") class AgentConfig(BaseModel): """Configuration for a custom agent.""" name: str description: str = "" model: str | None = None tool_groups: list[str] | None = None def load_agent_config(name: str | None) -> AgentConfig | None: """Load the custom or default agent's config from its directory. Args: name: The agent name. Returns: AgentConfig instance. Raises: FileNotFoundError: If the agent directory or config.yaml does not exist. ValueError: If config.yaml cannot be parsed. """ if name is None: return None if not AGENT_NAME_PATTERN.match(name): raise ValueError(f"Invalid agent name '{name}'. Must match pattern: {AGENT_NAME_PATTERN.pattern}") agent_dir = get_paths().agent_dir(name) config_file = agent_dir / "config.yaml" if not agent_dir.exists(): raise FileNotFoundError(f"Agent directory not found: {agent_dir}") if not config_file.exists(): raise FileNotFoundError(f"Agent config not found: {config_file}") try: with open(config_file, encoding="utf-8") as f: data: dict[str, Any] = yaml.safe_load(f) or {} except yaml.YAMLError as e: raise ValueError(f"Failed to parse agent config {config_file}: {e}") from e # Ensure name is set from directory name if not in file if "name" not in data: data["name"] = name # Strip unknown fields before passing to Pydantic (e.g. legacy prompt_file) known_fields = set(AgentConfig.model_fields.keys()) data = {k: v for k, v in data.items() if k in known_fields} return AgentConfig(**data) def load_agent_soul(agent_name: str | None) -> str | None: """Read the SOUL.md file for a custom agent, if it exists. SOUL.md defines the agent's personality, values, and behavioral guardrails. It is injected into the lead agent's system prompt as additional context. Args: agent_name: The name of the agent or None for the default agent. Returns: The SOUL.md content as a string, or None if the file does not exist. """ agent_dir = get_paths().agent_dir(agent_name) if agent_name else get_paths().base_dir soul_path = agent_dir / SOUL_FILENAME if not soul_path.exists(): return None content = soul_path.read_text(encoding="utf-8").strip() return content or None def list_custom_agents() -> list[AgentConfig]: """Scan the agents directory and return all valid custom agents. Returns: List of AgentConfig for each valid agent directory found. """ agents_dir = get_paths().agents_dir if not agents_dir.exists(): return [] agents: list[AgentConfig] = [] for entry in sorted(agents_dir.iterdir()): if not entry.is_dir(): continue config_file = entry / "config.yaml" if not config_file.exists(): logger.debug(f"Skipping {entry.name}: no config.yaml") continue try: agent_cfg = load_agent_config(entry.name) agents.append(agent_cfg) except Exception as e: logger.warning(f"Skipping agent '{entry.name}': {e}") return agents
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/agents_config.py", "license": "MIT License", "lines": 85, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/gateway/routers/agents.py
"""CRUD API for custom agents.""" import logging import re import shutil import yaml from fastapi import APIRouter, HTTPException from pydantic import BaseModel, Field from src.config.agents_config import AgentConfig, list_custom_agents, load_agent_config, load_agent_soul from src.config.paths import get_paths logger = logging.getLogger(__name__) router = APIRouter(prefix="/api", tags=["agents"]) AGENT_NAME_PATTERN = re.compile(r"^[A-Za-z0-9-]+$") class AgentResponse(BaseModel): """Response model for a custom agent.""" name: str = Field(..., description="Agent name (hyphen-case)") description: str = Field(default="", description="Agent description") model: str | None = Field(default=None, description="Optional model override") tool_groups: list[str] | None = Field(default=None, description="Optional tool group whitelist") soul: str | None = Field(default=None, description="SOUL.md content (included on GET /{name})") class AgentsListResponse(BaseModel): """Response model for listing all custom agents.""" agents: list[AgentResponse] class AgentCreateRequest(BaseModel): """Request body for creating a custom agent.""" name: str = Field(..., description="Agent name (must match ^[A-Za-z0-9-]+$, stored as lowercase)") description: str = Field(default="", description="Agent description") model: str | None = Field(default=None, description="Optional model override") tool_groups: list[str] | None = Field(default=None, description="Optional tool group whitelist") soul: str = Field(default="", description="SOUL.md content β€” agent personality and behavioral guardrails") class AgentUpdateRequest(BaseModel): """Request body for updating a custom agent.""" description: str | None = Field(default=None, description="Updated description") model: str | None = Field(default=None, description="Updated model override") tool_groups: list[str] | None = Field(default=None, description="Updated tool group whitelist") soul: str | None = Field(default=None, description="Updated SOUL.md content") def _validate_agent_name(name: str) -> None: """Validate agent name against allowed pattern. Args: name: The agent name to validate. Raises: HTTPException: 422 if the name is invalid. """ if not AGENT_NAME_PATTERN.match(name): raise HTTPException( status_code=422, detail=f"Invalid agent name '{name}'. Must match ^[A-Za-z0-9-]+$ (letters, digits, and hyphens only).", ) def _normalize_agent_name(name: str) -> str: """Normalize agent name to lowercase for filesystem storage.""" return name.lower() def _agent_config_to_response(agent_cfg: AgentConfig, include_soul: bool = False) -> AgentResponse: """Convert AgentConfig to AgentResponse.""" soul: str | None = None if include_soul: soul = load_agent_soul(agent_cfg.name) or "" return AgentResponse( name=agent_cfg.name, description=agent_cfg.description, model=agent_cfg.model, tool_groups=agent_cfg.tool_groups, soul=soul, ) @router.get( "/agents", response_model=AgentsListResponse, summary="List Custom Agents", description="List all custom agents available in the agents directory.", ) async def list_agents() -> AgentsListResponse: """List all custom agents. Returns: List of all custom agents with their metadata (without soul content). """ try: agents = list_custom_agents() return AgentsListResponse(agents=[_agent_config_to_response(a) for a in agents]) except Exception as e: logger.error(f"Failed to list agents: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to list agents: {str(e)}") @router.get( "/agents/check", summary="Check Agent Name", description="Validate an agent name and check if it is available (case-insensitive).", ) async def check_agent_name(name: str) -> dict: """Check whether an agent name is valid and not yet taken. Args: name: The agent name to check. Returns: ``{"available": true/false, "name": "<normalized>"}`` Raises: HTTPException: 422 if the name is invalid. """ _validate_agent_name(name) normalized = _normalize_agent_name(name) available = not get_paths().agent_dir(normalized).exists() return {"available": available, "name": normalized} @router.get( "/agents/{name}", response_model=AgentResponse, summary="Get Custom Agent", description="Retrieve details and SOUL.md content for a specific custom agent.", ) async def get_agent(name: str) -> AgentResponse: """Get a specific custom agent by name. Args: name: The agent name. Returns: Agent details including SOUL.md content. Raises: HTTPException: 404 if agent not found. """ _validate_agent_name(name) name = _normalize_agent_name(name) try: agent_cfg = load_agent_config(name) return _agent_config_to_response(agent_cfg, include_soul=True) except FileNotFoundError: raise HTTPException(status_code=404, detail=f"Agent '{name}' not found") except Exception as e: logger.error(f"Failed to get agent '{name}': {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to get agent: {str(e)}") @router.post( "/agents", response_model=AgentResponse, status_code=201, summary="Create Custom Agent", description="Create a new custom agent with its config and SOUL.md.", ) async def create_agent_endpoint(request: AgentCreateRequest) -> AgentResponse: """Create a new custom agent. Args: request: The agent creation request. Returns: The created agent details. Raises: HTTPException: 409 if agent already exists, 422 if name is invalid. """ _validate_agent_name(request.name) normalized_name = _normalize_agent_name(request.name) agent_dir = get_paths().agent_dir(normalized_name) if agent_dir.exists(): raise HTTPException(status_code=409, detail=f"Agent '{normalized_name}' already exists") try: agent_dir.mkdir(parents=True, exist_ok=True) # Write config.yaml config_data: dict = {"name": normalized_name} if request.description: config_data["description"] = request.description if request.model is not None: config_data["model"] = request.model if request.tool_groups is not None: config_data["tool_groups"] = request.tool_groups config_file = agent_dir / "config.yaml" with open(config_file, "w", encoding="utf-8") as f: yaml.dump(config_data, f, default_flow_style=False, allow_unicode=True) # Write SOUL.md soul_file = agent_dir / "SOUL.md" soul_file.write_text(request.soul, encoding="utf-8") logger.info(f"Created agent '{normalized_name}' at {agent_dir}") agent_cfg = load_agent_config(normalized_name) return _agent_config_to_response(agent_cfg, include_soul=True) except HTTPException: raise except Exception as e: # Clean up on failure if agent_dir.exists(): shutil.rmtree(agent_dir) logger.error(f"Failed to create agent '{request.name}': {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to create agent: {str(e)}") @router.put( "/agents/{name}", response_model=AgentResponse, summary="Update Custom Agent", description="Update an existing custom agent's config and/or SOUL.md.", ) async def update_agent(name: str, request: AgentUpdateRequest) -> AgentResponse: """Update an existing custom agent. Args: name: The agent name. request: The update request (all fields optional). Returns: The updated agent details. Raises: HTTPException: 404 if agent not found. """ _validate_agent_name(name) name = _normalize_agent_name(name) try: agent_cfg = load_agent_config(name) except FileNotFoundError: raise HTTPException(status_code=404, detail=f"Agent '{name}' not found") agent_dir = get_paths().agent_dir(name) try: # Update config if any config fields changed config_changed = any(v is not None for v in [request.description, request.model, request.tool_groups]) if config_changed: updated: dict = { "name": agent_cfg.name, "description": request.description if request.description is not None else agent_cfg.description, } new_model = request.model if request.model is not None else agent_cfg.model if new_model is not None: updated["model"] = new_model new_tool_groups = request.tool_groups if request.tool_groups is not None else agent_cfg.tool_groups if new_tool_groups is not None: updated["tool_groups"] = new_tool_groups config_file = agent_dir / "config.yaml" with open(config_file, "w", encoding="utf-8") as f: yaml.dump(updated, f, default_flow_style=False, allow_unicode=True) # Update SOUL.md if provided if request.soul is not None: soul_path = agent_dir / "SOUL.md" soul_path.write_text(request.soul, encoding="utf-8") logger.info(f"Updated agent '{name}'") refreshed_cfg = load_agent_config(name) return _agent_config_to_response(refreshed_cfg, include_soul=True) except HTTPException: raise except Exception as e: logger.error(f"Failed to update agent '{name}': {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to update agent: {str(e)}") class UserProfileResponse(BaseModel): """Response model for the global user profile (USER.md).""" content: str | None = Field(default=None, description="USER.md content, or null if not yet created") class UserProfileUpdateRequest(BaseModel): """Request body for setting the global user profile.""" content: str = Field(default="", description="USER.md content β€” describes the user's background and preferences") @router.get( "/user-profile", response_model=UserProfileResponse, summary="Get User Profile", description="Read the global USER.md file that is injected into all custom agents.", ) async def get_user_profile() -> UserProfileResponse: """Return the current USER.md content. Returns: UserProfileResponse with content=None if USER.md does not exist yet. """ try: user_md_path = get_paths().user_md_file if not user_md_path.exists(): return UserProfileResponse(content=None) raw = user_md_path.read_text(encoding="utf-8").strip() return UserProfileResponse(content=raw or None) except Exception as e: logger.error(f"Failed to read user profile: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to read user profile: {str(e)}") @router.put( "/user-profile", response_model=UserProfileResponse, summary="Update User Profile", description="Write the global USER.md file that is injected into all custom agents.", ) async def update_user_profile(request: UserProfileUpdateRequest) -> UserProfileResponse: """Create or overwrite the global USER.md. Args: request: The update request with the new USER.md content. Returns: UserProfileResponse with the saved content. """ try: paths = get_paths() paths.base_dir.mkdir(parents=True, exist_ok=True) paths.user_md_file.write_text(request.content, encoding="utf-8") logger.info(f"Updated USER.md at {paths.user_md_file}") return UserProfileResponse(content=request.content or None) except Exception as e: logger.error(f"Failed to update user profile: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to update user profile: {str(e)}") @router.delete( "/agents/{name}", status_code=204, summary="Delete Custom Agent", description="Delete a custom agent and all its files (config, SOUL.md, memory).", ) async def delete_agent(name: str) -> None: """Delete a custom agent. Args: name: The agent name. Raises: HTTPException: 404 if agent not found. """ _validate_agent_name(name) name = _normalize_agent_name(name) agent_dir = get_paths().agent_dir(name) if not agent_dir.exists(): raise HTTPException(status_code=404, detail=f"Agent '{name}' not found") try: shutil.rmtree(agent_dir) logger.info(f"Deleted agent '{name}' from {agent_dir}") except Exception as e: logger.error(f"Failed to delete agent '{name}': {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to delete agent: {str(e)}")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/agents.py", "license": "MIT License", "lines": 294, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/tools/builtins/setup_agent_tool.py
import logging import yaml from langchain_core.messages import ToolMessage from langchain_core.tools import tool from langgraph.prebuilt import ToolRuntime from langgraph.types import Command from src.config.paths import get_paths logger = logging.getLogger(__name__) @tool def setup_agent( soul: str, description: str, runtime: ToolRuntime, ) -> Command: """Setup the custom DeerFlow agent. Args: soul: Full SOUL.md content defining the agent's personality and behavior. description: One-line description of what the agent does. """ agent_name: str | None = runtime.context.get("agent_name") try: paths = get_paths() agent_dir = paths.agent_dir(agent_name) if agent_name else paths.base_dir agent_dir.mkdir(parents=True, exist_ok=True) if agent_name: # If agent_name is provided, we are creating a custom agent in the agents/ directory config_data: dict = {"name": agent_name} if description: config_data["description"] = description config_file = agent_dir / "config.yaml" with open(config_file, "w", encoding="utf-8") as f: yaml.dump(config_data, f, default_flow_style=False, allow_unicode=True) soul_file = agent_dir / "SOUL.md" soul_file.write_text(soul, encoding="utf-8") logger.info(f"[agent_creator] Created agent '{agent_name}' at {agent_dir}") return Command( update={ "created_agent_name": agent_name, "messages": [ToolMessage(content=f"Agent '{agent_name}' created successfully!", tool_call_id=runtime.tool_call_id)], } ) except Exception as e: import shutil if agent_name and agent_dir.exists(): # Cleanup the custom agent directory only if it was created but an error occurred during setup shutil.rmtree(agent_dir) logger.error(f"[agent_creator] Failed to create agent '{agent_name}': {e}", exc_info=True) return Command(update={"messages": [ToolMessage(content=f"Error: {e}", tool_call_id=runtime.tool_call_id)]})
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/tools/builtins/setup_agent_tool.py", "license": "MIT License", "lines": 48, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/tests/test_custom_agent.py
"""Tests for custom agent support.""" from __future__ import annotations from pathlib import Path from unittest.mock import patch import pytest import yaml from fastapi.testclient import TestClient # --------------------------------------------------------------------------- # Helpers # --------------------------------------------------------------------------- def _make_paths(base_dir: Path): """Return a Paths instance pointing to base_dir.""" from src.config.paths import Paths return Paths(base_dir=base_dir) def _write_agent(base_dir: Path, name: str, config: dict, soul: str = "You are helpful.") -> None: """Write an agent directory with config.yaml and SOUL.md.""" agent_dir = base_dir / "agents" / name agent_dir.mkdir(parents=True, exist_ok=True) config_copy = dict(config) if "name" not in config_copy: config_copy["name"] = name with open(agent_dir / "config.yaml", "w") as f: yaml.dump(config_copy, f) (agent_dir / "SOUL.md").write_text(soul, encoding="utf-8") # =========================================================================== # 1. Paths class – agent path methods # =========================================================================== class TestPaths: def test_agents_dir(self, tmp_path): paths = _make_paths(tmp_path) assert paths.agents_dir == tmp_path / "agents" def test_agent_dir(self, tmp_path): paths = _make_paths(tmp_path) assert paths.agent_dir("code-reviewer") == tmp_path / "agents" / "code-reviewer" def test_agent_memory_file(self, tmp_path): paths = _make_paths(tmp_path) assert paths.agent_memory_file("code-reviewer") == tmp_path / "agents" / "code-reviewer" / "memory.json" def test_user_md_file(self, tmp_path): paths = _make_paths(tmp_path) assert paths.user_md_file == tmp_path / "USER.md" def test_paths_are_different_from_global(self, tmp_path): paths = _make_paths(tmp_path) assert paths.memory_file != paths.agent_memory_file("my-agent") assert paths.memory_file == tmp_path / "memory.json" assert paths.agent_memory_file("my-agent") == tmp_path / "agents" / "my-agent" / "memory.json" # =========================================================================== # 2. AgentConfig – Pydantic parsing # =========================================================================== class TestAgentConfig: def test_minimal_config(self): from src.config.agents_config import AgentConfig cfg = AgentConfig(name="my-agent") assert cfg.name == "my-agent" assert cfg.description == "" assert cfg.model is None assert cfg.tool_groups is None def test_full_config(self): from src.config.agents_config import AgentConfig cfg = AgentConfig( name="code-reviewer", description="Specialized for code review", model="deepseek-v3", tool_groups=["file:read", "bash"], ) assert cfg.name == "code-reviewer" assert cfg.model == "deepseek-v3" assert cfg.tool_groups == ["file:read", "bash"] def test_config_from_dict(self): from src.config.agents_config import AgentConfig data = {"name": "test-agent", "description": "A test", "model": "gpt-4"} cfg = AgentConfig(**data) assert cfg.name == "test-agent" assert cfg.model == "gpt-4" assert cfg.tool_groups is None # =========================================================================== # 3. load_agent_config # =========================================================================== class TestLoadAgentConfig: def test_load_valid_config(self, tmp_path): config_dict = {"name": "code-reviewer", "description": "Code review agent", "model": "deepseek-v3"} _write_agent(tmp_path, "code-reviewer", config_dict) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import load_agent_config cfg = load_agent_config("code-reviewer") assert cfg.name == "code-reviewer" assert cfg.description == "Code review agent" assert cfg.model == "deepseek-v3" def test_load_missing_agent_raises(self, tmp_path): with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import load_agent_config with pytest.raises(FileNotFoundError): load_agent_config("nonexistent-agent") def test_load_missing_config_yaml_raises(self, tmp_path): # Create directory without config.yaml (tmp_path / "agents" / "broken-agent").mkdir(parents=True) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import load_agent_config with pytest.raises(FileNotFoundError): load_agent_config("broken-agent") def test_load_config_infers_name_from_dir(self, tmp_path): """Config without 'name' field should use directory name.""" agent_dir = tmp_path / "agents" / "inferred-name" agent_dir.mkdir(parents=True) (agent_dir / "config.yaml").write_text("description: My agent\n") (agent_dir / "SOUL.md").write_text("Hello") with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import load_agent_config cfg = load_agent_config("inferred-name") assert cfg.name == "inferred-name" def test_load_config_with_tool_groups(self, tmp_path): config_dict = {"name": "restricted", "tool_groups": ["file:read", "file:write"]} _write_agent(tmp_path, "restricted", config_dict) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import load_agent_config cfg = load_agent_config("restricted") assert cfg.tool_groups == ["file:read", "file:write"] def test_legacy_prompt_file_field_ignored(self, tmp_path): """Unknown fields like the old prompt_file should be silently ignored.""" agent_dir = tmp_path / "agents" / "legacy-agent" agent_dir.mkdir(parents=True) (agent_dir / "config.yaml").write_text("name: legacy-agent\nprompt_file: system.md\n") (agent_dir / "SOUL.md").write_text("Soul content") with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import load_agent_config cfg = load_agent_config("legacy-agent") assert cfg.name == "legacy-agent" # =========================================================================== # 4. load_agent_soul # =========================================================================== class TestLoadAgentSoul: def test_reads_soul_file(self, tmp_path): expected_soul = "You are a specialized code review expert." _write_agent(tmp_path, "code-reviewer", {"name": "code-reviewer"}, soul=expected_soul) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import AgentConfig, load_agent_soul cfg = AgentConfig(name="code-reviewer") soul = load_agent_soul(cfg.name) assert soul == expected_soul def test_missing_soul_file_returns_none(self, tmp_path): agent_dir = tmp_path / "agents" / "no-soul" agent_dir.mkdir(parents=True) (agent_dir / "config.yaml").write_text("name: no-soul\n") # No SOUL.md created with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import AgentConfig, load_agent_soul cfg = AgentConfig(name="no-soul") soul = load_agent_soul(cfg.name) assert soul is None def test_empty_soul_file_returns_none(self, tmp_path): agent_dir = tmp_path / "agents" / "empty-soul" agent_dir.mkdir(parents=True) (agent_dir / "config.yaml").write_text("name: empty-soul\n") (agent_dir / "SOUL.md").write_text(" \n ") with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import AgentConfig, load_agent_soul cfg = AgentConfig(name="empty-soul") soul = load_agent_soul(cfg.name) assert soul is None # =========================================================================== # 5. list_custom_agents # =========================================================================== class TestListCustomAgents: def test_empty_when_no_agents_dir(self, tmp_path): with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import list_custom_agents agents = list_custom_agents() assert agents == [] def test_discovers_multiple_agents(self, tmp_path): _write_agent(tmp_path, "agent-a", {"name": "agent-a"}) _write_agent(tmp_path, "agent-b", {"name": "agent-b", "description": "B"}) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import list_custom_agents agents = list_custom_agents() names = [a.name for a in agents] assert "agent-a" in names assert "agent-b" in names def test_skips_dirs_without_config_yaml(self, tmp_path): # Valid agent _write_agent(tmp_path, "valid-agent", {"name": "valid-agent"}) # Invalid dir (no config.yaml) (tmp_path / "agents" / "invalid-dir").mkdir(parents=True) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import list_custom_agents agents = list_custom_agents() assert len(agents) == 1 assert agents[0].name == "valid-agent" def test_skips_non_directory_entries(self, tmp_path): # Create the agents dir with a file (not a dir) agents_dir = tmp_path / "agents" agents_dir.mkdir(parents=True) (agents_dir / "not-a-dir.txt").write_text("hello") _write_agent(tmp_path, "real-agent", {"name": "real-agent"}) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import list_custom_agents agents = list_custom_agents() assert len(agents) == 1 assert agents[0].name == "real-agent" def test_returns_sorted_by_name(self, tmp_path): _write_agent(tmp_path, "z-agent", {"name": "z-agent"}) _write_agent(tmp_path, "a-agent", {"name": "a-agent"}) _write_agent(tmp_path, "m-agent", {"name": "m-agent"}) with patch("src.config.agents_config.get_paths", return_value=_make_paths(tmp_path)): from src.config.agents_config import list_custom_agents agents = list_custom_agents() names = [a.name for a in agents] assert names == sorted(names) # =========================================================================== # 7. Memory isolation: _get_memory_file_path # =========================================================================== class TestMemoryFilePath: def test_global_memory_path(self, tmp_path): """None agent_name should return global memory file.""" import src.agents.memory.updater as updater_mod from src.config.memory_config import MemoryConfig with ( patch("src.agents.memory.updater.get_paths", return_value=_make_paths(tmp_path)), patch("src.agents.memory.updater.get_memory_config", return_value=MemoryConfig(storage_path="")), ): path = updater_mod._get_memory_file_path(None) assert path == tmp_path / "memory.json" def test_agent_memory_path(self, tmp_path): """Providing agent_name should return per-agent memory file.""" import src.agents.memory.updater as updater_mod from src.config.memory_config import MemoryConfig with ( patch("src.agents.memory.updater.get_paths", return_value=_make_paths(tmp_path)), patch("src.agents.memory.updater.get_memory_config", return_value=MemoryConfig(storage_path="")), ): path = updater_mod._get_memory_file_path("code-reviewer") assert path == tmp_path / "agents" / "code-reviewer" / "memory.json" def test_different_paths_for_different_agents(self, tmp_path): import src.agents.memory.updater as updater_mod from src.config.memory_config import MemoryConfig with ( patch("src.agents.memory.updater.get_paths", return_value=_make_paths(tmp_path)), patch("src.agents.memory.updater.get_memory_config", return_value=MemoryConfig(storage_path="")), ): path_global = updater_mod._get_memory_file_path(None) path_a = updater_mod._get_memory_file_path("agent-a") path_b = updater_mod._get_memory_file_path("agent-b") assert path_global != path_a assert path_global != path_b assert path_a != path_b # =========================================================================== # 8. Gateway API – Agents endpoints # =========================================================================== def _make_test_app(tmp_path: Path): """Create a FastAPI app with the agents router, patching paths to tmp_path.""" from fastapi import FastAPI from src.gateway.routers.agents import router app = FastAPI() app.include_router(router) return app @pytest.fixture() def agent_client(tmp_path): """TestClient with agents router, using tmp_path as base_dir.""" paths_instance = _make_paths(tmp_path) with patch("src.config.agents_config.get_paths", return_value=paths_instance), patch("src.gateway.routers.agents.get_paths", return_value=paths_instance): app = _make_test_app(tmp_path) with TestClient(app) as client: client._tmp_path = tmp_path # type: ignore[attr-defined] yield client class TestAgentsAPI: def test_list_agents_empty(self, agent_client): response = agent_client.get("/api/agents") assert response.status_code == 200 data = response.json() assert data["agents"] == [] def test_create_agent(self, agent_client): payload = { "name": "code-reviewer", "description": "Reviews code", "soul": "You are a code reviewer.", } response = agent_client.post("/api/agents", json=payload) assert response.status_code == 201 data = response.json() assert data["name"] == "code-reviewer" assert data["description"] == "Reviews code" assert data["soul"] == "You are a code reviewer." def test_create_agent_invalid_name(self, agent_client): payload = {"name": "Code Reviewer!", "soul": "test"} response = agent_client.post("/api/agents", json=payload) assert response.status_code == 422 def test_create_duplicate_agent_409(self, agent_client): payload = {"name": "my-agent", "soul": "test"} agent_client.post("/api/agents", json=payload) # Second create should fail response = agent_client.post("/api/agents", json=payload) assert response.status_code == 409 def test_list_agents_after_create(self, agent_client): agent_client.post("/api/agents", json={"name": "agent-one", "soul": "p1"}) agent_client.post("/api/agents", json={"name": "agent-two", "soul": "p2"}) response = agent_client.get("/api/agents") assert response.status_code == 200 names = [a["name"] for a in response.json()["agents"]] assert "agent-one" in names assert "agent-two" in names def test_get_agent(self, agent_client): agent_client.post("/api/agents", json={"name": "test-agent", "soul": "Hello world"}) response = agent_client.get("/api/agents/test-agent") assert response.status_code == 200 data = response.json() assert data["name"] == "test-agent" assert data["soul"] == "Hello world" def test_get_missing_agent_404(self, agent_client): response = agent_client.get("/api/agents/nonexistent") assert response.status_code == 404 def test_update_agent_soul(self, agent_client): agent_client.post("/api/agents", json={"name": "update-me", "soul": "original"}) response = agent_client.put("/api/agents/update-me", json={"soul": "updated"}) assert response.status_code == 200 assert response.json()["soul"] == "updated" def test_update_agent_description(self, agent_client): agent_client.post("/api/agents", json={"name": "desc-agent", "description": "old desc", "soul": "p"}) response = agent_client.put("/api/agents/desc-agent", json={"description": "new desc"}) assert response.status_code == 200 assert response.json()["description"] == "new desc" def test_update_missing_agent_404(self, agent_client): response = agent_client.put("/api/agents/ghost-agent", json={"soul": "new"}) assert response.status_code == 404 def test_delete_agent(self, agent_client): agent_client.post("/api/agents", json={"name": "del-me", "soul": "bye"}) response = agent_client.delete("/api/agents/del-me") assert response.status_code == 204 # Verify it's gone response = agent_client.get("/api/agents/del-me") assert response.status_code == 404 def test_delete_missing_agent_404(self, agent_client): response = agent_client.delete("/api/agents/does-not-exist") assert response.status_code == 404 def test_create_agent_with_model_and_tool_groups(self, agent_client): payload = { "name": "specialized", "description": "Specialized agent", "model": "deepseek-v3", "tool_groups": ["file:read", "bash"], "soul": "You are specialized.", } response = agent_client.post("/api/agents", json=payload) assert response.status_code == 201 data = response.json() assert data["model"] == "deepseek-v3" assert data["tool_groups"] == ["file:read", "bash"] def test_create_persists_files_on_disk(self, agent_client, tmp_path): agent_client.post("/api/agents", json={"name": "disk-check", "soul": "disk soul"}) agent_dir = tmp_path / "agents" / "disk-check" assert agent_dir.exists() assert (agent_dir / "config.yaml").exists() assert (agent_dir / "SOUL.md").exists() assert (agent_dir / "SOUL.md").read_text() == "disk soul" def test_delete_removes_files_from_disk(self, agent_client, tmp_path): agent_client.post("/api/agents", json={"name": "remove-me", "soul": "bye"}) agent_dir = tmp_path / "agents" / "remove-me" assert agent_dir.exists() agent_client.delete("/api/agents/remove-me") assert not agent_dir.exists() # =========================================================================== # 9. Gateway API – User Profile endpoints # =========================================================================== class TestUserProfileAPI: def test_get_user_profile_empty(self, agent_client): response = agent_client.get("/api/user-profile") assert response.status_code == 200 assert response.json()["content"] is None def test_put_user_profile(self, agent_client, tmp_path): content = "# User Profile\n\nI am a developer." response = agent_client.put("/api/user-profile", json={"content": content}) assert response.status_code == 200 assert response.json()["content"] == content # File should be written to disk user_md = tmp_path / "USER.md" assert user_md.exists() assert user_md.read_text(encoding="utf-8") == content def test_get_user_profile_after_put(self, agent_client): content = "# Profile\n\nI work on data science." agent_client.put("/api/user-profile", json={"content": content}) response = agent_client.get("/api/user-profile") assert response.status_code == 200 assert response.json()["content"] == content def test_put_empty_user_profile_returns_none(self, agent_client): response = agent_client.put("/api/user-profile", json={"content": ""}) assert response.status_code == 200 assert response.json()["content"] is None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_custom_agent.py", "license": "MIT License", "lines": 384, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_reflection_resolvers.py
"""Tests for reflection resolvers.""" import pytest from src.reflection import resolvers from src.reflection.resolvers import resolve_variable def test_resolve_variable_reports_install_hint_for_missing_google_provider(monkeypatch: pytest.MonkeyPatch): """Missing google provider should return actionable install guidance.""" def fake_import_module(module_path: str): raise ModuleNotFoundError(f"No module named '{module_path}'", name=module_path) monkeypatch.setattr(resolvers, "import_module", fake_import_module) with pytest.raises(ImportError) as exc_info: resolve_variable("langchain_google_genai:ChatGoogleGenerativeAI") message = str(exc_info.value) assert "Could not import module langchain_google_genai" in message assert "uv add langchain-google-genai" in message def test_resolve_variable_reports_install_hint_for_missing_google_transitive_dependency( monkeypatch: pytest.MonkeyPatch, ) -> None: """Missing transitive dependency should still return actionable install guidance.""" def fake_import_module(module_path: str): # Simulate provider module existing but a transitive dependency (e.g. `google`) missing. raise ModuleNotFoundError("No module named 'google'", name="google") monkeypatch.setattr(resolvers, "import_module", fake_import_module) with pytest.raises(ImportError) as exc_info: resolve_variable("langchain_google_genai:ChatGoogleGenerativeAI") message = str(exc_info.value) # Even when a transitive dependency is missing, the hint should still point to the provider package. assert "uv add langchain-google-genai" in message def test_resolve_variable_invalid_path_format(): """Invalid variable path should fail with format guidance.""" with pytest.raises(ImportError) as exc_info: resolve_variable("invalid.variable.path") assert "doesn't look like a variable path" in str(exc_info.value)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_reflection_resolvers.py", "license": "MIT License", "lines": 32, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_skills_loader.py
"""Tests for recursive skills loading.""" from pathlib import Path from src.skills.loader import load_skills def _write_skill(skill_dir: Path, name: str, description: str) -> None: """Write a minimal SKILL.md for tests.""" skill_dir.mkdir(parents=True, exist_ok=True) content = f"---\nname: {name}\ndescription: {description}\n---\n\n# {name}\n" (skill_dir / "SKILL.md").write_text(content, encoding="utf-8") def test_load_skills_discovers_nested_skills_and_sets_container_paths(tmp_path: Path): """Nested skills should be discovered recursively with correct container paths.""" skills_root = tmp_path / "skills" _write_skill(skills_root / "public" / "root-skill", "root-skill", "Root skill") _write_skill(skills_root / "public" / "parent" / "child-skill", "child-skill", "Child skill") _write_skill(skills_root / "custom" / "team" / "helper", "team-helper", "Team helper") skills = load_skills(skills_path=skills_root, use_config=False, enabled_only=False) by_name = {skill.name: skill for skill in skills} assert {"root-skill", "child-skill", "team-helper"} <= set(by_name) root_skill = by_name["root-skill"] child_skill = by_name["child-skill"] team_skill = by_name["team-helper"] assert root_skill.skill_path == "root-skill" assert root_skill.get_container_file_path() == "/mnt/skills/public/root-skill/SKILL.md" assert child_skill.skill_path == "parent/child-skill" assert child_skill.get_container_file_path() == "/mnt/skills/public/parent/child-skill/SKILL.md" assert team_skill.skill_path == "team/helper" assert team_skill.get_container_file_path() == "/mnt/skills/custom/team/helper/SKILL.md" def test_load_skills_skips_hidden_directories(tmp_path: Path): """Hidden directories should be excluded from recursive discovery.""" skills_root = tmp_path / "skills" _write_skill(skills_root / "public" / "visible" / "ok-skill", "ok-skill", "Visible skill") _write_skill( skills_root / "public" / "visible" / ".hidden" / "secret-skill", "secret-skill", "Hidden skill", ) skills = load_skills(skills_path=skills_root, use_config=False, enabled_only=False) names = {skill.name for skill in skills} assert "ok-skill" in names assert "secret-skill" not in names
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_skills_loader.py", "license": "MIT License", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/mcp/oauth.py
"""OAuth token support for MCP HTTP/SSE servers.""" from __future__ import annotations import asyncio import logging from dataclasses import dataclass from datetime import UTC, datetime, timedelta from typing import Any from src.config.extensions_config import ExtensionsConfig, McpOAuthConfig logger = logging.getLogger(__name__) @dataclass class _OAuthToken: """Cached OAuth token.""" access_token: str token_type: str expires_at: datetime class OAuthTokenManager: """Acquire/cache/refresh OAuth tokens for MCP servers.""" def __init__(self, oauth_by_server: dict[str, McpOAuthConfig]): self._oauth_by_server = oauth_by_server self._tokens: dict[str, _OAuthToken] = {} self._locks: dict[str, asyncio.Lock] = {name: asyncio.Lock() for name in oauth_by_server} @classmethod def from_extensions_config(cls, extensions_config: ExtensionsConfig) -> OAuthTokenManager: oauth_by_server: dict[str, McpOAuthConfig] = {} for server_name, server_config in extensions_config.get_enabled_mcp_servers().items(): if server_config.oauth and server_config.oauth.enabled: oauth_by_server[server_name] = server_config.oauth return cls(oauth_by_server) def has_oauth_servers(self) -> bool: return bool(self._oauth_by_server) def oauth_server_names(self) -> list[str]: return list(self._oauth_by_server.keys()) async def get_authorization_header(self, server_name: str) -> str | None: oauth = self._oauth_by_server.get(server_name) if not oauth: return None token = self._tokens.get(server_name) if token and not self._is_expiring(token, oauth): return f"{token.token_type} {token.access_token}" lock = self._locks[server_name] async with lock: token = self._tokens.get(server_name) if token and not self._is_expiring(token, oauth): return f"{token.token_type} {token.access_token}" fresh = await self._fetch_token(oauth) self._tokens[server_name] = fresh logger.info(f"Refreshed OAuth access token for MCP server: {server_name}") return f"{fresh.token_type} {fresh.access_token}" @staticmethod def _is_expiring(token: _OAuthToken, oauth: McpOAuthConfig) -> bool: now = datetime.now(UTC) return token.expires_at <= now + timedelta(seconds=max(oauth.refresh_skew_seconds, 0)) async def _fetch_token(self, oauth: McpOAuthConfig) -> _OAuthToken: import httpx # pyright: ignore[reportMissingImports] data: dict[str, str] = { "grant_type": oauth.grant_type, **oauth.extra_token_params, } if oauth.scope: data["scope"] = oauth.scope if oauth.audience: data["audience"] = oauth.audience if oauth.grant_type == "client_credentials": if not oauth.client_id or not oauth.client_secret: raise ValueError("OAuth client_credentials requires client_id and client_secret") data["client_id"] = oauth.client_id data["client_secret"] = oauth.client_secret elif oauth.grant_type == "refresh_token": if not oauth.refresh_token: raise ValueError("OAuth refresh_token grant requires refresh_token") data["refresh_token"] = oauth.refresh_token if oauth.client_id: data["client_id"] = oauth.client_id if oauth.client_secret: data["client_secret"] = oauth.client_secret else: raise ValueError(f"Unsupported OAuth grant type: {oauth.grant_type}") async with httpx.AsyncClient(timeout=15.0) as client: response = await client.post(oauth.token_url, data=data) response.raise_for_status() payload = response.json() access_token = payload.get(oauth.token_field) if not access_token: raise ValueError(f"OAuth token response missing '{oauth.token_field}'") token_type = str(payload.get(oauth.token_type_field, oauth.default_token_type) or oauth.default_token_type) expires_in_raw = payload.get(oauth.expires_in_field, 3600) try: expires_in = int(expires_in_raw) except (TypeError, ValueError): expires_in = 3600 expires_at = datetime.now(UTC) + timedelta(seconds=max(expires_in, 1)) return _OAuthToken(access_token=access_token, token_type=token_type, expires_at=expires_at) def build_oauth_tool_interceptor(extensions_config: ExtensionsConfig) -> Any | None: """Build a tool interceptor that injects OAuth Authorization headers.""" token_manager = OAuthTokenManager.from_extensions_config(extensions_config) if not token_manager.has_oauth_servers(): return None async def oauth_interceptor(request: Any, handler: Any) -> Any: header = await token_manager.get_authorization_header(request.server_name) if not header: return await handler(request) updated_headers = dict(request.headers or {}) updated_headers["Authorization"] = header return await handler(request.override(headers=updated_headers)) return oauth_interceptor async def get_initial_oauth_headers(extensions_config: ExtensionsConfig) -> dict[str, str]: """Get initial OAuth Authorization headers for MCP server connections.""" token_manager = OAuthTokenManager.from_extensions_config(extensions_config) if not token_manager.has_oauth_servers(): return {} headers: dict[str, str] = {} for server_name in token_manager.oauth_server_names(): headers[server_name] = await token_manager.get_authorization_header(server_name) or "" return {name: value for name, value in headers.items() if value}
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/mcp/oauth.py", "license": "MIT License", "lines": 114, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/tests/test_mcp_oauth.py
"""Tests for MCP OAuth support.""" from __future__ import annotations import asyncio from typing import Any from src.config.extensions_config import ExtensionsConfig from src.mcp.oauth import OAuthTokenManager, build_oauth_tool_interceptor, get_initial_oauth_headers class _MockResponse: def __init__(self, payload: dict[str, Any]): self._payload = payload def raise_for_status(self) -> None: return None def json(self) -> dict[str, Any]: return self._payload class _MockAsyncClient: def __init__(self, payload: dict[str, Any], post_calls: list[dict[str, Any]], **kwargs): self._payload = payload self._post_calls = post_calls async def __aenter__(self): return self async def __aexit__(self, exc_type, exc, tb): return False async def post(self, url: str, data: dict[str, Any]): self._post_calls.append({"url": url, "data": data}) return _MockResponse(self._payload) def test_oauth_token_manager_fetches_and_caches_token(monkeypatch): post_calls: list[dict[str, Any]] = [] def _client_factory(*args, **kwargs): return _MockAsyncClient( payload={ "access_token": "token-123", "token_type": "Bearer", "expires_in": 3600, }, post_calls=post_calls, **kwargs, ) monkeypatch.setattr("httpx.AsyncClient", _client_factory) config = ExtensionsConfig.model_validate( { "mcpServers": { "secure-http": { "enabled": True, "type": "http", "url": "https://api.example.com/mcp", "oauth": { "enabled": True, "token_url": "https://auth.example.com/oauth/token", "grant_type": "client_credentials", "client_id": "client-id", "client_secret": "client-secret", }, } } } ) manager = OAuthTokenManager.from_extensions_config(config) first = asyncio.run(manager.get_authorization_header("secure-http")) second = asyncio.run(manager.get_authorization_header("secure-http")) assert first == "Bearer token-123" assert second == "Bearer token-123" assert len(post_calls) == 1 assert post_calls[0]["url"] == "https://auth.example.com/oauth/token" assert post_calls[0]["data"]["grant_type"] == "client_credentials" def test_build_oauth_interceptor_injects_authorization_header(monkeypatch): post_calls: list[dict[str, Any]] = [] def _client_factory(*args, **kwargs): return _MockAsyncClient( payload={ "access_token": "token-abc", "token_type": "Bearer", "expires_in": 3600, }, post_calls=post_calls, **kwargs, ) monkeypatch.setattr("httpx.AsyncClient", _client_factory) config = ExtensionsConfig.model_validate( { "mcpServers": { "secure-sse": { "enabled": True, "type": "sse", "url": "https://api.example.com/mcp", "oauth": { "enabled": True, "token_url": "https://auth.example.com/oauth/token", "grant_type": "client_credentials", "client_id": "client-id", "client_secret": "client-secret", }, } } } ) interceptor = build_oauth_tool_interceptor(config) assert interceptor is not None class _Request: def __init__(self): self.server_name = "secure-sse" self.headers = {"X-Test": "1"} def override(self, **kwargs): updated = _Request() updated.server_name = self.server_name updated.headers = kwargs.get("headers") return updated captured: dict[str, Any] = {} async def _handler(request): captured["headers"] = request.headers return "ok" result = asyncio.run(interceptor(_Request(), _handler)) assert result == "ok" assert captured["headers"]["Authorization"] == "Bearer token-abc" assert captured["headers"]["X-Test"] == "1" def test_get_initial_oauth_headers(monkeypatch): post_calls: list[dict[str, Any]] = [] def _client_factory(*args, **kwargs): return _MockAsyncClient( payload={ "access_token": "token-initial", "token_type": "Bearer", "expires_in": 3600, }, post_calls=post_calls, **kwargs, ) monkeypatch.setattr("httpx.AsyncClient", _client_factory) config = ExtensionsConfig.model_validate( { "mcpServers": { "secure-http": { "enabled": True, "type": "http", "url": "https://api.example.com/mcp", "oauth": { "enabled": True, "token_url": "https://auth.example.com/oauth/token", "grant_type": "client_credentials", "client_id": "client-id", "client_secret": "client-secret", }, }, "no-oauth": { "enabled": True, "type": "http", "url": "https://example.com/mcp", }, } } ) headers = asyncio.run(get_initial_oauth_headers(config)) assert headers == {"secure-http": "Bearer token-initial"} assert len(post_calls) == 1
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_mcp_oauth.py", "license": "MIT License", "lines": 152, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_readability.py
"""Tests for readability extraction fallback behavior.""" import subprocess import pytest from src.utils.readability import ReadabilityExtractor def test_extract_article_falls_back_when_readability_js_fails(monkeypatch): """When Node-based readability fails, extraction should fall back to Python mode.""" calls: list[bool] = [] def _fake_simple_json_from_html_string(html: str, use_readability: bool = False): calls.append(use_readability) if use_readability: raise subprocess.CalledProcessError( returncode=1, cmd=["node", "ExtractArticle.js"], stderr="boom", ) return {"title": "Fallback Title", "content": "<p>Fallback Content</p>"} monkeypatch.setattr( "src.utils.readability.simple_json_from_html_string", _fake_simple_json_from_html_string, ) article = ReadabilityExtractor().extract_article("<html><body>test</body></html>") assert calls == [True, False] assert article.title == "Fallback Title" assert article.html_content == "<p>Fallback Content</p>" def test_extract_article_re_raises_unexpected_exception(monkeypatch): """Unexpected errors should be surfaced instead of silently falling back.""" calls: list[bool] = [] def _fake_simple_json_from_html_string(html: str, use_readability: bool = False): calls.append(use_readability) if use_readability: raise RuntimeError("unexpected parser failure") return {"title": "Should Not Reach Fallback", "content": "<p>Fallback</p>"} monkeypatch.setattr( "src.utils.readability.simple_json_from_html_string", _fake_simple_json_from_html_string, ) with pytest.raises(RuntimeError, match="unexpected parser failure"): ReadabilityExtractor().extract_article("<html><body>test</body></html>") assert calls == [True]
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_readability.py", "license": "MIT License", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_uploads_router.py
import asyncio from io import BytesIO from pathlib import Path from unittest.mock import AsyncMock, MagicMock, patch from fastapi import UploadFile from src.gateway.routers import uploads def test_upload_files_writes_thread_storage_and_skips_local_sandbox_sync(tmp_path): thread_uploads_dir = tmp_path / "uploads" thread_uploads_dir.mkdir(parents=True) provider = MagicMock() provider.acquire.return_value = "local" sandbox = MagicMock() provider.get.return_value = sandbox with ( patch.object(uploads, "get_uploads_dir", return_value=thread_uploads_dir), patch.object(uploads, "get_sandbox_provider", return_value=provider), ): file = UploadFile(filename="notes.txt", file=BytesIO(b"hello uploads")) result = asyncio.run(uploads.upload_files("thread-local", files=[file])) assert result.success is True assert len(result.files) == 1 assert result.files[0]["filename"] == "notes.txt" assert (thread_uploads_dir / "notes.txt").read_bytes() == b"hello uploads" sandbox.update_file.assert_not_called() def test_upload_files_syncs_non_local_sandbox_and_marks_markdown_file(tmp_path): thread_uploads_dir = tmp_path / "uploads" thread_uploads_dir.mkdir(parents=True) provider = MagicMock() provider.acquire.return_value = "aio-1" sandbox = MagicMock() provider.get.return_value = sandbox async def fake_convert(file_path: Path) -> Path: md_path = file_path.with_suffix(".md") md_path.write_text("converted", encoding="utf-8") return md_path with ( patch.object(uploads, "get_uploads_dir", return_value=thread_uploads_dir), patch.object(uploads, "get_sandbox_provider", return_value=provider), patch.object(uploads, "convert_file_to_markdown", AsyncMock(side_effect=fake_convert)), ): file = UploadFile(filename="report.pdf", file=BytesIO(b"pdf-bytes")) result = asyncio.run(uploads.upload_files("thread-aio", files=[file])) assert result.success is True assert len(result.files) == 1 file_info = result.files[0] assert file_info["filename"] == "report.pdf" assert file_info["markdown_file"] == "report.md" assert (thread_uploads_dir / "report.pdf").read_bytes() == b"pdf-bytes" assert (thread_uploads_dir / "report.md").read_text(encoding="utf-8") == "converted" sandbox.update_file.assert_any_call("/mnt/user-data/uploads/report.pdf", b"pdf-bytes") sandbox.update_file.assert_any_call("/mnt/user-data/uploads/report.md", b"converted") def test_upload_files_rejects_dotdot_and_dot_filenames(tmp_path): thread_uploads_dir = tmp_path / "uploads" thread_uploads_dir.mkdir(parents=True) provider = MagicMock() provider.acquire.return_value = "local" sandbox = MagicMock() provider.get.return_value = sandbox with ( patch.object(uploads, "get_uploads_dir", return_value=thread_uploads_dir), patch.object(uploads, "get_sandbox_provider", return_value=provider), ): # These filenames must be rejected outright for bad_name in ["..", "."]: file = UploadFile(filename=bad_name, file=BytesIO(b"data")) result = asyncio.run(uploads.upload_files("thread-local", files=[file])) assert result.success is True assert result.files == [], f"Expected no files for unsafe filename {bad_name!r}" # Path-traversal prefixes are stripped to the basename and accepted safely file = UploadFile(filename="../etc/passwd", file=BytesIO(b"data")) result = asyncio.run(uploads.upload_files("thread-local", files=[file])) assert result.success is True assert len(result.files) == 1 assert result.files[0]["filename"] == "passwd" # Only the safely normalised file should exist assert [f.name for f in thread_uploads_dir.iterdir()] == ["passwd"]
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_uploads_router.py", "license": "MIT License", "lines": 76, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_mcp_client_config.py
"""Core behavior tests for MCP client server config building.""" import pytest from src.config.extensions_config import ExtensionsConfig, McpServerConfig from src.mcp.client import build_server_params, build_servers_config def test_build_server_params_stdio_success(): config = McpServerConfig( type="stdio", command="npx", args=["-y", "my-mcp-server"], env={"API_KEY": "secret"}, ) params = build_server_params("my-server", config) assert params == { "transport": "stdio", "command": "npx", "args": ["-y", "my-mcp-server"], "env": {"API_KEY": "secret"}, } def test_build_server_params_stdio_requires_command(): config = McpServerConfig(type="stdio", command=None) with pytest.raises(ValueError, match="requires 'command' field"): build_server_params("broken-stdio", config) @pytest.mark.parametrize("transport", ["sse", "http"]) def test_build_server_params_http_like_success(transport: str): config = McpServerConfig( type=transport, url="https://example.com/mcp", headers={"Authorization": "Bearer token"}, ) params = build_server_params("remote-server", config) assert params == { "transport": transport, "url": "https://example.com/mcp", "headers": {"Authorization": "Bearer token"}, } @pytest.mark.parametrize("transport", ["sse", "http"]) def test_build_server_params_http_like_requires_url(transport: str): config = McpServerConfig(type=transport, url=None) with pytest.raises(ValueError, match="requires 'url' field"): build_server_params("broken-remote", config) def test_build_server_params_rejects_unsupported_transport(): config = McpServerConfig(type="websocket") with pytest.raises(ValueError, match="unsupported transport type"): build_server_params("bad-transport", config) def test_build_servers_config_returns_empty_when_no_enabled_servers(): extensions = ExtensionsConfig( mcp_servers={ "disabled-a": McpServerConfig(enabled=False, type="stdio", command="echo"), "disabled-b": McpServerConfig(enabled=False, type="http", url="https://example.com"), }, skills={}, ) assert build_servers_config(extensions) == {} def test_build_servers_config_skips_invalid_server_and_keeps_valid_ones(): extensions = ExtensionsConfig( mcp_servers={ "valid-stdio": McpServerConfig(enabled=True, type="stdio", command="npx", args=["server"]), "invalid-stdio": McpServerConfig(enabled=True, type="stdio", command=None), "disabled-http": McpServerConfig(enabled=False, type="http", url="https://disabled.example.com"), }, skills={}, ) result = build_servers_config(extensions) assert "valid-stdio" in result assert result["valid-stdio"]["transport"] == "stdio" assert "invalid-stdio" not in result assert "disabled-http" not in result
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_mcp_client_config.py", "license": "MIT License", "lines": 67, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_task_tool_core_logic.py
"""Core behavior tests for task tool orchestration.""" import importlib from enum import Enum from types import SimpleNamespace from unittest.mock import MagicMock from src.subagents.config import SubagentConfig # Use module import so tests can patch the exact symbols referenced inside task_tool(). task_tool_module = importlib.import_module("src.tools.builtins.task_tool") class FakeSubagentStatus(Enum): # Match production enum values so branch comparisons behave identically. PENDING = "pending" RUNNING = "running" COMPLETED = "completed" FAILED = "failed" TIMED_OUT = "timed_out" def _make_runtime() -> SimpleNamespace: # Minimal ToolRuntime-like object; task_tool only reads these three attributes. return SimpleNamespace( state={ "sandbox": {"sandbox_id": "local"}, "thread_data": { "workspace_path": "/tmp/workspace", "uploads_path": "/tmp/uploads", "outputs_path": "/tmp/outputs", }, }, context={"thread_id": "thread-1"}, config={"metadata": {"model_name": "ark-model", "trace_id": "trace-1"}}, ) def _make_subagent_config() -> SubagentConfig: return SubagentConfig( name="general-purpose", description="General helper", system_prompt="Base system prompt", max_turns=50, timeout_seconds=10, ) def _make_result( status: FakeSubagentStatus, *, ai_messages: list[dict] | None = None, result: str | None = None, error: str | None = None, ) -> SimpleNamespace: return SimpleNamespace( status=status, ai_messages=ai_messages or [], result=result, error=error, ) def test_task_tool_returns_error_for_unknown_subagent(monkeypatch): monkeypatch.setattr(task_tool_module, "get_subagent_config", lambda _: None) result = task_tool_module.task_tool.func( runtime=None, description="ζ‰§θ‘Œδ»»εŠ‘", prompt="do work", subagent_type="general-purpose", tool_call_id="tc-1", ) assert result.startswith("Error: Unknown subagent type") def test_task_tool_emits_running_and_completed_events(monkeypatch): config = _make_subagent_config() runtime = _make_runtime() events = [] captured = {} get_available_tools = MagicMock(return_value=["tool-a", "tool-b"]) class DummyExecutor: def __init__(self, **kwargs): captured["executor_kwargs"] = kwargs def execute_async(self, prompt, task_id=None): captured["prompt"] = prompt captured["task_id"] = task_id return task_id or "generated-task-id" # Simulate two polling rounds: first running (with one message), then completed. responses = iter( [ _make_result(FakeSubagentStatus.RUNNING, ai_messages=[{"id": "m1", "content": "phase-1"}]), _make_result( FakeSubagentStatus.COMPLETED, ai_messages=[{"id": "m1", "content": "phase-1"}, {"id": "m2", "content": "phase-2"}], result="all done", ), ] ) monkeypatch.setattr(task_tool_module, "SubagentStatus", FakeSubagentStatus) monkeypatch.setattr(task_tool_module, "SubagentExecutor", DummyExecutor) monkeypatch.setattr(task_tool_module, "get_subagent_config", lambda _: config) monkeypatch.setattr(task_tool_module, "get_skills_prompt_section", lambda: "Skills Appendix") monkeypatch.setattr(task_tool_module, "get_background_task_result", lambda _: next(responses)) monkeypatch.setattr(task_tool_module, "get_stream_writer", lambda: events.append) monkeypatch.setattr(task_tool_module.time, "sleep", lambda _: None) # task_tool lazily imports from src.tools at call time, so patch that module-level function. monkeypatch.setattr("src.tools.get_available_tools", get_available_tools) output = task_tool_module.task_tool.func( runtime=runtime, description="运葌子任劑", prompt="collect diagnostics", subagent_type="general-purpose", tool_call_id="tc-123", max_turns=7, ) assert output == "Task Succeeded. Result: all done" assert captured["prompt"] == "collect diagnostics" assert captured["task_id"] == "tc-123" assert captured["executor_kwargs"]["thread_id"] == "thread-1" assert captured["executor_kwargs"]["parent_model"] == "ark-model" assert captured["executor_kwargs"]["config"].max_turns == 7 assert "Skills Appendix" in captured["executor_kwargs"]["config"].system_prompt get_available_tools.assert_called_once_with(model_name="ark-model", subagent_enabled=False) event_types = [e["type"] for e in events] assert event_types == ["task_started", "task_running", "task_running", "task_completed"] assert events[-1]["result"] == "all done" def test_task_tool_returns_failed_message(monkeypatch): config = _make_subagent_config() events = [] monkeypatch.setattr(task_tool_module, "SubagentStatus", FakeSubagentStatus) monkeypatch.setattr( task_tool_module, "SubagentExecutor", type("DummyExecutor", (), {"__init__": lambda self, **kwargs: None, "execute_async": lambda self, prompt, task_id=None: task_id}), ) monkeypatch.setattr(task_tool_module, "get_subagent_config", lambda _: config) monkeypatch.setattr(task_tool_module, "get_skills_prompt_section", lambda: "") monkeypatch.setattr( task_tool_module, "get_background_task_result", lambda _: _make_result(FakeSubagentStatus.FAILED, error="subagent crashed"), ) monkeypatch.setattr(task_tool_module, "get_stream_writer", lambda: events.append) monkeypatch.setattr(task_tool_module.time, "sleep", lambda _: None) monkeypatch.setattr("src.tools.get_available_tools", lambda **kwargs: []) output = task_tool_module.task_tool.func( runtime=_make_runtime(), description="ζ‰§θ‘Œδ»»εŠ‘", prompt="do fail", subagent_type="general-purpose", tool_call_id="tc-fail", ) assert output == "Task failed. Error: subagent crashed" assert events[-1]["type"] == "task_failed" assert events[-1]["error"] == "subagent crashed" def test_task_tool_returns_timed_out_message(monkeypatch): config = _make_subagent_config() events = [] monkeypatch.setattr(task_tool_module, "SubagentStatus", FakeSubagentStatus) monkeypatch.setattr( task_tool_module, "SubagentExecutor", type("DummyExecutor", (), {"__init__": lambda self, **kwargs: None, "execute_async": lambda self, prompt, task_id=None: task_id}), ) monkeypatch.setattr(task_tool_module, "get_subagent_config", lambda _: config) monkeypatch.setattr(task_tool_module, "get_skills_prompt_section", lambda: "") monkeypatch.setattr( task_tool_module, "get_background_task_result", lambda _: _make_result(FakeSubagentStatus.TIMED_OUT, error="timeout"), ) monkeypatch.setattr(task_tool_module, "get_stream_writer", lambda: events.append) monkeypatch.setattr(task_tool_module.time, "sleep", lambda _: None) monkeypatch.setattr("src.tools.get_available_tools", lambda **kwargs: []) output = task_tool_module.task_tool.func( runtime=_make_runtime(), description="ζ‰§θ‘Œδ»»εŠ‘", prompt="do timeout", subagent_type="general-purpose", tool_call_id="tc-timeout", ) assert output == "Task timed out. Error: timeout" assert events[-1]["type"] == "task_timed_out" assert events[-1]["error"] == "timeout" def test_task_tool_polling_safety_timeout(monkeypatch): config = _make_subagent_config() # Keep max_poll_count small for test speed: (1 + 60) // 5 = 12 config.timeout_seconds = 1 events = [] monkeypatch.setattr(task_tool_module, "SubagentStatus", FakeSubagentStatus) monkeypatch.setattr( task_tool_module, "SubagentExecutor", type("DummyExecutor", (), {"__init__": lambda self, **kwargs: None, "execute_async": lambda self, prompt, task_id=None: task_id}), ) monkeypatch.setattr(task_tool_module, "get_subagent_config", lambda _: config) monkeypatch.setattr(task_tool_module, "get_skills_prompt_section", lambda: "") monkeypatch.setattr( task_tool_module, "get_background_task_result", lambda _: _make_result(FakeSubagentStatus.RUNNING, ai_messages=[]), ) monkeypatch.setattr(task_tool_module, "get_stream_writer", lambda: events.append) monkeypatch.setattr(task_tool_module.time, "sleep", lambda _: None) monkeypatch.setattr("src.tools.get_available_tools", lambda **kwargs: []) output = task_tool_module.task_tool.func( runtime=_make_runtime(), description="ζ‰§θ‘Œδ»»εŠ‘", prompt="never finish", subagent_type="general-purpose", tool_call_id="tc-safety-timeout", ) assert output.startswith("Task polling timed out after 0 minutes") assert events[0]["type"] == "task_started" assert events[-1]["type"] == "task_timed_out"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_task_tool_core_logic.py", "license": "MIT License", "lines": 201, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/client.py
"""DeerFlowClient β€” Embedded Python client for DeerFlow agent system. Provides direct programmatic access to DeerFlow's agent capabilities without requiring LangGraph Server or Gateway API processes. Usage: from src.client import DeerFlowClient client = DeerFlowClient() response = client.chat("Analyze this paper for me", thread_id="my-thread") print(response) # Streaming for event in client.stream("hello"): print(event) """ import asyncio import json import logging import mimetypes import re import shutil import tempfile import uuid import zipfile from collections.abc import Generator from dataclasses import dataclass, field from pathlib import Path from typing import Any from langchain.agents import create_agent from langchain_core.messages import AIMessage, HumanMessage, SystemMessage, ToolMessage from langchain_core.runnables import RunnableConfig from src.agents.lead_agent.agent import _build_middlewares from src.agents.lead_agent.prompt import apply_prompt_template from src.agents.thread_state import ThreadState from src.config.app_config import get_app_config, reload_app_config from src.config.extensions_config import ExtensionsConfig, SkillStateConfig, get_extensions_config, reload_extensions_config from src.config.paths import get_paths from src.models import create_chat_model logger = logging.getLogger(__name__) @dataclass class StreamEvent: """A single event from the streaming agent response. Event types align with the LangGraph SSE protocol: - ``"values"``: Full state snapshot (title, messages, artifacts). - ``"messages-tuple"``: Per-message update (AI text, tool calls, tool results). - ``"end"``: Stream finished. Attributes: type: Event type. data: Event payload. Contents vary by type. """ type: str data: dict[str, Any] = field(default_factory=dict) class DeerFlowClient: """Embedded Python client for DeerFlow agent system. Provides direct programmatic access to DeerFlow's agent capabilities without requiring LangGraph Server or Gateway API processes. Note: Multi-turn conversations require a ``checkpointer``. Without one, each ``stream()`` / ``chat()`` call is stateless β€” ``thread_id`` is only used for file isolation (uploads / artifacts). The system prompt (including date, memory, and skills context) is generated when the internal agent is first created and cached until the configuration key changes. Call :meth:`reset_agent` to force a refresh in long-running processes. Example:: from src.client import DeerFlowClient client = DeerFlowClient() # Simple one-shot print(client.chat("hello")) # Streaming for event in client.stream("hello"): print(event.type, event.data) # Configuration queries print(client.list_models()) print(client.list_skills()) """ def __init__( self, config_path: str | None = None, checkpointer=None, *, model_name: str | None = None, thinking_enabled: bool = True, subagent_enabled: bool = False, plan_mode: bool = False, ): """Initialize the client. Loads configuration but defers agent creation to first use. Args: config_path: Path to config.yaml. Uses default resolution if None. checkpointer: LangGraph checkpointer instance for state persistence. Required for multi-turn conversations on the same thread_id. Without a checkpointer, each call is stateless. model_name: Override the default model name from config. thinking_enabled: Enable model's extended thinking. subagent_enabled: Enable subagent delegation. plan_mode: Enable TodoList middleware for plan mode. """ if config_path is not None: reload_app_config(config_path) self._app_config = get_app_config() self._checkpointer = checkpointer self._model_name = model_name self._thinking_enabled = thinking_enabled self._subagent_enabled = subagent_enabled self._plan_mode = plan_mode # Lazy agent β€” created on first call, recreated when config changes. self._agent = None self._agent_config_key: tuple | None = None def reset_agent(self) -> None: """Force the internal agent to be recreated on the next call. Use this after external changes (e.g. memory updates, skill installations) that should be reflected in the system prompt or tool set. """ self._agent = None self._agent_config_key = None # ------------------------------------------------------------------ # Internal helpers # ------------------------------------------------------------------ @staticmethod def _atomic_write_json(path: Path, data: dict) -> None: """Write JSON to *path* atomically (temp file + replace).""" fd = tempfile.NamedTemporaryFile( mode="w", dir=path.parent, suffix=".tmp", delete=False, ) try: json.dump(data, fd, indent=2) fd.close() Path(fd.name).replace(path) except BaseException: fd.close() Path(fd.name).unlink(missing_ok=True) raise def _get_runnable_config(self, thread_id: str, **overrides) -> RunnableConfig: """Build a RunnableConfig for agent invocation.""" configurable = { "thread_id": thread_id, "model_name": overrides.get("model_name", self._model_name), "thinking_enabled": overrides.get("thinking_enabled", self._thinking_enabled), "is_plan_mode": overrides.get("plan_mode", self._plan_mode), "subagent_enabled": overrides.get("subagent_enabled", self._subagent_enabled), } return RunnableConfig( configurable=configurable, recursion_limit=overrides.get("recursion_limit", 100), ) def _ensure_agent(self, config: RunnableConfig): """Create (or recreate) the agent when config-dependent params change.""" cfg = config.get("configurable", {}) key = ( cfg.get("model_name"), cfg.get("thinking_enabled"), cfg.get("is_plan_mode"), cfg.get("subagent_enabled"), ) if self._agent is not None and self._agent_config_key == key: return thinking_enabled = cfg.get("thinking_enabled", True) model_name = cfg.get("model_name") subagent_enabled = cfg.get("subagent_enabled", False) max_concurrent_subagents = cfg.get("max_concurrent_subagents", 3) kwargs: dict[str, Any] = { "model": create_chat_model(name=model_name, thinking_enabled=thinking_enabled), "tools": self._get_tools(model_name=model_name, subagent_enabled=subagent_enabled), "middleware": _build_middlewares(config, model_name=model_name), "system_prompt": apply_prompt_template( subagent_enabled=subagent_enabled, max_concurrent_subagents=max_concurrent_subagents, ), "state_schema": ThreadState, } if self._checkpointer is not None: kwargs["checkpointer"] = self._checkpointer self._agent = create_agent(**kwargs) self._agent_config_key = key logger.info("Agent created: model=%s, thinking=%s", model_name, thinking_enabled) @staticmethod def _get_tools(*, model_name: str | None, subagent_enabled: bool): """Lazy import to avoid circular dependency at module level.""" from src.tools import get_available_tools return get_available_tools(model_name=model_name, subagent_enabled=subagent_enabled) @staticmethod def _serialize_message(msg) -> dict: """Serialize a LangChain message to a plain dict for values events.""" if isinstance(msg, AIMessage): d: dict[str, Any] = {"type": "ai", "content": msg.content, "id": getattr(msg, "id", None)} if msg.tool_calls: d["tool_calls"] = [{"name": tc["name"], "args": tc["args"], "id": tc.get("id")} for tc in msg.tool_calls] return d if isinstance(msg, ToolMessage): return { "type": "tool", "content": msg.content if isinstance(msg.content, str) else str(msg.content), "name": getattr(msg, "name", None), "tool_call_id": getattr(msg, "tool_call_id", None), "id": getattr(msg, "id", None), } if isinstance(msg, HumanMessage): return {"type": "human", "content": msg.content, "id": getattr(msg, "id", None)} if isinstance(msg, SystemMessage): return {"type": "system", "content": msg.content, "id": getattr(msg, "id", None)} return {"type": "unknown", "content": str(msg), "id": getattr(msg, "id", None)} @staticmethod def _extract_text(content) -> str: """Extract plain text from AIMessage content (str or list of blocks).""" if isinstance(content, str): return content if isinstance(content, list): parts = [] for block in content: if isinstance(block, str): parts.append(block) elif isinstance(block, dict) and block.get("type") == "text": parts.append(block["text"]) return "\n".join(parts) if parts else "" return str(content) # ------------------------------------------------------------------ # Public API β€” conversation # ------------------------------------------------------------------ def stream( self, message: str, *, thread_id: str | None = None, **kwargs, ) -> Generator[StreamEvent, None, None]: """Stream a conversation turn, yielding events incrementally. Each call sends one user message and yields events until the agent finishes its turn. A ``checkpointer`` must be provided at init time for multi-turn context to be preserved across calls. Event types align with the LangGraph SSE protocol so that consumers can switch between HTTP streaming and embedded mode without changing their event-handling logic. Args: message: User message text. thread_id: Thread ID for conversation context. Auto-generated if None. **kwargs: Override client defaults (model_name, thinking_enabled, plan_mode, subagent_enabled, recursion_limit). Yields: StreamEvent with one of: - type="values" data={"title": str|None, "messages": [...], "artifacts": [...]} - type="messages-tuple" data={"type": "ai", "content": str, "id": str} - type="messages-tuple" data={"type": "ai", "content": "", "id": str, "tool_calls": [...]} - type="messages-tuple" data={"type": "tool", "content": str, "name": str, "tool_call_id": str, "id": str} - type="end" data={} """ if thread_id is None: thread_id = str(uuid.uuid4()) config = self._get_runnable_config(thread_id, **kwargs) self._ensure_agent(config) state: dict[str, Any] = {"messages": [HumanMessage(content=message)]} context = {"thread_id": thread_id} seen_ids: set[str] = set() for chunk in self._agent.stream(state, config=config, context=context, stream_mode="values"): messages = chunk.get("messages", []) for msg in messages: msg_id = getattr(msg, "id", None) if msg_id and msg_id in seen_ids: continue if msg_id: seen_ids.add(msg_id) if isinstance(msg, AIMessage): if msg.tool_calls: yield StreamEvent( type="messages-tuple", data={ "type": "ai", "content": "", "id": msg_id, "tool_calls": [ {"name": tc["name"], "args": tc["args"], "id": tc.get("id")} for tc in msg.tool_calls ], }, ) text = self._extract_text(msg.content) if text: yield StreamEvent( type="messages-tuple", data={"type": "ai", "content": text, "id": msg_id}, ) elif isinstance(msg, ToolMessage): yield StreamEvent( type="messages-tuple", data={ "type": "tool", "content": msg.content if isinstance(msg.content, str) else str(msg.content), "name": getattr(msg, "name", None), "tool_call_id": getattr(msg, "tool_call_id", None), "id": msg_id, }, ) # Emit a values event for each state snapshot yield StreamEvent( type="values", data={ "title": chunk.get("title"), "messages": [self._serialize_message(m) for m in messages], "artifacts": chunk.get("artifacts", []), }, ) yield StreamEvent(type="end", data={}) def chat(self, message: str, *, thread_id: str | None = None, **kwargs) -> str: """Send a message and return the final text response. Convenience wrapper around :meth:`stream` that returns only the **last** AI text from ``messages-tuple`` events. If the agent emits multiple text segments in one turn, intermediate segments are discarded. Use :meth:`stream` directly to capture all events. Args: message: User message text. thread_id: Thread ID for conversation context. Auto-generated if None. **kwargs: Override client defaults (same as stream()). Returns: The last AI message text, or empty string if no response. """ last_text = "" for event in self.stream(message, thread_id=thread_id, **kwargs): if event.type == "messages-tuple" and event.data.get("type") == "ai": content = event.data.get("content", "") if content: last_text = content return last_text # ------------------------------------------------------------------ # Public API β€” configuration queries # ------------------------------------------------------------------ def list_models(self) -> dict: """List available models from configuration. Returns: Dict with "models" key containing list of model info dicts, matching the Gateway API ``ModelsListResponse`` schema. """ return { "models": [ { "name": model.name, "display_name": getattr(model, "display_name", None), "description": getattr(model, "description", None), "supports_thinking": getattr(model, "supports_thinking", False), "supports_reasoning_effort": getattr(model, "supports_reasoning_effort", False), } for model in self._app_config.models ] } def list_skills(self, enabled_only: bool = False) -> dict: """List available skills. Args: enabled_only: If True, only return enabled skills. Returns: Dict with "skills" key containing list of skill info dicts, matching the Gateway API ``SkillsListResponse`` schema. """ from src.skills.loader import load_skills return { "skills": [ { "name": s.name, "description": s.description, "license": s.license, "category": s.category, "enabled": s.enabled, } for s in load_skills(enabled_only=enabled_only) ] } def get_memory(self) -> dict: """Get current memory data. Returns: Memory data dict (see src/agents/memory/updater.py for structure). """ from src.agents.memory.updater import get_memory_data return get_memory_data() def get_model(self, name: str) -> dict | None: """Get a specific model's configuration by name. Args: name: Model name. Returns: Model info dict matching the Gateway API ``ModelResponse`` schema, or None if not found. """ model = self._app_config.get_model_config(name) if model is None: return None return { "name": model.name, "display_name": getattr(model, "display_name", None), "description": getattr(model, "description", None), "supports_thinking": getattr(model, "supports_thinking", False), "supports_reasoning_effort": getattr(model, "supports_reasoning_effort", False), } # ------------------------------------------------------------------ # Public API β€” MCP configuration # ------------------------------------------------------------------ def get_mcp_config(self) -> dict: """Get MCP server configurations. Returns: Dict with "mcp_servers" key mapping server name to config, matching the Gateway API ``McpConfigResponse`` schema. """ config = get_extensions_config() return {"mcp_servers": {name: server.model_dump() for name, server in config.mcp_servers.items()}} def update_mcp_config(self, mcp_servers: dict[str, dict]) -> dict: """Update MCP server configurations. Writes to extensions_config.json and reloads the cache. Args: mcp_servers: Dict mapping server name to config dict. Each value should contain keys like enabled, type, command, args, env, url, etc. Returns: Dict with "mcp_servers" key, matching the Gateway API ``McpConfigResponse`` schema. Raises: OSError: If the config file cannot be written. """ config_path = ExtensionsConfig.resolve_config_path() if config_path is None: raise FileNotFoundError( "Cannot locate extensions_config.json. " "Set DEER_FLOW_EXTENSIONS_CONFIG_PATH or ensure it exists in the project root." ) current_config = get_extensions_config() config_data = { "mcpServers": mcp_servers, "skills": {name: {"enabled": skill.enabled} for name, skill in current_config.skills.items()}, } self._atomic_write_json(config_path, config_data) self._agent = None reloaded = reload_extensions_config() return {"mcp_servers": {name: server.model_dump() for name, server in reloaded.mcp_servers.items()}} # ------------------------------------------------------------------ # Public API β€” skills management # ------------------------------------------------------------------ def get_skill(self, name: str) -> dict | None: """Get a specific skill by name. Args: name: Skill name. Returns: Skill info dict, or None if not found. """ from src.skills.loader import load_skills skill = next((s for s in load_skills(enabled_only=False) if s.name == name), None) if skill is None: return None return { "name": skill.name, "description": skill.description, "license": skill.license, "category": skill.category, "enabled": skill.enabled, } def update_skill(self, name: str, *, enabled: bool) -> dict: """Update a skill's enabled status. Args: name: Skill name. enabled: New enabled status. Returns: Updated skill info dict. Raises: ValueError: If the skill is not found. OSError: If the config file cannot be written. """ from src.skills.loader import load_skills skills = load_skills(enabled_only=False) skill = next((s for s in skills if s.name == name), None) if skill is None: raise ValueError(f"Skill '{name}' not found") config_path = ExtensionsConfig.resolve_config_path() if config_path is None: raise FileNotFoundError( "Cannot locate extensions_config.json. " "Set DEER_FLOW_EXTENSIONS_CONFIG_PATH or ensure it exists in the project root." ) extensions_config = get_extensions_config() extensions_config.skills[name] = SkillStateConfig(enabled=enabled) config_data = { "mcpServers": {n: s.model_dump() for n, s in extensions_config.mcp_servers.items()}, "skills": {n: {"enabled": sc.enabled} for n, sc in extensions_config.skills.items()}, } self._atomic_write_json(config_path, config_data) self._agent = None reload_extensions_config() updated = next((s for s in load_skills(enabled_only=False) if s.name == name), None) if updated is None: raise RuntimeError(f"Skill '{name}' disappeared after update") return { "name": updated.name, "description": updated.description, "license": updated.license, "category": updated.category, "enabled": updated.enabled, } def install_skill(self, skill_path: str | Path) -> dict: """Install a skill from a .skill archive (ZIP). Args: skill_path: Path to the .skill file. Returns: Dict with success, skill_name, message. Raises: FileNotFoundError: If the file does not exist. ValueError: If the file is invalid. """ from src.gateway.routers.skills import _validate_skill_frontmatter from src.skills.loader import get_skills_root_path path = Path(skill_path) if not path.exists(): raise FileNotFoundError(f"Skill file not found: {skill_path}") if not path.is_file(): raise ValueError(f"Path is not a file: {skill_path}") if path.suffix != ".skill": raise ValueError("File must have .skill extension") if not zipfile.is_zipfile(path): raise ValueError("File is not a valid ZIP archive") skills_root = get_skills_root_path() custom_dir = skills_root / "custom" custom_dir.mkdir(parents=True, exist_ok=True) with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) with zipfile.ZipFile(path, "r") as zf: total_size = sum(info.file_size for info in zf.infolist()) if total_size > 100 * 1024 * 1024: raise ValueError("Skill archive too large when extracted (>100MB)") for info in zf.infolist(): if Path(info.filename).is_absolute() or ".." in Path(info.filename).parts: raise ValueError(f"Unsafe path in archive: {info.filename}") zf.extractall(tmp_path) for p in tmp_path.rglob("*"): if p.is_symlink(): p.unlink() items = list(tmp_path.iterdir()) if not items: raise ValueError("Skill archive is empty") skill_dir = items[0] if len(items) == 1 and items[0].is_dir() else tmp_path is_valid, message, skill_name = _validate_skill_frontmatter(skill_dir) if not is_valid: raise ValueError(f"Invalid skill: {message}") if not re.fullmatch(r"[a-zA-Z0-9_-]+", skill_name): raise ValueError(f"Invalid skill name: {skill_name}") target = custom_dir / skill_name if target.exists(): raise ValueError(f"Skill '{skill_name}' already exists") shutil.copytree(skill_dir, target) return {"success": True, "skill_name": skill_name, "message": f"Skill '{skill_name}' installed successfully"} # ------------------------------------------------------------------ # Public API β€” memory management # ------------------------------------------------------------------ def reload_memory(self) -> dict: """Reload memory data from file, forcing cache invalidation. Returns: The reloaded memory data dict. """ from src.agents.memory.updater import reload_memory_data return reload_memory_data() def get_memory_config(self) -> dict: """Get memory system configuration. Returns: Memory config dict. """ from src.config.memory_config import get_memory_config config = get_memory_config() return { "enabled": config.enabled, "storage_path": config.storage_path, "debounce_seconds": config.debounce_seconds, "max_facts": config.max_facts, "fact_confidence_threshold": config.fact_confidence_threshold, "injection_enabled": config.injection_enabled, "max_injection_tokens": config.max_injection_tokens, } def get_memory_status(self) -> dict: """Get memory status: config + current data. Returns: Dict with "config" and "data" keys. """ return { "config": self.get_memory_config(), "data": self.get_memory(), } # ------------------------------------------------------------------ # Public API β€” file uploads # ------------------------------------------------------------------ @staticmethod def _get_uploads_dir(thread_id: str) -> Path: """Get (and create) the uploads directory for a thread.""" base = get_paths().sandbox_uploads_dir(thread_id) base.mkdir(parents=True, exist_ok=True) return base def upload_files(self, thread_id: str, files: list[str | Path]) -> dict: """Upload local files into a thread's uploads directory. For PDF, PPT, Excel, and Word files, they are also converted to Markdown. Args: thread_id: Target thread ID. files: List of local file paths to upload. Returns: Dict with success, files, message β€” matching the Gateway API ``UploadResponse`` schema. Raises: FileNotFoundError: If any file does not exist. """ from src.gateway.routers.uploads import CONVERTIBLE_EXTENSIONS, convert_file_to_markdown # Validate all files upfront to avoid partial uploads. resolved_files = [] for f in files: p = Path(f) if not p.exists(): raise FileNotFoundError(f"File not found: {f}") resolved_files.append(p) uploads_dir = self._get_uploads_dir(thread_id) uploaded_files: list[dict] = [] for src_path in resolved_files: dest = uploads_dir / src_path.name shutil.copy2(src_path, dest) info: dict[str, Any] = { "filename": src_path.name, "size": str(dest.stat().st_size), "path": str(dest), "virtual_path": f"/mnt/user-data/uploads/{src_path.name}", "artifact_url": f"/api/threads/{thread_id}/artifacts/mnt/user-data/uploads/{src_path.name}", } if src_path.suffix.lower() in CONVERTIBLE_EXTENSIONS: try: try: asyncio.get_running_loop() import concurrent.futures with concurrent.futures.ThreadPoolExecutor() as pool: md_path = pool.submit(lambda: asyncio.run(convert_file_to_markdown(dest))).result() except RuntimeError: md_path = asyncio.run(convert_file_to_markdown(dest)) except Exception: logger.warning("Failed to convert %s to markdown", src_path.name, exc_info=True) md_path = None if md_path is not None: info["markdown_file"] = md_path.name info["markdown_virtual_path"] = f"/mnt/user-data/uploads/{md_path.name}" info["markdown_artifact_url"] = f"/api/threads/{thread_id}/artifacts/mnt/user-data/uploads/{md_path.name}" uploaded_files.append(info) return { "success": True, "files": uploaded_files, "message": f"Successfully uploaded {len(uploaded_files)} file(s)", } def list_uploads(self, thread_id: str) -> dict: """List files in a thread's uploads directory. Args: thread_id: Thread ID. Returns: Dict with "files" and "count" keys, matching the Gateway API ``list_uploaded_files`` response. """ uploads_dir = self._get_uploads_dir(thread_id) if not uploads_dir.exists(): return {"files": [], "count": 0} files = [] for fp in sorted(uploads_dir.iterdir()): if fp.is_file(): stat = fp.stat() files.append({ "filename": fp.name, "size": str(stat.st_size), "path": str(fp), "virtual_path": f"/mnt/user-data/uploads/{fp.name}", "artifact_url": f"/api/threads/{thread_id}/artifacts/mnt/user-data/uploads/{fp.name}", "extension": fp.suffix, "modified": stat.st_mtime, }) return {"files": files, "count": len(files)} def delete_upload(self, thread_id: str, filename: str) -> dict: """Delete a file from a thread's uploads directory. Args: thread_id: Thread ID. filename: Filename to delete. Returns: Dict with success and message, matching the Gateway API ``delete_uploaded_file`` response. Raises: FileNotFoundError: If the file does not exist. PermissionError: If path traversal is detected. """ uploads_dir = self._get_uploads_dir(thread_id) file_path = (uploads_dir / filename).resolve() try: file_path.relative_to(uploads_dir.resolve()) except ValueError as exc: raise PermissionError("Access denied: path traversal detected") from exc if not file_path.is_file(): raise FileNotFoundError(f"File not found: {filename}") file_path.unlink() return {"success": True, "message": f"Deleted {filename}"} # ------------------------------------------------------------------ # Public API β€” artifacts # ------------------------------------------------------------------ def get_artifact(self, thread_id: str, path: str) -> tuple[bytes, str]: """Read an artifact file produced by the agent. Args: thread_id: Thread ID. path: Virtual path (e.g. "mnt/user-data/outputs/file.txt"). Returns: Tuple of (file_bytes, mime_type). Raises: FileNotFoundError: If the artifact does not exist. ValueError: If the path is invalid. """ virtual_prefix = "mnt/user-data" clean_path = path.lstrip("/") if not clean_path.startswith(virtual_prefix): raise ValueError(f"Path must start with /{virtual_prefix}") relative = clean_path[len(virtual_prefix):].lstrip("/") base_dir = get_paths().sandbox_user_data_dir(thread_id) actual = (base_dir / relative).resolve() try: actual.relative_to(base_dir.resolve()) except ValueError as exc: raise PermissionError("Access denied: path traversal detected") from exc if not actual.exists(): raise FileNotFoundError(f"Artifact not found: {path}") if not actual.is_file(): raise ValueError(f"Path is not a file: {path}") mime_type, _ = mimetypes.guess_type(actual) return actual.read_bytes(), mime_type or "application/octet-stream"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/client.py", "license": "MIT License", "lines": 714, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/tests/test_client.py
"""Tests for DeerFlowClient.""" import json import tempfile import zipfile from pathlib import Path from unittest.mock import MagicMock, patch import pytest from langchain_core.messages import AIMessage, HumanMessage, ToolMessage # noqa: F401 from src.client import DeerFlowClient from src.gateway.routers.mcp import McpConfigResponse from src.gateway.routers.memory import MemoryConfigResponse, MemoryStatusResponse from src.gateway.routers.models import ModelResponse, ModelsListResponse from src.gateway.routers.skills import SkillInstallResponse, SkillResponse, SkillsListResponse from src.gateway.routers.uploads import UploadResponse # --------------------------------------------------------------------------- # Fixtures # --------------------------------------------------------------------------- @pytest.fixture def mock_app_config(): """Provide a minimal AppConfig mock.""" model = MagicMock() model.name = "test-model" model.supports_thinking = False model.supports_reasoning_effort = False model.model_dump.return_value = {"name": "test-model", "use": "langchain_openai:ChatOpenAI"} config = MagicMock() config.models = [model] return config @pytest.fixture def client(mock_app_config): """Create a DeerFlowClient with mocked config loading.""" with patch("src.client.get_app_config", return_value=mock_app_config): return DeerFlowClient() # --------------------------------------------------------------------------- # __init__ # --------------------------------------------------------------------------- class TestClientInit: def test_default_params(self, client): assert client._model_name is None assert client._thinking_enabled is True assert client._subagent_enabled is False assert client._plan_mode is False assert client._checkpointer is None assert client._agent is None def test_custom_params(self, mock_app_config): with patch("src.client.get_app_config", return_value=mock_app_config): c = DeerFlowClient( model_name="gpt-4", thinking_enabled=False, subagent_enabled=True, plan_mode=True, ) assert c._model_name == "gpt-4" assert c._thinking_enabled is False assert c._subagent_enabled is True assert c._plan_mode is True def test_custom_config_path(self, mock_app_config): with ( patch("src.client.reload_app_config") as mock_reload, patch("src.client.get_app_config", return_value=mock_app_config), ): DeerFlowClient(config_path="/tmp/custom.yaml") mock_reload.assert_called_once_with("/tmp/custom.yaml") def test_checkpointer_stored(self, mock_app_config): cp = MagicMock() with patch("src.client.get_app_config", return_value=mock_app_config): c = DeerFlowClient(checkpointer=cp) assert c._checkpointer is cp # --------------------------------------------------------------------------- # list_models / list_skills / get_memory # --------------------------------------------------------------------------- class TestConfigQueries: def test_list_models(self, client): result = client.list_models() assert "models" in result assert len(result["models"]) == 1 assert result["models"][0]["name"] == "test-model" # Verify Gateway-aligned fields are present assert "display_name" in result["models"][0] assert "supports_thinking" in result["models"][0] def test_list_skills(self, client): skill = MagicMock() skill.name = "web-search" skill.description = "Search the web" skill.license = "MIT" skill.category = "public" skill.enabled = True with patch("src.skills.loader.load_skills", return_value=[skill]) as mock_load: result = client.list_skills() mock_load.assert_called_once_with(enabled_only=False) assert "skills" in result assert len(result["skills"]) == 1 assert result["skills"][0] == { "name": "web-search", "description": "Search the web", "license": "MIT", "category": "public", "enabled": True, } def test_list_skills_enabled_only(self, client): with patch("src.skills.loader.load_skills", return_value=[]) as mock_load: client.list_skills(enabled_only=True) mock_load.assert_called_once_with(enabled_only=True) def test_get_memory(self, client): memory = {"version": "1.0", "facts": []} with patch("src.agents.memory.updater.get_memory_data", return_value=memory) as mock_mem: result = client.get_memory() mock_mem.assert_called_once() assert result == memory # --------------------------------------------------------------------------- # stream / chat # --------------------------------------------------------------------------- def _make_agent_mock(chunks: list[dict]): """Create a mock agent whose .stream() yields the given chunks.""" agent = MagicMock() agent.stream.return_value = iter(chunks) return agent def _ai_events(events): """Filter messages-tuple events with type=ai and non-empty content.""" return [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "ai" and e.data.get("content")] def _tool_call_events(events): """Filter messages-tuple events with type=ai and tool_calls.""" return [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "ai" and "tool_calls" in e.data] def _tool_result_events(events): """Filter messages-tuple events with type=tool.""" return [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "tool"] class TestStream: def test_basic_message(self, client): """stream() emits messages-tuple + values + end for a simple AI reply.""" ai = AIMessage(content="Hello!", id="ai-1") chunks = [ {"messages": [HumanMessage(content="hi", id="h-1")]}, {"messages": [HumanMessage(content="hi", id="h-1"), ai]}, ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi", thread_id="t1")) types = [e.type for e in events] assert "messages-tuple" in types assert "values" in types assert types[-1] == "end" msg_events = _ai_events(events) assert msg_events[0].data["content"] == "Hello!" def test_tool_call_and_result(self, client): """stream() emits messages-tuple events for tool calls and results.""" ai = AIMessage(content="", id="ai-1", tool_calls=[{"name": "bash", "args": {"cmd": "ls"}, "id": "tc-1"}]) tool = ToolMessage(content="file.txt", id="tm-1", tool_call_id="tc-1", name="bash") ai2 = AIMessage(content="Here are the files.", id="ai-2") chunks = [ {"messages": [HumanMessage(content="list files", id="h-1"), ai]}, {"messages": [HumanMessage(content="list files", id="h-1"), ai, tool]}, {"messages": [HumanMessage(content="list files", id="h-1"), ai, tool, ai2]}, ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("list files", thread_id="t2")) assert len(_tool_call_events(events)) >= 1 assert len(_tool_result_events(events)) >= 1 assert len(_ai_events(events)) >= 1 assert events[-1].type == "end" def test_values_event_with_title(self, client): """stream() emits values event containing title when present in state.""" ai = AIMessage(content="ok", id="ai-1") chunks = [ {"messages": [HumanMessage(content="hi", id="h-1"), ai], "title": "Greeting"}, ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi", thread_id="t3")) values_events = [e for e in events if e.type == "values"] assert len(values_events) >= 1 assert values_events[-1].data["title"] == "Greeting" assert "messages" in values_events[-1].data def test_deduplication(self, client): """Messages with the same id are not emitted twice.""" ai = AIMessage(content="Hello!", id="ai-1") chunks = [ {"messages": [HumanMessage(content="hi", id="h-1"), ai]}, {"messages": [HumanMessage(content="hi", id="h-1"), ai]}, # duplicate ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi", thread_id="t4")) msg_events = _ai_events(events) assert len(msg_events) == 1 def test_auto_thread_id(self, client): """stream() auto-generates a thread_id if not provided.""" agent = _make_agent_mock([{"messages": [AIMessage(content="ok", id="ai-1")]}]) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi")) # Should not raise; end event proves it completed assert events[-1].type == "end" def test_list_content_blocks(self, client): """stream() handles AIMessage with list-of-blocks content.""" ai = AIMessage( content=[ {"type": "thinking", "thinking": "hmm"}, {"type": "text", "text": "result"}, ], id="ai-1", ) chunks = [{"messages": [ai]}] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi", thread_id="t5")) msg_events = _ai_events(events) assert len(msg_events) == 1 assert msg_events[0].data["content"] == "result" class TestChat: def test_returns_last_message(self, client): """chat() returns the last AI message text.""" ai1 = AIMessage(content="thinking...", id="ai-1") ai2 = AIMessage(content="final answer", id="ai-2") chunks = [ {"messages": [HumanMessage(content="q", id="h-1"), ai1]}, {"messages": [HumanMessage(content="q", id="h-1"), ai1, ai2]}, ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): result = client.chat("q", thread_id="t6") assert result == "final answer" def test_empty_response(self, client): """chat() returns empty string if no AI message produced.""" chunks = [{"messages": []}] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): result = client.chat("q", thread_id="t7") assert result == "" # --------------------------------------------------------------------------- # _extract_text # --------------------------------------------------------------------------- class TestExtractText: def test_string(self): assert DeerFlowClient._extract_text("hello") == "hello" def test_list_text_blocks(self): content = [ {"type": "text", "text": "first"}, {"type": "thinking", "thinking": "skip"}, {"type": "text", "text": "second"}, ] assert DeerFlowClient._extract_text(content) == "first\nsecond" def test_list_plain_strings(self): assert DeerFlowClient._extract_text(["a", "b"]) == "a\nb" def test_empty_list(self): assert DeerFlowClient._extract_text([]) == "" def test_other_type(self): assert DeerFlowClient._extract_text(42) == "42" # --------------------------------------------------------------------------- # _ensure_agent # --------------------------------------------------------------------------- class TestEnsureAgent: def test_creates_agent(self, client): """_ensure_agent creates an agent on first call.""" mock_agent = MagicMock() config = client._get_runnable_config("t1") with ( patch("src.client.create_chat_model"), patch("src.client.create_agent", return_value=mock_agent), patch("src.client._build_middlewares", return_value=[]), patch("src.client.apply_prompt_template", return_value="prompt"), patch.object(client, "_get_tools", return_value=[]), ): client._ensure_agent(config) assert client._agent is mock_agent def test_reuses_agent_same_config(self, client): """_ensure_agent does not recreate if config key unchanged.""" mock_agent = MagicMock() client._agent = mock_agent client._agent_config_key = (None, True, False, False) config = client._get_runnable_config("t1") client._ensure_agent(config) # Should still be the same mock β€” no recreation assert client._agent is mock_agent # --------------------------------------------------------------------------- # get_model # --------------------------------------------------------------------------- class TestGetModel: def test_found(self, client): model_cfg = MagicMock() model_cfg.name = "test-model" model_cfg.display_name = "Test Model" model_cfg.description = "A test model" model_cfg.supports_thinking = True model_cfg.supports_reasoning_effort = True client._app_config.get_model_config.return_value = model_cfg result = client.get_model("test-model") assert result == { "name": "test-model", "display_name": "Test Model", "description": "A test model", "supports_thinking": True, "supports_reasoning_effort": True, } def test_not_found(self, client): client._app_config.get_model_config.return_value = None assert client.get_model("nonexistent") is None # --------------------------------------------------------------------------- # MCP config # --------------------------------------------------------------------------- class TestMcpConfig: def test_get_mcp_config(self, client): server = MagicMock() server.model_dump.return_value = {"enabled": True, "type": "stdio"} ext_config = MagicMock() ext_config.mcp_servers = {"github": server} with patch("src.client.get_extensions_config", return_value=ext_config): result = client.get_mcp_config() assert "mcp_servers" in result assert "github" in result["mcp_servers"] assert result["mcp_servers"]["github"]["enabled"] is True def test_update_mcp_config(self, client): # Set up current config with skills current_config = MagicMock() current_config.skills = {} reloaded_server = MagicMock() reloaded_server.model_dump.return_value = {"enabled": True, "type": "sse"} reloaded_config = MagicMock() reloaded_config.mcp_servers = {"new-server": reloaded_server} with tempfile.NamedTemporaryFile(mode="w", suffix=".json", delete=False) as f: json.dump({}, f) tmp_path = Path(f.name) try: # Pre-set agent to verify it gets invalidated client._agent = MagicMock() with ( patch("src.client.ExtensionsConfig.resolve_config_path", return_value=tmp_path), patch("src.client.get_extensions_config", return_value=current_config), patch("src.client.reload_extensions_config", return_value=reloaded_config), ): result = client.update_mcp_config({"new-server": {"enabled": True, "type": "sse"}}) assert "mcp_servers" in result assert "new-server" in result["mcp_servers"] assert client._agent is None # M2: agent invalidated # Verify file was actually written with open(tmp_path) as f: saved = json.load(f) assert "mcpServers" in saved finally: tmp_path.unlink() # --------------------------------------------------------------------------- # Skills management # --------------------------------------------------------------------------- class TestSkillsManagement: def _make_skill(self, name="test-skill", enabled=True): s = MagicMock() s.name = name s.description = "A test skill" s.license = "MIT" s.category = "public" s.enabled = enabled return s def test_get_skill_found(self, client): skill = self._make_skill() with patch("src.skills.loader.load_skills", return_value=[skill]): result = client.get_skill("test-skill") assert result is not None assert result["name"] == "test-skill" def test_get_skill_not_found(self, client): with patch("src.skills.loader.load_skills", return_value=[]): result = client.get_skill("nonexistent") assert result is None def test_update_skill(self, client): skill = self._make_skill(enabled=True) updated_skill = self._make_skill(enabled=False) ext_config = MagicMock() ext_config.mcp_servers = {} ext_config.skills = {} with tempfile.NamedTemporaryFile(mode="w", suffix=".json", delete=False) as f: json.dump({}, f) tmp_path = Path(f.name) try: # Pre-set agent to verify it gets invalidated client._agent = MagicMock() with ( patch("src.skills.loader.load_skills", side_effect=[[skill], [updated_skill]]), patch("src.client.ExtensionsConfig.resolve_config_path", return_value=tmp_path), patch("src.client.get_extensions_config", return_value=ext_config), patch("src.client.reload_extensions_config"), ): result = client.update_skill("test-skill", enabled=False) assert result["enabled"] is False assert client._agent is None # M2: agent invalidated finally: tmp_path.unlink() def test_update_skill_not_found(self, client): with patch("src.skills.loader.load_skills", return_value=[]): with pytest.raises(ValueError, match="not found"): client.update_skill("nonexistent", enabled=True) def test_install_skill(self, client): with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) # Create a valid .skill archive skill_dir = tmp_path / "my-skill" skill_dir.mkdir() (skill_dir / "SKILL.md").write_text("---\nname: my-skill\ndescription: A skill\n---\nContent") archive_path = tmp_path / "my-skill.skill" with zipfile.ZipFile(archive_path, "w") as zf: zf.write(skill_dir / "SKILL.md", "my-skill/SKILL.md") skills_root = tmp_path / "skills" (skills_root / "custom").mkdir(parents=True) with ( patch("src.skills.loader.get_skills_root_path", return_value=skills_root), patch("src.gateway.routers.skills._validate_skill_frontmatter", return_value=(True, "OK", "my-skill")), ): result = client.install_skill(archive_path) assert result["success"] is True assert result["skill_name"] == "my-skill" assert (skills_root / "custom" / "my-skill").exists() def test_install_skill_not_found(self, client): with pytest.raises(FileNotFoundError): client.install_skill("/nonexistent/path.skill") def test_install_skill_bad_extension(self, client): with tempfile.NamedTemporaryFile(suffix=".zip", delete=False) as f: tmp_path = Path(f.name) try: with pytest.raises(ValueError, match=".skill extension"): client.install_skill(tmp_path) finally: tmp_path.unlink() # --------------------------------------------------------------------------- # Memory management # --------------------------------------------------------------------------- class TestMemoryManagement: def test_reload_memory(self, client): data = {"version": "1.0", "facts": []} with patch("src.agents.memory.updater.reload_memory_data", return_value=data): result = client.reload_memory() assert result == data def test_get_memory_config(self, client): config = MagicMock() config.enabled = True config.storage_path = ".deer-flow/memory.json" config.debounce_seconds = 30 config.max_facts = 100 config.fact_confidence_threshold = 0.7 config.injection_enabled = True config.max_injection_tokens = 2000 with patch("src.config.memory_config.get_memory_config", return_value=config): result = client.get_memory_config() assert result["enabled"] is True assert result["max_facts"] == 100 def test_get_memory_status(self, client): config = MagicMock() config.enabled = True config.storage_path = ".deer-flow/memory.json" config.debounce_seconds = 30 config.max_facts = 100 config.fact_confidence_threshold = 0.7 config.injection_enabled = True config.max_injection_tokens = 2000 data = {"version": "1.0", "facts": []} with ( patch("src.config.memory_config.get_memory_config", return_value=config), patch("src.agents.memory.updater.get_memory_data", return_value=data), ): result = client.get_memory_status() assert "config" in result assert "data" in result # --------------------------------------------------------------------------- # Uploads # --------------------------------------------------------------------------- class TestUploads: def test_upload_files(self, client): with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) # Create a source file src_file = tmp_path / "test.txt" src_file.write_text("hello") uploads_dir = tmp_path / "uploads" uploads_dir.mkdir() with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): result = client.upload_files("thread-1", [src_file]) assert result["success"] is True assert len(result["files"]) == 1 assert result["files"][0]["filename"] == "test.txt" assert "artifact_url" in result["files"][0] assert "message" in result assert (uploads_dir / "test.txt").exists() def test_upload_files_not_found(self, client): with pytest.raises(FileNotFoundError): client.upload_files("thread-1", ["/nonexistent/file.txt"]) def test_list_uploads(self, client): with tempfile.TemporaryDirectory() as tmp: uploads_dir = Path(tmp) (uploads_dir / "a.txt").write_text("a") (uploads_dir / "b.txt").write_text("bb") with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): result = client.list_uploads("thread-1") assert result["count"] == 2 assert len(result["files"]) == 2 names = {f["filename"] for f in result["files"]} assert names == {"a.txt", "b.txt"} # Verify artifact_url is present for f in result["files"]: assert "artifact_url" in f def test_delete_upload(self, client): with tempfile.TemporaryDirectory() as tmp: uploads_dir = Path(tmp) (uploads_dir / "delete-me.txt").write_text("gone") with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): result = client.delete_upload("thread-1", "delete-me.txt") assert result["success"] is True assert "delete-me.txt" in result["message"] assert not (uploads_dir / "delete-me.txt").exists() def test_delete_upload_not_found(self, client): with tempfile.TemporaryDirectory() as tmp: with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=Path(tmp)): with pytest.raises(FileNotFoundError): client.delete_upload("thread-1", "nope.txt") def test_delete_upload_path_traversal(self, client): with tempfile.TemporaryDirectory() as tmp: uploads_dir = Path(tmp) with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): with pytest.raises(PermissionError): client.delete_upload("thread-1", "../../etc/passwd") # --------------------------------------------------------------------------- # Artifacts # --------------------------------------------------------------------------- class TestArtifacts: def test_get_artifact(self, client): with tempfile.TemporaryDirectory() as tmp: user_data_dir = Path(tmp) / "user-data" outputs = user_data_dir / "outputs" outputs.mkdir(parents=True) (outputs / "result.txt").write_text("artifact content") mock_paths = MagicMock() mock_paths.sandbox_user_data_dir.return_value = user_data_dir with patch("src.client.get_paths", return_value=mock_paths): content, mime = client.get_artifact("t1", "mnt/user-data/outputs/result.txt") assert content == b"artifact content" assert "text" in mime def test_get_artifact_not_found(self, client): with tempfile.TemporaryDirectory() as tmp: user_data_dir = Path(tmp) / "user-data" user_data_dir.mkdir() mock_paths = MagicMock() mock_paths.sandbox_user_data_dir.return_value = user_data_dir with patch("src.client.get_paths", return_value=mock_paths): with pytest.raises(FileNotFoundError): client.get_artifact("t1", "mnt/user-data/outputs/nope.txt") def test_get_artifact_bad_prefix(self, client): with pytest.raises(ValueError, match="must start with"): client.get_artifact("t1", "bad/path/file.txt") def test_get_artifact_path_traversal(self, client): with tempfile.TemporaryDirectory() as tmp: user_data_dir = Path(tmp) / "user-data" user_data_dir.mkdir() mock_paths = MagicMock() mock_paths.sandbox_user_data_dir.return_value = user_data_dir with patch("src.client.get_paths", return_value=mock_paths): with pytest.raises(PermissionError): client.get_artifact("t1", "mnt/user-data/../../../etc/passwd") # =========================================================================== # Scenario-based integration tests # =========================================================================== # These tests simulate realistic user workflows end-to-end, exercising # multiple methods in sequence to verify they compose correctly. class TestScenarioMultiTurnConversation: """Scenario: User has a multi-turn conversation within a single thread.""" def test_two_turn_conversation(self, client): """Two sequential chat() calls on the same thread_id produce independent results (without checkpointer, each call is stateless).""" ai1 = AIMessage(content="I'm a helpful assistant.", id="ai-1") ai2 = AIMessage(content="Python is great!", id="ai-2") agent = MagicMock() agent.stream.side_effect = [ iter([{"messages": [HumanMessage(content="who are you?", id="h-1"), ai1]}]), iter([{"messages": [HumanMessage(content="what language?", id="h-2"), ai2]}]), ] with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): r1 = client.chat("who are you?", thread_id="thread-multi") r2 = client.chat("what language?", thread_id="thread-multi") assert r1 == "I'm a helpful assistant." assert r2 == "Python is great!" assert agent.stream.call_count == 2 def test_stream_collects_all_event_types_across_turns(self, client): """A full turn emits messages-tuple (tool_call, tool_result, ai text) + values + end.""" ai_tc = AIMessage(content="", id="ai-1", tool_calls=[ {"name": "web_search", "args": {"query": "LangGraph"}, "id": "tc-1"}, ]) tool_r = ToolMessage(content="LangGraph is a framework...", id="tm-1", tool_call_id="tc-1", name="web_search") ai_final = AIMessage(content="LangGraph is a framework for building agents.", id="ai-2") chunks = [ {"messages": [HumanMessage(content="search", id="h-1"), ai_tc]}, {"messages": [HumanMessage(content="search", id="h-1"), ai_tc, tool_r]}, {"messages": [HumanMessage(content="search", id="h-1"), ai_tc, tool_r, ai_final], "title": "LangGraph Search"}, ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("search", thread_id="t-full")) # Verify expected event types types = set(e.type for e in events) assert types == {"messages-tuple", "values", "end"} assert events[-1].type == "end" # Verify tool_call data tc_events = _tool_call_events(events) assert len(tc_events) == 1 assert tc_events[0].data["tool_calls"][0]["name"] == "web_search" assert tc_events[0].data["tool_calls"][0]["args"] == {"query": "LangGraph"} # Verify tool_result data tr_events = _tool_result_events(events) assert len(tr_events) == 1 assert tr_events[0].data["tool_call_id"] == "tc-1" assert "LangGraph" in tr_events[0].data["content"] # Verify AI text msg_events = _ai_events(events) assert any("framework" in e.data["content"] for e in msg_events) # Verify values event contains title values_events = [e for e in events if e.type == "values"] assert any(e.data.get("title") == "LangGraph Search" for e in values_events) class TestScenarioToolChain: """Scenario: Agent chains multiple tool calls in sequence.""" def test_multi_tool_chain(self, client): """Agent calls bash β†’ reads output β†’ calls write_file β†’ responds.""" ai_bash = AIMessage(content="", id="ai-1", tool_calls=[ {"name": "bash", "args": {"cmd": "ls /mnt/user-data/workspace"}, "id": "tc-1"}, ]) bash_result = ToolMessage(content="README.md\nsrc/", id="tm-1", tool_call_id="tc-1", name="bash") ai_write = AIMessage(content="", id="ai-2", tool_calls=[ {"name": "write_file", "args": {"path": "/mnt/user-data/outputs/listing.txt", "content": "README.md\nsrc/"}, "id": "tc-2"}, ]) write_result = ToolMessage(content="File written successfully.", id="tm-2", tool_call_id="tc-2", name="write_file") ai_final = AIMessage(content="I listed the workspace and saved the output.", id="ai-3") chunks = [ {"messages": [HumanMessage(content="list and save", id="h-1"), ai_bash]}, {"messages": [HumanMessage(content="list and save", id="h-1"), ai_bash, bash_result]}, {"messages": [HumanMessage(content="list and save", id="h-1"), ai_bash, bash_result, ai_write]}, {"messages": [HumanMessage(content="list and save", id="h-1"), ai_bash, bash_result, ai_write, write_result]}, {"messages": [HumanMessage(content="list and save", id="h-1"), ai_bash, bash_result, ai_write, write_result, ai_final]}, ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("list and save", thread_id="t-chain")) tool_calls = _tool_call_events(events) tool_results = _tool_result_events(events) messages = _ai_events(events) assert len(tool_calls) == 2 assert tool_calls[0].data["tool_calls"][0]["name"] == "bash" assert tool_calls[1].data["tool_calls"][0]["name"] == "write_file" assert len(tool_results) == 2 assert len(messages) == 1 assert events[-1].type == "end" class TestScenarioFileLifecycle: """Scenario: Upload files β†’ list them β†’ use in chat β†’ download artifact.""" def test_upload_list_delete_lifecycle(self, client): """Upload β†’ list β†’ verify β†’ delete β†’ list again.""" with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) uploads_dir = tmp_path / "uploads" uploads_dir.mkdir() # Create source files (tmp_path / "report.txt").write_text("quarterly report data") (tmp_path / "data.csv").write_text("a,b,c\n1,2,3") with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): # Step 1: Upload result = client.upload_files("t-lifecycle", [ tmp_path / "report.txt", tmp_path / "data.csv", ]) assert result["success"] is True assert len(result["files"]) == 2 assert {f["filename"] for f in result["files"]} == {"report.txt", "data.csv"} # Step 2: List listed = client.list_uploads("t-lifecycle") assert listed["count"] == 2 assert all("virtual_path" in f for f in listed["files"]) # Step 3: Delete one del_result = client.delete_upload("t-lifecycle", "report.txt") assert del_result["success"] is True # Step 4: Verify deletion listed = client.list_uploads("t-lifecycle") assert listed["count"] == 1 assert listed["files"][0]["filename"] == "data.csv" def test_upload_then_read_artifact(self, client): """Upload a file, simulate agent producing artifact, read it back.""" with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) uploads_dir = tmp_path / "uploads" uploads_dir.mkdir() user_data_dir = tmp_path / "user-data" outputs_dir = user_data_dir / "outputs" outputs_dir.mkdir(parents=True) # Upload phase src_file = tmp_path / "input.txt" src_file.write_text("raw data to process") with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): uploaded = client.upload_files("t-artifact", [src_file]) assert len(uploaded["files"]) == 1 # Simulate agent writing an artifact (outputs_dir / "analysis.json").write_text('{"result": "processed"}') # Retrieve artifact mock_paths = MagicMock() mock_paths.sandbox_user_data_dir.return_value = user_data_dir with patch("src.client.get_paths", return_value=mock_paths): content, mime = client.get_artifact("t-artifact", "mnt/user-data/outputs/analysis.json") assert json.loads(content) == {"result": "processed"} assert "json" in mime class TestScenarioConfigManagement: """Scenario: Query and update configuration through a management session.""" def test_model_and_skill_discovery(self, client): """List models β†’ get specific model β†’ list skills β†’ get specific skill.""" # List models result = client.list_models() assert len(result["models"]) >= 1 model_name = result["models"][0]["name"] # Get specific model model_cfg = MagicMock() model_cfg.name = model_name model_cfg.display_name = None model_cfg.description = None model_cfg.supports_thinking = False model_cfg.supports_reasoning_effort = False client._app_config.get_model_config.return_value = model_cfg detail = client.get_model(model_name) assert detail["name"] == model_name # List skills skill = MagicMock() skill.name = "web-search" skill.description = "Search the web" skill.license = "MIT" skill.category = "public" skill.enabled = True with patch("src.skills.loader.load_skills", return_value=[skill]): skills_result = client.list_skills() assert len(skills_result["skills"]) == 1 # Get specific skill with patch("src.skills.loader.load_skills", return_value=[skill]): detail = client.get_skill("web-search") assert detail is not None assert detail["enabled"] is True def test_mcp_update_then_skill_toggle(self, client): """Update MCP config β†’ toggle skill β†’ verify both invalidate agent.""" with tempfile.TemporaryDirectory() as tmp: config_file = Path(tmp) / "extensions_config.json" config_file.write_text("{}") # --- MCP update --- current_config = MagicMock() current_config.skills = {} reloaded_server = MagicMock() reloaded_server.model_dump.return_value = {"enabled": True, "type": "sse"} reloaded_config = MagicMock() reloaded_config.mcp_servers = {"my-mcp": reloaded_server} client._agent = MagicMock() # Simulate existing agent with ( patch("src.client.ExtensionsConfig.resolve_config_path", return_value=config_file), patch("src.client.get_extensions_config", return_value=current_config), patch("src.client.reload_extensions_config", return_value=reloaded_config), ): mcp_result = client.update_mcp_config({"my-mcp": {"enabled": True}}) assert "my-mcp" in mcp_result["mcp_servers"] assert client._agent is None # Agent invalidated # --- Skill toggle --- skill = MagicMock() skill.name = "code-gen" skill.description = "Generate code" skill.license = "MIT" skill.category = "custom" skill.enabled = True toggled = MagicMock() toggled.name = "code-gen" toggled.description = "Generate code" toggled.license = "MIT" toggled.category = "custom" toggled.enabled = False ext_config = MagicMock() ext_config.mcp_servers = {} ext_config.skills = {} client._agent = MagicMock() # Simulate re-created agent with ( patch("src.skills.loader.load_skills", side_effect=[[skill], [toggled]]), patch("src.client.ExtensionsConfig.resolve_config_path", return_value=config_file), patch("src.client.get_extensions_config", return_value=ext_config), patch("src.client.reload_extensions_config"), ): skill_result = client.update_skill("code-gen", enabled=False) assert skill_result["enabled"] is False assert client._agent is None # Agent invalidated again class TestScenarioAgentRecreation: """Scenario: Config changes trigger agent recreation at the right times.""" def test_different_model_triggers_rebuild(self, client): """Switching model_name between calls forces agent rebuild.""" agents_created = [] def fake_create_agent(**kwargs): agent = MagicMock() agents_created.append(agent) return agent config_a = client._get_runnable_config("t1", model_name="gpt-4") config_b = client._get_runnable_config("t1", model_name="claude-3") with ( patch("src.client.create_chat_model"), patch("src.client.create_agent", side_effect=fake_create_agent), patch("src.client._build_middlewares", return_value=[]), patch("src.client.apply_prompt_template", return_value="prompt"), patch.object(client, "_get_tools", return_value=[]), ): client._ensure_agent(config_a) first_agent = client._agent client._ensure_agent(config_b) second_agent = client._agent assert len(agents_created) == 2 assert first_agent is not second_agent def test_same_config_reuses_agent(self, client): """Repeated calls with identical config do not rebuild.""" agents_created = [] def fake_create_agent(**kwargs): agent = MagicMock() agents_created.append(agent) return agent config = client._get_runnable_config("t1", model_name="gpt-4") with ( patch("src.client.create_chat_model"), patch("src.client.create_agent", side_effect=fake_create_agent), patch("src.client._build_middlewares", return_value=[]), patch("src.client.apply_prompt_template", return_value="prompt"), patch.object(client, "_get_tools", return_value=[]), ): client._ensure_agent(config) client._ensure_agent(config) client._ensure_agent(config) assert len(agents_created) == 1 def test_reset_agent_forces_rebuild(self, client): """reset_agent() clears cache, next call rebuilds.""" agents_created = [] def fake_create_agent(**kwargs): agent = MagicMock() agents_created.append(agent) return agent config = client._get_runnable_config("t1") with ( patch("src.client.create_chat_model"), patch("src.client.create_agent", side_effect=fake_create_agent), patch("src.client._build_middlewares", return_value=[]), patch("src.client.apply_prompt_template", return_value="prompt"), patch.object(client, "_get_tools", return_value=[]), ): client._ensure_agent(config) client.reset_agent() client._ensure_agent(config) assert len(agents_created) == 2 def test_per_call_override_triggers_rebuild(self, client): """stream() with model_name override creates a different agent config.""" ai = AIMessage(content="ok", id="ai-1") agent = _make_agent_mock([{"messages": [ai]}]) agents_created = [] def fake_ensure(config): key = tuple(config.get("configurable", {}).get(k) for k in ["model_name", "thinking_enabled", "is_plan_mode", "subagent_enabled"]) agents_created.append(key) client._agent = agent with patch.object(client, "_ensure_agent", side_effect=fake_ensure): list(client.stream("hi", thread_id="t1")) list(client.stream("hi", thread_id="t1", model_name="other-model")) # Two different config keys should have been created assert len(agents_created) == 2 assert agents_created[0] != agents_created[1] class TestScenarioThreadIsolation: """Scenario: Operations on different threads don't interfere.""" def test_uploads_isolated_per_thread(self, client): """Files uploaded to thread-A are not visible in thread-B.""" with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) uploads_a = tmp_path / "thread-a" / "uploads" uploads_b = tmp_path / "thread-b" / "uploads" uploads_a.mkdir(parents=True) uploads_b.mkdir(parents=True) src_file = tmp_path / "secret.txt" src_file.write_text("thread-a only") def get_dir(thread_id): return uploads_a if thread_id == "thread-a" else uploads_b with patch.object(DeerFlowClient, "_get_uploads_dir", side_effect=get_dir): client.upload_files("thread-a", [src_file]) files_a = client.list_uploads("thread-a") files_b = client.list_uploads("thread-b") assert files_a["count"] == 1 assert files_b["count"] == 0 def test_artifacts_isolated_per_thread(self, client): """Artifacts in thread-A are not accessible from thread-B.""" with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) data_a = tmp_path / "thread-a" data_b = tmp_path / "thread-b" (data_a / "outputs").mkdir(parents=True) (data_b / "outputs").mkdir(parents=True) (data_a / "outputs" / "result.txt").write_text("thread-a artifact") mock_paths = MagicMock() mock_paths.sandbox_user_data_dir.side_effect = lambda tid: data_a if tid == "thread-a" else data_b with patch("src.client.get_paths", return_value=mock_paths): content, _ = client.get_artifact("thread-a", "mnt/user-data/outputs/result.txt") assert content == b"thread-a artifact" with pytest.raises(FileNotFoundError): client.get_artifact("thread-b", "mnt/user-data/outputs/result.txt") class TestScenarioMemoryWorkflow: """Scenario: Memory query β†’ reload β†’ status check.""" def test_memory_full_lifecycle(self, client): """get_memory β†’ reload β†’ get_status covers the full memory API.""" initial_data = {"version": "1.0", "facts": [{"id": "f1", "content": "User likes Python"}]} updated_data = {"version": "1.0", "facts": [ {"id": "f1", "content": "User likes Python"}, {"id": "f2", "content": "User prefers dark mode"}, ]} config = MagicMock() config.enabled = True config.storage_path = ".deer-flow/memory.json" config.debounce_seconds = 30 config.max_facts = 100 config.fact_confidence_threshold = 0.7 config.injection_enabled = True config.max_injection_tokens = 2000 with patch("src.agents.memory.updater.get_memory_data", return_value=initial_data): mem = client.get_memory() assert len(mem["facts"]) == 1 with patch("src.agents.memory.updater.reload_memory_data", return_value=updated_data): refreshed = client.reload_memory() assert len(refreshed["facts"]) == 2 with ( patch("src.config.memory_config.get_memory_config", return_value=config), patch("src.agents.memory.updater.get_memory_data", return_value=updated_data), ): status = client.get_memory_status() assert status["config"]["enabled"] is True assert len(status["data"]["facts"]) == 2 class TestScenarioSkillInstallAndUse: """Scenario: Install a skill β†’ verify it appears β†’ toggle it.""" def test_install_then_toggle(self, client): """Install .skill archive β†’ list to verify β†’ disable β†’ verify disabled.""" with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) # Create .skill archive skill_src = tmp_path / "my-analyzer" skill_src.mkdir() (skill_src / "SKILL.md").write_text( "---\nname: my-analyzer\ndescription: Analyze code\nlicense: MIT\n---\nAnalysis skill" ) archive = tmp_path / "my-analyzer.skill" with zipfile.ZipFile(archive, "w") as zf: zf.write(skill_src / "SKILL.md", "my-analyzer/SKILL.md") skills_root = tmp_path / "skills" (skills_root / "custom").mkdir(parents=True) # Step 1: Install with ( patch("src.skills.loader.get_skills_root_path", return_value=skills_root), patch("src.gateway.routers.skills._validate_skill_frontmatter", return_value=(True, "OK", "my-analyzer")), ): result = client.install_skill(archive) assert result["success"] is True assert (skills_root / "custom" / "my-analyzer" / "SKILL.md").exists() # Step 2: List and find it installed_skill = MagicMock() installed_skill.name = "my-analyzer" installed_skill.description = "Analyze code" installed_skill.license = "MIT" installed_skill.category = "custom" installed_skill.enabled = True with patch("src.skills.loader.load_skills", return_value=[installed_skill]): skills_result = client.list_skills() assert any(s["name"] == "my-analyzer" for s in skills_result["skills"]) # Step 3: Disable it disabled_skill = MagicMock() disabled_skill.name = "my-analyzer" disabled_skill.description = "Analyze code" disabled_skill.license = "MIT" disabled_skill.category = "custom" disabled_skill.enabled = False ext_config = MagicMock() ext_config.mcp_servers = {} ext_config.skills = {} config_file = tmp_path / "extensions_config.json" config_file.write_text("{}") with ( patch("src.skills.loader.load_skills", side_effect=[[installed_skill], [disabled_skill]]), patch("src.client.ExtensionsConfig.resolve_config_path", return_value=config_file), patch("src.client.get_extensions_config", return_value=ext_config), patch("src.client.reload_extensions_config"), ): toggled = client.update_skill("my-analyzer", enabled=False) assert toggled["enabled"] is False class TestScenarioEdgeCases: """Scenario: Edge cases and error boundaries in realistic workflows.""" def test_empty_stream_response(self, client): """Agent produces no messages β€” only values + end events.""" agent = _make_agent_mock([{"messages": []}]) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi", thread_id="t-empty")) # values event (empty messages) + end assert len(events) == 2 assert events[0].type == "values" assert events[-1].type == "end" def test_chat_on_empty_response(self, client): """chat() returns empty string for no-message response.""" agent = _make_agent_mock([{"messages": []}]) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): result = client.chat("hi", thread_id="t-empty-chat") assert result == "" def test_multiple_title_changes(self, client): """Title changes are carried in values events.""" ai = AIMessage(content="ok", id="ai-1") chunks = [ {"messages": [ai], "title": "First Title"}, {"messages": [], "title": "First Title"}, # same title repeated {"messages": [], "title": "Second Title"}, # different title ] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("hi", thread_id="t-titles")) # Every chunk produces a values event with the title values_events = [e for e in events if e.type == "values"] assert len(values_events) == 3 assert values_events[0].data["title"] == "First Title" assert values_events[1].data["title"] == "First Title" assert values_events[2].data["title"] == "Second Title" def test_concurrent_tool_calls_in_single_message(self, client): """Agent produces multiple tool_calls in one AIMessage β€” emitted as single messages-tuple.""" ai = AIMessage(content="", id="ai-1", tool_calls=[ {"name": "web_search", "args": {"q": "a"}, "id": "tc-1"}, {"name": "web_search", "args": {"q": "b"}, "id": "tc-2"}, {"name": "bash", "args": {"cmd": "echo hi"}, "id": "tc-3"}, ]) chunks = [{"messages": [ai]}] agent = _make_agent_mock(chunks) with ( patch.object(client, "_ensure_agent"), patch.object(client, "_agent", agent), ): events = list(client.stream("do things", thread_id="t-parallel")) tc_events = _tool_call_events(events) assert len(tc_events) == 1 # One messages-tuple event for the AIMessage tool_calls = tc_events[0].data["tool_calls"] assert len(tool_calls) == 3 assert {tc["id"] for tc in tool_calls} == {"tc-1", "tc-2", "tc-3"} def test_upload_convertible_file_conversion_failure(self, client): """Upload a .pdf file where conversion fails β€” file still uploaded, no markdown.""" with tempfile.TemporaryDirectory() as tmp: tmp_path = Path(tmp) uploads_dir = tmp_path / "uploads" uploads_dir.mkdir() pdf_file = tmp_path / "doc.pdf" pdf_file.write_bytes(b"%PDF-1.4 fake content") with ( patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir), patch("src.gateway.routers.uploads.CONVERTIBLE_EXTENSIONS", {".pdf"}), patch("src.gateway.routers.uploads.convert_file_to_markdown", side_effect=Exception("conversion failed")), ): result = client.upload_files("t-pdf-fail", [pdf_file]) assert result["success"] is True assert len(result["files"]) == 1 assert result["files"][0]["filename"] == "doc.pdf" assert "markdown_file" not in result["files"][0] # Conversion failed gracefully assert (uploads_dir / "doc.pdf").exists() # File still uploaded # --------------------------------------------------------------------------- # Gateway conformance β€” validate client output against Gateway Pydantic models # --------------------------------------------------------------------------- class TestGatewayConformance: """Validate that DeerFlowClient return dicts conform to Gateway Pydantic response models. Each test calls a client method, then parses the result through the corresponding Gateway response model. If the client drifts (missing or wrong-typed fields), Pydantic raises ``ValidationError`` and CI catches it. """ def test_list_models(self, mock_app_config): model = MagicMock() model.name = "test-model" model.display_name = "Test Model" model.description = "A test model" model.supports_thinking = False mock_app_config.models = [model] with patch("src.client.get_app_config", return_value=mock_app_config): client = DeerFlowClient() result = client.list_models() parsed = ModelsListResponse(**result) assert len(parsed.models) == 1 assert parsed.models[0].name == "test-model" def test_get_model(self, mock_app_config): model = MagicMock() model.name = "test-model" model.display_name = "Test Model" model.description = "A test model" model.supports_thinking = True mock_app_config.models = [model] mock_app_config.get_model_config.return_value = model with patch("src.client.get_app_config", return_value=mock_app_config): client = DeerFlowClient() result = client.get_model("test-model") assert result is not None parsed = ModelResponse(**result) assert parsed.name == "test-model" def test_list_skills(self, client): skill = MagicMock() skill.name = "web-search" skill.description = "Search the web" skill.license = "MIT" skill.category = "public" skill.enabled = True with patch("src.skills.loader.load_skills", return_value=[skill]): result = client.list_skills() parsed = SkillsListResponse(**result) assert len(parsed.skills) == 1 assert parsed.skills[0].name == "web-search" def test_get_skill(self, client): skill = MagicMock() skill.name = "web-search" skill.description = "Search the web" skill.license = "MIT" skill.category = "public" skill.enabled = True with patch("src.skills.loader.load_skills", return_value=[skill]): result = client.get_skill("web-search") assert result is not None parsed = SkillResponse(**result) assert parsed.name == "web-search" def test_install_skill(self, client, tmp_path): skill_dir = tmp_path / "my-skill" skill_dir.mkdir() (skill_dir / "SKILL.md").write_text( "---\nname: my-skill\ndescription: A test skill\n---\nBody\n" ) archive = tmp_path / "my-skill.skill" with zipfile.ZipFile(archive, "w") as zf: zf.write(skill_dir / "SKILL.md", "my-skill/SKILL.md") custom_dir = tmp_path / "custom" custom_dir.mkdir() with patch("src.skills.loader.get_skills_root_path", return_value=tmp_path): result = client.install_skill(archive) parsed = SkillInstallResponse(**result) assert parsed.success is True assert parsed.skill_name == "my-skill" def test_get_mcp_config(self, client): server = MagicMock() server.model_dump.return_value = { "enabled": True, "type": "stdio", "command": "npx", "args": ["-y", "server"], "env": {}, "url": None, "headers": {}, "description": "test server", } ext_config = MagicMock() ext_config.mcp_servers = {"test": server} with patch("src.client.get_extensions_config", return_value=ext_config): result = client.get_mcp_config() parsed = McpConfigResponse(**result) assert "test" in parsed.mcp_servers def test_update_mcp_config(self, client, tmp_path): server = MagicMock() server.model_dump.return_value = { "enabled": True, "type": "stdio", "command": "npx", "args": [], "env": {}, "url": None, "headers": {}, "description": "", } ext_config = MagicMock() ext_config.mcp_servers = {"srv": server} ext_config.skills = {} config_file = tmp_path / "extensions_config.json" config_file.write_text("{}") with ( patch("src.client.get_extensions_config", return_value=ext_config), patch("src.client.ExtensionsConfig.resolve_config_path", return_value=config_file), patch("src.client.reload_extensions_config", return_value=ext_config), ): result = client.update_mcp_config({"srv": server.model_dump.return_value}) parsed = McpConfigResponse(**result) assert "srv" in parsed.mcp_servers def test_upload_files(self, client, tmp_path): uploads_dir = tmp_path / "uploads" uploads_dir.mkdir() src_file = tmp_path / "hello.txt" src_file.write_text("hello") with patch.object(DeerFlowClient, "_get_uploads_dir", return_value=uploads_dir): result = client.upload_files("t-conform", [src_file]) parsed = UploadResponse(**result) assert parsed.success is True assert len(parsed.files) == 1 def test_get_memory_config(self, client): mem_cfg = MagicMock() mem_cfg.enabled = True mem_cfg.storage_path = ".deer-flow/memory.json" mem_cfg.debounce_seconds = 30 mem_cfg.max_facts = 100 mem_cfg.fact_confidence_threshold = 0.7 mem_cfg.injection_enabled = True mem_cfg.max_injection_tokens = 2000 with patch("src.config.memory_config.get_memory_config", return_value=mem_cfg): result = client.get_memory_config() parsed = MemoryConfigResponse(**result) assert parsed.enabled is True assert parsed.max_facts == 100 def test_get_memory_status(self, client): mem_cfg = MagicMock() mem_cfg.enabled = True mem_cfg.storage_path = ".deer-flow/memory.json" mem_cfg.debounce_seconds = 30 mem_cfg.max_facts = 100 mem_cfg.fact_confidence_threshold = 0.7 mem_cfg.injection_enabled = True mem_cfg.max_injection_tokens = 2000 memory_data = { "version": "1.0", "lastUpdated": "", "user": { "workContext": {"summary": "", "updatedAt": ""}, "personalContext": {"summary": "", "updatedAt": ""}, "topOfMind": {"summary": "", "updatedAt": ""}, }, "history": { "recentMonths": {"summary": "", "updatedAt": ""}, "earlierContext": {"summary": "", "updatedAt": ""}, "longTermBackground": {"summary": "", "updatedAt": ""}, }, "facts": [], } with ( patch("src.config.memory_config.get_memory_config", return_value=mem_cfg), patch("src.agents.memory.updater.get_memory_data", return_value=memory_data), ): result = client.get_memory_status() parsed = MemoryStatusResponse(**result) assert parsed.config.enabled is True assert parsed.data.version == "1.0"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_client.py", "license": "MIT License", "lines": 1264, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_client_live.py
"""Live integration tests for DeerFlowClient with real API. These tests require a working config.yaml with valid API credentials. They are skipped in CI and must be run explicitly: PYTHONPATH=. uv run pytest tests/test_client_live.py -v -s """ import json import os from pathlib import Path import pytest from src.client import DeerFlowClient, StreamEvent # Skip entire module in CI or when no config.yaml exists _skip_reason = None if os.environ.get("CI"): _skip_reason = "Live tests skipped in CI" elif not Path(__file__).resolve().parents[2].joinpath("config.yaml").exists(): _skip_reason = "No config.yaml found β€” live tests require valid API credentials" if _skip_reason: pytest.skip(_skip_reason, allow_module_level=True) # --------------------------------------------------------------------------- # Fixtures # --------------------------------------------------------------------------- @pytest.fixture(scope="module") def client(): """Create a real DeerFlowClient (no mocks).""" return DeerFlowClient(thinking_enabled=False) @pytest.fixture def thread_tmp(tmp_path): """Provide a unique thread_id + tmp directory for file operations.""" import uuid tid = f"live-test-{uuid.uuid4().hex[:8]}" return tid, tmp_path # =========================================================================== # Scenario 1: Basic chat β€” model responds coherently # =========================================================================== class TestLiveBasicChat: def test_chat_returns_nonempty_string(self, client): """chat() returns a non-empty response from the real model.""" response = client.chat("Reply with exactly: HELLO") assert isinstance(response, str) assert len(response) > 0 print(f" chat response: {response}") def test_chat_follows_instruction(self, client): """Model can follow a simple instruction.""" response = client.chat("What is 7 * 8? Reply with just the number.") assert "56" in response print(f" math response: {response}") # =========================================================================== # Scenario 2: Streaming β€” events arrive in correct order # =========================================================================== class TestLiveStreaming: def test_stream_yields_messages_tuple_and_end(self, client): """stream() produces at least one messages-tuple event and ends with end.""" events = list(client.stream("Say hi in one word.")) types = [e.type for e in events] assert "messages-tuple" in types, f"Expected 'messages-tuple' event, got: {types}" assert "values" in types, f"Expected 'values' event, got: {types}" assert types[-1] == "end" for e in events: assert isinstance(e, StreamEvent) print(f" [{e.type}] {e.data}") def test_stream_ai_content_nonempty(self, client): """Streamed messages-tuple AI events contain non-empty content.""" ai_messages = [ e for e in client.stream("What color is the sky? One word.") if e.type == "messages-tuple" and e.data.get("type") == "ai" and e.data.get("content") ] assert len(ai_messages) >= 1 for m in ai_messages: assert len(m.data.get("content", "")) > 0 # =========================================================================== # Scenario 3: Tool use β€” agent calls a tool and returns result # =========================================================================== class TestLiveToolUse: def test_agent_uses_bash_tool(self, client): """Agent uses bash tool when asked to run a command.""" events = list(client.stream( "Use the bash tool to run: echo 'LIVE_TEST_OK'. " "Then tell me the output." )) types = [e.type for e in events] print(f" event types: {types}") for e in events: print(f" [{e.type}] {e.data}") # All message events are now messages-tuple mt_events = [e for e in events if e.type == "messages-tuple"] tc_events = [e for e in mt_events if e.data.get("type") == "ai" and "tool_calls" in e.data] tr_events = [e for e in mt_events if e.data.get("type") == "tool"] ai_events = [e for e in mt_events if e.data.get("type") == "ai" and e.data.get("content")] assert len(tc_events) >= 1, f"Expected tool_call event, got types: {types}" assert len(tr_events) >= 1, f"Expected tool result event, got types: {types}" assert len(ai_events) >= 1 assert tc_events[0].data["tool_calls"][0]["name"] == "bash" assert "LIVE_TEST_OK" in tr_events[0].data["content"] def test_agent_uses_ls_tool(self, client): """Agent uses ls tool to list a directory.""" events = list(client.stream( "Use the ls tool to list the contents of /mnt/user-data/workspace. " "Just report what you see." )) types = [e.type for e in events] print(f" event types: {types}") tc_events = [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "ai" and "tool_calls" in e.data] assert len(tc_events) >= 1 assert tc_events[0].data["tool_calls"][0]["name"] == "ls" # =========================================================================== # Scenario 4: Multi-tool chain β€” agent chains tools in sequence # =========================================================================== class TestLiveMultiToolChain: def test_write_then_read(self, client): """Agent writes a file, then reads it back.""" events = list(client.stream( "Step 1: Use write_file to write 'integration_test_content' to " "/mnt/user-data/outputs/live_test.txt. " "Step 2: Use read_file to read that file back. " "Step 3: Tell me the content you read." )) types = [e.type for e in events] print(f" event types: {types}") for e in events: print(f" [{e.type}] {e.data}") tc_events = [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "ai" and "tool_calls" in e.data] tool_names = [tc.data["tool_calls"][0]["name"] for tc in tc_events] assert "write_file" in tool_names, f"Expected write_file, got: {tool_names}" assert "read_file" in tool_names, f"Expected read_file, got: {tool_names}" # Final AI message or tool result should mention the content ai_events = [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "ai" and e.data.get("content")] tr_events = [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "tool"] final_text = ai_events[-1].data["content"] if ai_events else "" assert "integration_test_content" in final_text.lower() or any( "integration_test_content" in e.data.get("content", "") for e in tr_events ) # =========================================================================== # Scenario 5: File upload lifecycle with real filesystem # =========================================================================== class TestLiveFileUpload: def test_upload_list_delete(self, client, thread_tmp): """Upload β†’ list β†’ delete β†’ verify deletion.""" thread_id, tmp_path = thread_tmp # Create test files f1 = tmp_path / "test_upload_a.txt" f1.write_text("content A") f2 = tmp_path / "test_upload_b.txt" f2.write_text("content B") # Upload result = client.upload_files(thread_id, [f1, f2]) assert result["success"] is True assert len(result["files"]) == 2 filenames = {r["filename"] for r in result["files"]} assert filenames == {"test_upload_a.txt", "test_upload_b.txt"} for r in result["files"]: assert int(r["size"]) > 0 assert r["virtual_path"].startswith("/mnt/user-data/uploads/") assert "artifact_url" in r print(f" uploaded: {filenames}") # List listed = client.list_uploads(thread_id) assert listed["count"] == 2 print(f" listed: {[f['filename'] for f in listed['files']]}") # Delete one del_result = client.delete_upload(thread_id, "test_upload_a.txt") assert del_result["success"] is True remaining = client.list_uploads(thread_id) assert remaining["count"] == 1 assert remaining["files"][0]["filename"] == "test_upload_b.txt" print(f" after delete: {[f['filename'] for f in remaining['files']]}") # Delete the other client.delete_upload(thread_id, "test_upload_b.txt") empty = client.list_uploads(thread_id) assert empty["count"] == 0 assert empty["files"] == [] def test_upload_nonexistent_file_raises(self, client): with pytest.raises(FileNotFoundError): client.upload_files("t-fail", ["/nonexistent/path/file.txt"]) # =========================================================================== # Scenario 6: Configuration query β€” real config loading # =========================================================================== class TestLiveConfigQueries: def test_list_models_returns_configured_model(self, client): """list_models() returns at least one configured model with Gateway-aligned fields.""" result = client.list_models() assert "models" in result assert len(result["models"]) >= 1 names = [m["name"] for m in result["models"]] # Verify Gateway-aligned fields for m in result["models"]: assert "display_name" in m assert "supports_thinking" in m print(f" models: {names}") def test_get_model_found(self, client): """get_model() returns details for the first configured model.""" result = client.list_models() first_model_name = result["models"][0]["name"] model = client.get_model(first_model_name) assert model is not None assert model["name"] == first_model_name assert "display_name" in model assert "supports_thinking" in model print(f" model detail: {model}") def test_get_model_not_found(self, client): assert client.get_model("nonexistent-model-xyz") is None def test_list_skills(self, client): """list_skills() runs without error.""" result = client.list_skills() assert "skills" in result assert isinstance(result["skills"], list) print(f" skills count: {len(result['skills'])}") for s in result["skills"][:3]: print(f" - {s['name']}: {s['enabled']}") # =========================================================================== # Scenario 7: Artifact read after agent writes # =========================================================================== class TestLiveArtifact: def test_get_artifact_after_write(self, client): """Agent writes a file β†’ client reads it back via get_artifact().""" import uuid thread_id = f"live-artifact-{uuid.uuid4().hex[:8]}" # Ask agent to write a file events = list(client.stream( "Use write_file to create /mnt/user-data/outputs/artifact_test.json " "with content: {\"status\": \"ok\", \"source\": \"live_test\"}", thread_id=thread_id, )) # Verify write happened tc_events = [e for e in events if e.type == "messages-tuple" and e.data.get("type") == "ai" and "tool_calls" in e.data] assert any( any(tc["name"] == "write_file" for tc in e.data["tool_calls"]) for e in tc_events ) # Read artifact content, mime = client.get_artifact(thread_id, "mnt/user-data/outputs/artifact_test.json") data = json.loads(content) assert data["status"] == "ok" assert data["source"] == "live_test" assert "json" in mime print(f" artifact: {data}, mime: {mime}") def test_get_artifact_not_found(self, client): with pytest.raises(FileNotFoundError): client.get_artifact("nonexistent-thread", "mnt/user-data/outputs/nope.txt") # =========================================================================== # Scenario 8: Per-call overrides # =========================================================================== class TestLiveOverrides: def test_thinking_disabled_still_works(self, client): """Explicit thinking_enabled=False override produces a response.""" response = client.chat( "Say OK.", thinking_enabled=False, ) assert len(response) > 0 print(f" response: {response}") # =========================================================================== # Scenario 9: Error resilience # =========================================================================== class TestLiveErrorResilience: def test_delete_nonexistent_upload(self, client): with pytest.raises(FileNotFoundError): client.delete_upload("nonexistent-thread", "ghost.txt") def test_bad_artifact_path(self, client): with pytest.raises(ValueError): client.get_artifact("t", "invalid/path") def test_path_traversal_blocked(self, client): with pytest.raises(PermissionError): client.delete_upload("t", "../../etc/passwd")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_client_live.py", "license": "MIT License", "lines": 263, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_lead_agent_model_resolution.py
"""Tests for lead agent runtime model resolution behavior.""" from __future__ import annotations import pytest from src.agents.lead_agent import agent as lead_agent_module from src.config.app_config import AppConfig from src.config.model_config import ModelConfig from src.config.sandbox_config import SandboxConfig def _make_app_config(models: list[ModelConfig]) -> AppConfig: return AppConfig( models=models, sandbox=SandboxConfig(use="src.sandbox.local:LocalSandboxProvider"), ) def _make_model(name: str, *, supports_thinking: bool) -> ModelConfig: return ModelConfig( name=name, display_name=name, description=None, use="langchain_openai:ChatOpenAI", model=name, supports_thinking=supports_thinking, supports_vision=False, ) def test_resolve_model_name_falls_back_to_default(monkeypatch, caplog): app_config = _make_app_config( [ _make_model("default-model", supports_thinking=False), _make_model("other-model", supports_thinking=True), ] ) monkeypatch.setattr(lead_agent_module, "get_app_config", lambda: app_config) with caplog.at_level("WARNING"): resolved = lead_agent_module._resolve_model_name("missing-model") assert resolved == "default-model" assert "fallback to default model 'default-model'" in caplog.text def test_resolve_model_name_uses_default_when_none(monkeypatch): app_config = _make_app_config( [ _make_model("default-model", supports_thinking=False), _make_model("other-model", supports_thinking=True), ] ) monkeypatch.setattr(lead_agent_module, "get_app_config", lambda: app_config) resolved = lead_agent_module._resolve_model_name(None) assert resolved == "default-model" def test_resolve_model_name_raises_when_no_models_configured(monkeypatch): app_config = _make_app_config([]) monkeypatch.setattr(lead_agent_module, "get_app_config", lambda: app_config) with pytest.raises( ValueError, match="No chat models are configured", ): lead_agent_module._resolve_model_name("missing-model") def test_make_lead_agent_disables_thinking_when_model_does_not_support_it(monkeypatch): app_config = _make_app_config([_make_model("safe-model", supports_thinking=False)]) import src.tools as tools_module monkeypatch.setattr(lead_agent_module, "get_app_config", lambda: app_config) monkeypatch.setattr(tools_module, "get_available_tools", lambda **kwargs: []) monkeypatch.setattr(lead_agent_module, "_build_middlewares", lambda config, model_name, agent_name=None: []) captured: dict[str, object] = {} def _fake_create_chat_model(*, name, thinking_enabled, reasoning_effort=None): captured["name"] = name captured["thinking_enabled"] = thinking_enabled captured["reasoning_effort"] = reasoning_effort return object() monkeypatch.setattr(lead_agent_module, "create_chat_model", _fake_create_chat_model) monkeypatch.setattr(lead_agent_module, "create_agent", lambda **kwargs: kwargs) result = lead_agent_module.make_lead_agent( { "configurable": { "model_name": "safe-model", "thinking_enabled": True, "is_plan_mode": False, "subagent_enabled": False, } } ) assert captured["name"] == "safe-model" assert captured["thinking_enabled"] is False assert result["model"] is not None def test_build_middlewares_uses_resolved_model_name_for_vision(monkeypatch): app_config = _make_app_config( [ _make_model("stale-model", supports_thinking=False), ModelConfig( name="vision-model", display_name="vision-model", description=None, use="langchain_openai:ChatOpenAI", model="vision-model", supports_thinking=False, supports_vision=True, ), ] ) monkeypatch.setattr(lead_agent_module, "get_app_config", lambda: app_config) monkeypatch.setattr(lead_agent_module, "_create_summarization_middleware", lambda: None) monkeypatch.setattr(lead_agent_module, "_create_todo_list_middleware", lambda is_plan_mode: None) middlewares = lead_agent_module._build_middlewares( {"configurable": {"model_name": "stale-model", "is_plan_mode": False, "subagent_enabled": False}}, model_name="vision-model", ) assert any(isinstance(m, lead_agent_module.ViewImageMiddleware) for m in middlewares)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_lead_agent_model_resolution.py", "license": "MIT License", "lines": 102, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/config/paths.py
import os import re from pathlib import Path # Virtual path prefix seen by agents inside the sandbox VIRTUAL_PATH_PREFIX = "/mnt/user-data" _SAFE_THREAD_ID_RE = re.compile(r"^[A-Za-z0-9_\-]+$") class Paths: """ Centralized path configuration for DeerFlow application data. Directory layout (host side): {base_dir}/ β”œβ”€β”€ memory.json β”œβ”€β”€ USER.md <-- global user profile (injected into all agents) β”œβ”€β”€ agents/ β”‚ └── {agent_name}/ β”‚ β”œβ”€β”€ config.yaml β”‚ β”œβ”€β”€ SOUL.md <-- agent personality/identity (injected alongside lead prompt) β”‚ └── memory.json └── threads/ └── {thread_id}/ └── user-data/ <-- mounted as /mnt/user-data/ inside sandbox β”œβ”€β”€ workspace/ <-- /mnt/user-data/workspace/ β”œβ”€β”€ uploads/ <-- /mnt/user-data/uploads/ └── outputs/ <-- /mnt/user-data/outputs/ BaseDir resolution (in priority order): 1. Constructor argument `base_dir` 2. DEER_FLOW_HOME environment variable 3. Local dev fallback: cwd/.deer-flow (when cwd is the backend/ dir) 4. Default: $HOME/.deer-flow """ def __init__(self, base_dir: str | Path | None = None) -> None: self._base_dir = Path(base_dir).resolve() if base_dir is not None else None @property def base_dir(self) -> Path: """Root directory for all application data.""" if self._base_dir is not None: return self._base_dir if env_home := os.getenv("DEER_FLOW_HOME"): return Path(env_home).resolve() cwd = Path.cwd() if cwd.name == "backend" or (cwd / "pyproject.toml").exists(): return cwd / ".deer-flow" return Path.home() / ".deer-flow" @property def memory_file(self) -> Path: """Path to the persisted memory file: `{base_dir}/memory.json`.""" return self.base_dir / "memory.json" @property def user_md_file(self) -> Path: """Path to the global user profile file: `{base_dir}/USER.md`.""" return self.base_dir / "USER.md" @property def agents_dir(self) -> Path: """Root directory for all custom agents: `{base_dir}/agents/`.""" return self.base_dir / "agents" def agent_dir(self, name: str) -> Path: """Directory for a specific agent: `{base_dir}/agents/{name}/`.""" return self.agents_dir / name.lower() def agent_memory_file(self, name: str) -> Path: """Per-agent memory file: `{base_dir}/agents/{name}/memory.json`.""" return self.agent_dir(name) / "memory.json" def thread_dir(self, thread_id: str) -> Path: """ Host path for a thread's data: `{base_dir}/threads/{thread_id}/` This directory contains a `user-data/` subdirectory that is mounted as `/mnt/user-data/` inside the sandbox. Raises: ValueError: If `thread_id` contains unsafe characters (path separators or `..`) that could cause directory traversal. """ if not _SAFE_THREAD_ID_RE.match(thread_id): raise ValueError(f"Invalid thread_id {thread_id!r}: only alphanumeric characters, hyphens, and underscores are allowed.") return self.base_dir / "threads" / thread_id def sandbox_work_dir(self, thread_id: str) -> Path: """ Host path for the agent's workspace directory. Host: `{base_dir}/threads/{thread_id}/user-data/workspace/` Sandbox: `/mnt/user-data/workspace/` """ return self.thread_dir(thread_id) / "user-data" / "workspace" def sandbox_uploads_dir(self, thread_id: str) -> Path: """ Host path for user-uploaded files. Host: `{base_dir}/threads/{thread_id}/user-data/uploads/` Sandbox: `/mnt/user-data/uploads/` """ return self.thread_dir(thread_id) / "user-data" / "uploads" def sandbox_outputs_dir(self, thread_id: str) -> Path: """ Host path for agent-generated artifacts. Host: `{base_dir}/threads/{thread_id}/user-data/outputs/` Sandbox: `/mnt/user-data/outputs/` """ return self.thread_dir(thread_id) / "user-data" / "outputs" def sandbox_user_data_dir(self, thread_id: str) -> Path: """ Host path for the user-data root. Host: `{base_dir}/threads/{thread_id}/user-data/` Sandbox: `/mnt/user-data/` """ return self.thread_dir(thread_id) / "user-data" def ensure_thread_dirs(self, thread_id: str) -> None: """Create all standard sandbox directories for a thread.""" self.sandbox_work_dir(thread_id).mkdir(parents=True, exist_ok=True) self.sandbox_uploads_dir(thread_id).mkdir(parents=True, exist_ok=True) self.sandbox_outputs_dir(thread_id).mkdir(parents=True, exist_ok=True) def resolve_virtual_path(self, thread_id: str, virtual_path: str) -> Path: """Resolve a sandbox virtual path to the actual host filesystem path. Args: thread_id: The thread ID. virtual_path: Virtual path as seen inside the sandbox, e.g. ``/mnt/user-data/outputs/report.pdf``. Leading slashes are stripped before matching. Returns: The resolved absolute host filesystem path. Raises: ValueError: If the path does not start with the expected virtual prefix or a path-traversal attempt is detected. """ stripped = virtual_path.lstrip("/") prefix = VIRTUAL_PATH_PREFIX.lstrip("/") # Require an exact segment-boundary match to avoid prefix confusion # (e.g. reject paths like "mnt/user-dataX/..."). if stripped != prefix and not stripped.startswith(prefix + "/"): raise ValueError(f"Path must start with /{prefix}") relative = stripped[len(prefix) :].lstrip("/") base = self.sandbox_user_data_dir(thread_id).resolve() actual = (base / relative).resolve() try: actual.relative_to(base) except ValueError: raise ValueError("Access denied: path traversal detected") return actual # ── Singleton ──────────────────────────────────────────────────────────── _paths: Paths | None = None def get_paths() -> Paths: """Return the global Paths singleton (lazy-initialized).""" global _paths if _paths is None: _paths = Paths() return _paths
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/paths.py", "license": "MIT License", "lines": 141, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/config/subagents_config.py
"""Configuration for the subagent system loaded from config.yaml.""" import logging from pydantic import BaseModel, Field logger = logging.getLogger(__name__) class SubagentOverrideConfig(BaseModel): """Per-agent configuration overrides.""" timeout_seconds: int | None = Field( default=None, ge=1, description="Timeout in seconds for this subagent (None = use global default)", ) class SubagentsAppConfig(BaseModel): """Configuration for the subagent system.""" timeout_seconds: int = Field( default=900, ge=1, description="Default timeout in seconds for all subagents (default: 900 = 15 minutes)", ) agents: dict[str, SubagentOverrideConfig] = Field( default_factory=dict, description="Per-agent configuration overrides keyed by agent name", ) def get_timeout_for(self, agent_name: str) -> int: """Get the effective timeout for a specific agent. Args: agent_name: The name of the subagent. Returns: The timeout in seconds, using per-agent override if set, otherwise global default. """ override = self.agents.get(agent_name) if override is not None and override.timeout_seconds is not None: return override.timeout_seconds return self.timeout_seconds _subagents_config: SubagentsAppConfig = SubagentsAppConfig() def get_subagents_app_config() -> SubagentsAppConfig: """Get the current subagents configuration.""" return _subagents_config def load_subagents_config_from_dict(config_dict: dict) -> None: """Load subagents configuration from a dictionary.""" global _subagents_config _subagents_config = SubagentsAppConfig(**config_dict) overrides_summary = {name: f"{override.timeout_seconds}s" for name, override in _subagents_config.agents.items() if override.timeout_seconds is not None} if overrides_summary: logger.info(f"Subagents config loaded: default timeout={_subagents_config.timeout_seconds}s, per-agent overrides={overrides_summary}") else: logger.info(f"Subagents config loaded: default timeout={_subagents_config.timeout_seconds}s, no per-agent overrides")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/subagents_config.py", "license": "MIT License", "lines": 46, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/tests/test_subagent_timeout_config.py
"""Tests for subagent timeout configuration. Covers: - SubagentsAppConfig / SubagentOverrideConfig model validation and defaults - get_timeout_for() resolution logic (global vs per-agent) - load_subagents_config_from_dict() and get_subagents_app_config() singleton - registry.get_subagent_config() applies config overrides - registry.list_subagents() applies overrides for all agents - Polling timeout calculation in task_tool is consistent with config """ import pytest from src.config.subagents_config import ( SubagentOverrideConfig, SubagentsAppConfig, get_subagents_app_config, load_subagents_config_from_dict, ) from src.subagents.config import SubagentConfig # --------------------------------------------------------------------------- # Helpers # --------------------------------------------------------------------------- def _reset_subagents_config(timeout_seconds: int = 900, agents: dict | None = None) -> None: """Reset global subagents config to a known state.""" load_subagents_config_from_dict({"timeout_seconds": timeout_seconds, "agents": agents or {}}) # --------------------------------------------------------------------------- # SubagentOverrideConfig # --------------------------------------------------------------------------- class TestSubagentOverrideConfig: def test_default_is_none(self): override = SubagentOverrideConfig() assert override.timeout_seconds is None def test_explicit_value(self): override = SubagentOverrideConfig(timeout_seconds=300) assert override.timeout_seconds == 300 def test_rejects_zero(self): with pytest.raises(ValueError): SubagentOverrideConfig(timeout_seconds=0) def test_rejects_negative(self): with pytest.raises(ValueError): SubagentOverrideConfig(timeout_seconds=-1) def test_minimum_valid_value(self): override = SubagentOverrideConfig(timeout_seconds=1) assert override.timeout_seconds == 1 # --------------------------------------------------------------------------- # SubagentsAppConfig – defaults and validation # --------------------------------------------------------------------------- class TestSubagentsAppConfigDefaults: def test_default_timeout(self): config = SubagentsAppConfig() assert config.timeout_seconds == 900 def test_default_agents_empty(self): config = SubagentsAppConfig() assert config.agents == {} def test_custom_global_timeout(self): config = SubagentsAppConfig(timeout_seconds=1800) assert config.timeout_seconds == 1800 def test_rejects_zero_timeout(self): with pytest.raises(ValueError): SubagentsAppConfig(timeout_seconds=0) def test_rejects_negative_timeout(self): with pytest.raises(ValueError): SubagentsAppConfig(timeout_seconds=-60) # --------------------------------------------------------------------------- # SubagentsAppConfig.get_timeout_for() # --------------------------------------------------------------------------- class TestGetTimeoutFor: def test_returns_global_default_when_no_override(self): config = SubagentsAppConfig(timeout_seconds=600) assert config.get_timeout_for("general-purpose") == 600 assert config.get_timeout_for("bash") == 600 assert config.get_timeout_for("unknown-agent") == 600 def test_returns_per_agent_override_when_set(self): config = SubagentsAppConfig( timeout_seconds=900, agents={"bash": SubagentOverrideConfig(timeout_seconds=300)}, ) assert config.get_timeout_for("bash") == 300 def test_other_agents_still_use_global_default(self): config = SubagentsAppConfig( timeout_seconds=900, agents={"bash": SubagentOverrideConfig(timeout_seconds=300)}, ) assert config.get_timeout_for("general-purpose") == 900 def test_agent_with_none_override_falls_back_to_global(self): config = SubagentsAppConfig( timeout_seconds=900, agents={"general-purpose": SubagentOverrideConfig(timeout_seconds=None)}, ) assert config.get_timeout_for("general-purpose") == 900 def test_multiple_per_agent_overrides(self): config = SubagentsAppConfig( timeout_seconds=900, agents={ "general-purpose": SubagentOverrideConfig(timeout_seconds=1800), "bash": SubagentOverrideConfig(timeout_seconds=120), }, ) assert config.get_timeout_for("general-purpose") == 1800 assert config.get_timeout_for("bash") == 120 # --------------------------------------------------------------------------- # load_subagents_config_from_dict / get_subagents_app_config singleton # --------------------------------------------------------------------------- class TestLoadSubagentsConfig: def teardown_method(self): """Restore defaults after each test.""" _reset_subagents_config() def test_load_global_timeout(self): load_subagents_config_from_dict({"timeout_seconds": 300}) assert get_subagents_app_config().timeout_seconds == 300 def test_load_with_per_agent_overrides(self): load_subagents_config_from_dict( { "timeout_seconds": 900, "agents": { "general-purpose": {"timeout_seconds": 1800}, "bash": {"timeout_seconds": 60}, }, } ) cfg = get_subagents_app_config() assert cfg.get_timeout_for("general-purpose") == 1800 assert cfg.get_timeout_for("bash") == 60 def test_load_partial_override(self): load_subagents_config_from_dict( { "timeout_seconds": 600, "agents": {"bash": {"timeout_seconds": 120}}, } ) cfg = get_subagents_app_config() assert cfg.get_timeout_for("general-purpose") == 600 assert cfg.get_timeout_for("bash") == 120 def test_load_empty_dict_uses_defaults(self): load_subagents_config_from_dict({}) cfg = get_subagents_app_config() assert cfg.timeout_seconds == 900 assert cfg.agents == {} def test_load_replaces_previous_config(self): load_subagents_config_from_dict({"timeout_seconds": 100}) assert get_subagents_app_config().timeout_seconds == 100 load_subagents_config_from_dict({"timeout_seconds": 200}) assert get_subagents_app_config().timeout_seconds == 200 def test_singleton_returns_same_instance_between_calls(self): load_subagents_config_from_dict({"timeout_seconds": 777}) assert get_subagents_app_config() is get_subagents_app_config() # --------------------------------------------------------------------------- # registry.get_subagent_config – timeout override applied # --------------------------------------------------------------------------- class TestRegistryGetSubagentConfig: def teardown_method(self): _reset_subagents_config() def test_returns_none_for_unknown_agent(self): from src.subagents.registry import get_subagent_config assert get_subagent_config("nonexistent") is None def test_returns_config_for_builtin_agents(self): from src.subagents.registry import get_subagent_config assert get_subagent_config("general-purpose") is not None assert get_subagent_config("bash") is not None def test_default_timeout_preserved_when_no_config(self): from src.subagents.registry import get_subagent_config _reset_subagents_config(timeout_seconds=900) config = get_subagent_config("general-purpose") assert config.timeout_seconds == 900 def test_global_timeout_override_applied(self): from src.subagents.registry import get_subagent_config _reset_subagents_config(timeout_seconds=1800) config = get_subagent_config("general-purpose") assert config.timeout_seconds == 1800 def test_per_agent_timeout_override_applied(self): from src.subagents.registry import get_subagent_config load_subagents_config_from_dict( { "timeout_seconds": 900, "agents": {"bash": {"timeout_seconds": 120}}, } ) bash_config = get_subagent_config("bash") assert bash_config.timeout_seconds == 120 def test_per_agent_override_does_not_affect_other_agents(self): from src.subagents.registry import get_subagent_config load_subagents_config_from_dict( { "timeout_seconds": 900, "agents": {"bash": {"timeout_seconds": 120}}, } ) gp_config = get_subagent_config("general-purpose") assert gp_config.timeout_seconds == 900 def test_builtin_config_object_is_not_mutated(self): """Registry must return a new object, leaving the builtin default intact.""" from src.subagents.builtins import BUILTIN_SUBAGENTS from src.subagents.registry import get_subagent_config original_timeout = BUILTIN_SUBAGENTS["bash"].timeout_seconds load_subagents_config_from_dict({"timeout_seconds": 42}) returned = get_subagent_config("bash") assert returned.timeout_seconds == 42 assert BUILTIN_SUBAGENTS["bash"].timeout_seconds == original_timeout def test_config_preserves_other_fields(self): """Applying timeout override must not change other SubagentConfig fields.""" from src.subagents.builtins import BUILTIN_SUBAGENTS from src.subagents.registry import get_subagent_config _reset_subagents_config(timeout_seconds=300) original = BUILTIN_SUBAGENTS["general-purpose"] overridden = get_subagent_config("general-purpose") assert overridden.name == original.name assert overridden.description == original.description assert overridden.max_turns == original.max_turns assert overridden.model == original.model assert overridden.tools == original.tools assert overridden.disallowed_tools == original.disallowed_tools # --------------------------------------------------------------------------- # registry.list_subagents – all agents get overrides # --------------------------------------------------------------------------- class TestRegistryListSubagents: def teardown_method(self): _reset_subagents_config() def test_lists_both_builtin_agents(self): from src.subagents.registry import list_subagents names = {cfg.name for cfg in list_subagents()} assert "general-purpose" in names assert "bash" in names def test_all_returned_configs_get_global_override(self): from src.subagents.registry import list_subagents _reset_subagents_config(timeout_seconds=123) for cfg in list_subagents(): assert cfg.timeout_seconds == 123, f"{cfg.name} has wrong timeout" def test_per_agent_overrides_reflected_in_list(self): from src.subagents.registry import list_subagents load_subagents_config_from_dict( { "timeout_seconds": 900, "agents": { "general-purpose": {"timeout_seconds": 1800}, "bash": {"timeout_seconds": 60}, }, } ) by_name = {cfg.name: cfg for cfg in list_subagents()} assert by_name["general-purpose"].timeout_seconds == 1800 assert by_name["bash"].timeout_seconds == 60 # --------------------------------------------------------------------------- # Polling timeout calculation (logic extracted from task_tool) # --------------------------------------------------------------------------- class TestPollingTimeoutCalculation: """Verify the formula (timeout_seconds + 60) // 5 is correct for various inputs.""" @pytest.mark.parametrize( "timeout_seconds, expected_max_polls", [ (900, 192), # default 15 min β†’ (900+60)//5 = 192 (300, 72), # 5 min β†’ (300+60)//5 = 72 (1800, 372), # 30 min β†’ (1800+60)//5 = 372 (60, 24), # 1 min β†’ (60+60)//5 = 24 (1, 12), # minimum β†’ (1+60)//5 = 12 ], ) def test_polling_timeout_formula(self, timeout_seconds: int, expected_max_polls: int): dummy_config = SubagentConfig( name="test", description="test", system_prompt="test", timeout_seconds=timeout_seconds, ) max_poll_count = (dummy_config.timeout_seconds + 60) // 5 assert max_poll_count == expected_max_polls def test_polling_timeout_exceeds_execution_timeout(self): """Safety-net polling window must always be longer than the execution timeout.""" for timeout_seconds in [60, 300, 900, 1800]: dummy_config = SubagentConfig( name="test", description="test", system_prompt="test", timeout_seconds=timeout_seconds, ) max_poll_count = (dummy_config.timeout_seconds + 60) // 5 polling_window_seconds = max_poll_count * 5 assert polling_window_seconds > timeout_seconds
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_subagent_timeout_config.py", "license": "MIT License", "lines": 275, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_docker_sandbox_mode_detection.py
"""Regression tests for docker sandbox mode detection logic.""" from __future__ import annotations import subprocess import tempfile from pathlib import Path REPO_ROOT = Path(__file__).resolve().parents[2] SCRIPT_PATH = REPO_ROOT / "scripts" / "docker.sh" def _detect_mode_with_config(config_content: str) -> str: """Write config content into a temp project root and execute detect_sandbox_mode.""" with tempfile.TemporaryDirectory() as tmpdir: tmp_root = Path(tmpdir) (tmp_root / "config.yaml").write_text(config_content) command = f"source '{SCRIPT_PATH}' && PROJECT_ROOT='{tmp_root}' && detect_sandbox_mode" output = subprocess.check_output( ["bash", "-lc", command], text=True, ).strip() return output def test_detect_mode_defaults_to_local_when_config_missing(): """No config file should default to local mode.""" with tempfile.TemporaryDirectory() as tmpdir: command = f"source '{SCRIPT_PATH}' && PROJECT_ROOT='{tmpdir}' && detect_sandbox_mode" output = subprocess.check_output(["bash", "-lc", command], text=True).strip() assert output == "local" def test_detect_mode_local_provider(): """Local sandbox provider should map to local mode.""" config = """ sandbox: use: src.sandbox.local:LocalSandboxProvider """.strip() assert _detect_mode_with_config(config) == "local" def test_detect_mode_aio_without_provisioner_url(): """AIO sandbox without provisioner_url should map to aio mode.""" config = """ sandbox: use: src.community.aio_sandbox:AioSandboxProvider """.strip() assert _detect_mode_with_config(config) == "aio" def test_detect_mode_provisioner_with_url(): """AIO sandbox with provisioner_url should map to provisioner mode.""" config = """ sandbox: use: src.community.aio_sandbox:AioSandboxProvider provisioner_url: http://provisioner:8002 """.strip() assert _detect_mode_with_config(config) == "provisioner" def test_detect_mode_ignores_commented_provisioner_url(): """Commented provisioner_url should not activate provisioner mode.""" config = """ sandbox: use: src.community.aio_sandbox:AioSandboxProvider # provisioner_url: http://provisioner:8002 """.strip() assert _detect_mode_with_config(config) == "aio" def test_detect_mode_unknown_provider_falls_back_to_local(): """Unknown sandbox provider should default to local mode.""" config = """ sandbox: use: custom.module:UnknownProvider """.strip() assert _detect_mode_with_config(config) == "local"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_docker_sandbox_mode_detection.py", "license": "MIT License", "lines": 61, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/tests/test_provisioner_kubeconfig.py
"""Regression tests for provisioner kubeconfig path handling.""" from __future__ import annotations import importlib.util from pathlib import Path def _load_provisioner_module(): """Load docker/provisioner/app.py as an importable test module.""" repo_root = Path(__file__).resolve().parents[2] module_path = repo_root / "docker" / "provisioner" / "app.py" spec = importlib.util.spec_from_file_location("provisioner_app_test", module_path) assert spec is not None assert spec.loader is not None module = importlib.util.module_from_spec(spec) spec.loader.exec_module(module) return module def test_wait_for_kubeconfig_rejects_directory(tmp_path): """Directory mount at kubeconfig path should fail fast with clear error.""" provisioner_module = _load_provisioner_module() kubeconfig_dir = tmp_path / "config_dir" kubeconfig_dir.mkdir() provisioner_module.KUBECONFIG_PATH = str(kubeconfig_dir) try: provisioner_module._wait_for_kubeconfig(timeout=1) raise AssertionError("Expected RuntimeError for directory kubeconfig path") except RuntimeError as exc: assert "directory" in str(exc) def test_wait_for_kubeconfig_accepts_file(tmp_path): """Regular file mount should pass readiness wait.""" provisioner_module = _load_provisioner_module() kubeconfig_file = tmp_path / "config" kubeconfig_file.write_text("apiVersion: v1\n") provisioner_module.KUBECONFIG_PATH = str(kubeconfig_file) # Should return immediately without raising. provisioner_module._wait_for_kubeconfig(timeout=1) def test_init_k8s_client_rejects_directory_path(tmp_path): """KUBECONFIG_PATH that resolves to a directory should be rejected.""" provisioner_module = _load_provisioner_module() kubeconfig_dir = tmp_path / "config_dir" kubeconfig_dir.mkdir() provisioner_module.KUBECONFIG_PATH = str(kubeconfig_dir) try: provisioner_module._init_k8s_client() raise AssertionError("Expected RuntimeError for directory kubeconfig path") except RuntimeError as exc: assert "expected a file" in str(exc) def test_init_k8s_client_uses_file_kubeconfig(tmp_path, monkeypatch): """When file exists, provisioner should load kubeconfig file path.""" provisioner_module = _load_provisioner_module() kubeconfig_file = tmp_path / "config" kubeconfig_file.write_text("apiVersion: v1\n") called: dict[str, object] = {} def fake_load_kube_config(config_file: str): called["config_file"] = config_file monkeypatch.setattr( provisioner_module.k8s_config, "load_kube_config", fake_load_kube_config, ) monkeypatch.setattr( provisioner_module.k8s_client, "CoreV1Api", lambda *args, **kwargs: "core-v1", ) provisioner_module.KUBECONFIG_PATH = str(kubeconfig_file) result = provisioner_module._init_k8s_client() assert called["config_file"] == str(kubeconfig_file) assert result == "core-v1" def test_init_k8s_client_falls_back_to_incluster_when_missing(tmp_path, monkeypatch): """When kubeconfig file is missing, in-cluster config should be attempted.""" provisioner_module = _load_provisioner_module() missing_path = tmp_path / "missing-config" calls: dict[str, int] = {"incluster": 0} def fake_load_incluster_config(): calls["incluster"] += 1 monkeypatch.setattr( provisioner_module.k8s_config, "load_incluster_config", fake_load_incluster_config, ) monkeypatch.setattr( provisioner_module.k8s_client, "CoreV1Api", lambda *args, **kwargs: "core-v1", ) provisioner_module.KUBECONFIG_PATH = str(missing_path) result = provisioner_module._init_k8s_client() assert calls["incluster"] == 1 assert result == "core-v1"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/tests/test_provisioner_kubeconfig.py", "license": "MIT License", "lines": 87, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
bytedance/deer-flow:backend/src/config/tracing_config.py
import logging import os import threading from pydantic import BaseModel, Field logger = logging.getLogger(__name__) _config_lock = threading.Lock() class TracingConfig(BaseModel): """Configuration for LangSmith tracing.""" enabled: bool = Field(...) api_key: str | None = Field(...) project: str = Field(...) endpoint: str = Field(...) @property def is_configured(self) -> bool: """Check if tracing is fully configured (enabled and has API key).""" return self.enabled and bool(self.api_key) _tracing_config: TracingConfig | None = None def get_tracing_config() -> TracingConfig: """Get the current tracing configuration from environment variables. Returns: TracingConfig with current settings. """ global _tracing_config if _tracing_config is not None: return _tracing_config with _config_lock: if _tracing_config is not None: # Double-check after acquiring lock return _tracing_config _tracing_config = TracingConfig( enabled=os.environ.get("LANGSMITH_TRACING", "").lower() == "true", api_key=os.environ.get("LANGSMITH_API_KEY"), project=os.environ.get("LANGSMITH_PROJECT", "deer-flow"), endpoint=os.environ.get("LANGSMITH_ENDPOINT", "https://api.smith.langchain.com"), ) return _tracing_config def is_tracing_enabled() -> bool: """Check if LangSmith tracing is enabled and configured. Returns: True if tracing is enabled and has an API key. """ return get_tracing_config().is_configured
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/tracing_config.py", "license": "MIT License", "lines": 41, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/debug.py
#!/usr/bin/env python """ Debug script for lead_agent. Run this file directly in VS Code with breakpoints. Usage: 1. Set breakpoints in agent.py or other files 2. Press F5 or use "Run and Debug" panel 3. Input messages in the terminal to interact with the agent """ import asyncio import logging import os import sys # Ensure we can import from src sys.path.insert(0, os.path.dirname(os.path.abspath(__file__))) # Load environment variables from dotenv import load_dotenv from langchain_core.messages import HumanMessage from src.agents import make_lead_agent load_dotenv() # Configure logging logging.basicConfig( level=logging.INFO, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", ) async def main(): # Initialize MCP tools at startup try: from src.mcp import initialize_mcp_tools await initialize_mcp_tools() except Exception as e: print(f"Warning: Failed to initialize MCP tools: {e}") # Create agent with default config config = { "configurable": { "thread_id": "debug-thread-001", "thinking_enabled": True, "is_plan_mode": True, # Uncomment to use a specific model "model_name": "kimi-k2.5", } } agent = make_lead_agent(config) print("=" * 50) print("Lead Agent Debug Mode") print("Type 'quit' or 'exit' to stop") print("=" * 50) while True: try: user_input = input("\nYou: ").strip() if not user_input: continue if user_input.lower() in ("quit", "exit"): print("Goodbye!") break # Invoke the agent state = {"messages": [HumanMessage(content=user_input)]} result = await agent.ainvoke(state, config=config, context={"thread_id": "debug-thread-001"}) # Print the response if result.get("messages"): last_message = result["messages"][-1] print(f"\nAgent: {last_message.content}") except KeyboardInterrupt: print("\nInterrupted. Goodbye!") break except Exception as e: print(f"\nError: {e}") import traceback traceback.print_exc() if __name__ == "__main__": asyncio.run(main())
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/debug.py", "license": "MIT License", "lines": 72, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/lead_agent/agent.py
import logging from langchain.agents import create_agent from langchain.agents.middleware import SummarizationMiddleware, TodoListMiddleware from langchain_core.runnables import RunnableConfig from src.agents.lead_agent.prompt import apply_prompt_template from src.agents.middlewares.clarification_middleware import ClarificationMiddleware from src.agents.middlewares.dangling_tool_call_middleware import DanglingToolCallMiddleware from src.agents.middlewares.memory_middleware import MemoryMiddleware from src.agents.middlewares.subagent_limit_middleware import SubagentLimitMiddleware from src.agents.middlewares.thread_data_middleware import ThreadDataMiddleware from src.agents.middlewares.title_middleware import TitleMiddleware from src.agents.middlewares.uploads_middleware import UploadsMiddleware from src.agents.middlewares.view_image_middleware import ViewImageMiddleware from src.agents.thread_state import ThreadState from src.config.agents_config import load_agent_config from src.config.app_config import get_app_config from src.config.summarization_config import get_summarization_config from src.models import create_chat_model from src.sandbox.middleware import SandboxMiddleware logger = logging.getLogger(__name__) def _resolve_model_name(requested_model_name: str | None = None) -> str: """Resolve a runtime model name safely, falling back to default if invalid. Returns None if no models are configured.""" app_config = get_app_config() default_model_name = app_config.models[0].name if app_config.models else None if default_model_name is None: raise ValueError("No chat models are configured. Please configure at least one model in config.yaml.") if requested_model_name and app_config.get_model_config(requested_model_name): return requested_model_name if requested_model_name and requested_model_name != default_model_name: logger.warning(f"Model '{requested_model_name}' not found in config; fallback to default model '{default_model_name}'.") return default_model_name def _create_summarization_middleware() -> SummarizationMiddleware | None: """Create and configure the summarization middleware from config.""" config = get_summarization_config() if not config.enabled: return None # Prepare trigger parameter trigger = None if config.trigger is not None: if isinstance(config.trigger, list): trigger = [t.to_tuple() for t in config.trigger] else: trigger = config.trigger.to_tuple() # Prepare keep parameter keep = config.keep.to_tuple() # Prepare model parameter if config.model_name: model = config.model_name else: # Use a lightweight model for summarization to save costs # Falls back to default model if not explicitly specified model = create_chat_model(thinking_enabled=False) # Prepare kwargs kwargs = { "model": model, "trigger": trigger, "keep": keep, } if config.trim_tokens_to_summarize is not None: kwargs["trim_tokens_to_summarize"] = config.trim_tokens_to_summarize if config.summary_prompt is not None: kwargs["summary_prompt"] = config.summary_prompt return SummarizationMiddleware(**kwargs) def _create_todo_list_middleware(is_plan_mode: bool) -> TodoListMiddleware | None: """Create and configure the TodoList middleware. Args: is_plan_mode: Whether to enable plan mode with TodoList middleware. Returns: TodoListMiddleware instance if plan mode is enabled, None otherwise. """ if not is_plan_mode: return None # Custom prompts matching DeerFlow's style system_prompt = """ <todo_list_system> You have access to the `write_todos` tool to help you manage and track complex multi-step objectives. **CRITICAL RULES:** - Mark todos as completed IMMEDIATELY after finishing each step - do NOT batch completions - Keep EXACTLY ONE task as `in_progress` at any time (unless tasks can run in parallel) - Update the todo list in REAL-TIME as you work - this gives users visibility into your progress - DO NOT use this tool for simple tasks (< 3 steps) - just complete them directly **When to Use:** This tool is designed for complex objectives that require systematic tracking: - Complex multi-step tasks requiring 3+ distinct steps - Non-trivial tasks needing careful planning and execution - User explicitly requests a todo list - User provides multiple tasks (numbered or comma-separated list) - The plan may need revisions based on intermediate results **When NOT to Use:** - Single, straightforward tasks - Trivial tasks (< 3 steps) - Purely conversational or informational requests - Simple tool calls where the approach is obvious **Best Practices:** - Break down complex tasks into smaller, actionable steps - Use clear, descriptive task names - Remove tasks that become irrelevant - Add new tasks discovered during implementation - Don't be afraid to revise the todo list as you learn more **Task Management:** Writing todos takes time and tokens - use it when helpful for managing complex problems, not for simple requests. </todo_list_system> """ tool_description = """Use this tool to create and manage a structured task list for complex work sessions. **IMPORTANT: Only use this tool for complex tasks (3+ steps). For simple requests, just do the work directly.** ## When to Use Use this tool in these scenarios: 1. **Complex multi-step tasks**: When a task requires 3 or more distinct steps or actions 2. **Non-trivial tasks**: Tasks requiring careful planning or multiple operations 3. **User explicitly requests todo list**: When the user directly asks you to track tasks 4. **Multiple tasks**: When users provide a list of things to be done 5. **Dynamic planning**: When the plan may need updates based on intermediate results ## When NOT to Use Skip this tool when: 1. The task is straightforward and takes less than 3 steps 2. The task is trivial and tracking provides no benefit 3. The task is purely conversational or informational 4. It's clear what needs to be done and you can just do it ## How to Use 1. **Starting a task**: Mark it as `in_progress` BEFORE beginning work 2. **Completing a task**: Mark it as `completed` IMMEDIATELY after finishing 3. **Updating the list**: Add new tasks, remove irrelevant ones, or update descriptions as needed 4. **Multiple updates**: You can make several updates at once (e.g., complete one task and start the next) ## Task States - `pending`: Task not yet started - `in_progress`: Currently working on (can have multiple if tasks run in parallel) - `completed`: Task finished successfully ## Task Completion Requirements **CRITICAL: Only mark a task as completed when you have FULLY accomplished it.** Never mark a task as completed if: - There are unresolved issues or errors - Work is partial or incomplete - You encountered blockers preventing completion - You couldn't find necessary resources or dependencies - Quality standards haven't been met If blocked, keep the task as `in_progress` and create a new task describing what needs to be resolved. ## Best Practices - Create specific, actionable items - Break complex tasks into smaller, manageable steps - Use clear, descriptive task names - Update task status in real-time as you work - Mark tasks complete IMMEDIATELY after finishing (don't batch completions) - Remove tasks that are no longer relevant - **IMPORTANT**: When you write the todo list, mark your first task(s) as `in_progress` immediately - **IMPORTANT**: Unless all tasks are completed, always have at least one task `in_progress` to show progress Being proactive with task management demonstrates thoroughness and ensures all requirements are completed successfully. **Remember**: If you only need a few tool calls to complete a task and it's clear what to do, it's better to just do the task directly and NOT use this tool at all. """ return TodoListMiddleware(system_prompt=system_prompt, tool_description=tool_description) # ThreadDataMiddleware must be before SandboxMiddleware to ensure thread_id is available # UploadsMiddleware should be after ThreadDataMiddleware to access thread_id # DanglingToolCallMiddleware patches missing ToolMessages before model sees the history # SummarizationMiddleware should be early to reduce context before other processing # TodoListMiddleware should be before ClarificationMiddleware to allow todo management # TitleMiddleware generates title after first exchange # MemoryMiddleware queues conversation for memory update (after TitleMiddleware) # ViewImageMiddleware should be before ClarificationMiddleware to inject image details before LLM # ClarificationMiddleware should be last to intercept clarification requests after model calls def _build_middlewares(config: RunnableConfig, model_name: str | None, agent_name: str | None = None): """Build middleware chain based on runtime configuration. Args: config: Runtime configuration containing configurable options like is_plan_mode. agent_name: If provided, MemoryMiddleware will use per-agent memory storage. Returns: List of middleware instances. """ middlewares = [ThreadDataMiddleware(), UploadsMiddleware(), SandboxMiddleware(), DanglingToolCallMiddleware()] # Add summarization middleware if enabled summarization_middleware = _create_summarization_middleware() if summarization_middleware is not None: middlewares.append(summarization_middleware) # Add TodoList middleware if plan mode is enabled is_plan_mode = config.get("configurable", {}).get("is_plan_mode", False) todo_list_middleware = _create_todo_list_middleware(is_plan_mode) if todo_list_middleware is not None: middlewares.append(todo_list_middleware) # Add TitleMiddleware middlewares.append(TitleMiddleware()) # Add MemoryMiddleware (after TitleMiddleware) middlewares.append(MemoryMiddleware(agent_name=agent_name)) # Add ViewImageMiddleware only if the current model supports vision. # Use the resolved runtime model_name from make_lead_agent to avoid stale config values. app_config = get_app_config() model_config = app_config.get_model_config(model_name) if model_name else None if model_config is not None and model_config.supports_vision: middlewares.append(ViewImageMiddleware()) # Add SubagentLimitMiddleware to truncate excess parallel task calls subagent_enabled = config.get("configurable", {}).get("subagent_enabled", False) if subagent_enabled: max_concurrent_subagents = config.get("configurable", {}).get("max_concurrent_subagents", 3) middlewares.append(SubagentLimitMiddleware(max_concurrent=max_concurrent_subagents)) # ClarificationMiddleware should always be last middlewares.append(ClarificationMiddleware()) return middlewares def make_lead_agent(config: RunnableConfig): # Lazy import to avoid circular dependency from src.tools import get_available_tools from src.tools.builtins import setup_agent cfg = config.get("configurable", {}) thinking_enabled = cfg.get("thinking_enabled", True) reasoning_effort = cfg.get("reasoning_effort", None) requested_model_name: str | None = cfg.get("model_name") or cfg.get("model") is_plan_mode = cfg.get("is_plan_mode", False) subagent_enabled = cfg.get("subagent_enabled", False) max_concurrent_subagents = cfg.get("max_concurrent_subagents", 3) is_bootstrap = cfg.get("is_bootstrap", False) agent_name = cfg.get("agent_name") agent_config = load_agent_config(agent_name) if not is_bootstrap else None # Custom agent model or fallback to global/default model resolution agent_model_name = agent_config.model if agent_config and agent_config.model else _resolve_model_name() # Final model name resolution with request override, then agent config, then global default model_name = requested_model_name or agent_model_name app_config = get_app_config() model_config = app_config.get_model_config(model_name) if model_name else None if model_config is None: raise ValueError("No chat model could be resolved. Please configure at least one model in config.yaml or provide a valid 'model_name'/'model' in the request.") if thinking_enabled and not model_config.supports_thinking: logger.warning(f"Thinking mode is enabled but model '{model_name}' does not support it; fallback to non-thinking mode.") thinking_enabled = False logger.info( "Create Agent(%s) -> thinking_enabled: %s, reasoning_effort: %s, model_name: %s, is_plan_mode: %s, subagent_enabled: %s, max_concurrent_subagents: %s", agent_name or "default", thinking_enabled, reasoning_effort, model_name, is_plan_mode, subagent_enabled, max_concurrent_subagents, ) # Inject run metadata for LangSmith trace tagging if "metadata" not in config: config["metadata"] = {} config["metadata"].update( { "agent_name": agent_name or "default", "model_name": model_name or "default", "thinking_enabled": thinking_enabled, "reasoning_effort": reasoning_effort, "is_plan_mode": is_plan_mode, "subagent_enabled": subagent_enabled, } ) if is_bootstrap: # Special bootstrap agent with minimal prompt for initial custom agent creation flow system_prompt = apply_prompt_template(subagent_enabled=subagent_enabled, max_concurrent_subagents=max_concurrent_subagents, available_skills=set(["bootstrap"])) return create_agent( model=create_chat_model(name=model_name, thinking_enabled=thinking_enabled), tools=get_available_tools(model_name=model_name, subagent_enabled=subagent_enabled) + [setup_agent], middleware=_build_middlewares(config, model_name=model_name), system_prompt=system_prompt, state_schema=ThreadState, ) # Default lead agent (unchanged behavior) return create_agent( model=create_chat_model(name=model_name, thinking_enabled=thinking_enabled, reasoning_effort=reasoning_effort), tools=get_available_tools(model_name=model_name, groups=agent_config.tool_groups if agent_config else None, subagent_enabled=subagent_enabled), middleware=_build_middlewares(config, model_name=model_name, agent_name=agent_name), system_prompt=apply_prompt_template(subagent_enabled=subagent_enabled, max_concurrent_subagents=max_concurrent_subagents, agent_name=agent_name), state_schema=ThreadState, )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/lead_agent/agent.py", "license": "MIT License", "lines": 260, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/lead_agent/prompt.py
from datetime import datetime from src.config.agents_config import load_agent_soul from src.skills import load_skills def _build_subagent_section(max_concurrent: int) -> str: """Build the subagent system prompt section with dynamic concurrency limit. Args: max_concurrent: Maximum number of concurrent subagent calls allowed per response. Returns: Formatted subagent section string. """ n = max_concurrent return f"""<subagent_system> **πŸš€ SUBAGENT MODE ACTIVE - DECOMPOSE, DELEGATE, SYNTHESIZE** You are running with subagent capabilities enabled. Your role is to be a **task orchestrator**: 1. **DECOMPOSE**: Break complex tasks into parallel sub-tasks 2. **DELEGATE**: Launch multiple subagents simultaneously using parallel `task` calls 3. **SYNTHESIZE**: Collect and integrate results into a coherent answer **CORE PRINCIPLE: Complex tasks should be decomposed and distributed across multiple subagents for parallel execution.** **β›” HARD CONCURRENCY LIMIT: MAXIMUM {n} `task` CALLS PER RESPONSE. THIS IS NOT OPTIONAL.** - Each response, you may include **at most {n}** `task` tool calls. Any excess calls are **silently discarded** by the system β€” you will lose that work. - **Before launching subagents, you MUST count your sub-tasks in your thinking:** - If count ≀ {n}: Launch all in this response. - If count > {n}: **Pick the {n} most important/foundational sub-tasks for this turn.** Save the rest for the next turn. - **Multi-batch execution** (for >{n} sub-tasks): - Turn 1: Launch sub-tasks 1-{n} in parallel β†’ wait for results - Turn 2: Launch next batch in parallel β†’ wait for results - ... continue until all sub-tasks are complete - Final turn: Synthesize ALL results into a coherent answer - **Example thinking pattern**: "I identified 6 sub-tasks. Since the limit is {n} per turn, I will launch the first {n} now, and the rest in the next turn." **Available Subagents:** - **general-purpose**: For ANY non-trivial task - web research, code exploration, file operations, analysis, etc. - **bash**: For command execution (git, build, test, deploy operations) **Your Orchestration Strategy:** βœ… **DECOMPOSE + PARALLEL EXECUTION (Preferred Approach):** For complex queries, break them down into focused sub-tasks and execute in parallel batches (max {n} per turn): **Example 1: "Why is Tencent's stock price declining?" (3 sub-tasks β†’ 1 batch)** β†’ Turn 1: Launch 3 subagents in parallel: - Subagent 1: Recent financial reports, earnings data, and revenue trends - Subagent 2: Negative news, controversies, and regulatory issues - Subagent 3: Industry trends, competitor performance, and market sentiment β†’ Turn 2: Synthesize results **Example 2: "Compare 5 cloud providers" (5 sub-tasks β†’ multi-batch)** β†’ Turn 1: Launch {n} subagents in parallel (first batch) β†’ Turn 2: Launch remaining subagents in parallel β†’ Final turn: Synthesize ALL results into comprehensive comparison **Example 3: "Refactor the authentication system"** β†’ Turn 1: Launch 3 subagents in parallel: - Subagent 1: Analyze current auth implementation and technical debt - Subagent 2: Research best practices and security patterns - Subagent 3: Review related tests, documentation, and vulnerabilities β†’ Turn 2: Synthesize results βœ… **USE Parallel Subagents (max {n} per turn) when:** - **Complex research questions**: Requires multiple information sources or perspectives - **Multi-aspect analysis**: Task has several independent dimensions to explore - **Large codebases**: Need to analyze different parts simultaneously - **Comprehensive investigations**: Questions requiring thorough coverage from multiple angles ❌ **DO NOT use subagents (execute directly) when:** - **Task cannot be decomposed**: If you can't break it into 2+ meaningful parallel sub-tasks, execute directly - **Ultra-simple actions**: Read one file, quick edits, single commands - **Need immediate clarification**: Must ask user before proceeding - **Meta conversation**: Questions about conversation history - **Sequential dependencies**: Each step depends on previous results (do steps yourself sequentially) **CRITICAL WORKFLOW** (STRICTLY follow this before EVERY action): 1. **COUNT**: In your thinking, list all sub-tasks and count them explicitly: "I have N sub-tasks" 2. **PLAN BATCHES**: If N > {n}, explicitly plan which sub-tasks go in which batch: - "Batch 1 (this turn): first {n} sub-tasks" - "Batch 2 (next turn): next batch of sub-tasks" 3. **EXECUTE**: Launch ONLY the current batch (max {n} `task` calls). Do NOT launch sub-tasks from future batches. 4. **REPEAT**: After results return, launch the next batch. Continue until all batches complete. 5. **SYNTHESIZE**: After ALL batches are done, synthesize all results. 6. **Cannot decompose** β†’ Execute directly using available tools (bash, read_file, web_search, etc.) **β›” VIOLATION: Launching more than {n} `task` calls in a single response is a HARD ERROR. The system WILL discard excess calls and you WILL lose work. Always batch.** **Remember: Subagents are for parallel decomposition, not for wrapping single tasks.** **How It Works:** - The task tool runs subagents asynchronously in the background - The backend automatically polls for completion (you don't need to poll) - The tool call will block until the subagent completes its work - Once complete, the result is returned to you directly **Usage Example 1 - Single Batch (≀{n} sub-tasks):** ```python # User asks: "Why is Tencent's stock price declining?" # Thinking: 3 sub-tasks β†’ fits in 1 batch # Turn 1: Launch 3 subagents in parallel task(description="Tencent financial data", prompt="...", subagent_type="general-purpose") task(description="Tencent news & regulation", prompt="...", subagent_type="general-purpose") task(description="Industry & market trends", prompt="...", subagent_type="general-purpose") # All 3 run in parallel β†’ synthesize results ``` **Usage Example 2 - Multiple Batches (>{n} sub-tasks):** ```python # User asks: "Compare AWS, Azure, GCP, Alibaba Cloud, and Oracle Cloud" # Thinking: 5 sub-tasks β†’ need multiple batches (max {n} per batch) # Turn 1: Launch first batch of {n} task(description="AWS analysis", prompt="...", subagent_type="general-purpose") task(description="Azure analysis", prompt="...", subagent_type="general-purpose") task(description="GCP analysis", prompt="...", subagent_type="general-purpose") # Turn 2: Launch remaining batch (after first batch completes) task(description="Alibaba Cloud analysis", prompt="...", subagent_type="general-purpose") task(description="Oracle Cloud analysis", prompt="...", subagent_type="general-purpose") # Turn 3: Synthesize ALL results from both batches ``` **Counter-Example - Direct Execution (NO subagents):** ```python # User asks: "Run the tests" # Thinking: Cannot decompose into parallel sub-tasks # β†’ Execute directly bash("npm test") # Direct execution, not task() ``` **CRITICAL**: - **Max {n} `task` calls per turn** - the system enforces this, excess calls are discarded - Only use `task` when you can launch 2+ subagents in parallel - Single task = No value from subagents = Execute directly - For >{n} sub-tasks, use sequential batches of {n} across multiple turns </subagent_system>""" SYSTEM_PROMPT_TEMPLATE = """ <role> You are {agent_name}, an open-source super agent. </role> {soul} {memory_context} <thinking_style> - Think concisely and strategically about the user's request BEFORE taking action - Break down the task: What is clear? What is ambiguous? What is missing? - **PRIORITY CHECK: If anything is unclear, missing, or has multiple interpretations, you MUST ask for clarification FIRST - do NOT proceed with work** {subagent_thinking}- Never write down your full final answer or report in thinking process, but only outline - CRITICAL: After thinking, you MUST provide your actual response to the user. Thinking is for planning, the response is for delivery. - Your response must contain the actual answer, not just a reference to what you thought about </thinking_style> <clarification_system> **WORKFLOW PRIORITY: CLARIFY β†’ PLAN β†’ ACT** 1. **FIRST**: Analyze the request in your thinking - identify what's unclear, missing, or ambiguous 2. **SECOND**: If clarification is needed, call `ask_clarification` tool IMMEDIATELY - do NOT start working 3. **THIRD**: Only after all clarifications are resolved, proceed with planning and execution **CRITICAL RULE: Clarification ALWAYS comes BEFORE action. Never start working and clarify mid-execution.** **MANDATORY Clarification Scenarios - You MUST call ask_clarification BEFORE starting work when:** 1. **Missing Information** (`missing_info`): Required details not provided - Example: User says "create a web scraper" but doesn't specify the target website - Example: "Deploy the app" without specifying environment - **REQUIRED ACTION**: Call ask_clarification to get the missing information 2. **Ambiguous Requirements** (`ambiguous_requirement`): Multiple valid interpretations exist - Example: "Optimize the code" could mean performance, readability, or memory usage - Example: "Make it better" is unclear what aspect to improve - **REQUIRED ACTION**: Call ask_clarification to clarify the exact requirement 3. **Approach Choices** (`approach_choice`): Several valid approaches exist - Example: "Add authentication" could use JWT, OAuth, session-based, or API keys - Example: "Store data" could use database, files, cache, etc. - **REQUIRED ACTION**: Call ask_clarification to let user choose the approach 4. **Risky Operations** (`risk_confirmation`): Destructive actions need confirmation - Example: Deleting files, modifying production configs, database operations - Example: Overwriting existing code or data - **REQUIRED ACTION**: Call ask_clarification to get explicit confirmation 5. **Suggestions** (`suggestion`): You have a recommendation but want approval - Example: "I recommend refactoring this code. Should I proceed?" - **REQUIRED ACTION**: Call ask_clarification to get approval **STRICT ENFORCEMENT:** - ❌ DO NOT start working and then ask for clarification mid-execution - clarify FIRST - ❌ DO NOT skip clarification for "efficiency" - accuracy matters more than speed - ❌ DO NOT make assumptions when information is missing - ALWAYS ask - ❌ DO NOT proceed with guesses - STOP and call ask_clarification first - βœ… Analyze the request in thinking β†’ Identify unclear aspects β†’ Ask BEFORE any action - βœ… If you identify the need for clarification in your thinking, you MUST call the tool IMMEDIATELY - βœ… After calling ask_clarification, execution will be interrupted automatically - βœ… Wait for user response - do NOT continue with assumptions **How to Use:** ```python ask_clarification( question="Your specific question here?", clarification_type="missing_info", # or other type context="Why you need this information", # optional but recommended options=["option1", "option2"] # optional, for choices ) ``` **Example:** User: "Deploy the application" You (thinking): Missing environment info - I MUST ask for clarification You (action): ask_clarification( question="Which environment should I deploy to?", clarification_type="approach_choice", context="I need to know the target environment for proper configuration", options=["development", "staging", "production"] ) [Execution stops - wait for user response] User: "staging" You: "Deploying to staging..." [proceed] </clarification_system> {skills_section} {subagent_section} <working_directory existed="true"> - User uploads: `/mnt/user-data/uploads` - Files uploaded by the user (automatically listed in context) - User workspace: `/mnt/user-data/workspace` - Working directory for temporary files - Output files: `/mnt/user-data/outputs` - Final deliverables must be saved here **File Management:** - Uploaded files are automatically listed in the <uploaded_files> section before each request - Use `read_file` tool to read uploaded files using their paths from the list - For PDF, PPT, Excel, and Word files, converted Markdown versions (*.md) are available alongside originals - All temporary work happens in `/mnt/user-data/workspace` - Final deliverables must be copied to `/mnt/user-data/outputs` and presented using `present_file` tool </working_directory> <response_style> - Clear and Concise: Avoid over-formatting unless requested - Natural Tone: Use paragraphs and prose, not bullet points by default - Action-Oriented: Focus on delivering results, not explaining processes </response_style> <citations> - When to Use: After web_search, include citations if applicable - Format: Use Markdown link format `[citation:TITLE](URL)` - Example: ```markdown The key AI trends for 2026 include enhanced reasoning capabilities and multimodal integration [citation:AI Trends 2026](https://techcrunch.com/ai-trends). Recent breakthroughs in language models have also accelerated progress [citation:OpenAI Research](https://openai.com/research). ``` </citations> <critical_reminders> - **Clarification First**: ALWAYS clarify unclear/missing/ambiguous requirements BEFORE starting work - never assume or guess {subagent_reminder}- Skill First: Always load the relevant skill before starting **complex** tasks. - Progressive Loading: Load resources incrementally as referenced in skills - Output Files: Final deliverables must be in `/mnt/user-data/outputs` - Clarity: Be direct and helpful, avoid unnecessary meta-commentary - Including Images and Mermaid: Images and Mermaid diagrams are always welcomed in the Markdown format, and you're encouraged to use `![Image Description](image_path)\n\n` or "```mermaid" to display images in response or Markdown files - Multi-task: Better utilize parallel tool calling to call multiple tools at one time for better performance - Language Consistency: Keep using the same language as user's - Always Respond: Your thinking is internal. You MUST always provide a visible response to the user after thinking. </critical_reminders> """ def _get_memory_context(agent_name: str | None = None) -> str: """Get memory context for injection into system prompt. Args: agent_name: If provided, loads per-agent memory. If None, loads global memory. Returns: Formatted memory context string wrapped in XML tags, or empty string if disabled. """ try: from src.agents.memory import format_memory_for_injection, get_memory_data from src.config.memory_config import get_memory_config config = get_memory_config() if not config.enabled or not config.injection_enabled: return "" memory_data = get_memory_data(agent_name) memory_content = format_memory_for_injection(memory_data, max_tokens=config.max_injection_tokens) if not memory_content.strip(): return "" return f"""<memory> {memory_content} </memory> """ except Exception as e: print(f"Failed to load memory context: {e}") return "" def get_skills_prompt_section(available_skills: set[str] | None = None) -> str: """Generate the skills prompt section with available skills list. Returns the <skill_system>...</skill_system> block listing all enabled skills, suitable for injection into any agent's system prompt. """ skills = load_skills(enabled_only=True) try: from src.config import get_app_config config = get_app_config() container_base_path = config.skills.container_path except Exception: container_base_path = "/mnt/skills" if not skills: return "" if available_skills is not None: skills = [skill for skill in skills if skill.name in available_skills] skill_items = "\n".join( f" <skill>\n <name>{skill.name}</name>\n <description>{skill.description}</description>\n <location>{skill.get_container_file_path(container_base_path)}</location>\n </skill>" for skill in skills ) skills_list = f"<available_skills>\n{skill_items}\n</available_skills>" return f"""<skill_system> You have access to skills that provide optimized workflows for specific tasks. Each skill contains best practices, frameworks, and references to additional resources. **Progressive Loading Pattern:** 1. When a user query matches a skill's use case, immediately call `read_file` on the skill's main file using the path attribute provided in the skill tag below 2. Read and understand the skill's workflow and instructions 3. The skill file contains references to external resources under the same folder 4. Load referenced resources only when needed during execution 5. Follow the skill's instructions precisely **Skills are located at:** {container_base_path} {skills_list} </skill_system>""" def get_agent_soul(agent_name: str | None) -> str: # Append SOUL.md (agent personality) if present soul = load_agent_soul(agent_name) if soul: return f"<soul>\n{soul}\n</soul>\n" if soul else "" return "" def apply_prompt_template(subagent_enabled: bool = False, max_concurrent_subagents: int = 3, *, agent_name: str | None = None, available_skills: set[str] | None = None) -> str: # Get memory context memory_context = _get_memory_context(agent_name) # Include subagent section only if enabled (from runtime parameter) n = max_concurrent_subagents subagent_section = _build_subagent_section(n) if subagent_enabled else "" # Add subagent reminder to critical_reminders if enabled subagent_reminder = ( "- **Orchestrator Mode**: You are a task orchestrator - decompose complex tasks into parallel sub-tasks. " f"**HARD LIMIT: max {n} `task` calls per response.** " f"If >{n} sub-tasks, split into sequential batches of ≀{n}. Synthesize after ALL batches complete.\n" if subagent_enabled else "" ) # Add subagent thinking guidance if enabled subagent_thinking = ( "- **DECOMPOSITION CHECK: Can this task be broken into 2+ parallel sub-tasks? If YES, COUNT them. " f"If count > {n}, you MUST plan batches of ≀{n} and only launch the FIRST batch now. " f"NEVER launch more than {n} `task` calls in one response.**\n" if subagent_enabled else "" ) # Get skills section skills_section = get_skills_prompt_section(available_skills) # Format the prompt with dynamic skills and memory prompt = SYSTEM_PROMPT_TEMPLATE.format( agent_name=agent_name or "DeerFlow 2.0", soul=get_agent_soul(agent_name), skills_section=skills_section, memory_context=memory_context, subagent_section=subagent_section, subagent_reminder=subagent_reminder, subagent_thinking=subagent_thinking, ) return prompt + f"\n<current_date>{datetime.now().strftime('%Y-%m-%d, %A')}</current_date>"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/lead_agent/prompt.py", "license": "MIT License", "lines": 322, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/agents/memory/prompt.py
"""Prompt templates for memory update and injection.""" import re from typing import Any try: import tiktoken TIKTOKEN_AVAILABLE = True except ImportError: TIKTOKEN_AVAILABLE = False # Prompt template for updating memory based on conversation MEMORY_UPDATE_PROMPT = """You are a memory management system. Your task is to analyze a conversation and update the user's memory profile. Current Memory State: <current_memory> {current_memory} </current_memory> New Conversation to Process: <conversation> {conversation} </conversation> Instructions: 1. Analyze the conversation for important information about the user 2. Extract relevant facts, preferences, and context with specific details (numbers, names, technologies) 3. Update the memory sections as needed following the detailed length guidelines below Memory Section Guidelines: **User Context** (Current state - concise summaries): - workContext: Professional role, company, key projects, main technologies (2-3 sentences) Example: Core contributor, project names with metrics (16k+ stars), technical stack - personalContext: Languages, communication preferences, key interests (1-2 sentences) Example: Bilingual capabilities, specific interest areas, expertise domains - topOfMind: Multiple ongoing focus areas and priorities (3-5 sentences, detailed paragraph) Example: Primary project work, parallel technical investigations, ongoing learning/tracking Include: Active implementation work, troubleshooting issues, market/research interests Note: This captures SEVERAL concurrent focus areas, not just one task **History** (Temporal context - rich paragraphs): - recentMonths: Detailed summary of recent activities (4-6 sentences or 1-2 paragraphs) Timeline: Last 1-3 months of interactions Include: Technologies explored, projects worked on, problems solved, interests demonstrated - earlierContext: Important historical patterns (3-5 sentences or 1 paragraph) Timeline: 3-12 months ago Include: Past projects, learning journeys, established patterns - longTermBackground: Persistent background and foundational context (2-4 sentences) Timeline: Overall/foundational information Include: Core expertise, longstanding interests, fundamental working style **Facts Extraction**: - Extract specific, quantifiable details (e.g., "16k+ GitHub stars", "200+ datasets") - Include proper nouns (company names, project names, technology names) - Preserve technical terminology and version numbers - Categories: * preference: Tools, styles, approaches user prefers/dislikes * knowledge: Specific expertise, technologies mastered, domain knowledge * context: Background facts (job title, projects, locations, languages) * behavior: Working patterns, communication habits, problem-solving approaches * goal: Stated objectives, learning targets, project ambitions - Confidence levels: * 0.9-1.0: Explicitly stated facts ("I work on X", "My role is Y") * 0.7-0.8: Strongly implied from actions/discussions * 0.5-0.6: Inferred patterns (use sparingly, only for clear patterns) **What Goes Where**: - workContext: Current job, active projects, primary tech stack - personalContext: Languages, personality, interests outside direct work tasks - topOfMind: Multiple ongoing priorities and focus areas user cares about recently (gets updated most frequently) Should capture 3-5 concurrent themes: main work, side explorations, learning/tracking interests - recentMonths: Detailed account of recent technical explorations and work - earlierContext: Patterns from slightly older interactions still relevant - longTermBackground: Unchanging foundational facts about the user **Multilingual Content**: - Preserve original language for proper nouns and company names - Keep technical terms in their original form (DeepSeek, LangGraph, etc.) - Note language capabilities in personalContext Output Format (JSON): {{ "user": {{ "workContext": {{ "summary": "...", "shouldUpdate": true/false }}, "personalContext": {{ "summary": "...", "shouldUpdate": true/false }}, "topOfMind": {{ "summary": "...", "shouldUpdate": true/false }} }}, "history": {{ "recentMonths": {{ "summary": "...", "shouldUpdate": true/false }}, "earlierContext": {{ "summary": "...", "shouldUpdate": true/false }}, "longTermBackground": {{ "summary": "...", "shouldUpdate": true/false }} }}, "newFacts": [ {{ "content": "...", "category": "preference|knowledge|context|behavior|goal", "confidence": 0.0-1.0 }} ], "factsToRemove": ["fact_id_1", "fact_id_2"] }} Important Rules: - Only set shouldUpdate=true if there's meaningful new information - Follow length guidelines: workContext/personalContext are concise (1-3 sentences), topOfMind and history sections are detailed (paragraphs) - Include specific metrics, version numbers, and proper nouns in facts - Only add facts that are clearly stated (0.9+) or strongly implied (0.7+) - Remove facts that are contradicted by new information - When updating topOfMind, integrate new focus areas while removing completed/abandoned ones Keep 3-5 concurrent focus themes that are still active and relevant - For history sections, integrate new information chronologically into appropriate time period - Preserve technical accuracy - keep exact names of technologies, companies, projects - Focus on information useful for future interactions and personalization - IMPORTANT: Do NOT record file upload events in memory. Uploaded files are session-specific and ephemeral β€” they will not be accessible in future sessions. Recording upload events causes confusion in subsequent conversations. Return ONLY valid JSON, no explanation or markdown.""" # Prompt template for extracting facts from a single message FACT_EXTRACTION_PROMPT = """Extract factual information about the user from this message. Message: {message} Extract facts in this JSON format: {{ "facts": [ {{ "content": "...", "category": "preference|knowledge|context|behavior|goal", "confidence": 0.0-1.0 }} ] }} Categories: - preference: User preferences (likes/dislikes, styles, tools) - knowledge: User's expertise or knowledge areas - context: Background context (location, job, projects) - behavior: Behavioral patterns - goal: User's goals or objectives Rules: - Only extract clear, specific facts - Confidence should reflect certainty (explicit statement = 0.9+, implied = 0.6-0.8) - Skip vague or temporary information Return ONLY valid JSON.""" def _count_tokens(text: str, encoding_name: str = "cl100k_base") -> int: """Count tokens in text using tiktoken. Args: text: The text to count tokens for. encoding_name: The encoding to use (default: cl100k_base for GPT-4/3.5). Returns: The number of tokens in the text. """ if not TIKTOKEN_AVAILABLE: # Fallback to character-based estimation if tiktoken is not available return len(text) // 4 try: encoding = tiktoken.get_encoding(encoding_name) return len(encoding.encode(text)) except Exception: # Fallback to character-based estimation on error return len(text) // 4 def format_memory_for_injection(memory_data: dict[str, Any], max_tokens: int = 2000) -> str: """Format memory data for injection into system prompt. Args: memory_data: The memory data dictionary. max_tokens: Maximum tokens to use (counted via tiktoken for accuracy). Returns: Formatted memory string for system prompt injection. """ if not memory_data: return "" sections = [] # Format user context user_data = memory_data.get("user", {}) if user_data: user_sections = [] work_ctx = user_data.get("workContext", {}) if work_ctx.get("summary"): user_sections.append(f"Work: {work_ctx['summary']}") personal_ctx = user_data.get("personalContext", {}) if personal_ctx.get("summary"): user_sections.append(f"Personal: {personal_ctx['summary']}") top_of_mind = user_data.get("topOfMind", {}) if top_of_mind.get("summary"): user_sections.append(f"Current Focus: {top_of_mind['summary']}") if user_sections: sections.append("User Context:\n" + "\n".join(f"- {s}" for s in user_sections)) # Format history history_data = memory_data.get("history", {}) if history_data: history_sections = [] recent = history_data.get("recentMonths", {}) if recent.get("summary"): history_sections.append(f"Recent: {recent['summary']}") earlier = history_data.get("earlierContext", {}) if earlier.get("summary"): history_sections.append(f"Earlier: {earlier['summary']}") if history_sections: sections.append("History:\n" + "\n".join(f"- {s}" for s in history_sections)) if not sections: return "" result = "\n\n".join(sections) # Use accurate token counting with tiktoken token_count = _count_tokens(result) if token_count > max_tokens: # Truncate to fit within token limit # Estimate characters to remove based on token ratio char_per_token = len(result) / token_count target_chars = int(max_tokens * char_per_token * 0.95) # 95% to leave margin result = result[:target_chars] + "\n..." return result def format_conversation_for_update(messages: list[Any]) -> str: """Format conversation messages for memory update prompt. Args: messages: List of conversation messages. Returns: Formatted conversation string. """ lines = [] for msg in messages: role = getattr(msg, "type", "unknown") content = getattr(msg, "content", str(msg)) # Handle content that might be a list (multimodal) if isinstance(content, list): text_parts = [p.get("text", "") for p in content if isinstance(p, dict) and "text" in p] content = " ".join(text_parts) if text_parts else str(content) # Strip uploaded_files tags from human messages to avoid persisting # ephemeral file path info into long-term memory. Skip the turn entirely # when nothing remains after stripping (upload-only message). if role == "human": content = re.sub( r"<uploaded_files>[\s\S]*?</uploaded_files>\n*", "", str(content) ).strip() if not content: continue # Truncate very long messages if len(str(content)) > 1000: content = str(content)[:1000] + "..." if role == "human": lines.append(f"User: {content}") elif role == "ai": lines.append(f"Assistant: {content}") return "\n\n".join(lines)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/memory/prompt.py", "license": "MIT License", "lines": 220, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/agents/memory/queue.py
"""Memory update queue with debounce mechanism.""" import threading import time from dataclasses import dataclass, field from datetime import datetime from typing import Any from src.config.memory_config import get_memory_config @dataclass class ConversationContext: """Context for a conversation to be processed for memory update.""" thread_id: str messages: list[Any] timestamp: datetime = field(default_factory=datetime.utcnow) agent_name: str | None = None class MemoryUpdateQueue: """Queue for memory updates with debounce mechanism. This queue collects conversation contexts and processes them after a configurable debounce period. Multiple conversations received within the debounce window are batched together. """ def __init__(self): """Initialize the memory update queue.""" self._queue: list[ConversationContext] = [] self._lock = threading.Lock() self._timer: threading.Timer | None = None self._processing = False def add(self, thread_id: str, messages: list[Any], agent_name: str | None = None) -> None: """Add a conversation to the update queue. Args: thread_id: The thread ID. messages: The conversation messages. agent_name: If provided, memory is stored per-agent. If None, uses global memory. """ config = get_memory_config() if not config.enabled: return context = ConversationContext( thread_id=thread_id, messages=messages, agent_name=agent_name, ) with self._lock: # Check if this thread already has a pending update # If so, replace it with the newer one self._queue = [c for c in self._queue if c.thread_id != thread_id] self._queue.append(context) # Reset or start the debounce timer self._reset_timer() print(f"Memory update queued for thread {thread_id}, queue size: {len(self._queue)}") def _reset_timer(self) -> None: """Reset the debounce timer.""" config = get_memory_config() # Cancel existing timer if any if self._timer is not None: self._timer.cancel() # Start new timer self._timer = threading.Timer( config.debounce_seconds, self._process_queue, ) self._timer.daemon = True self._timer.start() print(f"Memory update timer set for {config.debounce_seconds}s") def _process_queue(self) -> None: """Process all queued conversation contexts.""" # Import here to avoid circular dependency from src.agents.memory.updater import MemoryUpdater with self._lock: if self._processing: # Already processing, reschedule self._reset_timer() return if not self._queue: return self._processing = True contexts_to_process = self._queue.copy() self._queue.clear() self._timer = None print(f"Processing {len(contexts_to_process)} queued memory updates") try: updater = MemoryUpdater() for context in contexts_to_process: try: print(f"Updating memory for thread {context.thread_id}") success = updater.update_memory( messages=context.messages, thread_id=context.thread_id, agent_name=context.agent_name, ) if success: print(f"Memory updated successfully for thread {context.thread_id}") else: print(f"Memory update skipped/failed for thread {context.thread_id}") except Exception as e: print(f"Error updating memory for thread {context.thread_id}: {e}") # Small delay between updates to avoid rate limiting if len(contexts_to_process) > 1: time.sleep(0.5) finally: with self._lock: self._processing = False def flush(self) -> None: """Force immediate processing of the queue. This is useful for testing or graceful shutdown. """ with self._lock: if self._timer is not None: self._timer.cancel() self._timer = None self._process_queue() def clear(self) -> None: """Clear the queue without processing. This is useful for testing. """ with self._lock: if self._timer is not None: self._timer.cancel() self._timer = None self._queue.clear() self._processing = False @property def pending_count(self) -> int: """Get the number of pending updates.""" with self._lock: return len(self._queue) @property def is_processing(self) -> bool: """Check if the queue is currently being processed.""" with self._lock: return self._processing # Global singleton instance _memory_queue: MemoryUpdateQueue | None = None _queue_lock = threading.Lock() def get_memory_queue() -> MemoryUpdateQueue: """Get the global memory update queue singleton. Returns: The memory update queue instance. """ global _memory_queue with _queue_lock: if _memory_queue is None: _memory_queue = MemoryUpdateQueue() return _memory_queue def reset_memory_queue() -> None: """Reset the global memory queue. This is useful for testing. """ global _memory_queue with _queue_lock: if _memory_queue is not None: _memory_queue.clear() _memory_queue = None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/memory/queue.py", "license": "MIT License", "lines": 152, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/memory/updater.py
"""Memory updater for reading, writing, and updating memory data.""" import json import re import uuid from datetime import datetime from pathlib import Path from typing import Any from src.agents.memory.prompt import ( MEMORY_UPDATE_PROMPT, format_conversation_for_update, ) from src.config.memory_config import get_memory_config from src.config.paths import get_paths from src.models import create_chat_model def _get_memory_file_path(agent_name: str | None = None) -> Path: """Get the path to the memory file. Args: agent_name: If provided, returns the per-agent memory file path. If None, returns the global memory file path. Returns: Path to the memory file. """ if agent_name is not None: return get_paths().agent_memory_file(agent_name) config = get_memory_config() if config.storage_path: p = Path(config.storage_path) # Absolute path: use as-is; relative path: resolve against base_dir return p if p.is_absolute() else get_paths().base_dir / p return get_paths().memory_file def _create_empty_memory() -> dict[str, Any]: """Create an empty memory structure.""" return { "version": "1.0", "lastUpdated": datetime.utcnow().isoformat() + "Z", "user": { "workContext": {"summary": "", "updatedAt": ""}, "personalContext": {"summary": "", "updatedAt": ""}, "topOfMind": {"summary": "", "updatedAt": ""}, }, "history": { "recentMonths": {"summary": "", "updatedAt": ""}, "earlierContext": {"summary": "", "updatedAt": ""}, "longTermBackground": {"summary": "", "updatedAt": ""}, }, "facts": [], } # Per-agent memory cache: keyed by agent_name (None = global) # Value: (memory_data, file_mtime) _memory_cache: dict[str | None, tuple[dict[str, Any], float | None]] = {} def get_memory_data(agent_name: str | None = None) -> dict[str, Any]: """Get the current memory data (cached with file modification time check). The cache is automatically invalidated if the memory file has been modified since the last load, ensuring fresh data is always returned. Args: agent_name: If provided, loads per-agent memory. If None, loads global memory. Returns: The memory data dictionary. """ file_path = _get_memory_file_path(agent_name) # Get current file modification time try: current_mtime = file_path.stat().st_mtime if file_path.exists() else None except OSError: current_mtime = None cached = _memory_cache.get(agent_name) # Invalidate cache if file has been modified or doesn't exist if cached is None or cached[1] != current_mtime: memory_data = _load_memory_from_file(agent_name) _memory_cache[agent_name] = (memory_data, current_mtime) return memory_data return cached[0] def reload_memory_data(agent_name: str | None = None) -> dict[str, Any]: """Reload memory data from file, forcing cache invalidation. Args: agent_name: If provided, reloads per-agent memory. If None, reloads global memory. Returns: The reloaded memory data dictionary. """ file_path = _get_memory_file_path(agent_name) memory_data = _load_memory_from_file(agent_name) try: mtime = file_path.stat().st_mtime if file_path.exists() else None except OSError: mtime = None _memory_cache[agent_name] = (memory_data, mtime) return memory_data def _load_memory_from_file(agent_name: str | None = None) -> dict[str, Any]: """Load memory data from file. Args: agent_name: If provided, loads per-agent memory file. If None, loads global. Returns: The memory data dictionary. """ file_path = _get_memory_file_path(agent_name) if not file_path.exists(): return _create_empty_memory() try: with open(file_path, encoding="utf-8") as f: data = json.load(f) return data except (json.JSONDecodeError, OSError) as e: print(f"Failed to load memory file: {e}") return _create_empty_memory() # Matches sentences that describe a file-upload *event* rather than general # file-related work. Deliberately narrow to avoid removing legitimate facts # such as "User works with CSV files" or "prefers PDF export". _UPLOAD_SENTENCE_RE = re.compile( r"[^.!?]*\b(?:" r"upload(?:ed|ing)?(?:\s+\w+){0,3}\s+(?:file|files?|document|documents?|attachment|attachments?)" r"|file\s+upload" r"|/mnt/user-data/uploads/" r"|<uploaded_files>" r")[^.!?]*[.!?]?\s*", re.IGNORECASE, ) def _strip_upload_mentions_from_memory(memory_data: dict[str, Any]) -> dict[str, Any]: """Remove sentences about file uploads from all memory summaries and facts. Uploaded files are session-scoped; persisting upload events in long-term memory causes the agent to search for non-existent files in future sessions. """ # Scrub summaries in user/history sections for section in ("user", "history"): section_data = memory_data.get(section, {}) for _key, val in section_data.items(): if isinstance(val, dict) and "summary" in val: cleaned = _UPLOAD_SENTENCE_RE.sub("", val["summary"]).strip() cleaned = re.sub(r" +", " ", cleaned) val["summary"] = cleaned # Also remove any facts that describe upload events facts = memory_data.get("facts", []) if facts: memory_data["facts"] = [ f for f in facts if not _UPLOAD_SENTENCE_RE.search(f.get("content", "")) ] return memory_data def _save_memory_to_file(memory_data: dict[str, Any], agent_name: str | None = None) -> bool: """Save memory data to file and update cache. Args: memory_data: The memory data to save. agent_name: If provided, saves to per-agent memory file. If None, saves to global. Returns: True if successful, False otherwise. """ file_path = _get_memory_file_path(agent_name) try: # Ensure directory exists file_path.parent.mkdir(parents=True, exist_ok=True) # Update lastUpdated timestamp memory_data["lastUpdated"] = datetime.utcnow().isoformat() + "Z" # Write atomically using temp file temp_path = file_path.with_suffix(".tmp") with open(temp_path, "w", encoding="utf-8") as f: json.dump(memory_data, f, indent=2, ensure_ascii=False) # Rename temp file to actual file (atomic on most systems) temp_path.replace(file_path) # Update cache and file modification time try: mtime = file_path.stat().st_mtime except OSError: mtime = None _memory_cache[agent_name] = (memory_data, mtime) print(f"Memory saved to {file_path}") return True except OSError as e: print(f"Failed to save memory file: {e}") return False class MemoryUpdater: """Updates memory using LLM based on conversation context.""" def __init__(self, model_name: str | None = None): """Initialize the memory updater. Args: model_name: Optional model name to use. If None, uses config or default. """ self._model_name = model_name def _get_model(self): """Get the model for memory updates.""" config = get_memory_config() model_name = self._model_name or config.model_name return create_chat_model(name=model_name, thinking_enabled=False) def update_memory(self, messages: list[Any], thread_id: str | None = None, agent_name: str | None = None) -> bool: """Update memory based on conversation messages. Args: messages: List of conversation messages. thread_id: Optional thread ID for tracking source. agent_name: If provided, updates per-agent memory. If None, updates global memory. Returns: True if update was successful, False otherwise. """ config = get_memory_config() if not config.enabled: return False if not messages: return False try: # Get current memory current_memory = get_memory_data(agent_name) # Format conversation for prompt conversation_text = format_conversation_for_update(messages) if not conversation_text.strip(): return False # Build prompt prompt = MEMORY_UPDATE_PROMPT.format( current_memory=json.dumps(current_memory, indent=2), conversation=conversation_text, ) # Call LLM model = self._get_model() response = model.invoke(prompt) response_text = str(response.content).strip() # Parse response # Remove markdown code blocks if present if response_text.startswith("```"): lines = response_text.split("\n") response_text = "\n".join(lines[1:-1] if lines[-1] == "```" else lines[1:]) update_data = json.loads(response_text) # Apply updates updated_memory = self._apply_updates(current_memory, update_data, thread_id) # Strip file-upload mentions from all summaries before saving. # Uploaded files are session-scoped and won't exist in future sessions, # so recording upload events in long-term memory causes the agent to # try (and fail) to locate those files in subsequent conversations. updated_memory = _strip_upload_mentions_from_memory(updated_memory) # Save return _save_memory_to_file(updated_memory, agent_name) except json.JSONDecodeError as e: print(f"Failed to parse LLM response for memory update: {e}") return False except Exception as e: print(f"Memory update failed: {e}") return False def _apply_updates( self, current_memory: dict[str, Any], update_data: dict[str, Any], thread_id: str | None = None, ) -> dict[str, Any]: """Apply LLM-generated updates to memory. Args: current_memory: Current memory data. update_data: Updates from LLM. thread_id: Optional thread ID for tracking. Returns: Updated memory data. """ config = get_memory_config() now = datetime.utcnow().isoformat() + "Z" # Update user sections user_updates = update_data.get("user", {}) for section in ["workContext", "personalContext", "topOfMind"]: section_data = user_updates.get(section, {}) if section_data.get("shouldUpdate") and section_data.get("summary"): current_memory["user"][section] = { "summary": section_data["summary"], "updatedAt": now, } # Update history sections history_updates = update_data.get("history", {}) for section in ["recentMonths", "earlierContext", "longTermBackground"]: section_data = history_updates.get(section, {}) if section_data.get("shouldUpdate") and section_data.get("summary"): current_memory["history"][section] = { "summary": section_data["summary"], "updatedAt": now, } # Remove facts facts_to_remove = set(update_data.get("factsToRemove", [])) if facts_to_remove: current_memory["facts"] = [f for f in current_memory.get("facts", []) if f.get("id") not in facts_to_remove] # Add new facts new_facts = update_data.get("newFacts", []) for fact in new_facts: confidence = fact.get("confidence", 0.5) if confidence >= config.fact_confidence_threshold: fact_entry = { "id": f"fact_{uuid.uuid4().hex[:8]}", "content": fact.get("content", ""), "category": fact.get("category", "context"), "confidence": confidence, "createdAt": now, "source": thread_id or "unknown", } current_memory["facts"].append(fact_entry) # Enforce max facts limit if len(current_memory["facts"]) > config.max_facts: # Sort by confidence and keep top ones current_memory["facts"] = sorted( current_memory["facts"], key=lambda f: f.get("confidence", 0), reverse=True, )[: config.max_facts] return current_memory def update_memory_from_conversation(messages: list[Any], thread_id: str | None = None, agent_name: str | None = None) -> bool: """Convenience function to update memory from a conversation. Args: messages: List of conversation messages. thread_id: Optional thread ID. agent_name: If provided, updates per-agent memory. If None, updates global memory. Returns: True if successful, False otherwise. """ updater = MemoryUpdater() return updater.update_memory(messages, thread_id, agent_name)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/memory/updater.py", "license": "MIT License", "lines": 305, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/middlewares/clarification_middleware.py
"""Middleware for intercepting clarification requests and presenting them to the user.""" from collections.abc import Callable from typing import override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langchain_core.messages import ToolMessage from langgraph.graph import END from langgraph.prebuilt.tool_node import ToolCallRequest from langgraph.types import Command class ClarificationMiddlewareState(AgentState): """Compatible with the `ThreadState` schema.""" pass class ClarificationMiddleware(AgentMiddleware[ClarificationMiddlewareState]): """Intercepts clarification tool calls and interrupts execution to present questions to the user. When the model calls the `ask_clarification` tool, this middleware: 1. Intercepts the tool call before execution 2. Extracts the clarification question and metadata 3. Formats a user-friendly message 4. Returns a Command that interrupts execution and presents the question 5. Waits for user response before continuing This replaces the tool-based approach where clarification continued the conversation flow. """ state_schema = ClarificationMiddlewareState def _is_chinese(self, text: str) -> bool: """Check if text contains Chinese characters. Args: text: Text to check Returns: True if text contains Chinese characters """ return any("\u4e00" <= char <= "\u9fff" for char in text) def _format_clarification_message(self, args: dict) -> str: """Format the clarification arguments into a user-friendly message. Args: args: The tool call arguments containing clarification details Returns: Formatted message string """ question = args.get("question", "") clarification_type = args.get("clarification_type", "missing_info") context = args.get("context") options = args.get("options", []) # Type-specific icons type_icons = { "missing_info": "❓", "ambiguous_requirement": "πŸ€”", "approach_choice": "πŸ”€", "risk_confirmation": "⚠️", "suggestion": "πŸ’‘", } icon = type_icons.get(clarification_type, "❓") # Build the message naturally message_parts = [] # Add icon and question together for a more natural flow if context: # If there's context, present it first as background message_parts.append(f"{icon} {context}") message_parts.append(f"\n{question}") else: # Just the question with icon message_parts.append(f"{icon} {question}") # Add options in a cleaner format if options and len(options) > 0: message_parts.append("") # blank line for spacing for i, option in enumerate(options, 1): message_parts.append(f" {i}. {option}") return "\n".join(message_parts) def _handle_clarification(self, request: ToolCallRequest) -> Command: """Handle clarification request and return command to interrupt execution. Args: request: Tool call request Returns: Command that interrupts execution with the formatted clarification message """ # Extract clarification arguments args = request.tool_call.get("args", {}) question = args.get("question", "") print("[ClarificationMiddleware] Intercepted clarification request") print(f"[ClarificationMiddleware] Question: {question}") # Format the clarification message formatted_message = self._format_clarification_message(args) # Get the tool call ID tool_call_id = request.tool_call.get("id", "") # Create a ToolMessage with the formatted question # This will be added to the message history tool_message = ToolMessage( content=formatted_message, tool_call_id=tool_call_id, name="ask_clarification", ) # Return a Command that: # 1. Adds the formatted tool message # 2. Interrupts execution by going to __end__ # Note: We don't add an extra AIMessage here - the frontend will detect # and display ask_clarification tool messages directly return Command( update={"messages": [tool_message]}, goto=END, ) @override def wrap_tool_call( self, request: ToolCallRequest, handler: Callable[[ToolCallRequest], ToolMessage | Command], ) -> ToolMessage | Command: """Intercept ask_clarification tool calls and interrupt execution (sync version). Args: request: Tool call request handler: Original tool execution handler Returns: Command that interrupts execution with the formatted clarification message """ # Check if this is an ask_clarification tool call if request.tool_call.get("name") != "ask_clarification": # Not a clarification call, execute normally return handler(request) return self._handle_clarification(request) @override async def awrap_tool_call( self, request: ToolCallRequest, handler: Callable[[ToolCallRequest], ToolMessage | Command], ) -> ToolMessage | Command: """Intercept ask_clarification tool calls and interrupt execution (async version). Args: request: Tool call request handler: Original tool execution handler (async) Returns: Command that interrupts execution with the formatted clarification message """ # Check if this is an ask_clarification tool call if request.tool_call.get("name") != "ask_clarification": # Not a clarification call, execute normally return await handler(request) return self._handle_clarification(request)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/clarification_middleware.py", "license": "MIT License", "lines": 135, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/middlewares/dangling_tool_call_middleware.py
"""Middleware to fix dangling tool calls in message history. A dangling tool call occurs when an AIMessage contains tool_calls but there are no corresponding ToolMessages in the history (e.g., due to user interruption or request cancellation). This causes LLM errors due to incomplete message format. This middleware intercepts the model call to detect and patch such gaps by inserting synthetic ToolMessages with an error indicator immediately after the AIMessage that made the tool calls, ensuring correct message ordering. Note: Uses wrap_model_call instead of before_model to ensure patches are inserted at the correct positions (immediately after each dangling AIMessage), not appended to the end of the message list as before_model + add_messages reducer would do. """ import logging from collections.abc import Awaitable, Callable from typing import override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langchain.agents.middleware.types import ModelCallResult, ModelRequest, ModelResponse from langchain_core.messages import ToolMessage logger = logging.getLogger(__name__) class DanglingToolCallMiddleware(AgentMiddleware[AgentState]): """Inserts placeholder ToolMessages for dangling tool calls before model invocation. Scans the message history for AIMessages whose tool_calls lack corresponding ToolMessages, and injects synthetic error responses immediately after the offending AIMessage so the LLM receives a well-formed conversation. """ def _build_patched_messages(self, messages: list) -> list | None: """Return a new message list with patches inserted at the correct positions. For each AIMessage with dangling tool_calls (no corresponding ToolMessage), a synthetic ToolMessage is inserted immediately after that AIMessage. Returns None if no patches are needed. """ # Collect IDs of all existing ToolMessages existing_tool_msg_ids: set[str] = set() for msg in messages: if isinstance(msg, ToolMessage): existing_tool_msg_ids.add(msg.tool_call_id) # Check if any patching is needed needs_patch = False for msg in messages: if getattr(msg, "type", None) != "ai": continue for tc in getattr(msg, "tool_calls", None) or []: tc_id = tc.get("id") if tc_id and tc_id not in existing_tool_msg_ids: needs_patch = True break if needs_patch: break if not needs_patch: return None # Build new list with patches inserted right after each dangling AIMessage patched: list = [] patched_ids: set[str] = set() patch_count = 0 for msg in messages: patched.append(msg) if getattr(msg, "type", None) != "ai": continue for tc in getattr(msg, "tool_calls", None) or []: tc_id = tc.get("id") if tc_id and tc_id not in existing_tool_msg_ids and tc_id not in patched_ids: patched.append( ToolMessage( content="[Tool call was interrupted and did not return a result.]", tool_call_id=tc_id, name=tc.get("name", "unknown"), status="error", ) ) patched_ids.add(tc_id) patch_count += 1 logger.warning(f"Injecting {patch_count} placeholder ToolMessage(s) for dangling tool calls") return patched @override def wrap_model_call( self, request: ModelRequest, handler: Callable[[ModelRequest], ModelResponse], ) -> ModelCallResult: patched = self._build_patched_messages(request.messages) if patched is not None: request = request.override(messages=patched) return handler(request) @override async def awrap_model_call( self, request: ModelRequest, handler: Callable[[ModelRequest], Awaitable[ModelResponse]], ) -> ModelCallResult: patched = self._build_patched_messages(request.messages) if patched is not None: request = request.override(messages=patched) return await handler(request)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/dangling_tool_call_middleware.py", "license": "MIT License", "lines": 93, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/middlewares/memory_middleware.py
"""Middleware for memory mechanism.""" import re from typing import Any, override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langgraph.runtime import Runtime from src.agents.memory.queue import get_memory_queue from src.config.memory_config import get_memory_config class MemoryMiddlewareState(AgentState): """Compatible with the `ThreadState` schema.""" pass def _filter_messages_for_memory(messages: list[Any]) -> list[Any]: """Filter messages to keep only user inputs and final assistant responses. This filters out: - Tool messages (intermediate tool call results) - AI messages with tool_calls (intermediate steps, not final responses) - The <uploaded_files> block injected by UploadsMiddleware into human messages (file paths are session-scoped and must not persist in long-term memory). The user's actual question is preserved; only turns whose content is entirely the upload block (nothing remains after stripping) are dropped along with their paired assistant response. Only keeps: - Human messages (with the ephemeral upload block removed) - AI messages without tool_calls (final assistant responses), unless the paired human turn was upload-only and had no real user text. Args: messages: List of all conversation messages. Returns: Filtered list containing only user inputs and final assistant responses. """ _UPLOAD_BLOCK_RE = re.compile( r"<uploaded_files>[\s\S]*?</uploaded_files>\n*", re.IGNORECASE ) filtered = [] skip_next_ai = False for msg in messages: msg_type = getattr(msg, "type", None) if msg_type == "human": content = getattr(msg, "content", "") if isinstance(content, list): content = " ".join( p.get("text", "") for p in content if isinstance(p, dict) ) content_str = str(content) if "<uploaded_files>" in content_str: # Strip the ephemeral upload block; keep the user's real question. stripped = _UPLOAD_BLOCK_RE.sub("", content_str).strip() if not stripped: # Nothing left β€” the entire turn was upload bookkeeping; # skip it and the paired assistant response. skip_next_ai = True continue # Rebuild the message with cleaned content so the user's question # is still available for memory summarisation. from copy import copy clean_msg = copy(msg) clean_msg.content = stripped filtered.append(clean_msg) skip_next_ai = False else: filtered.append(msg) skip_next_ai = False elif msg_type == "ai": tool_calls = getattr(msg, "tool_calls", None) if not tool_calls: if skip_next_ai: skip_next_ai = False continue filtered.append(msg) # Skip tool messages and AI messages with tool_calls return filtered class MemoryMiddleware(AgentMiddleware[MemoryMiddlewareState]): """Middleware that queues conversation for memory update after agent execution. This middleware: 1. After each agent execution, queues the conversation for memory update 2. Only includes user inputs and final assistant responses (ignores tool calls) 3. The queue uses debouncing to batch multiple updates together 4. Memory is updated asynchronously via LLM summarization """ state_schema = MemoryMiddlewareState def __init__(self, agent_name: str | None = None): """Initialize the MemoryMiddleware. Args: agent_name: If provided, memory is stored per-agent. If None, uses global memory. """ super().__init__() self._agent_name = agent_name @override def after_agent(self, state: MemoryMiddlewareState, runtime: Runtime) -> dict | None: """Queue conversation for memory update after agent completes. Args: state: The current agent state. runtime: The runtime context. Returns: None (no state changes needed from this middleware). """ config = get_memory_config() if not config.enabled: return None # Get thread ID from runtime context thread_id = runtime.context.get("thread_id") if not thread_id: print("MemoryMiddleware: No thread_id in context, skipping memory update") return None # Get messages from state messages = state.get("messages", []) if not messages: print("MemoryMiddleware: No messages in state, skipping memory update") return None # Filter to only keep user inputs and final assistant responses filtered_messages = _filter_messages_for_memory(messages) # Only queue if there's meaningful conversation # At minimum need one user message and one assistant response user_messages = [m for m in filtered_messages if getattr(m, "type", None) == "human"] assistant_messages = [m for m in filtered_messages if getattr(m, "type", None) == "ai"] if not user_messages or not assistant_messages: return None # Queue the filtered conversation for memory update queue = get_memory_queue() queue.add(thread_id=thread_id, messages=filtered_messages, agent_name=self._agent_name) return None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/memory_middleware.py", "license": "MIT License", "lines": 121, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/middlewares/subagent_limit_middleware.py
"""Middleware to enforce maximum concurrent subagent tool calls per model response.""" import logging from typing import override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langgraph.runtime import Runtime from src.subagents.executor import MAX_CONCURRENT_SUBAGENTS logger = logging.getLogger(__name__) # Valid range for max_concurrent_subagents MIN_SUBAGENT_LIMIT = 2 MAX_SUBAGENT_LIMIT = 4 def _clamp_subagent_limit(value: int) -> int: """Clamp subagent limit to valid range [2, 4].""" return max(MIN_SUBAGENT_LIMIT, min(MAX_SUBAGENT_LIMIT, value)) class SubagentLimitMiddleware(AgentMiddleware[AgentState]): """Truncates excess 'task' tool calls from a single model response. When an LLM generates more than max_concurrent parallel task tool calls in one response, this middleware keeps only the first max_concurrent and discards the rest. This is more reliable than prompt-based limits. Args: max_concurrent: Maximum number of concurrent subagent calls allowed. Defaults to MAX_CONCURRENT_SUBAGENTS (3). Clamped to [2, 4]. """ def __init__(self, max_concurrent: int = MAX_CONCURRENT_SUBAGENTS): super().__init__() self.max_concurrent = _clamp_subagent_limit(max_concurrent) def _truncate_task_calls(self, state: AgentState) -> dict | None: messages = state.get("messages", []) if not messages: return None last_msg = messages[-1] if getattr(last_msg, "type", None) != "ai": return None tool_calls = getattr(last_msg, "tool_calls", None) if not tool_calls: return None # Count task tool calls task_indices = [i for i, tc in enumerate(tool_calls) if tc.get("name") == "task"] if len(task_indices) <= self.max_concurrent: return None # Build set of indices to drop (excess task calls beyond the limit) indices_to_drop = set(task_indices[self.max_concurrent :]) truncated_tool_calls = [tc for i, tc in enumerate(tool_calls) if i not in indices_to_drop] dropped_count = len(indices_to_drop) logger.warning(f"Truncated {dropped_count} excess task tool call(s) from model response (limit: {self.max_concurrent})") # Replace the AIMessage with truncated tool_calls (same id triggers replacement) updated_msg = last_msg.model_copy(update={"tool_calls": truncated_tool_calls}) return {"messages": [updated_msg]} @override def after_model(self, state: AgentState, runtime: Runtime) -> dict | None: return self._truncate_task_calls(state) @override async def aafter_model(self, state: AgentState, runtime: Runtime) -> dict | None: return self._truncate_task_calls(state)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/subagent_limit_middleware.py", "license": "MIT License", "lines": 54, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/agents/middlewares/thread_data_middleware.py
from typing import NotRequired, override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langgraph.runtime import Runtime from src.agents.thread_state import ThreadDataState from src.config.paths import Paths, get_paths class ThreadDataMiddlewareState(AgentState): """Compatible with the `ThreadState` schema.""" thread_data: NotRequired[ThreadDataState | None] class ThreadDataMiddleware(AgentMiddleware[ThreadDataMiddlewareState]): """Create thread data directories for each thread execution. Creates the following directory structure: - {base_dir}/threads/{thread_id}/user-data/workspace - {base_dir}/threads/{thread_id}/user-data/uploads - {base_dir}/threads/{thread_id}/user-data/outputs Lifecycle Management: - With lazy_init=True (default): Only compute paths, directories created on-demand - With lazy_init=False: Eagerly create directories in before_agent() """ state_schema = ThreadDataMiddlewareState def __init__(self, base_dir: str | None = None, lazy_init: bool = True): """Initialize the middleware. Args: base_dir: Base directory for thread data. Defaults to Paths resolution. lazy_init: If True, defer directory creation until needed. If False, create directories eagerly in before_agent(). Default is True for optimal performance. """ super().__init__() self._paths = Paths(base_dir) if base_dir else get_paths() self._lazy_init = lazy_init def _get_thread_paths(self, thread_id: str) -> dict[str, str]: """Get the paths for a thread's data directories. Args: thread_id: The thread ID. Returns: Dictionary with workspace_path, uploads_path, and outputs_path. """ return { "workspace_path": str(self._paths.sandbox_work_dir(thread_id)), "uploads_path": str(self._paths.sandbox_uploads_dir(thread_id)), "outputs_path": str(self._paths.sandbox_outputs_dir(thread_id)), } def _create_thread_directories(self, thread_id: str) -> dict[str, str]: """Create the thread data directories. Args: thread_id: The thread ID. Returns: Dictionary with the created directory paths. """ self._paths.ensure_thread_dirs(thread_id) return self._get_thread_paths(thread_id) @override def before_agent(self, state: ThreadDataMiddlewareState, runtime: Runtime) -> dict | None: thread_id = runtime.context.get("thread_id") if thread_id is None: raise ValueError("Thread ID is required in the context") if self._lazy_init: # Lazy initialization: only compute paths, don't create directories paths = self._get_thread_paths(thread_id) else: # Eager initialization: create directories immediately paths = self._create_thread_directories(thread_id) print(f"Created thread data directories for thread {thread_id}") return { "thread_data": { **paths, } }
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/thread_data_middleware.py", "license": "MIT License", "lines": 69, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/agents/middlewares/title_middleware.py
"""Middleware for automatic thread title generation.""" from typing import NotRequired, override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langgraph.runtime import Runtime from src.config.title_config import get_title_config from src.models import create_chat_model class TitleMiddlewareState(AgentState): """Compatible with the `ThreadState` schema.""" title: NotRequired[str | None] class TitleMiddleware(AgentMiddleware[TitleMiddlewareState]): """Automatically generate a title for the thread after the first user message.""" state_schema = TitleMiddlewareState def _should_generate_title(self, state: TitleMiddlewareState) -> bool: """Check if we should generate a title for this thread.""" config = get_title_config() if not config.enabled: return False # Check if thread already has a title in state if state.get("title"): return False # Check if this is the first turn (has at least one user message and one assistant response) messages = state.get("messages", []) if len(messages) < 2: return False # Count user and assistant messages user_messages = [m for m in messages if m.type == "human"] assistant_messages = [m for m in messages if m.type == "ai"] # Generate title after first complete exchange return len(user_messages) == 1 and len(assistant_messages) >= 1 def _generate_title(self, state: TitleMiddlewareState) -> str: """Generate a concise title based on the conversation.""" config = get_title_config() messages = state.get("messages", []) # Get first user message and first assistant response user_msg_content = next((m.content for m in messages if m.type == "human"), "") assistant_msg_content = next((m.content for m in messages if m.type == "ai"), "") # Ensure content is string (LangChain messages can have list content) user_msg = str(user_msg_content) if user_msg_content else "" assistant_msg = str(assistant_msg_content) if assistant_msg_content else "" # Use a lightweight model to generate title model = create_chat_model(thinking_enabled=False) prompt = config.prompt_template.format( max_words=config.max_words, user_msg=user_msg[:500], assistant_msg=assistant_msg[:500], ) try: response = model.invoke(prompt) # Ensure response content is string title_content = str(response.content) if response.content else "" title = title_content.strip().strip('"').strip("'") # Limit to max characters return title[: config.max_chars] if len(title) > config.max_chars else title except Exception as e: print(f"Failed to generate title: {e}") # Fallback: use first part of user message (by character count) fallback_chars = min(config.max_chars, 50) # Use max_chars or 50, whichever is smaller if len(user_msg) > fallback_chars: return user_msg[:fallback_chars].rstrip() + "..." return user_msg if user_msg else "New Conversation" @override def after_agent(self, state: TitleMiddlewareState, runtime: Runtime) -> dict | None: """Generate and set thread title after the first agent response.""" if self._should_generate_title(state): title = self._generate_title(state) print(f"Generated thread title: {title}") # Store title in state (will be persisted by checkpointer if configured) return {"title": title} return None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/title_middleware.py", "license": "MIT License", "lines": 70, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/middlewares/uploads_middleware.py
"""Middleware to inject uploaded files information into agent context.""" import logging from pathlib import Path from typing import NotRequired, override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langchain_core.messages import HumanMessage from langgraph.runtime import Runtime from src.config.paths import Paths, get_paths logger = logging.getLogger(__name__) class UploadsMiddlewareState(AgentState): """State schema for uploads middleware.""" uploaded_files: NotRequired[list[dict] | None] class UploadsMiddleware(AgentMiddleware[UploadsMiddlewareState]): """Middleware to inject uploaded files information into the agent context. Reads file metadata from the current message's additional_kwargs.files (set by the frontend after upload) and prepends an <uploaded_files> block to the last human message so the model knows which files are available. """ state_schema = UploadsMiddlewareState def __init__(self, base_dir: str | None = None): """Initialize the middleware. Args: base_dir: Base directory for thread data. Defaults to Paths resolution. """ super().__init__() self._paths = Paths(base_dir) if base_dir else get_paths() def _create_files_message(self, new_files: list[dict], historical_files: list[dict]) -> str: """Create a formatted message listing uploaded files. Args: new_files: Files uploaded in the current message. historical_files: Files uploaded in previous messages. Returns: Formatted string inside <uploaded_files> tags. """ lines = ["<uploaded_files>"] lines.append("The following files were uploaded in this message:") lines.append("") if new_files: for file in new_files: size_kb = file["size"] / 1024 size_str = f"{size_kb:.1f} KB" if size_kb < 1024 else f"{size_kb / 1024:.1f} MB" lines.append(f"- {file['filename']} ({size_str})") lines.append(f" Path: {file['path']}") lines.append("") else: lines.append("(empty)") if historical_files: lines.append("The following files were uploaded in previous messages and are still available:") lines.append("") for file in historical_files: size_kb = file["size"] / 1024 size_str = f"{size_kb:.1f} KB" if size_kb < 1024 else f"{size_kb / 1024:.1f} MB" lines.append(f"- {file['filename']} ({size_str})") lines.append(f" Path: {file['path']}") lines.append("") lines.append("You can read these files using the `read_file` tool with the paths shown above.") lines.append("</uploaded_files>") return "\n".join(lines) def _files_from_kwargs(self, message: HumanMessage, uploads_dir: Path | None = None) -> list[dict] | None: """Extract file info from message additional_kwargs.files. The frontend sends uploaded file metadata in additional_kwargs.files after a successful upload. Each entry has: filename, size (bytes), path (virtual path), status. Args: message: The human message to inspect. uploads_dir: Physical uploads directory used to verify file existence. When provided, entries whose files no longer exist are skipped. Returns: List of file dicts with virtual paths, or None if the field is absent or empty. """ kwargs_files = (message.additional_kwargs or {}).get("files") if not isinstance(kwargs_files, list) or not kwargs_files: return None files = [] for f in kwargs_files: if not isinstance(f, dict): continue filename = f.get("filename") or "" if not filename or Path(filename).name != filename: continue if uploads_dir is not None and not (uploads_dir / filename).is_file(): continue files.append( { "filename": filename, "size": int(f.get("size") or 0), "path": f"/mnt/user-data/uploads/{filename}", "extension": Path(filename).suffix, } ) return files if files else None @override def before_agent(self, state: UploadsMiddlewareState, runtime: Runtime) -> dict | None: """Inject uploaded files information before agent execution. New files come from the current message's additional_kwargs.files. Historical files are scanned from the thread's uploads directory, excluding the new ones. Prepends <uploaded_files> context to the last human message content. The original additional_kwargs (including files metadata) is preserved on the updated message so the frontend can read it from the stream. Args: state: Current agent state. runtime: Runtime context containing thread_id. Returns: State updates including uploaded files list. """ messages = list(state.get("messages", [])) if not messages: return None last_message_index = len(messages) - 1 last_message = messages[last_message_index] if not isinstance(last_message, HumanMessage): return None # Resolve uploads directory for existence checks thread_id = runtime.context.get("thread_id") uploads_dir = self._paths.sandbox_uploads_dir(thread_id) if thread_id else None # Get newly uploaded files from the current message's additional_kwargs.files new_files = self._files_from_kwargs(last_message, uploads_dir) or [] # Collect historical files from the uploads directory (all except the new ones) new_filenames = {f["filename"] for f in new_files} historical_files: list[dict] = [] if uploads_dir and uploads_dir.exists(): for file_path in sorted(uploads_dir.iterdir()): if file_path.is_file() and file_path.name not in new_filenames: stat = file_path.stat() historical_files.append( { "filename": file_path.name, "size": stat.st_size, "path": f"/mnt/user-data/uploads/{file_path.name}", "extension": file_path.suffix, } ) if not new_files and not historical_files: return None logger.debug(f"New files: {[f['filename'] for f in new_files]}, historical: {[f['filename'] for f in historical_files]}") # Create files message and prepend to the last human message content files_message = self._create_files_message(new_files, historical_files) # Extract original content - handle both string and list formats original_content = "" if isinstance(last_message.content, str): original_content = last_message.content elif isinstance(last_message.content, list): text_parts = [] for block in last_message.content: if isinstance(block, dict) and block.get("type") == "text": text_parts.append(block.get("text", "")) original_content = "\n".join(text_parts) # Create new message with combined content. # Preserve additional_kwargs (including files metadata) so the frontend # can read structured file info from the streamed message. updated_message = HumanMessage( content=f"{files_message}\n\n{original_content}", id=last_message.id, additional_kwargs=last_message.additional_kwargs, ) messages[last_message_index] = updated_message return { "uploaded_files": new_files, "messages": messages, }
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/uploads_middleware.py", "license": "MIT License", "lines": 162, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/agents/middlewares/view_image_middleware.py
"""Middleware for injecting image details into conversation before LLM call.""" from typing import NotRequired, override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langchain_core.messages import AIMessage, HumanMessage, ToolMessage from langgraph.runtime import Runtime from src.agents.thread_state import ViewedImageData class ViewImageMiddlewareState(AgentState): """Compatible with the `ThreadState` schema.""" viewed_images: NotRequired[dict[str, ViewedImageData] | None] class ViewImageMiddleware(AgentMiddleware[ViewImageMiddlewareState]): """Injects image details as a human message before LLM calls when view_image tools have completed. This middleware: 1. Runs before each LLM call 2. Checks if the last assistant message contains view_image tool calls 3. Verifies all tool calls in that message have been completed (have corresponding ToolMessages) 4. If conditions are met, creates a human message with all viewed image details (including base64 data) 5. Adds the message to state so the LLM can see and analyze the images This enables the LLM to automatically receive and analyze images that were loaded via view_image tool, without requiring explicit user prompts to describe the images. """ state_schema = ViewImageMiddlewareState def _get_last_assistant_message(self, messages: list) -> AIMessage | None: """Get the last assistant message from the message list. Args: messages: List of messages Returns: Last AIMessage or None if not found """ for msg in reversed(messages): if isinstance(msg, AIMessage): return msg return None def _has_view_image_tool(self, message: AIMessage) -> bool: """Check if the assistant message contains view_image tool calls. Args: message: Assistant message to check Returns: True if message contains view_image tool calls """ if not hasattr(message, "tool_calls") or not message.tool_calls: return False return any(tool_call.get("name") == "view_image" for tool_call in message.tool_calls) def _all_tools_completed(self, messages: list, assistant_msg: AIMessage) -> bool: """Check if all tool calls in the assistant message have been completed. Args: messages: List of all messages assistant_msg: The assistant message containing tool calls Returns: True if all tool calls have corresponding ToolMessages """ if not hasattr(assistant_msg, "tool_calls") or not assistant_msg.tool_calls: return False # Get all tool call IDs from the assistant message tool_call_ids = {tool_call.get("id") for tool_call in assistant_msg.tool_calls if tool_call.get("id")} # Find the index of the assistant message try: assistant_idx = messages.index(assistant_msg) except ValueError: return False # Get all ToolMessages after the assistant message completed_tool_ids = set() for msg in messages[assistant_idx + 1 :]: if isinstance(msg, ToolMessage) and msg.tool_call_id: completed_tool_ids.add(msg.tool_call_id) # Check if all tool calls have been completed return tool_call_ids.issubset(completed_tool_ids) def _create_image_details_message(self, state: ViewImageMiddlewareState) -> list[str | dict]: """Create a formatted message with all viewed image details. Args: state: Current state containing viewed_images Returns: List of content blocks (text and images) for the HumanMessage """ viewed_images = state.get("viewed_images", {}) if not viewed_images: return ["No images have been viewed."] # Build the message with image information content_blocks: list[str | dict] = [{"type": "text", "text": "Here are the images you've viewed:"}] for image_path, image_data in viewed_images.items(): mime_type = image_data.get("mime_type", "unknown") base64_data = image_data.get("base64", "") # Add text description content_blocks.append({"type": "text", "text": f"\n- **{image_path}** ({mime_type})"}) # Add the actual image data so LLM can "see" it if base64_data: content_blocks.append( { "type": "image_url", "image_url": {"url": f"data:{mime_type};base64,{base64_data}"}, } ) return content_blocks def _should_inject_image_message(self, state: ViewImageMiddlewareState) -> bool: """Determine if we should inject an image details message. Args: state: Current state Returns: True if we should inject the message """ messages = state.get("messages", []) if not messages: return False # Get the last assistant message last_assistant_msg = self._get_last_assistant_message(messages) if not last_assistant_msg: return False # Check if it has view_image tool calls if not self._has_view_image_tool(last_assistant_msg): return False # Check if all tools have been completed if not self._all_tools_completed(messages, last_assistant_msg): return False # Check if we've already added an image details message # Look for a human message after the last assistant message that contains image details assistant_idx = messages.index(last_assistant_msg) for msg in messages[assistant_idx + 1 :]: if isinstance(msg, HumanMessage): content_str = str(msg.content) if "Here are the images you've viewed" in content_str or "Here are the details of the images you've viewed" in content_str: # Already added, don't add again return False return True def _inject_image_message(self, state: ViewImageMiddlewareState) -> dict | None: """Internal helper to inject image details message. Args: state: Current state Returns: State update with additional human message, or None if no update needed """ if not self._should_inject_image_message(state): return None # Create the image details message with text and image content image_content = self._create_image_details_message(state) # Create a new human message with mixed content (text + images) human_msg = HumanMessage(content=image_content) print("[ViewImageMiddleware] Injecting image details message with images before LLM call") # Return state update with the new message return {"messages": [human_msg]} @override def before_model(self, state: ViewImageMiddlewareState, runtime: Runtime) -> dict | None: """Inject image details message before LLM call if view_image tools have completed (sync version). This runs before each LLM call, checking if the previous turn included view_image tool calls that have all completed. If so, it injects a human message with the image details so the LLM can see and analyze the images. Args: state: Current state runtime: Runtime context (unused but required by interface) Returns: State update with additional human message, or None if no update needed """ return self._inject_image_message(state) @override async def abefore_model(self, state: ViewImageMiddlewareState, runtime: Runtime) -> dict | None: """Inject image details message before LLM call if view_image tools have completed (async version). This runs before each LLM call, checking if the previous turn included view_image tool calls that have all completed. If so, it injects a human message with the image details so the LLM can see and analyze the images. Args: state: Current state runtime: Runtime context (unused but required by interface) Returns: State update with additional human message, or None if no update needed """ return self._inject_image_message(state)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/middlewares/view_image_middleware.py", "license": "MIT License", "lines": 165, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/agents/thread_state.py
from typing import Annotated, NotRequired, TypedDict from langchain.agents import AgentState class SandboxState(TypedDict): sandbox_id: NotRequired[str | None] class ThreadDataState(TypedDict): workspace_path: NotRequired[str | None] uploads_path: NotRequired[str | None] outputs_path: NotRequired[str | None] class ViewedImageData(TypedDict): base64: str mime_type: str def merge_artifacts(existing: list[str] | None, new: list[str] | None) -> list[str]: """Reducer for artifacts list - merges and deduplicates artifacts.""" if existing is None: return new or [] if new is None: return existing # Use dict.fromkeys to deduplicate while preserving order return list(dict.fromkeys(existing + new)) def merge_viewed_images(existing: dict[str, ViewedImageData] | None, new: dict[str, ViewedImageData] | None) -> dict[str, ViewedImageData]: """Reducer for viewed_images dict - merges image dictionaries. Special case: If new is an empty dict {}, it clears the existing images. This allows middlewares to clear the viewed_images state after processing. """ if existing is None: return new or {} if new is None: return existing # Special case: empty dict means clear all viewed images if len(new) == 0: return {} # Merge dictionaries, new values override existing ones for same keys return {**existing, **new} class ThreadState(AgentState): sandbox: NotRequired[SandboxState | None] thread_data: NotRequired[ThreadDataState | None] title: NotRequired[str | None] artifacts: Annotated[list[str], merge_artifacts] todos: NotRequired[list | None] uploaded_files: NotRequired[list[dict] | None] viewed_images: Annotated[dict[str, ViewedImageData], merge_viewed_images] # image_path -> {base64, mime_type}
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/agents/thread_state.py", "license": "MIT License", "lines": 41, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/community/aio_sandbox/aio_sandbox.py
import base64 import logging from agent_sandbox import Sandbox as AioSandboxClient from src.sandbox.sandbox import Sandbox logger = logging.getLogger(__name__) class AioSandbox(Sandbox): """Sandbox implementation using the agent-infra/sandbox Docker container. This sandbox connects to a running AIO sandbox container via HTTP API. """ def __init__(self, id: str, base_url: str, home_dir: str | None = None): """Initialize the AIO sandbox. Args: id: Unique identifier for this sandbox instance. base_url: URL of the sandbox API (e.g., http://localhost:8080). home_dir: Home directory inside the sandbox. If None, will be fetched from the sandbox. """ super().__init__(id) self._base_url = base_url self._client = AioSandboxClient(base_url=base_url, timeout=600) self._home_dir = home_dir @property def base_url(self) -> str: return self._base_url @property def home_dir(self) -> str: """Get the home directory inside the sandbox.""" if self._home_dir is None: context = self._client.sandbox.get_context() self._home_dir = context.home_dir return self._home_dir def execute_command(self, command: str) -> str: """Execute a shell command in the sandbox. Args: command: The command to execute. Returns: The output of the command. """ try: result = self._client.shell.exec_command(command=command) output = result.data.output if result.data else "" return output if output else "(no output)" except Exception as e: logger.error(f"Failed to execute command in sandbox: {e}") return f"Error: {e}" def read_file(self, path: str) -> str: """Read the content of a file in the sandbox. Args: path: The absolute path of the file to read. Returns: The content of the file. """ try: result = self._client.file.read_file(file=path) return result.data.content if result.data else "" except Exception as e: logger.error(f"Failed to read file in sandbox: {e}") return f"Error: {e}" def list_dir(self, path: str, max_depth: int = 2) -> list[str]: """List the contents of a directory in the sandbox. Args: path: The absolute path of the directory to list. max_depth: The maximum depth to traverse. Default is 2. Returns: The contents of the directory. """ try: # Use shell command to list directory with depth limit # The -L flag limits the depth for the tree command result = self._client.shell.exec_command(command=f"find {path} -maxdepth {max_depth} -type f -o -type d 2>/dev/null | head -500") output = result.data.output if result.data else "" if output: return [line.strip() for line in output.strip().split("\n") if line.strip()] return [] except Exception as e: logger.error(f"Failed to list directory in sandbox: {e}") return [] def write_file(self, path: str, content: str, append: bool = False) -> None: """Write content to a file in the sandbox. Args: path: The absolute path of the file to write to. content: The text content to write to the file. append: Whether to append the content to the file. """ try: if append: # Read existing content first and append existing = self.read_file(path) if not existing.startswith("Error:"): content = existing + content self._client.file.write_file(file=path, content=content) except Exception as e: logger.error(f"Failed to write file in sandbox: {e}") raise def update_file(self, path: str, content: bytes) -> None: """Update a file with binary content in the sandbox. Args: path: The absolute path of the file to update. content: The binary content to write to the file. """ try: base64_content = base64.b64encode(content).decode("utf-8") self._client.file.write_file(file=path, content=base64_content, encoding="base64") except Exception as e: logger.error(f"Failed to update file in sandbox: {e}") raise
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/aio_sandbox.py", "license": "MIT License", "lines": 105, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/aio_sandbox/aio_sandbox_provider.py
"""AIO Sandbox Provider β€” orchestrates sandbox lifecycle with pluggable backends. This provider composes two abstractions: - SandboxBackend: how sandboxes are provisioned (local container vs remote/K8s) - SandboxStateStore: how threadβ†’sandbox mappings are persisted (file vs Redis) The provider itself handles: - In-process caching for fast repeated access - Thread-safe locking (in-process + cross-process via state store) - Idle timeout management - Graceful shutdown with signal handling - Mount computation (thread-specific, skills) """ import atexit import hashlib import logging import os import signal import threading import time import uuid from src.config import get_app_config from src.config.paths import VIRTUAL_PATH_PREFIX, get_paths from src.sandbox.sandbox import Sandbox from src.sandbox.sandbox_provider import SandboxProvider from .aio_sandbox import AioSandbox from .backend import SandboxBackend, wait_for_sandbox_ready from .file_state_store import FileSandboxStateStore from .local_backend import LocalContainerBackend from .remote_backend import RemoteSandboxBackend from .sandbox_info import SandboxInfo from .state_store import SandboxStateStore logger = logging.getLogger(__name__) # Default configuration DEFAULT_IMAGE = "enterprise-public-cn-beijing.cr.volces.com/vefaas-public/all-in-one-sandbox:latest" DEFAULT_PORT = 8080 DEFAULT_CONTAINER_PREFIX = "deer-flow-sandbox" DEFAULT_IDLE_TIMEOUT = 600 # 10 minutes in seconds IDLE_CHECK_INTERVAL = 60 # Check every 60 seconds class AioSandboxProvider(SandboxProvider): """Sandbox provider that manages containers running the AIO sandbox. Architecture: This provider composes a SandboxBackend (how to provision) and a SandboxStateStore (how to persist state), enabling: - Local Docker/Apple Container mode (auto-start containers) - Remote/K8s mode (connect to pre-existing sandbox URL) - Cross-process consistency via file-based or Redis state stores Configuration options in config.yaml under sandbox: use: src.community.aio_sandbox:AioSandboxProvider image: <container image> port: 8080 # Base port for local containers base_url: http://... # If set, uses remote backend (K8s/external) auto_start: true # Whether to auto-start local containers container_prefix: deer-flow-sandbox idle_timeout: 600 # Idle timeout in seconds (0 to disable) mounts: # Volume mounts for local containers - host_path: /path/on/host container_path: /path/in/container read_only: false environment: # Environment variables for containers NODE_ENV: production API_KEY: $MY_API_KEY """ def __init__(self): self._lock = threading.Lock() self._sandboxes: dict[str, AioSandbox] = {} # sandbox_id -> AioSandbox instance self._sandbox_infos: dict[str, SandboxInfo] = {} # sandbox_id -> SandboxInfo (for destroy) self._thread_sandboxes: dict[str, str] = {} # thread_id -> sandbox_id self._thread_locks: dict[str, threading.Lock] = {} # thread_id -> in-process lock self._last_activity: dict[str, float] = {} # sandbox_id -> last activity timestamp self._shutdown_called = False self._idle_checker_stop = threading.Event() self._idle_checker_thread: threading.Thread | None = None self._config = self._load_config() self._backend: SandboxBackend = self._create_backend() self._state_store: SandboxStateStore = self._create_state_store() # Register shutdown handler atexit.register(self.shutdown) self._register_signal_handlers() # Start idle checker if enabled if self._config.get("idle_timeout", DEFAULT_IDLE_TIMEOUT) > 0: self._start_idle_checker() # ── Factory methods ────────────────────────────────────────────────── def _create_backend(self) -> SandboxBackend: """Create the appropriate backend based on configuration. Selection logic (checked in order): 1. ``provisioner_url`` set β†’ RemoteSandboxBackend (provisioner mode) Provisioner dynamically creates Pods + Services in k3s. 2. ``auto_start`` β†’ LocalContainerBackend (Docker / Apple Container) """ provisioner_url = self._config.get("provisioner_url") if provisioner_url: logger.info(f"Using remote sandbox backend with provisioner at {provisioner_url}") return RemoteSandboxBackend(provisioner_url=provisioner_url) if not self._config.get("auto_start", True): raise RuntimeError("auto_start is disabled and no base_url is configured") logger.info("Using local container sandbox backend") return LocalContainerBackend( image=self._config["image"], base_port=self._config["port"], container_prefix=self._config["container_prefix"], config_mounts=self._config["mounts"], environment=self._config["environment"], ) def _create_state_store(self) -> SandboxStateStore: """Create the state store for cross-process sandbox mapping persistence. Currently uses file-based store. For distributed multi-host deployments, a Redis-based store can be plugged in here. """ # TODO: Support RedisSandboxStateStore for distributed deployments. # Configuration would be: # sandbox: # state_store: redis # redis_url: redis://localhost:6379/0 # This would enable cross-host sandbox discovery (e.g., multiple K8s pods # without shared PVC, or multi-node Docker Swarm). return FileSandboxStateStore(base_dir=str(get_paths().base_dir)) # ── Configuration ──────────────────────────────────────────────────── def _load_config(self) -> dict: """Load sandbox configuration from app config.""" config = get_app_config() sandbox_config = config.sandbox return { "image": sandbox_config.image or DEFAULT_IMAGE, "port": sandbox_config.port or DEFAULT_PORT, "base_url": sandbox_config.base_url, "auto_start": sandbox_config.auto_start if sandbox_config.auto_start is not None else True, "container_prefix": sandbox_config.container_prefix or DEFAULT_CONTAINER_PREFIX, "idle_timeout": getattr(sandbox_config, "idle_timeout", None) or DEFAULT_IDLE_TIMEOUT, "mounts": sandbox_config.mounts or [], "environment": self._resolve_env_vars(sandbox_config.environment or {}), # provisioner URL for dynamic pod management (e.g. http://provisioner:8002) "provisioner_url": getattr(sandbox_config, "provisioner_url", None) or "", } @staticmethod def _resolve_env_vars(env_config: dict[str, str]) -> dict[str, str]: """Resolve environment variable references (values starting with $).""" resolved = {} for key, value in env_config.items(): if isinstance(value, str) and value.startswith("$"): env_name = value[1:] resolved[key] = os.environ.get(env_name, "") else: resolved[key] = str(value) return resolved # ── Deterministic ID ───────────────────────────────────────────────── @staticmethod def _deterministic_sandbox_id(thread_id: str) -> str: """Generate a deterministic sandbox ID from a thread ID. Ensures all processes derive the same sandbox_id for a given thread, enabling cross-process sandbox discovery without shared memory. """ return hashlib.sha256(thread_id.encode()).hexdigest()[:8] # ── Mount helpers ──────────────────────────────────────────────────── def _get_extra_mounts(self, thread_id: str | None) -> list[tuple[str, str, bool]]: """Collect all extra mounts for a sandbox (thread-specific + skills).""" mounts: list[tuple[str, str, bool]] = [] if thread_id: mounts.extend(self._get_thread_mounts(thread_id)) logger.info(f"Adding thread mounts for thread {thread_id}: {mounts}") skills_mount = self._get_skills_mount() if skills_mount: mounts.append(skills_mount) logger.info(f"Adding skills mount: {skills_mount}") return mounts @staticmethod def _get_thread_mounts(thread_id: str) -> list[tuple[str, str, bool]]: """Get volume mounts for a thread's data directories. Creates directories if they don't exist (lazy initialization). """ paths = get_paths() paths.ensure_thread_dirs(thread_id) mounts = [ (str(paths.sandbox_work_dir(thread_id)), f"{VIRTUAL_PATH_PREFIX}/workspace", False), (str(paths.sandbox_uploads_dir(thread_id)), f"{VIRTUAL_PATH_PREFIX}/uploads", False), (str(paths.sandbox_outputs_dir(thread_id)), f"{VIRTUAL_PATH_PREFIX}/outputs", False), ] return mounts @staticmethod def _get_skills_mount() -> tuple[str, str, bool] | None: """Get the skills directory mount configuration.""" try: config = get_app_config() skills_path = config.skills.get_skills_path() container_path = config.skills.container_path if skills_path.exists(): return (str(skills_path), container_path, True) # Read-only for security except Exception as e: logger.warning(f"Could not setup skills mount: {e}") return None # ── Idle timeout management ────────────────────────────────────────── def _start_idle_checker(self) -> None: """Start the background thread that checks for idle sandboxes.""" self._idle_checker_thread = threading.Thread( target=self._idle_checker_loop, name="sandbox-idle-checker", daemon=True, ) self._idle_checker_thread.start() logger.info(f"Started idle checker thread (timeout: {self._config.get('idle_timeout', DEFAULT_IDLE_TIMEOUT)}s)") def _idle_checker_loop(self) -> None: idle_timeout = self._config.get("idle_timeout", DEFAULT_IDLE_TIMEOUT) while not self._idle_checker_stop.wait(timeout=IDLE_CHECK_INTERVAL): try: self._cleanup_idle_sandboxes(idle_timeout) except Exception as e: logger.error(f"Error in idle checker loop: {e}") def _cleanup_idle_sandboxes(self, idle_timeout: float) -> None: current_time = time.time() sandboxes_to_release = [] with self._lock: for sandbox_id, last_activity in self._last_activity.items(): idle_duration = current_time - last_activity if idle_duration > idle_timeout: sandboxes_to_release.append(sandbox_id) logger.info(f"Sandbox {sandbox_id} idle for {idle_duration:.1f}s, marking for release") for sandbox_id in sandboxes_to_release: try: logger.info(f"Releasing idle sandbox {sandbox_id}") self.release(sandbox_id) except Exception as e: logger.error(f"Failed to release idle sandbox {sandbox_id}: {e}") # ── Signal handling ────────────────────────────────────────────────── def _register_signal_handlers(self) -> None: """Register signal handlers for graceful shutdown.""" self._original_sigterm = signal.getsignal(signal.SIGTERM) self._original_sigint = signal.getsignal(signal.SIGINT) def signal_handler(signum, frame): self.shutdown() original = self._original_sigterm if signum == signal.SIGTERM else self._original_sigint if callable(original): original(signum, frame) elif original == signal.SIG_DFL: signal.signal(signum, signal.SIG_DFL) signal.raise_signal(signum) try: signal.signal(signal.SIGTERM, signal_handler) signal.signal(signal.SIGINT, signal_handler) except ValueError: logger.debug("Could not register signal handlers (not main thread)") # ── Thread locking (in-process) ────────────────────────────────────── def _get_thread_lock(self, thread_id: str) -> threading.Lock: """Get or create an in-process lock for a specific thread_id.""" with self._lock: if thread_id not in self._thread_locks: self._thread_locks[thread_id] = threading.Lock() return self._thread_locks[thread_id] # ── Core: acquire / get / release / shutdown ───────────────────────── def acquire(self, thread_id: str | None = None) -> str: """Acquire a sandbox environment and return its ID. For the same thread_id, this method will return the same sandbox_id across multiple turns, multiple processes, and (with shared storage) multiple pods. Thread-safe with both in-process and cross-process locking. Args: thread_id: Optional thread ID for thread-specific configurations. Returns: The ID of the acquired sandbox environment. """ if thread_id: thread_lock = self._get_thread_lock(thread_id) with thread_lock: return self._acquire_internal(thread_id) else: return self._acquire_internal(thread_id) def _acquire_internal(self, thread_id: str | None) -> str: """Internal sandbox acquisition with three-layer consistency. Layer 1: In-process cache (fastest, covers same-process repeated access) Layer 2: Cross-process state store + file lock (covers multi-process) Layer 3: Backend discovery (covers containers started by other processes) """ # ── Layer 1: In-process cache (fast path) ── if thread_id: with self._lock: if thread_id in self._thread_sandboxes: existing_id = self._thread_sandboxes[thread_id] if existing_id in self._sandboxes: logger.info(f"Reusing in-process sandbox {existing_id} for thread {thread_id}") self._last_activity[existing_id] = time.time() return existing_id else: del self._thread_sandboxes[thread_id] # Deterministic ID for thread-specific, random for anonymous sandbox_id = self._deterministic_sandbox_id(thread_id) if thread_id else str(uuid.uuid4())[:8] # ── Layer 2 & 3: Cross-process recovery + creation ── if thread_id: with self._state_store.lock(thread_id): # Try to recover from persisted state or discover existing container recovered_id = self._try_recover(thread_id) if recovered_id is not None: return recovered_id # Nothing to recover β€” create new sandbox (still under cross-process lock) return self._create_sandbox(thread_id, sandbox_id) else: return self._create_sandbox(thread_id, sandbox_id) def _try_recover(self, thread_id: str) -> str | None: """Try to recover a sandbox from persisted state or backend discovery. Called under cross-process lock for the given thread_id. Args: thread_id: The thread ID. Returns: The sandbox_id if recovery succeeded, None otherwise. """ info = self._state_store.load(thread_id) if info is None: return None # Re-discover: verifies sandbox is alive and gets current connection info # (handles cases like port changes after container restart) discovered = self._backend.discover(info.sandbox_id) if discovered is None: logger.info(f"Persisted sandbox {info.sandbox_id} for thread {thread_id} could not be recovered") self._state_store.remove(thread_id) return None # Adopt into this process's memory sandbox = AioSandbox(id=discovered.sandbox_id, base_url=discovered.sandbox_url) with self._lock: self._sandboxes[discovered.sandbox_id] = sandbox self._sandbox_infos[discovered.sandbox_id] = discovered self._last_activity[discovered.sandbox_id] = time.time() self._thread_sandboxes[thread_id] = discovered.sandbox_id # Update state if connection info changed if discovered.sandbox_url != info.sandbox_url: self._state_store.save(thread_id, discovered) logger.info(f"Recovered sandbox {discovered.sandbox_id} for thread {thread_id} at {discovered.sandbox_url}") return discovered.sandbox_id def _create_sandbox(self, thread_id: str | None, sandbox_id: str) -> str: """Create a new sandbox via the backend. Args: thread_id: Optional thread ID. sandbox_id: The sandbox ID to use. Returns: The sandbox_id. Raises: RuntimeError: If sandbox creation or readiness check fails. """ extra_mounts = self._get_extra_mounts(thread_id) info = self._backend.create(thread_id, sandbox_id, extra_mounts=extra_mounts or None) # Wait for sandbox to be ready if not wait_for_sandbox_ready(info.sandbox_url, timeout=60): self._backend.destroy(info) raise RuntimeError(f"Sandbox {sandbox_id} failed to become ready within timeout at {info.sandbox_url}") sandbox = AioSandbox(id=sandbox_id, base_url=info.sandbox_url) with self._lock: self._sandboxes[sandbox_id] = sandbox self._sandbox_infos[sandbox_id] = info self._last_activity[sandbox_id] = time.time() if thread_id: self._thread_sandboxes[thread_id] = sandbox_id # Persist for cross-process discovery if thread_id: self._state_store.save(thread_id, info) logger.info(f"Created sandbox {sandbox_id} for thread {thread_id} at {info.sandbox_url}") return sandbox_id def get(self, sandbox_id: str) -> Sandbox | None: """Get a sandbox by ID. Updates last activity timestamp. Args: sandbox_id: The ID of the sandbox. Returns: The sandbox instance if found, None otherwise. """ with self._lock: sandbox = self._sandboxes.get(sandbox_id) if sandbox is not None: self._last_activity[sandbox_id] = time.time() return sandbox def release(self, sandbox_id: str) -> None: """Release a sandbox: clean up in-memory state, persisted state, and backend resources. Args: sandbox_id: The ID of the sandbox to release. """ info = None thread_ids_to_remove: list[str] = [] with self._lock: self._sandboxes.pop(sandbox_id, None) info = self._sandbox_infos.pop(sandbox_id, None) thread_ids_to_remove = [tid for tid, sid in self._thread_sandboxes.items() if sid == sandbox_id] for tid in thread_ids_to_remove: del self._thread_sandboxes[tid] self._last_activity.pop(sandbox_id, None) # Clean up persisted state (outside lock, involves file I/O) for tid in thread_ids_to_remove: self._state_store.remove(tid) # Destroy backend resources (stop container, release port, etc.) if info: self._backend.destroy(info) logger.info(f"Released sandbox {sandbox_id}") def shutdown(self) -> None: """Shutdown all sandboxes. Thread-safe and idempotent.""" with self._lock: if self._shutdown_called: return self._shutdown_called = True sandbox_ids = list(self._sandboxes.keys()) # Stop idle checker self._idle_checker_stop.set() if self._idle_checker_thread is not None and self._idle_checker_thread.is_alive(): self._idle_checker_thread.join(timeout=5) logger.info("Stopped idle checker thread") logger.info(f"Shutting down {len(sandbox_ids)} sandbox(es)") for sandbox_id in sandbox_ids: try: self.release(sandbox_id) except Exception as e: logger.error(f"Failed to release sandbox {sandbox_id} during shutdown: {e}")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/aio_sandbox_provider.py", "license": "MIT License", "lines": 402, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/aio_sandbox/backend.py
"""Abstract base class for sandbox provisioning backends.""" from __future__ import annotations import logging import time from abc import ABC, abstractmethod import requests from .sandbox_info import SandboxInfo logger = logging.getLogger(__name__) def wait_for_sandbox_ready(sandbox_url: str, timeout: int = 30) -> bool: """Poll sandbox health endpoint until ready or timeout. Args: sandbox_url: URL of the sandbox (e.g. http://k3s:30001). timeout: Maximum time to wait in seconds. Returns: True if sandbox is ready, False otherwise. """ start_time = time.time() while time.time() - start_time < timeout: try: response = requests.get(f"{sandbox_url}/v1/sandbox", timeout=5) if response.status_code == 200: return True except requests.exceptions.RequestException: pass time.sleep(1) return False class SandboxBackend(ABC): """Abstract base for sandbox provisioning backends. Two implementations: - LocalContainerBackend: starts Docker/Apple Container locally, manages ports - RemoteSandboxBackend: connects to a pre-existing URL (K8s service, external) """ @abstractmethod def create(self, thread_id: str, sandbox_id: str, extra_mounts: list[tuple[str, str, bool]] | None = None) -> SandboxInfo: """Create/provision a new sandbox. Args: thread_id: Thread ID for which the sandbox is being created. Useful for backends that want to organize sandboxes by thread. sandbox_id: Deterministic sandbox identifier. extra_mounts: Additional volume mounts as (host_path, container_path, read_only) tuples. Ignored by backends that don't manage containers (e.g., remote). Returns: SandboxInfo with connection details. """ ... @abstractmethod def destroy(self, info: SandboxInfo) -> None: """Destroy/cleanup a sandbox and release its resources. Args: info: The sandbox metadata to destroy. """ ... @abstractmethod def is_alive(self, info: SandboxInfo) -> bool: """Quick check whether a sandbox is still alive. This should be a lightweight check (e.g., container inspect) rather than a full health check. Args: info: The sandbox metadata to check. Returns: True if the sandbox appears to be alive. """ ... @abstractmethod def discover(self, sandbox_id: str) -> SandboxInfo | None: """Try to discover an existing sandbox by its deterministic ID. Used for cross-process recovery: when another process started a sandbox, this process can discover it by the deterministic container name or URL. Args: sandbox_id: The deterministic sandbox ID to look for. Returns: SandboxInfo if found and healthy, None otherwise. """ ...
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/backend.py", "license": "MIT License", "lines": 73, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/community/aio_sandbox/file_state_store.py
"""File-based sandbox state store. Uses JSON files for persistence and fcntl file locking for cross-process mutual exclusion. Works across processes on the same machine or across K8s pods with a shared PVC mount. """ from __future__ import annotations import fcntl import json import logging import os from collections.abc import Generator from contextlib import contextmanager from pathlib import Path from src.config.paths import Paths from .sandbox_info import SandboxInfo from .state_store import SandboxStateStore logger = logging.getLogger(__name__) SANDBOX_STATE_FILE = "sandbox.json" SANDBOX_LOCK_FILE = "sandbox.lock" class FileSandboxStateStore(SandboxStateStore): """File-based state store using JSON files and fcntl file locking. State is stored at: {base_dir}/threads/{thread_id}/sandbox.json Lock files at: {base_dir}/threads/{thread_id}/sandbox.lock This works across processes on the same machine sharing a filesystem. For K8s multi-pod scenarios, requires a shared PVC mount at base_dir. """ def __init__(self, base_dir: str): """Initialize the file-based state store. Args: base_dir: Root directory for state files (typically Paths.base_dir). """ self._paths = Paths(base_dir) def _thread_dir(self, thread_id: str) -> Path: """Get the directory for a thread's state files.""" return self._paths.thread_dir(thread_id) def save(self, thread_id: str, info: SandboxInfo) -> None: thread_dir = self._thread_dir(thread_id) os.makedirs(thread_dir, exist_ok=True) state_file = thread_dir / SANDBOX_STATE_FILE try: state_file.write_text(json.dumps(info.to_dict())) logger.info(f"Saved sandbox state for thread {thread_id}: {info.sandbox_id}") except OSError as e: logger.warning(f"Failed to save sandbox state for thread {thread_id}: {e}") def load(self, thread_id: str) -> SandboxInfo | None: state_file = self._thread_dir(thread_id) / SANDBOX_STATE_FILE if not state_file.exists(): return None try: data = json.loads(state_file.read_text()) return SandboxInfo.from_dict(data) except (OSError, json.JSONDecodeError, KeyError) as e: logger.warning(f"Failed to load sandbox state for thread {thread_id}: {e}") return None def remove(self, thread_id: str) -> None: state_file = self._thread_dir(thread_id) / SANDBOX_STATE_FILE try: if state_file.exists(): state_file.unlink() logger.info(f"Removed sandbox state for thread {thread_id}") except OSError as e: logger.warning(f"Failed to remove sandbox state for thread {thread_id}: {e}") @contextmanager def lock(self, thread_id: str) -> Generator[None, None, None]: """Acquire a cross-process file lock using fcntl.flock. The lock is held for the duration of the context manager. Only one process can hold the lock at a time for a given thread_id. Note: fcntl.flock is available on macOS and Linux. """ thread_dir = self._thread_dir(thread_id) os.makedirs(thread_dir, exist_ok=True) lock_path = thread_dir / SANDBOX_LOCK_FILE lock_file = open(lock_path, "w") try: fcntl.flock(lock_file.fileno(), fcntl.LOCK_EX) yield finally: try: fcntl.flock(lock_file.fileno(), fcntl.LOCK_UN) lock_file.close() except OSError: pass
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/file_state_store.py", "license": "MIT License", "lines": 82, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/aio_sandbox/local_backend.py
"""Local container backend for sandbox provisioning. Manages sandbox containers using Docker or Apple Container on the local machine. Handles container lifecycle, port allocation, and cross-process container discovery. """ from __future__ import annotations import logging import subprocess from src.utils.network import get_free_port, release_port from .backend import SandboxBackend, wait_for_sandbox_ready from .sandbox_info import SandboxInfo logger = logging.getLogger(__name__) class LocalContainerBackend(SandboxBackend): """Backend that manages sandbox containers locally using Docker or Apple Container. On macOS, automatically prefers Apple Container if available, otherwise falls back to Docker. On other platforms, uses Docker. Features: - Deterministic container naming for cross-process discovery - Port allocation with thread-safe utilities - Container lifecycle management (start/stop with --rm) - Support for volume mounts and environment variables """ def __init__( self, *, image: str, base_port: int, container_prefix: str, config_mounts: list, environment: dict[str, str], ): """Initialize the local container backend. Args: image: Container image to use. base_port: Base port number to start searching for free ports. container_prefix: Prefix for container names (e.g., "deer-flow-sandbox"). config_mounts: Volume mount configurations from config (list of VolumeMountConfig). environment: Environment variables to inject into containers. """ self._image = image self._base_port = base_port self._container_prefix = container_prefix self._config_mounts = config_mounts self._environment = environment self._runtime = self._detect_runtime() @property def runtime(self) -> str: """The detected container runtime ("docker" or "container").""" return self._runtime def _detect_runtime(self) -> str: """Detect which container runtime to use. On macOS, prefer Apple Container if available, otherwise fall back to Docker. On other platforms, use Docker. Returns: "container" for Apple Container, "docker" for Docker. """ import platform if platform.system() == "Darwin": try: result = subprocess.run( ["container", "--version"], capture_output=True, text=True, check=True, timeout=5, ) logger.info(f"Detected Apple Container: {result.stdout.strip()}") return "container" except (FileNotFoundError, subprocess.CalledProcessError, subprocess.TimeoutExpired): logger.info("Apple Container not available, falling back to Docker") return "docker" # ── SandboxBackend interface ────────────────────────────────────────── def create(self, thread_id: str, sandbox_id: str, extra_mounts: list[tuple[str, str, bool]] | None = None) -> SandboxInfo: """Start a new container and return its connection info. Args: thread_id: Thread ID for which the sandbox is being created. Useful for backends that want to organize sandboxes by thread. sandbox_id: Deterministic sandbox identifier (used in container name). extra_mounts: Additional volume mounts as (host_path, container_path, read_only) tuples. Returns: SandboxInfo with container details. Raises: RuntimeError: If the container fails to start. """ container_name = f"{self._container_prefix}-{sandbox_id}" port = get_free_port(start_port=self._base_port) try: container_id = self._start_container(container_name, port, extra_mounts) except Exception: release_port(port) raise return SandboxInfo( sandbox_id=sandbox_id, sandbox_url=f"http://localhost:{port}", container_name=container_name, container_id=container_id, ) def destroy(self, info: SandboxInfo) -> None: """Stop the container and release its port.""" if info.container_id: self._stop_container(info.container_id) # Extract port from sandbox_url for release try: from urllib.parse import urlparse port = urlparse(info.sandbox_url).port if port: release_port(port) except Exception: pass def is_alive(self, info: SandboxInfo) -> bool: """Check if the container is still running (lightweight, no HTTP).""" if info.container_name: return self._is_container_running(info.container_name) return False def discover(self, sandbox_id: str) -> SandboxInfo | None: """Discover an existing container by its deterministic name. Checks if a container with the expected name is running, retrieves its port, and verifies it responds to health checks. Args: sandbox_id: The deterministic sandbox ID (determines container name). Returns: SandboxInfo if container found and healthy, None otherwise. """ container_name = f"{self._container_prefix}-{sandbox_id}" if not self._is_container_running(container_name): return None port = self._get_container_port(container_name) if port is None: return None sandbox_url = f"http://localhost:{port}" if not wait_for_sandbox_ready(sandbox_url, timeout=5): return None return SandboxInfo( sandbox_id=sandbox_id, sandbox_url=sandbox_url, container_name=container_name, ) # ── Container operations ───────────────────────────────────────────── def _start_container( self, container_name: str, port: int, extra_mounts: list[tuple[str, str, bool]] | None = None, ) -> str: """Start a new container. Args: container_name: Name for the container. port: Host port to map to container port 8080. extra_mounts: Additional volume mounts. Returns: The container ID. Raises: RuntimeError: If container fails to start. """ cmd = [self._runtime, "run"] # Docker-specific security options if self._runtime == "docker": cmd.extend(["--security-opt", "seccomp=unconfined"]) cmd.extend( [ "--rm", "-d", "-p", f"{port}:8080", "--name", container_name, ] ) # Environment variables for key, value in self._environment.items(): cmd.extend(["-e", f"{key}={value}"]) # Config-level volume mounts for mount in self._config_mounts: mount_spec = f"{mount.host_path}:{mount.container_path}" if mount.read_only: mount_spec += ":ro" cmd.extend(["-v", mount_spec]) # Extra mounts (thread-specific, skills, etc.) if extra_mounts: for host_path, container_path, read_only in extra_mounts: mount_spec = f"{host_path}:{container_path}" if read_only: mount_spec += ":ro" cmd.extend(["-v", mount_spec]) cmd.append(self._image) logger.info(f"Starting container using {self._runtime}: {' '.join(cmd)}") try: result = subprocess.run(cmd, capture_output=True, text=True, check=True) container_id = result.stdout.strip() logger.info(f"Started container {container_name} (ID: {container_id}) using {self._runtime}") return container_id except subprocess.CalledProcessError as e: logger.error(f"Failed to start container using {self._runtime}: {e.stderr}") raise RuntimeError(f"Failed to start sandbox container: {e.stderr}") def _stop_container(self, container_id: str) -> None: """Stop a container (--rm ensures automatic removal).""" try: subprocess.run( [self._runtime, "stop", container_id], capture_output=True, text=True, check=True, ) logger.info(f"Stopped container {container_id} using {self._runtime}") except subprocess.CalledProcessError as e: logger.warning(f"Failed to stop container {container_id}: {e.stderr}") def _is_container_running(self, container_name: str) -> bool: """Check if a named container is currently running. This enables cross-process container discovery β€” any process can detect containers started by another process via the deterministic container name. """ try: result = subprocess.run( [self._runtime, "inspect", "-f", "{{.State.Running}}", container_name], capture_output=True, text=True, timeout=5, ) return result.returncode == 0 and result.stdout.strip().lower() == "true" except (subprocess.CalledProcessError, subprocess.TimeoutExpired): return False def _get_container_port(self, container_name: str) -> int | None: """Get the host port of a running container. Args: container_name: The container name to inspect. Returns: The host port mapped to container port 8080, or None if not found. """ try: result = subprocess.run( [self._runtime, "port", container_name, "8080"], capture_output=True, text=True, timeout=5, ) if result.returncode == 0 and result.stdout.strip(): # Output format: "0.0.0.0:PORT" or ":::PORT" port_str = result.stdout.strip().split(":")[-1] return int(port_str) except (subprocess.CalledProcessError, subprocess.TimeoutExpired, ValueError): pass return None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/local_backend.py", "license": "MIT License", "lines": 240, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/aio_sandbox/remote_backend.py
"""Remote sandbox backend β€” delegates Pod lifecycle to the provisioner service. The provisioner dynamically creates per-sandbox-id Pods + NodePort Services in k3s. The backend accesses sandbox pods directly via ``k3s:{NodePort}``. Architecture: β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” HTTP β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” K8s API β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚ this file β”‚ ──────▸ β”‚ provisioner β”‚ ────────▸ β”‚ k3s β”‚ β”‚ (backend) β”‚ β”‚ :8002 β”‚ β”‚ :6443 β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”˜ β”‚ creates β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”Œβ”€β”€β”€β”€β”€β–Όβ”€β”€β”€β”€β”€β”€β” β”‚ backend β”‚ ────────▸ β”‚ sandbox β”‚ β”‚ β”‚ direct β”‚ Pod(s) β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ k3s:NPort β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ """ from __future__ import annotations import logging import requests from .backend import SandboxBackend from .sandbox_info import SandboxInfo logger = logging.getLogger(__name__) class RemoteSandboxBackend(SandboxBackend): """Backend that delegates sandbox lifecycle to the provisioner service. All Pod creation, destruction, and discovery are handled by the provisioner. This backend is a thin HTTP client. Typical config.yaml:: sandbox: use: src.community.aio_sandbox:AioSandboxProvider provisioner_url: http://provisioner:8002 """ def __init__(self, provisioner_url: str): """Initialize with the provisioner service URL. Args: provisioner_url: URL of the provisioner service (e.g., ``http://provisioner:8002``). """ self._provisioner_url = provisioner_url.rstrip("/") @property def provisioner_url(self) -> str: return self._provisioner_url # ── SandboxBackend interface ────────────────────────────────────────── def create( self, thread_id: str, sandbox_id: str, extra_mounts: list[tuple[str, str, bool]] | None = None, ) -> SandboxInfo: """Create a sandbox Pod + Service via the provisioner. Calls ``POST /api/sandboxes`` which creates a dedicated Pod + NodePort Service in k3s. """ return self._provisioner_create(thread_id, sandbox_id, extra_mounts) def destroy(self, info: SandboxInfo) -> None: """Destroy a sandbox Pod + Service via the provisioner.""" self._provisioner_destroy(info.sandbox_id) def is_alive(self, info: SandboxInfo) -> bool: """Check whether the sandbox Pod is running.""" return self._provisioner_is_alive(info.sandbox_id) def discover(self, sandbox_id: str) -> SandboxInfo | None: """Discover an existing sandbox via the provisioner. Calls ``GET /api/sandboxes/{sandbox_id}`` and returns info if the Pod exists. """ return self._provisioner_discover(sandbox_id) # ── Provisioner API calls ───────────────────────────────────────────── def _provisioner_create(self, thread_id: str, sandbox_id: str, extra_mounts: list[tuple[str, str, bool]] | None = None) -> SandboxInfo: """POST /api/sandboxes β†’ create Pod + Service.""" try: resp = requests.post( f"{self._provisioner_url}/api/sandboxes", json={ "sandbox_id": sandbox_id, "thread_id": thread_id, }, timeout=30, ) resp.raise_for_status() data = resp.json() logger.info(f"Provisioner created sandbox {sandbox_id}: sandbox_url={data['sandbox_url']}") return SandboxInfo( sandbox_id=sandbox_id, sandbox_url=data["sandbox_url"], ) except requests.RequestException as exc: logger.error(f"Provisioner create failed for {sandbox_id}: {exc}") raise RuntimeError(f"Provisioner create failed: {exc}") from exc def _provisioner_destroy(self, sandbox_id: str) -> None: """DELETE /api/sandboxes/{sandbox_id} β†’ destroy Pod + Service.""" try: resp = requests.delete( f"{self._provisioner_url}/api/sandboxes/{sandbox_id}", timeout=15, ) if resp.ok: logger.info(f"Provisioner destroyed sandbox {sandbox_id}") else: logger.warning(f"Provisioner destroy returned {resp.status_code}: {resp.text}") except requests.RequestException as exc: logger.warning(f"Provisioner destroy failed for {sandbox_id}: {exc}") def _provisioner_is_alive(self, sandbox_id: str) -> bool: """GET /api/sandboxes/{sandbox_id} β†’ check Pod phase.""" try: resp = requests.get( f"{self._provisioner_url}/api/sandboxes/{sandbox_id}", timeout=10, ) if resp.ok: data = resp.json() return data.get("status") == "Running" return False except requests.RequestException: return False def _provisioner_discover(self, sandbox_id: str) -> SandboxInfo | None: """GET /api/sandboxes/{sandbox_id} β†’ discover existing sandbox.""" try: resp = requests.get( f"{self._provisioner_url}/api/sandboxes/{sandbox_id}", timeout=10, ) if resp.status_code == 404: return None resp.raise_for_status() data = resp.json() return SandboxInfo( sandbox_id=sandbox_id, sandbox_url=data["sandbox_url"], ) except requests.RequestException as exc: logger.debug(f"Provisioner discover failed for {sandbox_id}: {exc}") return None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/remote_backend.py", "license": "MIT License", "lines": 129, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/aio_sandbox/sandbox_info.py
"""Sandbox metadata for cross-process discovery and state persistence.""" from __future__ import annotations import time from dataclasses import dataclass, field @dataclass class SandboxInfo: """Persisted sandbox metadata that enables cross-process discovery. This dataclass holds all the information needed to reconnect to an existing sandbox from a different process (e.g., gateway vs langgraph, multiple workers, or across K8s pods with shared storage). """ sandbox_id: str sandbox_url: str # e.g. http://localhost:8080 or http://k3s:30001 container_name: str | None = None # Only for local container backend container_id: str | None = None # Only for local container backend created_at: float = field(default_factory=time.time) def to_dict(self) -> dict: return { "sandbox_id": self.sandbox_id, "sandbox_url": self.sandbox_url, "container_name": self.container_name, "container_id": self.container_id, "created_at": self.created_at, } @classmethod def from_dict(cls, data: dict) -> SandboxInfo: return cls( sandbox_id=data["sandbox_id"], sandbox_url=data.get("sandbox_url", data.get("base_url", "")), container_name=data.get("container_name"), container_id=data.get("container_id"), created_at=data.get("created_at", time.time()), )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/sandbox_info.py", "license": "MIT License", "lines": 33, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/community/aio_sandbox/state_store.py
"""Abstract base class for sandbox state persistence. The state store handles cross-process persistence of thread_id β†’ sandbox mappings, enabling different processes (gateway, langgraph, multiple workers) to find the same sandbox for a given thread. """ from __future__ import annotations from abc import ABC, abstractmethod from collections.abc import Generator from contextlib import contextmanager from .sandbox_info import SandboxInfo class SandboxStateStore(ABC): """Abstract base for persisting thread_id β†’ sandbox mappings across processes. Implementations: - FileSandboxStateStore: JSON files + fcntl file locking (single-host) - TODO: RedisSandboxStateStore: Redis-based for distributed multi-host deployments """ @abstractmethod def save(self, thread_id: str, info: SandboxInfo) -> None: """Save sandbox state for a thread. Args: thread_id: The thread ID. info: Sandbox metadata to persist. """ ... @abstractmethod def load(self, thread_id: str) -> SandboxInfo | None: """Load sandbox state for a thread. Args: thread_id: The thread ID. Returns: SandboxInfo if found, None otherwise. """ ... @abstractmethod def remove(self, thread_id: str) -> None: """Remove sandbox state for a thread. Args: thread_id: The thread ID. """ ... @abstractmethod @contextmanager def lock(self, thread_id: str) -> Generator[None, None, None]: """Acquire a cross-process lock for a thread's sandbox operations. Ensures only one process can create/modify a sandbox for a given thread_id at a time, preventing duplicate sandbox creation. Args: thread_id: The thread ID to lock. Yields: None β€” use as a context manager. """ ...
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/aio_sandbox/state_store.py", "license": "MIT License", "lines": 52, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/community/firecrawl/tools.py
import json from firecrawl import FirecrawlApp from langchain.tools import tool from src.config import get_app_config def _get_firecrawl_client() -> FirecrawlApp: config = get_app_config().get_tool_config("web_search") api_key = None if config is not None: api_key = config.model_extra.get("api_key") return FirecrawlApp(api_key=api_key) # type: ignore[arg-type] @tool("web_search", parse_docstring=True) def web_search_tool(query: str) -> str: """Search the web. Args: query: The query to search for. """ try: config = get_app_config().get_tool_config("web_search") max_results = 5 if config is not None: max_results = config.model_extra.get("max_results", max_results) client = _get_firecrawl_client() result = client.search(query, limit=max_results) # result.web contains list of SearchResultWeb objects web_results = result.web or [] normalized_results = [ { "title": getattr(item, "title", "") or "", "url": getattr(item, "url", "") or "", "snippet": getattr(item, "description", "") or "", } for item in web_results ] json_results = json.dumps(normalized_results, indent=2, ensure_ascii=False) return json_results except Exception as e: return f"Error: {str(e)}" @tool("web_fetch", parse_docstring=True) def web_fetch_tool(url: str) -> str: """Fetch the contents of a web page at a given URL. Only fetch EXACT URLs that have been provided directly by the user or have been returned in results from the web_search and web_fetch tools. This tool can NOT access content that requires authentication, such as private Google Docs or pages behind login walls. Do NOT add www. to URLs that do NOT have them. URLs must include the schema: https://example.com is a valid URL while example.com is an invalid URL. Args: url: The URL to fetch the contents of. """ try: client = _get_firecrawl_client() result = client.scrape(url, formats=["markdown"]) markdown_content = result.markdown or "" metadata = result.metadata title = metadata.title if metadata and metadata.title else "Untitled" if not markdown_content: return "Error: No content found" except Exception as e: return f"Error: {str(e)}" return f"# {title}\n\n{markdown_content[:4096]}"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/firecrawl/tools.py", "license": "MIT License", "lines": 58, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/community/image_search/tools.py
""" Image Search Tool - Search images using DuckDuckGo for reference in image generation. """ import json import logging from langchain.tools import tool from src.config import get_app_config logger = logging.getLogger(__name__) def _search_images( query: str, max_results: int = 5, region: str = "wt-wt", safesearch: str = "moderate", size: str | None = None, color: str | None = None, type_image: str | None = None, layout: str | None = None, license_image: str | None = None, ) -> list[dict]: """ Execute image search using DuckDuckGo. Args: query: Search keywords max_results: Maximum number of results region: Search region safesearch: Safe search level size: Image size (Small/Medium/Large/Wallpaper) color: Color filter type_image: Image type (photo/clipart/gif/transparent/line) layout: Layout (Square/Tall/Wide) license_image: License filter Returns: List of search results """ try: from ddgs import DDGS except ImportError: logger.error("ddgs library not installed. Run: pip install ddgs") return [] ddgs = DDGS(timeout=30) try: kwargs = { "region": region, "safesearch": safesearch, "max_results": max_results, } if size: kwargs["size"] = size if color: kwargs["color"] = color if type_image: kwargs["type_image"] = type_image if layout: kwargs["layout"] = layout if license_image: kwargs["license_image"] = license_image results = ddgs.images(query, **kwargs) return list(results) if results else [] except Exception as e: logger.error(f"Failed to search images: {e}") return [] @tool("image_search", parse_docstring=True) def image_search_tool( query: str, max_results: int = 5, size: str | None = None, type_image: str | None = None, layout: str | None = None, ) -> str: """Search for images online. Use this tool BEFORE image generation to find reference images for characters, portraits, objects, scenes, or any content requiring visual accuracy. **When to use:** - Before generating character/portrait images: search for similar poses, expressions, styles - Before generating specific objects/products: search for accurate visual references - Before generating scenes/locations: search for architectural or environmental references - Before generating fashion/clothing: search for style and detail references The returned image URLs can be used as reference images in image generation to significantly improve quality. Args: query: Search keywords describing the images you want to find. Be specific for better results (e.g., "Japanese woman street photography 1990s" instead of just "woman"). max_results: Maximum number of images to return. Default is 5. size: Image size filter. Options: "Small", "Medium", "Large", "Wallpaper". Use "Large" for reference images. type_image: Image type filter. Options: "photo", "clipart", "gif", "transparent", "line". Use "photo" for realistic references. layout: Layout filter. Options: "Square", "Tall", "Wide". Choose based on your generation needs. """ config = get_app_config().get_tool_config("image_search") # Override max_results from config if set if config is not None and "max_results" in config.model_extra: max_results = config.model_extra.get("max_results", max_results) results = _search_images( query=query, max_results=max_results, size=size, type_image=type_image, layout=layout, ) if not results: return json.dumps({"error": "No images found", "query": query}, ensure_ascii=False) normalized_results = [ { "title": r.get("title", ""), "image_url": r.get("thumbnail", ""), "thumbnail_url": r.get("thumbnail", ""), } for r in results ] output = { "query": query, "total_results": len(normalized_results), "results": normalized_results, "usage_hint": "Use the 'image_url' values as reference images in image generation. Download them first if needed.", } return json.dumps(output, indent=2, ensure_ascii=False)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/image_search/tools.py", "license": "MIT License", "lines": 111, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/community/jina_ai/jina_client.py
import logging import os import requests logger = logging.getLogger(__name__) class JinaClient: def crawl(self, url: str, return_format: str = "html", timeout: int = 10) -> str: headers = { "Content-Type": "application/json", "X-Return-Format": return_format, "X-Timeout": str(timeout), } if os.getenv("JINA_API_KEY"): headers["Authorization"] = f"Bearer {os.getenv('JINA_API_KEY')}" else: logger.warning("Jina API key is not set. Provide your own key to access a higher rate limit. See https://jina.ai/reader for more information.") data = {"url": url} try: response = requests.post("https://r.jina.ai/", headers=headers, json=data) if response.status_code != 200: error_message = f"Jina API returned status {response.status_code}: {response.text}" logger.error(error_message) return f"Error: {error_message}" if not response.text or not response.text.strip(): error_message = "Jina API returned empty response" logger.error(error_message) return f"Error: {error_message}" return response.text except Exception as e: error_message = f"Request to Jina API failed: {str(e)}" logger.error(error_message) return f"Error: {error_message}"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/jina_ai/jina_client.py", "license": "MIT License", "lines": 31, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/community/jina_ai/tools.py
from langchain.tools import tool from src.community.jina_ai.jina_client import JinaClient from src.config import get_app_config from src.utils.readability import ReadabilityExtractor readability_extractor = ReadabilityExtractor() @tool("web_fetch", parse_docstring=True) def web_fetch_tool(url: str) -> str: """Fetch the contents of a web page at a given URL. Only fetch EXACT URLs that have been provided directly by the user or have been returned in results from the web_search and web_fetch tools. This tool can NOT access content that requires authentication, such as private Google Docs or pages behind login walls. Do NOT add www. to URLs that do NOT have them. URLs must include the schema: https://example.com is a valid URL while example.com is an invalid URL. Args: url: The URL to fetch the contents of. """ jina_client = JinaClient() timeout = 10 config = get_app_config().get_tool_config("web_fetch") if config is not None and "timeout" in config.model_extra: timeout = config.model_extra.get("timeout") html_content = jina_client.crawl(url, return_format="html", timeout=timeout) article = readability_extractor.extract_article(html_content) return article.to_markdown()[:4096]
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/jina_ai/tools.py", "license": "MIT License", "lines": 23, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/community/tavily/tools.py
import json from langchain.tools import tool from tavily import TavilyClient from src.config import get_app_config def _get_tavily_client() -> TavilyClient: config = get_app_config().get_tool_config("web_search") api_key = None if config is not None and "api_key" in config.model_extra: api_key = config.model_extra.get("api_key") return TavilyClient(api_key=api_key) @tool("web_search", parse_docstring=True) def web_search_tool(query: str) -> str: """Search the web. Args: query: The query to search for. """ config = get_app_config().get_tool_config("web_search") max_results = 5 if config is not None and "max_results" in config.model_extra: max_results = config.model_extra.get("max_results") client = _get_tavily_client() res = client.search(query, max_results=max_results) normalized_results = [ { "title": result["title"], "url": result["url"], "snippet": result["content"], } for result in res["results"] ] json_results = json.dumps(normalized_results, indent=2, ensure_ascii=False) return json_results @tool("web_fetch", parse_docstring=True) def web_fetch_tool(url: str) -> str: """Fetch the contents of a web page at a given URL. Only fetch EXACT URLs that have been provided directly by the user or have been returned in results from the web_search and web_fetch tools. This tool can NOT access content that requires authentication, such as private Google Docs or pages behind login walls. Do NOT add www. to URLs that do NOT have them. URLs must include the schema: https://example.com is a valid URL while example.com is an invalid URL. Args: url: The URL to fetch the contents of. """ client = _get_tavily_client() res = client.extract([url]) if "failed_results" in res and len(res["failed_results"]) > 0: return f"Error: {res['failed_results'][0]['error']}" elif "results" in res and len(res["results"]) > 0: result = res["results"][0] return f"# {result['title']}\n\n{result['raw_content'][:4096]}" else: return "Error: No results found"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/community/tavily/tools.py", "license": "MIT License", "lines": 51, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/app_config.py
import os from pathlib import Path from typing import Any, Self import yaml from dotenv import load_dotenv from pydantic import BaseModel, ConfigDict, Field from src.config.extensions_config import ExtensionsConfig from src.config.memory_config import load_memory_config_from_dict from src.config.model_config import ModelConfig from src.config.sandbox_config import SandboxConfig from src.config.skills_config import SkillsConfig from src.config.subagents_config import load_subagents_config_from_dict from src.config.summarization_config import load_summarization_config_from_dict from src.config.title_config import load_title_config_from_dict from src.config.tool_config import ToolConfig, ToolGroupConfig load_dotenv() class AppConfig(BaseModel): """Config for the DeerFlow application""" models: list[ModelConfig] = Field(default_factory=list, description="Available models") sandbox: SandboxConfig = Field(description="Sandbox configuration") tools: list[ToolConfig] = Field(default_factory=list, description="Available tools") tool_groups: list[ToolGroupConfig] = Field(default_factory=list, description="Available tool groups") skills: SkillsConfig = Field(default_factory=SkillsConfig, description="Skills configuration") extensions: ExtensionsConfig = Field(default_factory=ExtensionsConfig, description="Extensions configuration (MCP servers and skills state)") model_config = ConfigDict(extra="allow", frozen=False) @classmethod def resolve_config_path(cls, config_path: str | None = None) -> Path: """Resolve the config file path. Priority: 1. If provided `config_path` argument, use it. 2. If provided `DEER_FLOW_CONFIG_PATH` environment variable, use it. 3. Otherwise, first check the `config.yaml` in the current directory, then fallback to `config.yaml` in the parent directory. """ if config_path: path = Path(config_path) if not Path.exists(path): raise FileNotFoundError(f"Config file specified by param `config_path` not found at {path}") return path elif os.getenv("DEER_FLOW_CONFIG_PATH"): path = Path(os.getenv("DEER_FLOW_CONFIG_PATH")) if not Path.exists(path): raise FileNotFoundError(f"Config file specified by environment variable `DEER_FLOW_CONFIG_PATH` not found at {path}") return path else: # Check if the config.yaml is in the current directory path = Path(os.getcwd()) / "config.yaml" if not path.exists(): # Check if the config.yaml is in the parent directory of CWD path = Path(os.getcwd()).parent / "config.yaml" if not path.exists(): raise FileNotFoundError("`config.yaml` file not found at the current directory nor its parent directory") return path @classmethod def from_file(cls, config_path: str | None = None) -> Self: """Load config from YAML file. See `resolve_config_path` for more details. Args: config_path: Path to the config file. Returns: AppConfig: The loaded config. """ resolved_path = cls.resolve_config_path(config_path) with open(resolved_path, encoding="utf-8") as f: config_data = yaml.safe_load(f) config_data = cls.resolve_env_variables(config_data) # Load title config if present if "title" in config_data: load_title_config_from_dict(config_data["title"]) # Load summarization config if present if "summarization" in config_data: load_summarization_config_from_dict(config_data["summarization"]) # Load memory config if present if "memory" in config_data: load_memory_config_from_dict(config_data["memory"]) # Load subagents config if present if "subagents" in config_data: load_subagents_config_from_dict(config_data["subagents"]) # Load extensions config separately (it's in a different file) extensions_config = ExtensionsConfig.from_file() config_data["extensions"] = extensions_config.model_dump() result = cls.model_validate(config_data) return result @classmethod def resolve_env_variables(cls, config: Any) -> Any: """Recursively resolve environment variables in the config. Environment variables are resolved using the `os.getenv` function. Example: $OPENAI_API_KEY Args: config: The config to resolve environment variables in. Returns: The config with environment variables resolved. """ if isinstance(config, str): if config.startswith("$"): env_value = os.getenv(config[1:]) if env_value is None: raise ValueError(f"Environment variable {config[1:]} not found for config value {config}") return env_value return config elif isinstance(config, dict): return {k: cls.resolve_env_variables(v) for k, v in config.items()} elif isinstance(config, list): return [cls.resolve_env_variables(item) for item in config] return config def get_model_config(self, name: str) -> ModelConfig | None: """Get the model config by name. Args: name: The name of the model to get the config for. Returns: The model config if found, otherwise None. """ return next((model for model in self.models if model.name == name), None) def get_tool_config(self, name: str) -> ToolConfig | None: """Get the tool config by name. Args: name: The name of the tool to get the config for. Returns: The tool config if found, otherwise None. """ return next((tool for tool in self.tools if tool.name == name), None) def get_tool_group_config(self, name: str) -> ToolGroupConfig | None: """Get the tool group config by name. Args: name: The name of the tool group to get the config for. Returns: The tool group config if found, otherwise None. """ return next((group for group in self.tool_groups if group.name == name), None) _app_config: AppConfig | None = None def get_app_config() -> AppConfig: """Get the DeerFlow config instance. Returns a cached singleton instance. Use `reload_app_config()` to reload from file, or `reset_app_config()` to clear the cache. """ global _app_config if _app_config is None: _app_config = AppConfig.from_file() return _app_config def reload_app_config(config_path: str | None = None) -> AppConfig: """Reload the config from file and update the cached instance. This is useful when the config file has been modified and you want to pick up the changes without restarting the application. Args: config_path: Optional path to config file. If not provided, uses the default resolution strategy. Returns: The newly loaded AppConfig instance. """ global _app_config _app_config = AppConfig.from_file(config_path) return _app_config def reset_app_config() -> None: """Reset the cached config instance. This clears the singleton cache, causing the next call to `get_app_config()` to reload from file. Useful for testing or when switching between different configurations. """ global _app_config _app_config = None def set_app_config(config: AppConfig) -> None: """Set a custom config instance. This allows injecting a custom or mock config for testing purposes. Args: config: The AppConfig instance to use. """ global _app_config _app_config = config
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/app_config.py", "license": "MIT License", "lines": 166, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/config/extensions_config.py
"""Unified extensions configuration for MCP servers and skills.""" import json import os from pathlib import Path from typing import Any, Literal from pydantic import BaseModel, ConfigDict, Field class McpOAuthConfig(BaseModel): """OAuth configuration for an MCP server (HTTP/SSE transports).""" enabled: bool = Field(default=True, description="Whether OAuth token injection is enabled") token_url: str = Field(description="OAuth token endpoint URL") grant_type: Literal["client_credentials", "refresh_token"] = Field( default="client_credentials", description="OAuth grant type", ) client_id: str | None = Field(default=None, description="OAuth client ID") client_secret: str | None = Field(default=None, description="OAuth client secret") refresh_token: str | None = Field(default=None, description="OAuth refresh token (for refresh_token grant)") scope: str | None = Field(default=None, description="OAuth scope") audience: str | None = Field(default=None, description="OAuth audience (provider-specific)") token_field: str = Field(default="access_token", description="Field name containing access token in token response") token_type_field: str = Field(default="token_type", description="Field name containing token type in token response") expires_in_field: str = Field(default="expires_in", description="Field name containing expiry (seconds) in token response") default_token_type: str = Field(default="Bearer", description="Default token type when missing in token response") refresh_skew_seconds: int = Field(default=60, description="Refresh token this many seconds before expiry") extra_token_params: dict[str, str] = Field(default_factory=dict, description="Additional form params sent to token endpoint") model_config = ConfigDict(extra="allow") class McpServerConfig(BaseModel): """Configuration for a single MCP server.""" enabled: bool = Field(default=True, description="Whether this MCP server is enabled") type: str = Field(default="stdio", description="Transport type: 'stdio', 'sse', or 'http'") command: str | None = Field(default=None, description="Command to execute to start the MCP server (for stdio type)") args: list[str] = Field(default_factory=list, description="Arguments to pass to the command (for stdio type)") env: dict[str, str] = Field(default_factory=dict, description="Environment variables for the MCP server") url: str | None = Field(default=None, description="URL of the MCP server (for sse or http type)") headers: dict[str, str] = Field(default_factory=dict, description="HTTP headers to send (for sse or http type)") oauth: McpOAuthConfig | None = Field(default=None, description="OAuth configuration (for sse or http type)") description: str = Field(default="", description="Human-readable description of what this MCP server provides") model_config = ConfigDict(extra="allow") class SkillStateConfig(BaseModel): """Configuration for a single skill's state.""" enabled: bool = Field(default=True, description="Whether this skill is enabled") class ExtensionsConfig(BaseModel): """Unified configuration for MCP servers and skills.""" mcp_servers: dict[str, McpServerConfig] = Field( default_factory=dict, description="Map of MCP server name to configuration", alias="mcpServers", ) skills: dict[str, SkillStateConfig] = Field( default_factory=dict, description="Map of skill name to state configuration", ) model_config = ConfigDict(extra="allow", populate_by_name=True) @classmethod def resolve_config_path(cls, config_path: str | None = None) -> Path | None: """Resolve the extensions config file path. Priority: 1. If provided `config_path` argument, use it. 2. If provided `DEER_FLOW_EXTENSIONS_CONFIG_PATH` environment variable, use it. 3. Otherwise, check for `extensions_config.json` in the current directory, then in the parent directory. 4. For backward compatibility, also check for `mcp_config.json` if `extensions_config.json` is not found. 5. If not found, return None (extensions are optional). Args: config_path: Optional path to extensions config file. Returns: Path to the extensions config file if found, otherwise None. """ if config_path: path = Path(config_path) if not path.exists(): raise FileNotFoundError(f"Extensions config file specified by param `config_path` not found at {path}") return path elif os.getenv("DEER_FLOW_EXTENSIONS_CONFIG_PATH"): path = Path(os.getenv("DEER_FLOW_EXTENSIONS_CONFIG_PATH")) if not path.exists(): raise FileNotFoundError(f"Extensions config file specified by environment variable `DEER_FLOW_EXTENSIONS_CONFIG_PATH` not found at {path}") return path else: # Check if the extensions_config.json is in the current directory path = Path(os.getcwd()) / "extensions_config.json" if path.exists(): return path # Check if the extensions_config.json is in the parent directory of CWD path = Path(os.getcwd()).parent / "extensions_config.json" if path.exists(): return path # Backward compatibility: check for mcp_config.json path = Path(os.getcwd()) / "mcp_config.json" if path.exists(): return path path = Path(os.getcwd()).parent / "mcp_config.json" if path.exists(): return path # Extensions are optional, so return None if not found return None @classmethod def from_file(cls, config_path: str | None = None) -> "ExtensionsConfig": """Load extensions config from JSON file. See `resolve_config_path` for more details. Args: config_path: Path to the extensions config file. Returns: ExtensionsConfig: The loaded config, or empty config if file not found. """ resolved_path = cls.resolve_config_path(config_path) if resolved_path is None: # Return empty config if extensions config file is not found return cls(mcp_servers={}, skills={}) with open(resolved_path, encoding="utf-8") as f: config_data = json.load(f) cls.resolve_env_variables(config_data) return cls.model_validate(config_data) @classmethod def resolve_env_variables(cls, config: dict[str, Any]) -> dict[str, Any]: """Recursively resolve environment variables in the config. Environment variables are resolved using the `os.getenv` function. Example: $OPENAI_API_KEY Args: config: The config to resolve environment variables in. Returns: The config with environment variables resolved. """ for key, value in config.items(): if isinstance(value, str): if value.startswith("$"): env_value = os.getenv(value[1:]) if env_value is None: raise ValueError(f"Environment variable {value[1:]} not found for config value {value}") config[key] = env_value else: config[key] = value elif isinstance(value, dict): config[key] = cls.resolve_env_variables(value) elif isinstance(value, list): config[key] = [cls.resolve_env_variables(item) if isinstance(item, dict) else item for item in value] return config def get_enabled_mcp_servers(self) -> dict[str, McpServerConfig]: """Get only the enabled MCP servers. Returns: Dictionary of enabled MCP servers. """ return {name: config for name, config in self.mcp_servers.items() if config.enabled} def is_skill_enabled(self, skill_name: str, skill_category: str) -> bool: """Check if a skill is enabled. Args: skill_name: Name of the skill skill_category: Category of the skill Returns: True if enabled, False otherwise """ skill_config = self.skills.get(skill_name) if skill_config is None: # Default to enable for public & custom skill return skill_category in ("public", "custom") return skill_config.enabled _extensions_config: ExtensionsConfig | None = None def get_extensions_config() -> ExtensionsConfig: """Get the extensions config instance. Returns a cached singleton instance. Use `reload_extensions_config()` to reload from file, or `reset_extensions_config()` to clear the cache. Returns: The cached ExtensionsConfig instance. """ global _extensions_config if _extensions_config is None: _extensions_config = ExtensionsConfig.from_file() return _extensions_config def reload_extensions_config(config_path: str | None = None) -> ExtensionsConfig: """Reload the extensions config from file and update the cached instance. This is useful when the config file has been modified and you want to pick up the changes without restarting the application. Args: config_path: Optional path to extensions config file. If not provided, uses the default resolution strategy. Returns: The newly loaded ExtensionsConfig instance. """ global _extensions_config _extensions_config = ExtensionsConfig.from_file(config_path) return _extensions_config def reset_extensions_config() -> None: """Reset the cached extensions config instance. This clears the singleton cache, causing the next call to `get_extensions_config()` to reload from file. Useful for testing or when switching between different configurations. """ global _extensions_config _extensions_config = None def set_extensions_config(config: ExtensionsConfig) -> None: """Set a custom extensions config instance. This allows injecting a custom or mock config for testing purposes. Args: config: The ExtensionsConfig instance to use. """ global _extensions_config _extensions_config = config
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/extensions_config.py", "license": "MIT License", "lines": 195, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/config/memory_config.py
"""Configuration for memory mechanism.""" from pydantic import BaseModel, Field class MemoryConfig(BaseModel): """Configuration for global memory mechanism.""" enabled: bool = Field( default=True, description="Whether to enable memory mechanism", ) storage_path: str = Field( default="", description=( "Path to store memory data. " "If empty, defaults to `{base_dir}/memory.json` (see Paths.memory_file). " "Absolute paths are used as-is. " "Relative paths are resolved against `Paths.base_dir` " "(not the backend working directory). " "Note: if you previously set this to `.deer-flow/memory.json`, " "the file will now be resolved as `{base_dir}/.deer-flow/memory.json`; " "migrate existing data or use an absolute path to preserve the old location." ), ) debounce_seconds: int = Field( default=30, ge=1, le=300, description="Seconds to wait before processing queued updates (debounce)", ) model_name: str | None = Field( default=None, description="Model name to use for memory updates (None = use default model)", ) max_facts: int = Field( default=100, ge=10, le=500, description="Maximum number of facts to store", ) fact_confidence_threshold: float = Field( default=0.7, ge=0.0, le=1.0, description="Minimum confidence threshold for storing facts", ) injection_enabled: bool = Field( default=True, description="Whether to inject memory into system prompt", ) max_injection_tokens: int = Field( default=2000, ge=100, le=8000, description="Maximum tokens to use for memory injection", ) # Global configuration instance _memory_config: MemoryConfig = MemoryConfig() def get_memory_config() -> MemoryConfig: """Get the current memory configuration.""" return _memory_config def set_memory_config(config: MemoryConfig) -> None: """Set the memory configuration.""" global _memory_config _memory_config = config def load_memory_config_from_dict(config_dict: dict) -> None: """Load memory configuration from a dictionary.""" global _memory_config _memory_config = MemoryConfig(**config_dict)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/memory_config.py", "license": "MIT License", "lines": 66, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/model_config.py
from pydantic import BaseModel, ConfigDict, Field class ModelConfig(BaseModel): """Config section for a model""" name: str = Field(..., description="Unique name for the model") display_name: str | None = Field(..., default_factory=lambda: None, description="Display name for the model") description: str | None = Field(..., default_factory=lambda: None, description="Description for the model") use: str = Field( ..., description="Class path of the model provider(e.g. langchain_openai.ChatOpenAI)", ) model: str = Field(..., description="Model name") model_config = ConfigDict(extra="allow") supports_thinking: bool = Field(default_factory=lambda: False, description="Whether the model supports thinking") supports_reasoning_effort: bool = Field(default_factory=lambda: False, description="Whether the model supports reasoning effort") when_thinking_enabled: dict | None = Field( default_factory=lambda: None, description="Extra settings to be passed to the model when thinking is enabled", ) supports_vision: bool = Field(default_factory=lambda: False, description="Whether the model supports vision/image inputs")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/model_config.py", "license": "MIT License", "lines": 19, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/sandbox_config.py
from pydantic import BaseModel, ConfigDict, Field class VolumeMountConfig(BaseModel): """Configuration for a volume mount.""" host_path: str = Field(..., description="Path on the host machine") container_path: str = Field(..., description="Path inside the container") read_only: bool = Field(default=False, description="Whether the mount is read-only") class SandboxConfig(BaseModel): """Config section for a sandbox. Common options: use: Class path of the sandbox provider (required) AioSandboxProvider specific options: image: Docker image to use (default: enterprise-public-cn-beijing.cr.volces.com/vefaas-public/all-in-one-sandbox:latest) port: Base port for sandbox containers (default: 8080) base_url: If set, uses existing sandbox instead of starting new container auto_start: Whether to automatically start Docker container (default: true) container_prefix: Prefix for container names (default: deer-flow-sandbox) idle_timeout: Idle timeout in seconds before sandbox is released (default: 600 = 10 minutes). Set to 0 to disable. mounts: List of volume mounts to share directories with the container environment: Environment variables to inject into the container (values starting with $ are resolved from host env) """ use: str = Field( ..., description="Class path of the sandbox provider (e.g. src.sandbox.local:LocalSandboxProvider)", ) image: str | None = Field( default=None, description="Docker image to use for the sandbox container", ) port: int | None = Field( default=None, description="Base port for sandbox containers", ) base_url: str | None = Field( default=None, description="If set, uses existing sandbox at this URL instead of starting new container", ) auto_start: bool | None = Field( default=None, description="Whether to automatically start Docker container", ) container_prefix: str | None = Field( default=None, description="Prefix for container names", ) idle_timeout: int | None = Field( default=None, description="Idle timeout in seconds before sandbox is released (default: 600 = 10 minutes). Set to 0 to disable.", ) mounts: list[VolumeMountConfig] = Field( default_factory=list, description="List of volume mounts to share directories between host and container", ) environment: dict[str, str] = Field( default_factory=dict, description="Environment variables to inject into the sandbox container. Values starting with $ will be resolved from host environment variables.", ) model_config = ConfigDict(extra="allow")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/sandbox_config.py", "license": "MIT License", "lines": 57, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/skills_config.py
from pathlib import Path from pydantic import BaseModel, Field class SkillsConfig(BaseModel): """Configuration for skills system""" path: str | None = Field( default=None, description="Path to skills directory. If not specified, defaults to ../skills relative to backend directory", ) container_path: str = Field( default="/mnt/skills", description="Path where skills are mounted in the sandbox container", ) def get_skills_path(self) -> Path: """ Get the resolved skills directory path. Returns: Path to the skills directory """ if self.path: # Use configured path (can be absolute or relative) path = Path(self.path) if not path.is_absolute(): # If relative, resolve from current working directory path = Path.cwd() / path return path.resolve() else: # Default: ../skills relative to backend directory from src.skills.loader import get_skills_root_path return get_skills_root_path() def get_skill_container_path(self, skill_name: str, category: str = "public") -> str: """ Get the full container path for a specific skill. Args: skill_name: Name of the skill (directory name) category: Category of the skill (public or custom) Returns: Full path to the skill in the container """ return f"{self.container_path}/{category}/{skill_name}"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/skills_config.py", "license": "MIT License", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/summarization_config.py
"""Configuration for conversation summarization.""" from typing import Literal from pydantic import BaseModel, Field ContextSizeType = Literal["fraction", "tokens", "messages"] class ContextSize(BaseModel): """Context size specification for trigger or keep parameters.""" type: ContextSizeType = Field(description="Type of context size specification") value: int | float = Field(description="Value for the context size specification") def to_tuple(self) -> tuple[ContextSizeType, int | float]: """Convert to tuple format expected by SummarizationMiddleware.""" return (self.type, self.value) class SummarizationConfig(BaseModel): """Configuration for automatic conversation summarization.""" enabled: bool = Field( default=False, description="Whether to enable automatic conversation summarization", ) model_name: str | None = Field( default=None, description="Model name to use for summarization (None = use a lightweight model)", ) trigger: ContextSize | list[ContextSize] | None = Field( default=None, description="One or more thresholds that trigger summarization. When any threshold is met, summarization runs. " "Examples: {'type': 'messages', 'value': 50} triggers at 50 messages, " "{'type': 'tokens', 'value': 4000} triggers at 4000 tokens, " "{'type': 'fraction', 'value': 0.8} triggers at 80% of model's max input tokens", ) keep: ContextSize = Field( default_factory=lambda: ContextSize(type="messages", value=20), description="Context retention policy after summarization. Specifies how much history to preserve. " "Examples: {'type': 'messages', 'value': 20} keeps 20 messages, " "{'type': 'tokens', 'value': 3000} keeps 3000 tokens, " "{'type': 'fraction', 'value': 0.3} keeps 30% of model's max input tokens", ) trim_tokens_to_summarize: int | None = Field( default=4000, description="Maximum tokens to keep when preparing messages for summarization. Pass null to skip trimming.", ) summary_prompt: str | None = Field( default=None, description="Custom prompt template for generating summaries. If not provided, uses the default LangChain prompt.", ) # Global configuration instance _summarization_config: SummarizationConfig = SummarizationConfig() def get_summarization_config() -> SummarizationConfig: """Get the current summarization configuration.""" return _summarization_config def set_summarization_config(config: SummarizationConfig) -> None: """Set the summarization configuration.""" global _summarization_config _summarization_config = config def load_summarization_config_from_dict(config_dict: dict) -> None: """Load summarization configuration from a dictionary.""" global _summarization_config _summarization_config = SummarizationConfig(**config_dict)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/summarization_config.py", "license": "MIT License", "lines": 56, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/title_config.py
"""Configuration for automatic thread title generation.""" from pydantic import BaseModel, Field class TitleConfig(BaseModel): """Configuration for automatic thread title generation.""" enabled: bool = Field( default=True, description="Whether to enable automatic title generation", ) max_words: int = Field( default=6, ge=1, le=20, description="Maximum number of words in the generated title", ) max_chars: int = Field( default=60, ge=10, le=200, description="Maximum number of characters in the generated title", ) model_name: str | None = Field( default=None, description="Model name to use for title generation (None = use default model)", ) prompt_template: str = Field( default=("Generate a concise title (max {max_words} words) for this conversation.\nUser: {user_msg}\nAssistant: {assistant_msg}\n\nReturn ONLY the title, no quotes, no explanation."), description="Prompt template for title generation", ) # Global configuration instance _title_config: TitleConfig = TitleConfig() def get_title_config() -> TitleConfig: """Get the current title configuration.""" return _title_config def set_title_config(config: TitleConfig) -> None: """Set the title configuration.""" global _title_config _title_config = config def load_title_config_from_dict(config_dict: dict) -> None: """Load title configuration from a dictionary.""" global _title_config _title_config = TitleConfig(**config_dict)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/title_config.py", "license": "MIT License", "lines": 41, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/config/tool_config.py
from pydantic import BaseModel, ConfigDict, Field class ToolGroupConfig(BaseModel): """Config section for a tool group""" name: str = Field(..., description="Unique name for the tool group") model_config = ConfigDict(extra="allow") class ToolConfig(BaseModel): """Config section for a tool""" name: str = Field(..., description="Unique name for the tool") group: str = Field(..., description="Group name for the tool") use: str = Field( ..., description="Variable name of the tool provider(e.g. src.sandbox.tools:bash_tool)", ) model_config = ConfigDict(extra="allow")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/config/tool_config.py", "license": "MIT License", "lines": 14, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/gateway/app.py
import logging import sys from collections.abc import AsyncGenerator from contextlib import asynccontextmanager from fastapi import FastAPI from src.config.app_config import get_app_config from src.gateway.config import get_gateway_config from src.gateway.routers import ( agents, artifacts, mcp, memory, models, skills, suggestions, uploads, ) # Configure logging logging.basicConfig( level=logging.INFO, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", ) logger = logging.getLogger(__name__) @asynccontextmanager async def lifespan(app: FastAPI) -> AsyncGenerator[None, None]: """Application lifespan handler.""" # Load config and check necessary environment variables at startup try: get_app_config() logger.info("Configuration loaded successfully") except Exception as e: logger.error(f"Failed to load configuration: {e}") sys.exit(1) config = get_gateway_config() logger.info(f"Starting API Gateway on {config.host}:{config.port}") # NOTE: MCP tools initialization is NOT done here because: # 1. Gateway doesn't use MCP tools - they are used by Agents in the LangGraph Server # 2. Gateway and LangGraph Server are separate processes with independent caches # MCP tools are lazily initialized in LangGraph Server when first needed yield logger.info("Shutting down API Gateway") def create_app() -> FastAPI: """Create and configure the FastAPI application. Returns: Configured FastAPI application instance. """ app = FastAPI( title="DeerFlow API Gateway", description=""" ## DeerFlow API Gateway API Gateway for DeerFlow - A LangGraph-based AI agent backend with sandbox execution capabilities. ### Features - **Models Management**: Query and retrieve available AI models - **MCP Configuration**: Manage Model Context Protocol (MCP) server configurations - **Memory Management**: Access and manage global memory data for personalized conversations - **Skills Management**: Query and manage skills and their enabled status - **Artifacts**: Access thread artifacts and generated files - **Health Monitoring**: System health check endpoints ### Architecture LangGraph requests are handled by nginx reverse proxy. This gateway provides custom endpoints for models, MCP configuration, skills, and artifacts. """, version="0.1.0", lifespan=lifespan, docs_url="/docs", redoc_url="/redoc", openapi_url="/openapi.json", openapi_tags=[ { "name": "models", "description": "Operations for querying available AI models and their configurations", }, { "name": "mcp", "description": "Manage Model Context Protocol (MCP) server configurations", }, { "name": "memory", "description": "Access and manage global memory data for personalized conversations", }, { "name": "skills", "description": "Manage skills and their configurations", }, { "name": "artifacts", "description": "Access and download thread artifacts and generated files", }, { "name": "uploads", "description": "Upload and manage user files for threads", }, { "name": "agents", "description": "Create and manage custom agents with per-agent config and prompts", }, { "name": "suggestions", "description": "Generate follow-up question suggestions for conversations", }, { "name": "health", "description": "Health check and system status endpoints", }, ], ) # CORS is handled by nginx - no need for FastAPI middleware # Include routers # Models API is mounted at /api/models app.include_router(models.router) # MCP API is mounted at /api/mcp app.include_router(mcp.router) # Memory API is mounted at /api/memory app.include_router(memory.router) # Skills API is mounted at /api/skills app.include_router(skills.router) # Artifacts API is mounted at /api/threads/{thread_id}/artifacts app.include_router(artifacts.router) # Uploads API is mounted at /api/threads/{thread_id}/uploads app.include_router(uploads.router) # Agents API is mounted at /api/agents app.include_router(agents.router) # Suggestions API is mounted at /api/threads/{thread_id}/suggestions app.include_router(suggestions.router) @app.get("/health", tags=["health"]) async def health_check() -> dict: """Health check endpoint. Returns: Service health status information. """ return {"status": "healthy", "service": "deer-flow-gateway"} return app # Create app instance for uvicorn app = create_app()
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/app.py", "license": "MIT License", "lines": 135, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/gateway/config.py
import os from pydantic import BaseModel, Field class GatewayConfig(BaseModel): """Configuration for the API Gateway.""" host: str = Field(default="0.0.0.0", description="Host to bind the gateway server") port: int = Field(default=8001, description="Port to bind the gateway server") cors_origins: list[str] = Field(default_factory=lambda: ["http://localhost:3000"], description="Allowed CORS origins") _gateway_config: GatewayConfig | None = None def get_gateway_config() -> GatewayConfig: """Get gateway config, loading from environment if available.""" global _gateway_config if _gateway_config is None: cors_origins_str = os.getenv("CORS_ORIGINS", "http://localhost:3000") _gateway_config = GatewayConfig( host=os.getenv("GATEWAY_HOST", "0.0.0.0"), port=int(os.getenv("GATEWAY_PORT", "8001")), cors_origins=cors_origins_str.split(","), ) return _gateway_config
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/config.py", "license": "MIT License", "lines": 19, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/gateway/path_utils.py
"""Shared path resolution for thread virtual paths (e.g. mnt/user-data/outputs/...).""" from pathlib import Path from fastapi import HTTPException from src.config.paths import get_paths def resolve_thread_virtual_path(thread_id: str, virtual_path: str) -> Path: """Resolve a virtual path to the actual filesystem path under thread user-data. Args: thread_id: The thread ID. virtual_path: The virtual path as seen inside the sandbox (e.g., /mnt/user-data/outputs/file.txt). Returns: The resolved filesystem path. Raises: HTTPException: If the path is invalid or outside allowed directories. """ try: return get_paths().resolve_virtual_path(thread_id, virtual_path) except ValueError as e: status = 403 if "traversal" in str(e) else 400 raise HTTPException(status_code=status, detail=str(e))
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/path_utils.py", "license": "MIT License", "lines": 20, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/gateway/routers/artifacts.py
import logging import mimetypes import zipfile from pathlib import Path from urllib.parse import quote from fastapi import APIRouter, HTTPException, Request from fastapi.responses import FileResponse, HTMLResponse, PlainTextResponse, Response from src.gateway.path_utils import resolve_thread_virtual_path logger = logging.getLogger(__name__) router = APIRouter(prefix="/api", tags=["artifacts"]) def is_text_file_by_content(path: Path, sample_size: int = 8192) -> bool: """Check if file is text by examining content for null bytes.""" try: with open(path, "rb") as f: chunk = f.read(sample_size) # Text files shouldn't contain null bytes return b"\x00" not in chunk except Exception: return False def _extract_file_from_skill_archive(zip_path: Path, internal_path: str) -> bytes | None: """Extract a file from a .skill ZIP archive. Args: zip_path: Path to the .skill file (ZIP archive). internal_path: Path to the file inside the archive (e.g., "SKILL.md"). Returns: The file content as bytes, or None if not found. """ if not zipfile.is_zipfile(zip_path): return None try: with zipfile.ZipFile(zip_path, "r") as zip_ref: # List all files in the archive namelist = zip_ref.namelist() # Try direct path first if internal_path in namelist: return zip_ref.read(internal_path) # Try with any top-level directory prefix (e.g., "skill-name/SKILL.md") for name in namelist: if name.endswith("/" + internal_path) or name == internal_path: return zip_ref.read(name) # Not found return None except (zipfile.BadZipFile, KeyError): return None @router.get( "/threads/{thread_id}/artifacts/{path:path}", summary="Get Artifact File", description="Retrieve an artifact file generated by the AI agent. Supports text, HTML, and binary files.", ) async def get_artifact(thread_id: str, path: str, request: Request) -> FileResponse: """Get an artifact file by its path. The endpoint automatically detects file types and returns appropriate content types. Use the `?download=true` query parameter to force file download. Args: thread_id: The thread ID. path: The artifact path with virtual prefix (e.g., mnt/user-data/outputs/file.txt). request: FastAPI request object (automatically injected). Returns: The file content as a FileResponse with appropriate content type: - HTML files: Rendered as HTML - Text files: Plain text with proper MIME type - Binary files: Inline display with download option Raises: HTTPException: - 400 if path is invalid or not a file - 403 if access denied (path traversal detected) - 404 if file not found Query Parameters: download (bool): If true, returns file as attachment for download Example: - Get HTML file: `/api/threads/abc123/artifacts/mnt/user-data/outputs/index.html` - Download file: `/api/threads/abc123/artifacts/mnt/user-data/outputs/data.csv?download=true` """ # Check if this is a request for a file inside a .skill archive (e.g., xxx.skill/SKILL.md) if ".skill/" in path: # Split the path at ".skill/" to get the ZIP file path and internal path skill_marker = ".skill/" marker_pos = path.find(skill_marker) skill_file_path = path[: marker_pos + len(".skill")] # e.g., "mnt/user-data/outputs/my-skill.skill" internal_path = path[marker_pos + len(skill_marker) :] # e.g., "SKILL.md" actual_skill_path = resolve_thread_virtual_path(thread_id, skill_file_path) if not actual_skill_path.exists(): raise HTTPException(status_code=404, detail=f"Skill file not found: {skill_file_path}") if not actual_skill_path.is_file(): raise HTTPException(status_code=400, detail=f"Path is not a file: {skill_file_path}") # Extract the file from the .skill archive content = _extract_file_from_skill_archive(actual_skill_path, internal_path) if content is None: raise HTTPException(status_code=404, detail=f"File '{internal_path}' not found in skill archive") # Determine MIME type based on the internal file mime_type, _ = mimetypes.guess_type(internal_path) # Add cache headers to avoid repeated ZIP extraction (cache for 5 minutes) cache_headers = {"Cache-Control": "private, max-age=300"} if mime_type and mime_type.startswith("text/"): return PlainTextResponse(content=content.decode("utf-8"), media_type=mime_type, headers=cache_headers) # Default to plain text for unknown types that look like text try: return PlainTextResponse(content=content.decode("utf-8"), media_type="text/plain", headers=cache_headers) except UnicodeDecodeError: return Response(content=content, media_type=mime_type or "application/octet-stream", headers=cache_headers) actual_path = resolve_thread_virtual_path(thread_id, path) logger.info(f"Resolving artifact path: thread_id={thread_id}, requested_path={path}, actual_path={actual_path}") if not actual_path.exists(): raise HTTPException(status_code=404, detail=f"Artifact not found: {path}") if not actual_path.is_file(): raise HTTPException(status_code=400, detail=f"Path is not a file: {path}") mime_type, _ = mimetypes.guess_type(actual_path) # Encode filename for Content-Disposition header (RFC 5987) encoded_filename = quote(actual_path.name) # if `download` query parameter is true, return the file as a download if request.query_params.get("download"): return FileResponse(path=actual_path, filename=actual_path.name, media_type=mime_type, headers={"Content-Disposition": f"attachment; filename*=UTF-8''{encoded_filename}"}) if mime_type and mime_type == "text/html": return HTMLResponse(content=actual_path.read_text()) if mime_type and mime_type.startswith("text/"): return PlainTextResponse(content=actual_path.read_text(), media_type=mime_type) if is_text_file_by_content(actual_path): return PlainTextResponse(content=actual_path.read_text(), media_type=mime_type) return Response(content=actual_path.read_bytes(), media_type=mime_type, headers={"Content-Disposition": f"inline; filename*=UTF-8''{encoded_filename}"})
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/artifacts.py", "license": "MIT License", "lines": 119, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/gateway/routers/mcp.py
import json import logging from pathlib import Path from typing import Literal from fastapi import APIRouter, HTTPException from pydantic import BaseModel, Field from src.config.extensions_config import ExtensionsConfig, get_extensions_config, reload_extensions_config logger = logging.getLogger(__name__) router = APIRouter(prefix="/api", tags=["mcp"]) class McpOAuthConfigResponse(BaseModel): """OAuth configuration for an MCP server.""" enabled: bool = Field(default=True, description="Whether OAuth token injection is enabled") token_url: str = Field(default="", description="OAuth token endpoint URL") grant_type: Literal["client_credentials", "refresh_token"] = Field(default="client_credentials", description="OAuth grant type") client_id: str | None = Field(default=None, description="OAuth client ID") client_secret: str | None = Field(default=None, description="OAuth client secret") refresh_token: str | None = Field(default=None, description="OAuth refresh token") scope: str | None = Field(default=None, description="OAuth scope") audience: str | None = Field(default=None, description="OAuth audience") token_field: str = Field(default="access_token", description="Token response field containing access token") token_type_field: str = Field(default="token_type", description="Token response field containing token type") expires_in_field: str = Field(default="expires_in", description="Token response field containing expires-in seconds") default_token_type: str = Field(default="Bearer", description="Default token type when response omits token_type") refresh_skew_seconds: int = Field(default=60, description="Refresh this many seconds before expiry") extra_token_params: dict[str, str] = Field(default_factory=dict, description="Additional form params sent to token endpoint") class McpServerConfigResponse(BaseModel): """Response model for MCP server configuration.""" enabled: bool = Field(default=True, description="Whether this MCP server is enabled") type: str = Field(default="stdio", description="Transport type: 'stdio', 'sse', or 'http'") command: str | None = Field(default=None, description="Command to execute to start the MCP server (for stdio type)") args: list[str] = Field(default_factory=list, description="Arguments to pass to the command (for stdio type)") env: dict[str, str] = Field(default_factory=dict, description="Environment variables for the MCP server") url: str | None = Field(default=None, description="URL of the MCP server (for sse or http type)") headers: dict[str, str] = Field(default_factory=dict, description="HTTP headers to send (for sse or http type)") oauth: McpOAuthConfigResponse | None = Field(default=None, description="OAuth configuration for MCP HTTP/SSE servers") description: str = Field(default="", description="Human-readable description of what this MCP server provides") class McpConfigResponse(BaseModel): """Response model for MCP configuration.""" mcp_servers: dict[str, McpServerConfigResponse] = Field( default_factory=dict, description="Map of MCP server name to configuration", ) class McpConfigUpdateRequest(BaseModel): """Request model for updating MCP configuration.""" mcp_servers: dict[str, McpServerConfigResponse] = Field( ..., description="Map of MCP server name to configuration", ) @router.get( "/mcp/config", response_model=McpConfigResponse, summary="Get MCP Configuration", description="Retrieve the current Model Context Protocol (MCP) server configurations.", ) async def get_mcp_configuration() -> McpConfigResponse: """Get the current MCP configuration. Returns: The current MCP configuration with all servers. Example: ```json { "mcp_servers": { "github": { "enabled": true, "command": "npx", "args": ["-y", "@modelcontextprotocol/server-github"], "env": {"GITHUB_TOKEN": "ghp_xxx"}, "description": "GitHub MCP server for repository operations" } } } ``` """ config = get_extensions_config() return McpConfigResponse(mcp_servers={name: McpServerConfigResponse(**server.model_dump()) for name, server in config.mcp_servers.items()}) @router.put( "/mcp/config", response_model=McpConfigResponse, summary="Update MCP Configuration", description="Update Model Context Protocol (MCP) server configurations and save to file.", ) async def update_mcp_configuration(request: McpConfigUpdateRequest) -> McpConfigResponse: """Update the MCP configuration. This will: 1. Save the new configuration to the mcp_config.json file 2. Reload the configuration cache 3. Reset MCP tools cache to trigger reinitialization Args: request: The new MCP configuration to save. Returns: The updated MCP configuration. Raises: HTTPException: 500 if the configuration file cannot be written. Example Request: ```json { "mcp_servers": { "github": { "enabled": true, "command": "npx", "args": ["-y", "@modelcontextprotocol/server-github"], "env": {"GITHUB_TOKEN": "$GITHUB_TOKEN"}, "description": "GitHub MCP server for repository operations" } } } ``` """ try: # Get the current config path (or determine where to save it) config_path = ExtensionsConfig.resolve_config_path() # If no config file exists, create one in the parent directory (project root) if config_path is None: config_path = Path.cwd().parent / "extensions_config.json" logger.info(f"No existing extensions config found. Creating new config at: {config_path}") # Load current config to preserve skills configuration current_config = get_extensions_config() # Convert request to dict format for JSON serialization config_data = { "mcpServers": {name: server.model_dump() for name, server in request.mcp_servers.items()}, "skills": {name: {"enabled": skill.enabled} for name, skill in current_config.skills.items()}, } # Write the configuration to file with open(config_path, "w") as f: json.dump(config_data, f, indent=2) logger.info(f"MCP configuration updated and saved to: {config_path}") # NOTE: No need to reload/reset cache here - LangGraph Server (separate process) # will detect config file changes via mtime and reinitialize MCP tools automatically # Reload the configuration and update the global cache reloaded_config = reload_extensions_config() return McpConfigResponse(mcp_servers={name: McpServerConfigResponse(**server.model_dump()) for name, server in reloaded_config.mcp_servers.items()}) except Exception as e: logger.error(f"Failed to update MCP configuration: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to update MCP configuration: {str(e)}")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/mcp.py", "license": "MIT License", "lines": 134, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/gateway/routers/memory.py
"""Memory API router for retrieving and managing global memory data.""" from fastapi import APIRouter from pydantic import BaseModel, Field from src.agents.memory.updater import get_memory_data, reload_memory_data from src.config.memory_config import get_memory_config router = APIRouter(prefix="/api", tags=["memory"]) class ContextSection(BaseModel): """Model for context sections (user and history).""" summary: str = Field(default="", description="Summary content") updatedAt: str = Field(default="", description="Last update timestamp") class UserContext(BaseModel): """Model for user context.""" workContext: ContextSection = Field(default_factory=ContextSection) personalContext: ContextSection = Field(default_factory=ContextSection) topOfMind: ContextSection = Field(default_factory=ContextSection) class HistoryContext(BaseModel): """Model for history context.""" recentMonths: ContextSection = Field(default_factory=ContextSection) earlierContext: ContextSection = Field(default_factory=ContextSection) longTermBackground: ContextSection = Field(default_factory=ContextSection) class Fact(BaseModel): """Model for a memory fact.""" id: str = Field(..., description="Unique identifier for the fact") content: str = Field(..., description="Fact content") category: str = Field(default="context", description="Fact category") confidence: float = Field(default=0.5, description="Confidence score (0-1)") createdAt: str = Field(default="", description="Creation timestamp") source: str = Field(default="unknown", description="Source thread ID") class MemoryResponse(BaseModel): """Response model for memory data.""" version: str = Field(default="1.0", description="Memory schema version") lastUpdated: str = Field(default="", description="Last update timestamp") user: UserContext = Field(default_factory=UserContext) history: HistoryContext = Field(default_factory=HistoryContext) facts: list[Fact] = Field(default_factory=list) class MemoryConfigResponse(BaseModel): """Response model for memory configuration.""" enabled: bool = Field(..., description="Whether memory is enabled") storage_path: str = Field(..., description="Path to memory storage file") debounce_seconds: int = Field(..., description="Debounce time for memory updates") max_facts: int = Field(..., description="Maximum number of facts to store") fact_confidence_threshold: float = Field(..., description="Minimum confidence threshold for facts") injection_enabled: bool = Field(..., description="Whether memory injection is enabled") max_injection_tokens: int = Field(..., description="Maximum tokens for memory injection") class MemoryStatusResponse(BaseModel): """Response model for memory status.""" config: MemoryConfigResponse data: MemoryResponse @router.get( "/memory", response_model=MemoryResponse, summary="Get Memory Data", description="Retrieve the current global memory data including user context, history, and facts.", ) async def get_memory() -> MemoryResponse: """Get the current global memory data. Returns: The current memory data with user context, history, and facts. Example Response: ```json { "version": "1.0", "lastUpdated": "2024-01-15T10:30:00Z", "user": { "workContext": {"summary": "Working on DeerFlow project", "updatedAt": "..."}, "personalContext": {"summary": "Prefers concise responses", "updatedAt": "..."}, "topOfMind": {"summary": "Building memory API", "updatedAt": "..."} }, "history": { "recentMonths": {"summary": "Recent development activities", "updatedAt": "..."}, "earlierContext": {"summary": "", "updatedAt": ""}, "longTermBackground": {"summary": "", "updatedAt": ""} }, "facts": [ { "id": "fact_abc123", "content": "User prefers TypeScript over JavaScript", "category": "preference", "confidence": 0.9, "createdAt": "2024-01-15T10:30:00Z", "source": "thread_xyz" } ] } ``` """ memory_data = get_memory_data() return MemoryResponse(**memory_data) @router.post( "/memory/reload", response_model=MemoryResponse, summary="Reload Memory Data", description="Reload memory data from the storage file, refreshing the in-memory cache.", ) async def reload_memory() -> MemoryResponse: """Reload memory data from file. This forces a reload of the memory data from the storage file, useful when the file has been modified externally. Returns: The reloaded memory data. """ memory_data = reload_memory_data() return MemoryResponse(**memory_data) @router.get( "/memory/config", response_model=MemoryConfigResponse, summary="Get Memory Configuration", description="Retrieve the current memory system configuration.", ) async def get_memory_config_endpoint() -> MemoryConfigResponse: """Get the memory system configuration. Returns: The current memory configuration settings. Example Response: ```json { "enabled": true, "storage_path": ".deer-flow/memory.json", "debounce_seconds": 30, "max_facts": 100, "fact_confidence_threshold": 0.7, "injection_enabled": true, "max_injection_tokens": 2000 } ``` """ config = get_memory_config() return MemoryConfigResponse( enabled=config.enabled, storage_path=config.storage_path, debounce_seconds=config.debounce_seconds, max_facts=config.max_facts, fact_confidence_threshold=config.fact_confidence_threshold, injection_enabled=config.injection_enabled, max_injection_tokens=config.max_injection_tokens, ) @router.get( "/memory/status", response_model=MemoryStatusResponse, summary="Get Memory Status", description="Retrieve both memory configuration and current data in a single request.", ) async def get_memory_status() -> MemoryStatusResponse: """Get the memory system status including configuration and data. Returns: Combined memory configuration and current data. """ config = get_memory_config() memory_data = get_memory_data() return MemoryStatusResponse( config=MemoryConfigResponse( enabled=config.enabled, storage_path=config.storage_path, debounce_seconds=config.debounce_seconds, max_facts=config.max_facts, fact_confidence_threshold=config.fact_confidence_threshold, injection_enabled=config.injection_enabled, max_injection_tokens=config.max_injection_tokens, ), data=MemoryResponse(**memory_data), )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/memory.py", "license": "MIT License", "lines": 161, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/gateway/routers/models.py
from fastapi import APIRouter, HTTPException from pydantic import BaseModel, Field from src.config import get_app_config router = APIRouter(prefix="/api", tags=["models"]) class ModelResponse(BaseModel): """Response model for model information.""" name: str = Field(..., description="Unique identifier for the model") display_name: str | None = Field(None, description="Human-readable name") description: str | None = Field(None, description="Model description") supports_thinking: bool = Field(default=False, description="Whether model supports thinking mode") supports_reasoning_effort: bool = Field(default=False, description="Whether model supports reasoning effort") class ModelsListResponse(BaseModel): """Response model for listing all models.""" models: list[ModelResponse] @router.get( "/models", response_model=ModelsListResponse, summary="List All Models", description="Retrieve a list of all available AI models configured in the system.", ) async def list_models() -> ModelsListResponse: """List all available models from configuration. Returns model information suitable for frontend display, excluding sensitive fields like API keys and internal configuration. Returns: A list of all configured models with their metadata. Example Response: ```json { "models": [ { "name": "gpt-4", "display_name": "GPT-4", "description": "OpenAI GPT-4 model", "supports_thinking": false }, { "name": "claude-3-opus", "display_name": "Claude 3 Opus", "description": "Anthropic Claude 3 Opus model", "supports_thinking": true } ] } ``` """ config = get_app_config() models = [ ModelResponse( name=model.name, display_name=model.display_name, description=model.description, supports_thinking=model.supports_thinking, supports_reasoning_effort=model.supports_reasoning_effort, ) for model in config.models ] return ModelsListResponse(models=models) @router.get( "/models/{model_name}", response_model=ModelResponse, summary="Get Model Details", description="Retrieve detailed information about a specific AI model by its name.", ) async def get_model(model_name: str) -> ModelResponse: """Get a specific model by name. Args: model_name: The unique name of the model to retrieve. Returns: Model information if found. Raises: HTTPException: 404 if model not found. Example Response: ```json { "name": "gpt-4", "display_name": "GPT-4", "description": "OpenAI GPT-4 model", "supports_thinking": false } ``` """ config = get_app_config() model = config.get_model_config(model_name) if model is None: raise HTTPException(status_code=404, detail=f"Model '{model_name}' not found") return ModelResponse( name=model.name, display_name=model.display_name, description=model.description, supports_thinking=model.supports_thinking, supports_reasoning_effort=model.supports_reasoning_effort, )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/models.py", "license": "MIT License", "lines": 93, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/gateway/routers/skills.py
import json import logging import re import shutil import tempfile import zipfile from pathlib import Path import yaml from fastapi import APIRouter, HTTPException from pydantic import BaseModel, Field from src.config.extensions_config import ExtensionsConfig, SkillStateConfig, get_extensions_config, reload_extensions_config from src.gateway.path_utils import resolve_thread_virtual_path from src.skills import Skill, load_skills from src.skills.loader import get_skills_root_path logger = logging.getLogger(__name__) router = APIRouter(prefix="/api", tags=["skills"]) class SkillResponse(BaseModel): """Response model for skill information.""" name: str = Field(..., description="Name of the skill") description: str = Field(..., description="Description of what the skill does") license: str | None = Field(None, description="License information") category: str = Field(..., description="Category of the skill (public or custom)") enabled: bool = Field(default=True, description="Whether this skill is enabled") class SkillsListResponse(BaseModel): """Response model for listing all skills.""" skills: list[SkillResponse] class SkillUpdateRequest(BaseModel): """Request model for updating a skill.""" enabled: bool = Field(..., description="Whether to enable or disable the skill") class SkillInstallRequest(BaseModel): """Request model for installing a skill from a .skill file.""" thread_id: str = Field(..., description="The thread ID where the .skill file is located") path: str = Field(..., description="Virtual path to the .skill file (e.g., mnt/user-data/outputs/my-skill.skill)") class SkillInstallResponse(BaseModel): """Response model for skill installation.""" success: bool = Field(..., description="Whether the installation was successful") skill_name: str = Field(..., description="Name of the installed skill") message: str = Field(..., description="Installation result message") # Allowed properties in SKILL.md frontmatter ALLOWED_FRONTMATTER_PROPERTIES = {"name", "description", "license", "allowed-tools", "metadata"} def _validate_skill_frontmatter(skill_dir: Path) -> tuple[bool, str, str | None]: """Validate a skill directory's SKILL.md frontmatter. Args: skill_dir: Path to the skill directory containing SKILL.md. Returns: Tuple of (is_valid, message, skill_name). """ skill_md = skill_dir / "SKILL.md" if not skill_md.exists(): return False, "SKILL.md not found", None content = skill_md.read_text() if not content.startswith("---"): return False, "No YAML frontmatter found", None # Extract frontmatter match = re.match(r"^---\n(.*?)\n---", content, re.DOTALL) if not match: return False, "Invalid frontmatter format", None frontmatter_text = match.group(1) # Parse YAML frontmatter try: frontmatter = yaml.safe_load(frontmatter_text) if not isinstance(frontmatter, dict): return False, "Frontmatter must be a YAML dictionary", None except yaml.YAMLError as e: return False, f"Invalid YAML in frontmatter: {e}", None # Check for unexpected properties unexpected_keys = set(frontmatter.keys()) - ALLOWED_FRONTMATTER_PROPERTIES if unexpected_keys: return False, f"Unexpected key(s) in SKILL.md frontmatter: {', '.join(sorted(unexpected_keys))}", None # Check required fields if "name" not in frontmatter: return False, "Missing 'name' in frontmatter", None if "description" not in frontmatter: return False, "Missing 'description' in frontmatter", None # Validate name name = frontmatter.get("name", "") if not isinstance(name, str): return False, f"Name must be a string, got {type(name).__name__}", None name = name.strip() if not name: return False, "Name cannot be empty", None # Check naming convention (hyphen-case: lowercase with hyphens) if not re.match(r"^[a-z0-9-]+$", name): return False, f"Name '{name}' should be hyphen-case (lowercase letters, digits, and hyphens only)", None if name.startswith("-") or name.endswith("-") or "--" in name: return False, f"Name '{name}' cannot start/end with hyphen or contain consecutive hyphens", None if len(name) > 64: return False, f"Name is too long ({len(name)} characters). Maximum is 64 characters.", None # Validate description description = frontmatter.get("description", "") if not isinstance(description, str): return False, f"Description must be a string, got {type(description).__name__}", None description = description.strip() if description: if "<" in description or ">" in description: return False, "Description cannot contain angle brackets (< or >)", None if len(description) > 1024: return False, f"Description is too long ({len(description)} characters). Maximum is 1024 characters.", None return True, "Skill is valid!", name def _skill_to_response(skill: Skill) -> SkillResponse: """Convert a Skill object to a SkillResponse.""" return SkillResponse( name=skill.name, description=skill.description, license=skill.license, category=skill.category, enabled=skill.enabled, ) @router.get( "/skills", response_model=SkillsListResponse, summary="List All Skills", description="Retrieve a list of all available skills from both public and custom directories.", ) async def list_skills() -> SkillsListResponse: """List all available skills. Returns all skills regardless of their enabled status. Returns: A list of all skills with their metadata. Example Response: ```json { "skills": [ { "name": "PDF Processing", "description": "Extract and analyze PDF content", "license": "MIT", "category": "public", "enabled": true }, { "name": "Frontend Design", "description": "Generate frontend designs and components", "license": null, "category": "custom", "enabled": false } ] } ``` """ try: # Load all skills (including disabled ones) skills = load_skills(enabled_only=False) return SkillsListResponse(skills=[_skill_to_response(skill) for skill in skills]) except Exception as e: logger.error(f"Failed to load skills: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to load skills: {str(e)}") @router.get( "/skills/{skill_name}", response_model=SkillResponse, summary="Get Skill Details", description="Retrieve detailed information about a specific skill by its name.", ) async def get_skill(skill_name: str) -> SkillResponse: """Get a specific skill by name. Args: skill_name: The name of the skill to retrieve. Returns: Skill information if found. Raises: HTTPException: 404 if skill not found. Example Response: ```json { "name": "PDF Processing", "description": "Extract and analyze PDF content", "license": "MIT", "category": "public", "enabled": true } ``` """ try: skills = load_skills(enabled_only=False) skill = next((s for s in skills if s.name == skill_name), None) if skill is None: raise HTTPException(status_code=404, detail=f"Skill '{skill_name}' not found") return _skill_to_response(skill) except HTTPException: raise except Exception as e: logger.error(f"Failed to get skill {skill_name}: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to get skill: {str(e)}") @router.put( "/skills/{skill_name}", response_model=SkillResponse, summary="Update Skill", description="Update a skill's enabled status by modifying the skills_state_config.json file.", ) async def update_skill(skill_name: str, request: SkillUpdateRequest) -> SkillResponse: """Update a skill's enabled status. This will modify the skills_state_config.json file to update the enabled state. The SKILL.md file itself is not modified. Args: skill_name: The name of the skill to update. request: The update request containing the new enabled status. Returns: The updated skill information. Raises: HTTPException: 404 if skill not found, 500 if update fails. Example Request: ```json { "enabled": false } ``` Example Response: ```json { "name": "PDF Processing", "description": "Extract and analyze PDF content", "license": "MIT", "category": "public", "enabled": false } ``` """ try: # Find the skill to verify it exists skills = load_skills(enabled_only=False) skill = next((s for s in skills if s.name == skill_name), None) if skill is None: raise HTTPException(status_code=404, detail=f"Skill '{skill_name}' not found") # Get or create config path config_path = ExtensionsConfig.resolve_config_path() if config_path is None: # Create new config file in parent directory (project root) config_path = Path.cwd().parent / "extensions_config.json" logger.info(f"No existing extensions config found. Creating new config at: {config_path}") # Load current configuration extensions_config = get_extensions_config() # Update the skill's enabled status extensions_config.skills[skill_name] = SkillStateConfig(enabled=request.enabled) # Convert to JSON format (preserve MCP servers config) config_data = { "mcpServers": {name: server.model_dump() for name, server in extensions_config.mcp_servers.items()}, "skills": {name: {"enabled": skill_config.enabled} for name, skill_config in extensions_config.skills.items()}, } # Write the configuration to file with open(config_path, "w") as f: json.dump(config_data, f, indent=2) logger.info(f"Skills configuration updated and saved to: {config_path}") # Reload the extensions config to update the global cache reload_extensions_config() # Reload the skills to get the updated status (for API response) skills = load_skills(enabled_only=False) updated_skill = next((s for s in skills if s.name == skill_name), None) if updated_skill is None: raise HTTPException(status_code=500, detail=f"Failed to reload skill '{skill_name}' after update") logger.info(f"Skill '{skill_name}' enabled status updated to {request.enabled}") return _skill_to_response(updated_skill) except HTTPException: raise except Exception as e: logger.error(f"Failed to update skill {skill_name}: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to update skill: {str(e)}") @router.post( "/skills/install", response_model=SkillInstallResponse, summary="Install Skill", description="Install a skill from a .skill file (ZIP archive) located in the thread's user-data directory.", ) async def install_skill(request: SkillInstallRequest) -> SkillInstallResponse: """Install a skill from a .skill file. The .skill file is a ZIP archive containing a skill directory with SKILL.md and optional resources (scripts, references, assets). Args: request: The install request containing thread_id and virtual path to .skill file. Returns: Installation result with skill name and status message. Raises: HTTPException: - 400 if path is invalid or file is not a valid .skill file - 403 if access denied (path traversal detected) - 404 if file not found - 409 if skill already exists - 500 if installation fails Example Request: ```json { "thread_id": "abc123-def456", "path": "/mnt/user-data/outputs/my-skill.skill" } ``` Example Response: ```json { "success": true, "skill_name": "my-skill", "message": "Skill 'my-skill' installed successfully" } ``` """ try: # Resolve the virtual path to actual file path skill_file_path = resolve_thread_virtual_path(request.thread_id, request.path) # Check if file exists if not skill_file_path.exists(): raise HTTPException(status_code=404, detail=f"Skill file not found: {request.path}") # Check if it's a file if not skill_file_path.is_file(): raise HTTPException(status_code=400, detail=f"Path is not a file: {request.path}") # Check file extension if not skill_file_path.suffix == ".skill": raise HTTPException(status_code=400, detail="File must have .skill extension") # Verify it's a valid ZIP file if not zipfile.is_zipfile(skill_file_path): raise HTTPException(status_code=400, detail="File is not a valid ZIP archive") # Get the custom skills directory skills_root = get_skills_root_path() custom_skills_dir = skills_root / "custom" # Create custom directory if it doesn't exist custom_skills_dir.mkdir(parents=True, exist_ok=True) # Extract to a temporary directory first for validation with tempfile.TemporaryDirectory() as temp_dir: temp_path = Path(temp_dir) # Extract the .skill file with zipfile.ZipFile(skill_file_path, "r") as zip_ref: zip_ref.extractall(temp_path) # Find the skill directory (should be the only top-level directory) extracted_items = list(temp_path.iterdir()) if len(extracted_items) == 0: raise HTTPException(status_code=400, detail="Skill archive is empty") # Handle both cases: single directory or files directly in root if len(extracted_items) == 1 and extracted_items[0].is_dir(): skill_dir = extracted_items[0] else: # Files are directly in the archive root skill_dir = temp_path # Validate the skill is_valid, message, skill_name = _validate_skill_frontmatter(skill_dir) if not is_valid: raise HTTPException(status_code=400, detail=f"Invalid skill: {message}") if not skill_name: raise HTTPException(status_code=400, detail="Could not determine skill name") # Check if skill already exists target_dir = custom_skills_dir / skill_name if target_dir.exists(): raise HTTPException(status_code=409, detail=f"Skill '{skill_name}' already exists. Please remove it first or use a different name.") # Move the skill directory to the custom skills directory shutil.copytree(skill_dir, target_dir) logger.info(f"Skill '{skill_name}' installed successfully to {target_dir}") return SkillInstallResponse(success=True, skill_name=skill_name, message=f"Skill '{skill_name}' installed successfully") except HTTPException: raise except Exception as e: logger.error(f"Failed to install skill: {e}", exc_info=True) raise HTTPException(status_code=500, detail=f"Failed to install skill: {str(e)}")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/skills.py", "license": "MIT License", "lines": 349, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/gateway/routers/uploads.py
"""Upload router for handling file uploads.""" import logging from pathlib import Path from fastapi import APIRouter, File, HTTPException, UploadFile from pydantic import BaseModel from src.config.paths import VIRTUAL_PATH_PREFIX, get_paths from src.sandbox.sandbox_provider import get_sandbox_provider logger = logging.getLogger(__name__) router = APIRouter(prefix="/api/threads/{thread_id}/uploads", tags=["uploads"]) # File extensions that should be converted to markdown CONVERTIBLE_EXTENSIONS = { ".pdf", ".ppt", ".pptx", ".xls", ".xlsx", ".doc", ".docx", } class UploadResponse(BaseModel): """Response model for file upload.""" success: bool files: list[dict[str, str]] message: str def get_uploads_dir(thread_id: str) -> Path: """Get the uploads directory for a thread. Args: thread_id: The thread ID. Returns: Path to the uploads directory. """ base_dir = get_paths().sandbox_uploads_dir(thread_id) base_dir.mkdir(parents=True, exist_ok=True) return base_dir async def convert_file_to_markdown(file_path: Path) -> Path | None: """Convert a file to markdown using markitdown. Args: file_path: Path to the file to convert. Returns: Path to the markdown file if conversion was successful, None otherwise. """ try: from markitdown import MarkItDown md = MarkItDown() result = md.convert(str(file_path)) # Save as .md file with same name md_path = file_path.with_suffix(".md") md_path.write_text(result.text_content, encoding="utf-8") logger.info(f"Converted {file_path.name} to markdown: {md_path.name}") return md_path except Exception as e: logger.error(f"Failed to convert {file_path.name} to markdown: {e}") return None @router.post("", response_model=UploadResponse) async def upload_files( thread_id: str, files: list[UploadFile] = File(...), ) -> UploadResponse: """Upload multiple files to a thread's uploads directory. For PDF, PPT, Excel, and Word files, they will be converted to markdown using markitdown. All files (original and converted) are saved to /mnt/user-data/uploads. Args: thread_id: The thread ID to upload files to. files: List of files to upload. Returns: Upload response with success status and file information. """ if not files: raise HTTPException(status_code=400, detail="No files provided") uploads_dir = get_uploads_dir(thread_id) paths = get_paths() uploaded_files = [] sandbox_provider = get_sandbox_provider() sandbox_id = sandbox_provider.acquire(thread_id) sandbox = sandbox_provider.get(sandbox_id) for file in files: if not file.filename: continue try: # Normalize filename to prevent path traversal safe_filename = Path(file.filename).name if not safe_filename or safe_filename in {".", ".."} or "/" in safe_filename or "\\" in safe_filename: logger.warning(f"Skipping file with unsafe filename: {file.filename!r}") continue content = await file.read() file_path = uploads_dir / safe_filename file_path.write_bytes(content) # Build relative path from backend root relative_path = str(paths.sandbox_uploads_dir(thread_id) / safe_filename) virtual_path = f"{VIRTUAL_PATH_PREFIX}/uploads/{safe_filename}" # Keep local sandbox source of truth in thread-scoped host storage. # For non-local sandboxes, also sync to virtual path for runtime visibility. if sandbox_id != "local": sandbox.update_file(virtual_path, content) file_info = { "filename": safe_filename, "size": str(len(content)), "path": relative_path, # Actual filesystem path (relative to backend/) "virtual_path": virtual_path, # Path for Agent in sandbox "artifact_url": f"/api/threads/{thread_id}/artifacts/mnt/user-data/uploads/{safe_filename}", # HTTP URL } logger.info(f"Saved file: {safe_filename} ({len(content)} bytes) to {relative_path}") # Check if file should be converted to markdown file_ext = file_path.suffix.lower() if file_ext in CONVERTIBLE_EXTENSIONS: md_path = await convert_file_to_markdown(file_path) if md_path: md_relative_path = str(paths.sandbox_uploads_dir(thread_id) / md_path.name) md_virtual_path = f"{VIRTUAL_PATH_PREFIX}/uploads/{md_path.name}" if sandbox_id != "local": sandbox.update_file(md_virtual_path, md_path.read_bytes()) file_info["markdown_file"] = md_path.name file_info["markdown_path"] = md_relative_path file_info["markdown_virtual_path"] = md_virtual_path file_info["markdown_artifact_url"] = f"/api/threads/{thread_id}/artifacts/mnt/user-data/uploads/{md_path.name}" uploaded_files.append(file_info) except Exception as e: logger.error(f"Failed to upload {file.filename}: {e}") raise HTTPException(status_code=500, detail=f"Failed to upload {file.filename}: {str(e)}") return UploadResponse( success=True, files=uploaded_files, message=f"Successfully uploaded {len(uploaded_files)} file(s)", ) @router.get("/list", response_model=dict) async def list_uploaded_files(thread_id: str) -> dict: """List all files in a thread's uploads directory. Args: thread_id: The thread ID to list files for. Returns: Dictionary containing list of files with their metadata. """ uploads_dir = get_uploads_dir(thread_id) if not uploads_dir.exists(): return {"files": [], "count": 0} files = [] for file_path in sorted(uploads_dir.iterdir()): if file_path.is_file(): stat = file_path.stat() relative_path = str(get_paths().sandbox_uploads_dir(thread_id) / file_path.name) files.append( { "filename": file_path.name, "size": stat.st_size, "path": relative_path, # Actual filesystem path "virtual_path": f"{VIRTUAL_PATH_PREFIX}/uploads/{file_path.name}", # Path for Agent in sandbox "artifact_url": f"/api/threads/{thread_id}/artifacts/mnt/user-data/uploads/{file_path.name}", # HTTP URL "extension": file_path.suffix, "modified": stat.st_mtime, } ) return {"files": files, "count": len(files)} @router.delete("/{filename}") async def delete_uploaded_file(thread_id: str, filename: str) -> dict: """Delete a file from a thread's uploads directory. Args: thread_id: The thread ID. filename: The filename to delete. Returns: Success message. """ uploads_dir = get_uploads_dir(thread_id) file_path = uploads_dir / filename if not file_path.exists(): raise HTTPException(status_code=404, detail=f"File not found: {filename}") # Security check: ensure the path is within the uploads directory try: file_path.resolve().relative_to(uploads_dir.resolve()) except ValueError: raise HTTPException(status_code=403, detail="Access denied") try: file_path.unlink() logger.info(f"Deleted file: {filename}") return {"success": True, "message": f"Deleted {filename}"} except Exception as e: logger.error(f"Failed to delete {filename}: {e}") raise HTTPException(status_code=500, detail=f"Failed to delete {filename}: {str(e)}")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/gateway/routers/uploads.py", "license": "MIT License", "lines": 177, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/mcp/cache.py
"""Cache for MCP tools to avoid repeated loading.""" import asyncio import logging import os from langchain_core.tools import BaseTool logger = logging.getLogger(__name__) _mcp_tools_cache: list[BaseTool] | None = None _cache_initialized = False _initialization_lock = asyncio.Lock() _config_mtime: float | None = None # Track config file modification time def _get_config_mtime() -> float | None: """Get the modification time of the extensions config file. Returns: The modification time as a float, or None if the file doesn't exist. """ from src.config.extensions_config import ExtensionsConfig config_path = ExtensionsConfig.resolve_config_path() if config_path and config_path.exists(): return os.path.getmtime(config_path) return None def _is_cache_stale() -> bool: """Check if the cache is stale due to config file changes. Returns: True if the cache should be invalidated, False otherwise. """ global _config_mtime if not _cache_initialized: return False # Not initialized yet, not stale current_mtime = _get_config_mtime() # If we couldn't get mtime before or now, assume not stale if _config_mtime is None or current_mtime is None: return False # If the config file has been modified since we cached, it's stale if current_mtime > _config_mtime: logger.info(f"MCP config file has been modified (mtime: {_config_mtime} -> {current_mtime}), cache is stale") return True return False async def initialize_mcp_tools() -> list[BaseTool]: """Initialize and cache MCP tools. This should be called once at application startup. Returns: List of LangChain tools from all enabled MCP servers. """ global _mcp_tools_cache, _cache_initialized, _config_mtime async with _initialization_lock: if _cache_initialized: logger.info("MCP tools already initialized") return _mcp_tools_cache or [] from src.mcp.tools import get_mcp_tools logger.info("Initializing MCP tools...") _mcp_tools_cache = await get_mcp_tools() _cache_initialized = True _config_mtime = _get_config_mtime() # Record config file mtime logger.info(f"MCP tools initialized: {len(_mcp_tools_cache)} tool(s) loaded (config mtime: {_config_mtime})") return _mcp_tools_cache def get_cached_mcp_tools() -> list[BaseTool]: """Get cached MCP tools with lazy initialization. If tools are not initialized, automatically initializes them. This ensures MCP tools work in both FastAPI and LangGraph Studio contexts. Also checks if the config file has been modified since last initialization, and re-initializes if needed. This ensures that changes made through the Gateway API (which runs in a separate process) are reflected in the LangGraph Server. Returns: List of cached MCP tools. """ global _cache_initialized # Check if cache is stale due to config file changes if _is_cache_stale(): logger.info("MCP cache is stale, resetting for re-initialization...") reset_mcp_tools_cache() if not _cache_initialized: logger.info("MCP tools not initialized, performing lazy initialization...") try: # Try to initialize in the current event loop loop = asyncio.get_event_loop() if loop.is_running(): # If loop is already running (e.g., in LangGraph Studio), # we need to create a new loop in a thread import concurrent.futures with concurrent.futures.ThreadPoolExecutor() as executor: future = executor.submit(asyncio.run, initialize_mcp_tools()) future.result() else: # If no loop is running, we can use the current loop loop.run_until_complete(initialize_mcp_tools()) except RuntimeError: # No event loop exists, create one asyncio.run(initialize_mcp_tools()) except Exception as e: logger.error(f"Failed to lazy-initialize MCP tools: {e}") return [] return _mcp_tools_cache or [] def reset_mcp_tools_cache() -> None: """Reset the MCP tools cache. This is useful for testing or when you want to reload MCP tools. """ global _mcp_tools_cache, _cache_initialized, _config_mtime _mcp_tools_cache = None _cache_initialized = False _config_mtime = None logger.info("MCP tools cache reset")
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/mcp/cache.py", "license": "MIT License", "lines": 102, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/mcp/client.py
"""MCP client using langchain-mcp-adapters.""" import logging from typing import Any from src.config.extensions_config import ExtensionsConfig, McpServerConfig logger = logging.getLogger(__name__) def build_server_params(server_name: str, config: McpServerConfig) -> dict[str, Any]: """Build server parameters for MultiServerMCPClient. Args: server_name: Name of the MCP server. config: Configuration for the MCP server. Returns: Dictionary of server parameters for langchain-mcp-adapters. """ transport_type = config.type or "stdio" params: dict[str, Any] = {"transport": transport_type} if transport_type == "stdio": if not config.command: raise ValueError(f"MCP server '{server_name}' with stdio transport requires 'command' field") params["command"] = config.command params["args"] = config.args # Add environment variables if present if config.env: params["env"] = config.env elif transport_type in ("sse", "http"): if not config.url: raise ValueError(f"MCP server '{server_name}' with {transport_type} transport requires 'url' field") params["url"] = config.url # Add headers if present if config.headers: params["headers"] = config.headers else: raise ValueError(f"MCP server '{server_name}' has unsupported transport type: {transport_type}") return params def build_servers_config(extensions_config: ExtensionsConfig) -> dict[str, dict[str, Any]]: """Build servers configuration for MultiServerMCPClient. Args: extensions_config: Extensions configuration containing all MCP servers. Returns: Dictionary mapping server names to their parameters. """ enabled_servers = extensions_config.get_enabled_mcp_servers() if not enabled_servers: logger.info("No enabled MCP servers found") return {} servers_config = {} for server_name, server_config in enabled_servers.items(): try: servers_config[server_name] = build_server_params(server_name, server_config) logger.info(f"Configured MCP server: {server_name}") except Exception as e: logger.error(f"Failed to configure MCP server '{server_name}': {e}") return servers_config
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/mcp/client.py", "license": "MIT License", "lines": 52, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/mcp/tools.py
"""Load MCP tools using langchain-mcp-adapters.""" import logging from langchain_core.tools import BaseTool from src.config.extensions_config import ExtensionsConfig from src.mcp.client import build_servers_config from src.mcp.oauth import build_oauth_tool_interceptor, get_initial_oauth_headers logger = logging.getLogger(__name__) async def get_mcp_tools() -> list[BaseTool]: """Get all tools from enabled MCP servers. Returns: List of LangChain tools from all enabled MCP servers. """ try: from langchain_mcp_adapters.client import MultiServerMCPClient except ImportError: logger.warning("langchain-mcp-adapters not installed. Install it to enable MCP tools: pip install langchain-mcp-adapters") return [] # NOTE: We use ExtensionsConfig.from_file() instead of get_extensions_config() # to always read the latest configuration from disk. This ensures that changes # made through the Gateway API (which runs in a separate process) are immediately # reflected when initializing MCP tools. extensions_config = ExtensionsConfig.from_file() servers_config = build_servers_config(extensions_config) if not servers_config: logger.info("No enabled MCP servers configured") return [] try: # Create the multi-server MCP client logger.info(f"Initializing MCP client with {len(servers_config)} server(s)") # Inject initial OAuth headers for server connections (tool discovery/session init) initial_oauth_headers = await get_initial_oauth_headers(extensions_config) for server_name, auth_header in initial_oauth_headers.items(): if server_name not in servers_config: continue if servers_config[server_name].get("transport") in ("sse", "http"): existing_headers = dict(servers_config[server_name].get("headers", {})) existing_headers["Authorization"] = auth_header servers_config[server_name]["headers"] = existing_headers tool_interceptors = [] oauth_interceptor = build_oauth_tool_interceptor(extensions_config) if oauth_interceptor is not None: tool_interceptors.append(oauth_interceptor) client = MultiServerMCPClient(servers_config, tool_interceptors=tool_interceptors) # Get all tools from all servers tools = await client.get_tools() logger.info(f"Successfully loaded {len(tools)} tool(s) from MCP servers") return tools except Exception as e: logger.error(f"Failed to load MCP tools: {e}", exc_info=True) return []
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/mcp/tools.py", "license": "MIT License", "lines": 50, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/models/factory.py
import logging from langchain.chat_models import BaseChatModel from src.config import get_app_config, get_tracing_config, is_tracing_enabled from src.reflection import resolve_class logger = logging.getLogger(__name__) def create_chat_model(name: str | None = None, thinking_enabled: bool = False, **kwargs) -> BaseChatModel: """Create a chat model instance from the config. Args: name: The name of the model to create. If None, the first model in the config will be used. Returns: A chat model instance. """ config = get_app_config() if name is None: name = config.models[0].name model_config = config.get_model_config(name) if model_config is None: raise ValueError(f"Model {name} not found in config") from None model_class = resolve_class(model_config.use, BaseChatModel) model_settings_from_config = model_config.model_dump( exclude_none=True, exclude={ "use", "name", "display_name", "description", "supports_thinking", "supports_reasoning_effort", "when_thinking_enabled", "supports_vision", }, ) if thinking_enabled and model_config.when_thinking_enabled is not None: if not model_config.supports_thinking: raise ValueError(f"Model {name} does not support thinking. Set `supports_thinking` to true in the `config.yaml` to enable thinking.") from None model_settings_from_config.update(model_config.when_thinking_enabled) if not thinking_enabled and model_config.when_thinking_enabled and model_config.when_thinking_enabled.get("extra_body", {}).get("thinking", {}).get("type"): kwargs.update({"extra_body": {"thinking": {"type": "disabled"}}}) kwargs.update({"reasoning_effort": "minimal"}) if not model_config.supports_reasoning_effort: kwargs.update({"reasoning_effort": None}) model_instance = model_class(**kwargs, **model_settings_from_config) if is_tracing_enabled(): try: from langchain_core.tracers.langchain import LangChainTracer tracing_config = get_tracing_config() tracer = LangChainTracer( project_name=tracing_config.project, ) existing_callbacks = model_instance.callbacks or [] model_instance.callbacks = [*existing_callbacks, tracer] logger.debug(f"LangSmith tracing attached to model '{name}' (project='{tracing_config.project}')") except Exception as e: logger.warning(f"Failed to attach LangSmith tracing to model '{name}': {e}") return model_instance
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/models/factory.py", "license": "MIT License", "lines": 55, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/models/patched_deepseek.py
"""Patched ChatDeepSeek that preserves reasoning_content in multi-turn conversations. This module provides a patched version of ChatDeepSeek that properly handles reasoning_content when sending messages back to the API. The original implementation stores reasoning_content in additional_kwargs but doesn't include it when making subsequent API calls, which causes errors with APIs that require reasoning_content on all assistant messages when thinking mode is enabled. """ from typing import Any from langchain_core.language_models import LanguageModelInput from langchain_core.messages import AIMessage from langchain_deepseek import ChatDeepSeek class PatchedChatDeepSeek(ChatDeepSeek): """ChatDeepSeek with proper reasoning_content preservation. When using thinking/reasoning enabled models, the API expects reasoning_content to be present on ALL assistant messages in multi-turn conversations. This patched version ensures reasoning_content from additional_kwargs is included in the request payload. """ def _get_request_payload( self, input_: LanguageModelInput, *, stop: list[str] | None = None, **kwargs: Any, ) -> dict: """Get request payload with reasoning_content preserved. Overrides the parent method to inject reasoning_content from additional_kwargs into assistant messages in the payload. """ # Get the original messages before conversion original_messages = self._convert_input(input_).to_messages() # Call parent to get the base payload payload = super()._get_request_payload(input_, stop=stop, **kwargs) # Match payload messages with original messages to restore reasoning_content payload_messages = payload.get("messages", []) # The payload messages and original messages should be in the same order # Iterate through both and match by position if len(payload_messages) == len(original_messages): for payload_msg, orig_msg in zip(payload_messages, original_messages): if payload_msg.get("role") == "assistant" and isinstance(orig_msg, AIMessage): reasoning_content = orig_msg.additional_kwargs.get("reasoning_content") if reasoning_content is not None: payload_msg["reasoning_content"] = reasoning_content else: # Fallback: match by counting assistant messages ai_messages = [m for m in original_messages if isinstance(m, AIMessage)] assistant_payloads = [(i, m) for i, m in enumerate(payload_messages) if m.get("role") == "assistant"] for (idx, payload_msg), ai_msg in zip(assistant_payloads, ai_messages): reasoning_content = ai_msg.additional_kwargs.get("reasoning_content") if reasoning_content is not None: payload_messages[idx]["reasoning_content"] = reasoning_content return payload
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/models/patched_deepseek.py", "license": "MIT License", "lines": 52, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/reflection/resolvers.py
from importlib import import_module MODULE_TO_PACKAGE_HINTS = { "langchain_google_genai": "langchain-google-genai", "langchain_anthropic": "langchain-anthropic", "langchain_openai": "langchain-openai", "langchain_deepseek": "langchain-deepseek", } def _build_missing_dependency_hint(module_path: str, err: ImportError) -> str: """Build an actionable hint when module import fails.""" module_root = module_path.split(".", 1)[0] missing_module = getattr(err, "name", None) or module_root # Prefer provider package hints for known integrations, even when the import # error is triggered by a transitive dependency (e.g. `google`). package_name = MODULE_TO_PACKAGE_HINTS.get(module_root) if package_name is None: package_name = MODULE_TO_PACKAGE_HINTS.get(missing_module, missing_module.replace("_", "-")) return ( f"Missing dependency '{missing_module}'. " f"Install it with `uv add {package_name}` (or `pip install {package_name}`), then restart DeerFlow." ) def resolve_variable[T]( variable_path: str, expected_type: type[T] | tuple[type, ...] | None = None, ) -> T: """Resolve a variable from a path. Args: variable_path: The path to the variable (e.g. "parent_package_name.sub_package_name.module_name:variable_name"). expected_type: Optional type or tuple of types to validate the resolved variable against. If provided, uses isinstance() to check if the variable is an instance of the expected type(s). Returns: The resolved variable. Raises: ImportError: If the module path is invalid or the attribute doesn't exist. ValueError: If the resolved variable doesn't pass the validation checks. """ try: module_path, variable_name = variable_path.rsplit(":", 1) except ValueError as err: raise ImportError(f"{variable_path} doesn't look like a variable path. Example: parent_package_name.sub_package_name.module_name:variable_name") from err try: module = import_module(module_path) except ImportError as err: module_root = module_path.split(".", 1)[0] err_name = getattr(err, "name", None) if isinstance(err, ModuleNotFoundError) or err_name == module_root: hint = _build_missing_dependency_hint(module_path, err) raise ImportError(f"Could not import module {module_path}. {hint}") from err # Preserve the original ImportError message for non-missing-module failures. raise ImportError(f"Error importing module {module_path}: {err}") from err try: variable = getattr(module, variable_name) except AttributeError as err: raise ImportError(f"Module {module_path} does not define a {variable_name} attribute/class") from err # Type validation if expected_type is not None: if not isinstance(variable, expected_type): type_name = expected_type.__name__ if isinstance(expected_type, type) else " or ".join(t.__name__ for t in expected_type) raise ValueError(f"{variable_path} is not an instance of {type_name}, got {type(variable).__name__}") return variable def resolve_class[T](class_path: str, base_class: type[T] | None = None) -> type[T]: """Resolve a class from a module path and class name. Args: class_path: The path to the class (e.g. "langchain_openai:ChatOpenAI"). base_class: The base class to check if the resolved class is a subclass of. Returns: The resolved class. Raises: ImportError: If the module path is invalid or the attribute doesn't exist. ValueError: If the resolved object is not a class or not a subclass of base_class. """ model_class = resolve_variable(class_path, expected_type=type) if not isinstance(model_class, type): raise ValueError(f"{class_path} is not a valid class") if base_class is not None and not issubclass(model_class, base_class): raise ValueError(f"{class_path} is not a subclass of {base_class.__name__}") return model_class
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/reflection/resolvers.py", "license": "MIT License", "lines": 76, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/sandbox/exceptions.py
"""Sandbox-related exceptions with structured error information.""" class SandboxError(Exception): """Base exception for all sandbox-related errors.""" def __init__(self, message: str, details: dict | None = None): super().__init__(message) self.message = message self.details = details or {} def __str__(self) -> str: if self.details: detail_str = ", ".join(f"{k}={v}" for k, v in self.details.items()) return f"{self.message} ({detail_str})" return self.message class SandboxNotFoundError(SandboxError): """Raised when a sandbox cannot be found or is not available.""" def __init__(self, message: str = "Sandbox not found", sandbox_id: str | None = None): details = {"sandbox_id": sandbox_id} if sandbox_id else None super().__init__(message, details) self.sandbox_id = sandbox_id class SandboxRuntimeError(SandboxError): """Raised when sandbox runtime is not available or misconfigured.""" pass class SandboxCommandError(SandboxError): """Raised when a command execution fails in the sandbox.""" def __init__(self, message: str, command: str | None = None, exit_code: int | None = None): details = {} if command: details["command"] = command[:100] + "..." if len(command) > 100 else command if exit_code is not None: details["exit_code"] = exit_code super().__init__(message, details) self.command = command self.exit_code = exit_code class SandboxFileError(SandboxError): """Raised when a file operation fails in the sandbox.""" def __init__(self, message: str, path: str | None = None, operation: str | None = None): details = {} if path: details["path"] = path if operation: details["operation"] = operation super().__init__(message, details) self.path = path self.operation = operation class SandboxPermissionError(SandboxFileError): """Raised when a permission error occurs during file operations.""" pass class SandboxFileNotFoundError(SandboxFileError): """Raised when a file or directory is not found.""" pass
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/exceptions.py", "license": "MIT License", "lines": 49, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/sandbox/local/list_dir.py
import fnmatch from pathlib import Path IGNORE_PATTERNS = [ # Version Control ".git", ".svn", ".hg", ".bzr", # Dependencies "node_modules", "__pycache__", ".venv", "venv", ".env", "env", ".tox", ".nox", ".eggs", "*.egg-info", "site-packages", # Build outputs "dist", "build", ".next", ".nuxt", ".output", ".turbo", "target", "out", # IDE & Editor ".idea", ".vscode", "*.swp", "*.swo", "*~", ".project", ".classpath", ".settings", # OS generated ".DS_Store", "Thumbs.db", "desktop.ini", "*.lnk", # Logs & temp files "*.log", "*.tmp", "*.temp", "*.bak", "*.cache", ".cache", "logs", # Coverage & test artifacts ".coverage", "coverage", ".nyc_output", "htmlcov", ".pytest_cache", ".mypy_cache", ".ruff_cache", ] def _should_ignore(name: str) -> bool: """Check if a file/directory name matches any ignore pattern.""" for pattern in IGNORE_PATTERNS: if fnmatch.fnmatch(name, pattern): return True return False def list_dir(path: str, max_depth: int = 2) -> list[str]: """ List files and directories up to max_depth levels deep. Args: path: The root directory path to list. max_depth: Maximum depth to traverse (default: 2). 1 = only direct children, 2 = children + grandchildren, etc. Returns: A list of absolute paths for files and directories, excluding items matching IGNORE_PATTERNS. """ result: list[str] = [] root_path = Path(path).resolve() if not root_path.is_dir(): return result def _traverse(current_path: Path, current_depth: int) -> None: """Recursively traverse directories up to max_depth.""" if current_depth > max_depth: return try: for item in current_path.iterdir(): if _should_ignore(item.name): continue post_fix = "/" if item.is_dir() else "" result.append(str(item.resolve()) + post_fix) # Recurse into subdirectories if not at max depth if item.is_dir() and current_depth < max_depth: _traverse(item, current_depth + 1) except PermissionError: pass _traverse(root_path, 1) return sorted(result)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/local/list_dir.py", "license": "MIT License", "lines": 98, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/sandbox/local/local_sandbox.py
import os import shutil import subprocess from pathlib import Path from src.sandbox.local.list_dir import list_dir from src.sandbox.sandbox import Sandbox class LocalSandbox(Sandbox): def __init__(self, id: str, path_mappings: dict[str, str] | None = None): """ Initialize local sandbox with optional path mappings. Args: id: Sandbox identifier path_mappings: Dictionary mapping container paths to local paths Example: {"/mnt/skills": "/absolute/path/to/skills"} """ super().__init__(id) self.path_mappings = path_mappings or {} def _resolve_path(self, path: str) -> str: """ Resolve container path to actual local path using mappings. Args: path: Path that might be a container path Returns: Resolved local path """ path_str = str(path) # Try each mapping (longest prefix first for more specific matches) for container_path, local_path in sorted(self.path_mappings.items(), key=lambda x: len(x[0]), reverse=True): if path_str.startswith(container_path): # Replace the container path prefix with local path relative = path_str[len(container_path) :].lstrip("/") resolved = str(Path(local_path) / relative) if relative else local_path return resolved # No mapping found, return original path return path_str def _reverse_resolve_path(self, path: str) -> str: """ Reverse resolve local path back to container path using mappings. Args: path: Local path that might need to be mapped to container path Returns: Container path if mapping exists, otherwise original path """ path_str = str(Path(path).resolve()) # Try each mapping (longest local path first for more specific matches) for container_path, local_path in sorted(self.path_mappings.items(), key=lambda x: len(x[1]), reverse=True): local_path_resolved = str(Path(local_path).resolve()) if path_str.startswith(local_path_resolved): # Replace the local path prefix with container path relative = path_str[len(local_path_resolved) :].lstrip("/") resolved = f"{container_path}/{relative}" if relative else container_path return resolved # No mapping found, return original path return path_str def _reverse_resolve_paths_in_output(self, output: str) -> str: """ Reverse resolve local paths back to container paths in output string. Args: output: Output string that may contain local paths Returns: Output with local paths resolved to container paths """ import re # Sort mappings by local path length (longest first) for correct prefix matching sorted_mappings = sorted(self.path_mappings.items(), key=lambda x: len(x[1]), reverse=True) if not sorted_mappings: return output # Create pattern that matches absolute paths # Match paths like /Users/... or other absolute paths result = output for container_path, local_path in sorted_mappings: local_path_resolved = str(Path(local_path).resolve()) # Escape the local path for use in regex escaped_local = re.escape(local_path_resolved) # Match the local path followed by optional path components pattern = re.compile(escaped_local + r"(?:/[^\s\"';&|<>()]*)?") def replace_match(match: re.Match) -> str: matched_path = match.group(0) return self._reverse_resolve_path(matched_path) result = pattern.sub(replace_match, result) return result def _resolve_paths_in_command(self, command: str) -> str: """ Resolve container paths to local paths in a command string. Args: command: Command string that may contain container paths Returns: Command with container paths resolved to local paths """ import re # Sort mappings by length (longest first) for correct prefix matching sorted_mappings = sorted(self.path_mappings.items(), key=lambda x: len(x[0]), reverse=True) # Build regex pattern to match all container paths # Match container path followed by optional path components if not sorted_mappings: return command # Create pattern that matches any of the container paths patterns = [re.escape(container_path) + r"(?:/[^\s\"';&|<>()]*)??" for container_path, _ in sorted_mappings] pattern = re.compile("|".join(f"({p})" for p in patterns)) def replace_match(match: re.Match) -> str: matched_path = match.group(0) return self._resolve_path(matched_path) return pattern.sub(replace_match, command) @staticmethod def _get_shell() -> str: """Detect available shell executable with fallback. Returns the first available shell in order of preference: /bin/zsh β†’ /bin/bash β†’ /bin/sh β†’ first `sh` found on PATH. Raises a RuntimeError if no suitable shell is found. """ for shell in ("/bin/zsh", "/bin/bash", "/bin/sh"): if os.path.isfile(shell) and os.access(shell, os.X_OK): return shell shell_from_path = shutil.which("sh") if shell_from_path is not None: return shell_from_path raise RuntimeError( "No suitable shell executable found. Tried /bin/zsh, /bin/bash, " "/bin/sh, and `sh` on PATH." ) def execute_command(self, command: str) -> str: # Resolve container paths in command before execution resolved_command = self._resolve_paths_in_command(command) result = subprocess.run( resolved_command, executable=self._get_shell(), shell=True, capture_output=True, text=True, timeout=600, ) output = result.stdout if result.stderr: output += f"\nStd Error:\n{result.stderr}" if output else result.stderr if result.returncode != 0: output += f"\nExit Code: {result.returncode}" final_output = output if output else "(no output)" # Reverse resolve local paths back to container paths in output return self._reverse_resolve_paths_in_output(final_output) def list_dir(self, path: str, max_depth=2) -> list[str]: resolved_path = self._resolve_path(path) entries = list_dir(resolved_path, max_depth) # Reverse resolve local paths back to container paths in output return [self._reverse_resolve_paths_in_output(entry) for entry in entries] def read_file(self, path: str) -> str: resolved_path = self._resolve_path(path) with open(resolved_path) as f: return f.read() def write_file(self, path: str, content: str, append: bool = False) -> None: resolved_path = self._resolve_path(path) dir_path = os.path.dirname(resolved_path) if dir_path: os.makedirs(dir_path, exist_ok=True) mode = "a" if append else "w" with open(resolved_path, mode) as f: f.write(content) def update_file(self, path: str, content: bytes) -> None: resolved_path = self._resolve_path(path) dir_path = os.path.dirname(resolved_path) if dir_path: os.makedirs(dir_path, exist_ok=True) with open(resolved_path, "wb") as f: f.write(content)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/local/local_sandbox.py", "license": "MIT License", "lines": 163, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/sandbox/local/local_sandbox_provider.py
from src.sandbox.local.local_sandbox import LocalSandbox from src.sandbox.sandbox import Sandbox from src.sandbox.sandbox_provider import SandboxProvider _singleton: LocalSandbox | None = None class LocalSandboxProvider(SandboxProvider): def __init__(self): """Initialize the local sandbox provider with path mappings.""" self._path_mappings = self._setup_path_mappings() def _setup_path_mappings(self) -> dict[str, str]: """ Setup path mappings for local sandbox. Maps container paths to actual local paths, including skills directory. Returns: Dictionary of path mappings """ mappings = {} # Map skills container path to local skills directory try: from src.config import get_app_config config = get_app_config() skills_path = config.skills.get_skills_path() container_path = config.skills.container_path # Only add mapping if skills directory exists if skills_path.exists(): mappings[container_path] = str(skills_path) except Exception as e: # Log but don't fail if config loading fails print(f"Warning: Could not setup skills path mapping: {e}") return mappings def acquire(self, thread_id: str | None = None) -> str: global _singleton if _singleton is None: _singleton = LocalSandbox("local", path_mappings=self._path_mappings) return _singleton.id def get(self, sandbox_id: str) -> Sandbox | None: if sandbox_id == "local": if _singleton is None: self.acquire() return _singleton return None def release(self, sandbox_id: str) -> None: # LocalSandbox uses singleton pattern - no cleanup needed. # Note: This method is intentionally not called by SandboxMiddleware # to allow sandbox reuse across multiple turns in a thread. # For Docker-based providers (e.g., AioSandboxProvider), cleanup # happens at application shutdown via the shutdown() method. pass
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/local/local_sandbox_provider.py", "license": "MIT License", "lines": 47, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/sandbox/middleware.py
from typing import NotRequired, override from langchain.agents import AgentState from langchain.agents.middleware import AgentMiddleware from langgraph.runtime import Runtime from src.agents.thread_state import SandboxState, ThreadDataState from src.sandbox import get_sandbox_provider class SandboxMiddlewareState(AgentState): """Compatible with the `ThreadState` schema.""" sandbox: NotRequired[SandboxState | None] thread_data: NotRequired[ThreadDataState | None] class SandboxMiddleware(AgentMiddleware[SandboxMiddlewareState]): """Create a sandbox environment and assign it to an agent. Lifecycle Management: - With lazy_init=True (default): Sandbox is acquired on first tool call - With lazy_init=False: Sandbox is acquired on first agent invocation (before_agent) - Sandbox is reused across multiple turns within the same thread - Sandbox is NOT released after each agent call to avoid wasteful recreation - Cleanup happens at application shutdown via SandboxProvider.shutdown() """ state_schema = SandboxMiddlewareState def __init__(self, lazy_init: bool = True): """Initialize sandbox middleware. Args: lazy_init: If True, defer sandbox acquisition until first tool call. If False, acquire sandbox eagerly in before_agent(). Default is True for optimal performance. """ super().__init__() self._lazy_init = lazy_init def _acquire_sandbox(self, thread_id: str) -> str: provider = get_sandbox_provider() sandbox_id = provider.acquire(thread_id) print(f"Acquiring sandbox {sandbox_id}") return sandbox_id @override def before_agent(self, state: SandboxMiddlewareState, runtime: Runtime) -> dict | None: # Skip acquisition if lazy_init is enabled if self._lazy_init: return super().before_agent(state, runtime) # Eager initialization (original behavior) if "sandbox" not in state or state["sandbox"] is None: thread_id = runtime.context["thread_id"] print(f"Thread ID: {thread_id}") sandbox_id = self._acquire_sandbox(thread_id) return {"sandbox": {"sandbox_id": sandbox_id}} return super().before_agent(state, runtime)
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/middleware.py", "license": "MIT License", "lines": 46, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/sandbox/sandbox.py
from abc import ABC, abstractmethod class Sandbox(ABC): """Abstract base class for sandbox environments""" _id: str def __init__(self, id: str): self._id = id @property def id(self) -> str: return self._id @abstractmethod def execute_command(self, command: str) -> str: """Execute bash command in sandbox. Args: command: The command to execute. Returns: The standard or error output of the command. """ pass @abstractmethod def read_file(self, path: str) -> str: """Read the content of a file. Args: path: The absolute path of the file to read. Returns: The content of the file. """ pass @abstractmethod def list_dir(self, path: str, max_depth=2) -> list[str]: """List the contents of a directory. Args: path: The absolute path of the directory to list. max_depth: The maximum depth to traverse. Default is 2. Returns: The contents of the directory. """ pass @abstractmethod def write_file(self, path: str, content: str, append: bool = False) -> None: """Write content to a file. Args: path: The absolute path of the file to write to. content: The text content to write to the file. append: Whether to append the content to the file. If False, the file will be created or overwritten. """ pass @abstractmethod def update_file(self, path: str, content: bytes) -> None: """Update a file with binary content. Args: path: The absolute path of the file to update. content: The binary content to write to the file. """ pass
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/sandbox.py", "license": "MIT License", "lines": 54, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/sandbox/sandbox_provider.py
from abc import ABC, abstractmethod from src.config import get_app_config from src.reflection import resolve_class from src.sandbox.sandbox import Sandbox class SandboxProvider(ABC): """Abstract base class for sandbox providers""" @abstractmethod def acquire(self, thread_id: str | None = None) -> str: """Acquire a sandbox environment and return its ID. Returns: The ID of the acquired sandbox environment. """ pass @abstractmethod def get(self, sandbox_id: str) -> Sandbox | None: """Get a sandbox environment by ID. Args: sandbox_id: The ID of the sandbox environment to retain. """ pass @abstractmethod def release(self, sandbox_id: str) -> None: """Release a sandbox environment. Args: sandbox_id: The ID of the sandbox environment to destroy. """ pass _default_sandbox_provider: SandboxProvider | None = None def get_sandbox_provider(**kwargs) -> SandboxProvider: """Get the sandbox provider singleton. Returns a cached singleton instance. Use `reset_sandbox_provider()` to clear the cache, or `shutdown_sandbox_provider()` to properly shutdown and clear. Returns: A sandbox provider instance. """ global _default_sandbox_provider if _default_sandbox_provider is None: config = get_app_config() cls = resolve_class(config.sandbox.use, SandboxProvider) _default_sandbox_provider = cls(**kwargs) return _default_sandbox_provider def reset_sandbox_provider() -> None: """Reset the sandbox provider singleton. This clears the cached instance without calling shutdown. The next call to `get_sandbox_provider()` will create a new instance. Useful for testing or when switching configurations. Note: If the provider has active sandboxes, they will be orphaned. Use `shutdown_sandbox_provider()` for proper cleanup. """ global _default_sandbox_provider _default_sandbox_provider = None def shutdown_sandbox_provider() -> None: """Shutdown and reset the sandbox provider. This properly shuts down the provider (releasing all sandboxes) before clearing the singleton. Call this when the application is shutting down or when you need to completely reset the sandbox system. """ global _default_sandbox_provider if _default_sandbox_provider is not None: if hasattr(_default_sandbox_provider, "shutdown"): _default_sandbox_provider.shutdown() _default_sandbox_provider = None def set_sandbox_provider(provider: SandboxProvider) -> None: """Set a custom sandbox provider instance. This allows injecting a custom or mock provider for testing purposes. Args: provider: The SandboxProvider instance to use. """ global _default_sandbox_provider _default_sandbox_provider = provider
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/sandbox_provider.py", "license": "MIT License", "lines": 70, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/sandbox/tools.py
import re from langchain.tools import ToolRuntime, tool from langgraph.typing import ContextT from src.agents.thread_state import ThreadDataState, ThreadState from src.config.paths import VIRTUAL_PATH_PREFIX from src.sandbox.exceptions import ( SandboxError, SandboxNotFoundError, SandboxRuntimeError, ) from src.sandbox.sandbox import Sandbox from src.sandbox.sandbox_provider import get_sandbox_provider def replace_virtual_path(path: str, thread_data: ThreadDataState | None) -> str: """Replace virtual /mnt/user-data paths with actual thread data paths. Mapping: /mnt/user-data/workspace/* -> thread_data['workspace_path']/* /mnt/user-data/uploads/* -> thread_data['uploads_path']/* /mnt/user-data/outputs/* -> thread_data['outputs_path']/* Args: path: The path that may contain virtual path prefix. thread_data: The thread data containing actual paths. Returns: The path with virtual prefix replaced by actual path. """ if not path.startswith(VIRTUAL_PATH_PREFIX): return path if thread_data is None: return path # Map virtual subdirectories to thread_data keys path_mapping = { "workspace": thread_data.get("workspace_path"), "uploads": thread_data.get("uploads_path"), "outputs": thread_data.get("outputs_path"), } # Extract the subdirectory after /mnt/user-data/ relative_path = path[len(VIRTUAL_PATH_PREFIX) :].lstrip("/") if not relative_path: return path # Find which subdirectory this path belongs to parts = relative_path.split("/", 1) subdir = parts[0] rest = parts[1] if len(parts) > 1 else "" actual_base = path_mapping.get(subdir) if actual_base is None: return path if rest: return f"{actual_base}/{rest}" return actual_base def replace_virtual_paths_in_command(command: str, thread_data: ThreadDataState | None) -> str: """Replace all virtual /mnt/user-data paths in a command string. Args: command: The command string that may contain virtual paths. thread_data: The thread data containing actual paths. Returns: The command with all virtual paths replaced. """ if VIRTUAL_PATH_PREFIX not in command: return command if thread_data is None: return command # Pattern to match /mnt/user-data followed by path characters pattern = re.compile(rf"{re.escape(VIRTUAL_PATH_PREFIX)}(/[^\s\"';&|<>()]*)?") def replace_match(match: re.Match) -> str: full_path = match.group(0) return replace_virtual_path(full_path, thread_data) return pattern.sub(replace_match, command) def get_thread_data(runtime: ToolRuntime[ContextT, ThreadState] | None) -> ThreadDataState | None: """Extract thread_data from runtime state.""" if runtime is None: return None if runtime.state is None: return None return runtime.state.get("thread_data") def is_local_sandbox(runtime: ToolRuntime[ContextT, ThreadState] | None) -> bool: """Check if the current sandbox is a local sandbox. Path replacement is only needed for local sandbox since aio sandbox already has /mnt/user-data mounted in the container. """ if runtime is None: return False if runtime.state is None: return False sandbox_state = runtime.state.get("sandbox") if sandbox_state is None: return False return sandbox_state.get("sandbox_id") == "local" def sandbox_from_runtime(runtime: ToolRuntime[ContextT, ThreadState] | None = None) -> Sandbox: """Extract sandbox instance from tool runtime. DEPRECATED: Use ensure_sandbox_initialized() for lazy initialization support. This function assumes sandbox is already initialized and will raise error if not. Raises: SandboxRuntimeError: If runtime is not available or sandbox state is missing. SandboxNotFoundError: If sandbox with the given ID cannot be found. """ if runtime is None: raise SandboxRuntimeError("Tool runtime not available") if runtime.state is None: raise SandboxRuntimeError("Tool runtime state not available") sandbox_state = runtime.state.get("sandbox") if sandbox_state is None: raise SandboxRuntimeError("Sandbox state not initialized in runtime") sandbox_id = sandbox_state.get("sandbox_id") if sandbox_id is None: raise SandboxRuntimeError("Sandbox ID not found in state") sandbox = get_sandbox_provider().get(sandbox_id) if sandbox is None: raise SandboxNotFoundError(f"Sandbox with ID '{sandbox_id}' not found", sandbox_id=sandbox_id) return sandbox def ensure_sandbox_initialized(runtime: ToolRuntime[ContextT, ThreadState] | None = None) -> Sandbox: """Ensure sandbox is initialized, acquiring lazily if needed. On first call, acquires a sandbox from the provider and stores it in runtime state. Subsequent calls return the existing sandbox. Thread-safety is guaranteed by the provider's internal locking mechanism. Args: runtime: Tool runtime containing state and context. Returns: Initialized sandbox instance. Raises: SandboxRuntimeError: If runtime is not available or thread_id is missing. SandboxNotFoundError: If sandbox acquisition fails. """ if runtime is None: raise SandboxRuntimeError("Tool runtime not available") if runtime.state is None: raise SandboxRuntimeError("Tool runtime state not available") # Check if sandbox already exists in state sandbox_state = runtime.state.get("sandbox") if sandbox_state is not None: sandbox_id = sandbox_state.get("sandbox_id") if sandbox_id is not None: sandbox = get_sandbox_provider().get(sandbox_id) if sandbox is not None: return sandbox # Sandbox was released, fall through to acquire new one # Lazy acquisition: get thread_id and acquire sandbox thread_id = runtime.context.get("thread_id") if thread_id is None: raise SandboxRuntimeError("Thread ID not available in runtime context") provider = get_sandbox_provider() print(f"Lazy acquiring sandbox for thread {thread_id}") sandbox_id = provider.acquire(thread_id) # Update runtime state - this persists across tool calls runtime.state["sandbox"] = {"sandbox_id": sandbox_id} # Retrieve and return the sandbox sandbox = provider.get(sandbox_id) if sandbox is None: raise SandboxNotFoundError("Sandbox not found after acquisition", sandbox_id=sandbox_id) return sandbox def ensure_thread_directories_exist(runtime: ToolRuntime[ContextT, ThreadState] | None) -> None: """Ensure thread data directories (workspace, uploads, outputs) exist. This function is called lazily when any sandbox tool is first used. For local sandbox, it creates the directories on the filesystem. For other sandboxes (like aio), directories are already mounted in the container. Args: runtime: Tool runtime containing state and context. """ if runtime is None: return # Only create directories for local sandbox if not is_local_sandbox(runtime): return thread_data = get_thread_data(runtime) if thread_data is None: return # Check if directories have already been created if runtime.state.get("thread_directories_created"): return # Create the three directories import os for key in ["workspace_path", "uploads_path", "outputs_path"]: path = thread_data.get(key) if path: os.makedirs(path, exist_ok=True) # Mark as created to avoid redundant operations runtime.state["thread_directories_created"] = True @tool("bash", parse_docstring=True) def bash_tool(runtime: ToolRuntime[ContextT, ThreadState], description: str, command: str) -> str: """Execute a bash command in a Linux environment. - Use `python` to run Python code. - Use `pip install` to install Python packages. Args: description: Explain why you are running this command in short words. ALWAYS PROVIDE THIS PARAMETER FIRST. command: The bash command to execute. Always use absolute paths for files and directories. """ try: sandbox = ensure_sandbox_initialized(runtime) ensure_thread_directories_exist(runtime) if is_local_sandbox(runtime): thread_data = get_thread_data(runtime) command = replace_virtual_paths_in_command(command, thread_data) return sandbox.execute_command(command) except SandboxError as e: return f"Error: {e}" except Exception as e: return f"Error: Unexpected error executing command: {type(e).__name__}: {e}" @tool("ls", parse_docstring=True) def ls_tool(runtime: ToolRuntime[ContextT, ThreadState], description: str, path: str) -> str: """List the contents of a directory up to 2 levels deep in tree format. Args: description: Explain why you are listing this directory in short words. ALWAYS PROVIDE THIS PARAMETER FIRST. path: The **absolute** path to the directory to list. """ try: sandbox = ensure_sandbox_initialized(runtime) ensure_thread_directories_exist(runtime) if is_local_sandbox(runtime): thread_data = get_thread_data(runtime) path = replace_virtual_path(path, thread_data) children = sandbox.list_dir(path) if not children: return "(empty)" return "\n".join(children) except SandboxError as e: return f"Error: {e}" except FileNotFoundError: return f"Error: Directory not found: {path}" except PermissionError: return f"Error: Permission denied: {path}" except Exception as e: return f"Error: Unexpected error listing directory: {type(e).__name__}: {e}" @tool("read_file", parse_docstring=True) def read_file_tool( runtime: ToolRuntime[ContextT, ThreadState], description: str, path: str, start_line: int | None = None, end_line: int | None = None, ) -> str: """Read the contents of a text file. Use this to examine source code, configuration files, logs, or any text-based file. Args: description: Explain why you are reading this file in short words. ALWAYS PROVIDE THIS PARAMETER FIRST. path: The **absolute** path to the file to read. start_line: Optional starting line number (1-indexed, inclusive). Use with end_line to read a specific range. end_line: Optional ending line number (1-indexed, inclusive). Use with start_line to read a specific range. """ try: sandbox = ensure_sandbox_initialized(runtime) ensure_thread_directories_exist(runtime) if is_local_sandbox(runtime): thread_data = get_thread_data(runtime) path = replace_virtual_path(path, thread_data) content = sandbox.read_file(path) if not content: return "(empty)" if start_line is not None and end_line is not None: content = "\n".join(content.splitlines()[start_line - 1 : end_line]) return content except SandboxError as e: return f"Error: {e}" except FileNotFoundError: return f"Error: File not found: {path}" except PermissionError: return f"Error: Permission denied reading file: {path}" except IsADirectoryError: return f"Error: Path is a directory, not a file: {path}" except Exception as e: return f"Error: Unexpected error reading file: {type(e).__name__}: {e}" @tool("write_file", parse_docstring=True) def write_file_tool( runtime: ToolRuntime[ContextT, ThreadState], description: str, path: str, content: str, append: bool = False, ) -> str: """Write text content to a file. Args: description: Explain why you are writing to this file in short words. ALWAYS PROVIDE THIS PARAMETER FIRST. path: The **absolute** path to the file to write to. ALWAYS PROVIDE THIS PARAMETER SECOND. content: The content to write to the file. ALWAYS PROVIDE THIS PARAMETER THIRD. """ try: sandbox = ensure_sandbox_initialized(runtime) ensure_thread_directories_exist(runtime) if is_local_sandbox(runtime): thread_data = get_thread_data(runtime) path = replace_virtual_path(path, thread_data) sandbox.write_file(path, content, append) return "OK" except SandboxError as e: return f"Error: {e}" except PermissionError: return f"Error: Permission denied writing to file: {path}" except IsADirectoryError: return f"Error: Path is a directory, not a file: {path}" except OSError as e: return f"Error: Failed to write file '{path}': {e}" except Exception as e: return f"Error: Unexpected error writing file: {type(e).__name__}: {e}" @tool("str_replace", parse_docstring=True) def str_replace_tool( runtime: ToolRuntime[ContextT, ThreadState], description: str, path: str, old_str: str, new_str: str, replace_all: bool = False, ) -> str: """Replace a substring in a file with another substring. If `replace_all` is False (default), the substring to replace must appear **exactly once** in the file. Args: description: Explain why you are replacing the substring in short words. ALWAYS PROVIDE THIS PARAMETER FIRST. path: The **absolute** path to the file to replace the substring in. ALWAYS PROVIDE THIS PARAMETER SECOND. old_str: The substring to replace. ALWAYS PROVIDE THIS PARAMETER THIRD. new_str: The new substring. ALWAYS PROVIDE THIS PARAMETER FOURTH. replace_all: Whether to replace all occurrences of the substring. If False, only the first occurrence will be replaced. Default is False. """ try: sandbox = ensure_sandbox_initialized(runtime) ensure_thread_directories_exist(runtime) if is_local_sandbox(runtime): thread_data = get_thread_data(runtime) path = replace_virtual_path(path, thread_data) content = sandbox.read_file(path) if not content: return "OK" if old_str not in content: return f"Error: String to replace not found in file: {path}" if replace_all: content = content.replace(old_str, new_str) else: content = content.replace(old_str, new_str, 1) sandbox.write_file(path, content) return "OK" except SandboxError as e: return f"Error: {e}" except FileNotFoundError: return f"Error: File not found: {path}" except PermissionError: return f"Error: Permission denied accessing file: {path}" except Exception as e: return f"Error: Unexpected error replacing string: {type(e).__name__}: {e}"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/sandbox/tools.py", "license": "MIT License", "lines": 332, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/skills/loader.py
import os from pathlib import Path from .parser import parse_skill_file from .types import Skill def get_skills_root_path() -> Path: """ Get the root path of the skills directory. Returns: Path to the skills directory (deer-flow/skills) """ # backend directory is current file's parent's parent's parent backend_dir = Path(__file__).resolve().parent.parent.parent # skills directory is sibling to backend directory skills_dir = backend_dir.parent / "skills" return skills_dir def load_skills(skills_path: Path | None = None, use_config: bool = True, enabled_only: bool = False) -> list[Skill]: """ Load all skills from the skills directory. Scans both public and custom skill directories, parsing SKILL.md files to extract metadata. The enabled state is determined by the skills_state_config.json file. Args: skills_path: Optional custom path to skills directory. If not provided and use_config is True, uses path from config. Otherwise defaults to deer-flow/skills use_config: Whether to load skills path from config (default: True) enabled_only: If True, only return enabled skills (default: False) Returns: List of Skill objects, sorted by name """ if skills_path is None: if use_config: try: from src.config import get_app_config config = get_app_config() skills_path = config.skills.get_skills_path() except Exception: # Fallback to default if config fails skills_path = get_skills_root_path() else: skills_path = get_skills_root_path() if not skills_path.exists(): return [] skills = [] # Scan public and custom directories for category in ["public", "custom"]: category_path = skills_path / category if not category_path.exists() or not category_path.is_dir(): continue for current_root, dir_names, file_names in os.walk(category_path): # Keep traversal deterministic and skip hidden directories. dir_names[:] = sorted(name for name in dir_names if not name.startswith(".")) if "SKILL.md" not in file_names: continue skill_file = Path(current_root) / "SKILL.md" relative_path = skill_file.parent.relative_to(category_path) skill = parse_skill_file(skill_file, category=category, relative_path=relative_path) if skill: skills.append(skill) # Load skills state configuration and update enabled status # NOTE: We use ExtensionsConfig.from_file() instead of get_extensions_config() # to always read the latest configuration from disk. This ensures that changes # made through the Gateway API (which runs in a separate process) are immediately # reflected in the LangGraph Server when loading skills. try: from src.config.extensions_config import ExtensionsConfig extensions_config = ExtensionsConfig.from_file() for skill in skills: skill.enabled = extensions_config.is_skill_enabled(skill.name, skill.category) except Exception as e: # If config loading fails, default to all enabled print(f"Warning: Failed to load extensions config: {e}") # Filter by enabled status if requested if enabled_only: skills = [skill for skill in skills if skill.enabled] # Sort by name for consistent ordering skills.sort(key=lambda s: s.name) return skills
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/skills/loader.py", "license": "MIT License", "lines": 77, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
bytedance/deer-flow:backend/src/skills/parser.py
import re from pathlib import Path from .types import Skill def parse_skill_file(skill_file: Path, category: str, relative_path: Path | None = None) -> Skill | None: """ Parse a SKILL.md file and extract metadata. Args: skill_file: Path to the SKILL.md file category: Category of the skill ('public' or 'custom') Returns: Skill object if parsing succeeds, None otherwise """ if not skill_file.exists() or skill_file.name != "SKILL.md": return None try: content = skill_file.read_text(encoding="utf-8") # Extract YAML front matter # Pattern: ---\nkey: value\n--- front_matter_match = re.match(r"^---\s*\n(.*?)\n---\s*\n", content, re.DOTALL) if not front_matter_match: return None front_matter = front_matter_match.group(1) # Parse YAML front matter (simple key-value parsing) metadata = {} for line in front_matter.split("\n"): line = line.strip() if not line: continue if ":" in line: key, value = line.split(":", 1) metadata[key.strip()] = value.strip() # Extract required fields name = metadata.get("name") description = metadata.get("description") if not name or not description: return None license_text = metadata.get("license") return Skill( name=name, description=description, license=license_text, skill_dir=skill_file.parent, skill_file=skill_file, relative_path=relative_path or Path(skill_file.parent.name), category=category, enabled=True, # Default to enabled, actual state comes from config file ) except Exception as e: print(f"Error parsing skill file {skill_file}: {e}") return None
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/skills/parser.py", "license": "MIT License", "lines": 50, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/skills/types.py
from dataclasses import dataclass from pathlib import Path @dataclass class Skill: """Represents a skill with its metadata and file path""" name: str description: str license: str | None skill_dir: Path skill_file: Path relative_path: Path # Relative path from category root to skill directory category: str # 'public' or 'custom' enabled: bool = False # Whether this skill is enabled @property def skill_path(self) -> str: """Returns the relative path from the category root (skills/{category}) to this skill's directory""" path = self.relative_path.as_posix() return "" if path == "." else path def get_container_path(self, container_base_path: str = "/mnt/skills") -> str: """ Get the full path to this skill in the container. Args: container_base_path: Base path where skills are mounted in the container Returns: Full container path to the skill directory """ category_base = f"{container_base_path}/{self.category}" skill_path = self.skill_path if skill_path: return f"{category_base}/{skill_path}" return category_base def get_container_file_path(self, container_base_path: str = "/mnt/skills") -> str: """ Get the full path to this skill's main file (SKILL.md) in the container. Args: container_base_path: Base path where skills are mounted in the container Returns: Full container path to the skill's SKILL.md file """ return f"{self.get_container_path(container_base_path)}/SKILL.md" def __repr__(self) -> str: return f"Skill(name={self.name!r}, description={self.description!r}, category={self.category!r})"
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/skills/types.py", "license": "MIT License", "lines": 42, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
bytedance/deer-flow:backend/src/subagents/builtins/bash_agent.py
"""Bash command execution subagent configuration.""" from src.subagents.config import SubagentConfig BASH_AGENT_CONFIG = SubagentConfig( name="bash", description="""Command execution specialist for running bash commands in a separate context. Use this subagent when: - You need to run a series of related bash commands - Terminal operations like git, npm, docker, etc. - Command output is verbose and would clutter main context - Build, test, or deployment operations Do NOT use for simple single commands - use bash tool directly instead.""", system_prompt="""You are a bash command execution specialist. Execute the requested commands carefully and report results clearly. <guidelines> - Execute commands one at a time when they depend on each other - Use parallel execution when commands are independent - Report both stdout and stderr when relevant - Handle errors gracefully and explain what went wrong - Use absolute paths for file operations - Be cautious with destructive operations (rm, overwrite, etc.) </guidelines> <output_format> For each command or group of commands: 1. What was executed 2. The result (success/failure) 3. Relevant output (summarized if verbose) 4. Any errors or warnings </output_format> <working_directory> You have access to the sandbox environment: - User uploads: `/mnt/user-data/uploads` - User workspace: `/mnt/user-data/workspace` - Output files: `/mnt/user-data/outputs` </working_directory> """, tools=["bash", "ls", "read_file", "write_file", "str_replace"], # Sandbox tools only disallowed_tools=["task", "ask_clarification", "present_files"], model="inherit", max_turns=30, )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/subagents/builtins/bash_agent.py", "license": "MIT License", "lines": 39, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/subagents/builtins/general_purpose.py
"""General-purpose subagent configuration.""" from src.subagents.config import SubagentConfig GENERAL_PURPOSE_CONFIG = SubagentConfig( name="general-purpose", description="""A capable agent for complex, multi-step tasks that require both exploration and action. Use this subagent when: - The task requires both exploration and modification - Complex reasoning is needed to interpret results - Multiple dependent steps must be executed - The task would benefit from isolated context management Do NOT use for simple, single-step operations.""", system_prompt="""You are a general-purpose subagent working on a delegated task. Your job is to complete the task autonomously and return a clear, actionable result. <guidelines> - Focus on completing the delegated task efficiently - Use available tools as needed to accomplish the goal - Think step by step but act decisively - If you encounter issues, explain them clearly in your response - Return a concise summary of what you accomplished - Do NOT ask for clarification - work with the information provided </guidelines> <output_format> When you complete the task, provide: 1. A brief summary of what was accomplished 2. Key findings or results 3. Any relevant file paths, data, or artifacts created 4. Issues encountered (if any) 5. Citations: Use `[citation:Title](URL)` format for external sources </output_format> <working_directory> You have access to the same sandbox environment as the parent agent: - User uploads: `/mnt/user-data/uploads` - User workspace: `/mnt/user-data/workspace` - Output files: `/mnt/user-data/outputs` </working_directory> """, tools=None, # Inherit all tools from parent disallowed_tools=["task", "ask_clarification", "present_files"], # Prevent nesting and clarification model="inherit", max_turns=50, )
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/subagents/builtins/general_purpose.py", "license": "MIT License", "lines": 40, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
bytedance/deer-flow:backend/src/subagents/config.py
"""Subagent configuration definitions.""" from dataclasses import dataclass, field @dataclass class SubagentConfig: """Configuration for a subagent. Attributes: name: Unique identifier for the subagent. description: When Claude should delegate to this subagent. system_prompt: The system prompt that guides the subagent's behavior. tools: Optional list of tool names to allow. If None, inherits all tools. disallowed_tools: Optional list of tool names to deny. model: Model to use - 'inherit' uses parent's model. max_turns: Maximum number of agent turns before stopping. timeout_seconds: Maximum execution time in seconds (default: 900 = 15 minutes). """ name: str description: str system_prompt: str tools: list[str] | None = None disallowed_tools: list[str] | None = field(default_factory=lambda: ["task"]) model: str = "inherit" max_turns: int = 50 timeout_seconds: int = 900
{ "repo_id": "bytedance/deer-flow", "file_path": "backend/src/subagents/config.py", "license": "MIT License", "lines": 23, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation