repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/test_stress_api.py
tests/memory/test_stress_api.py
#!/usr/bin/env python3 """ Stress test for Crawl4AI's Docker API server (/crawl and /crawl/stream endpoints). This version targets a running Crawl4AI API server, sending concurrent requests to test its ability to handle multiple crawl jobs simultaneously. It uses httpx for async HTTP requests and logs results per batch of requests, including server-side memory usage reported by the API. """ import asyncio import time import uuid import argparse import json import sys import os import shutil from typing import List, Dict, Optional, Union, AsyncGenerator, Tuple import httpx import pathlib # Import pathlib explicitly from rich.console import Console from rich.panel import Panel from rich.syntax import Syntax # --- Constants --- DEFAULT_API_URL = "http://localhost:11235" # Default port DEFAULT_API_URL = "http://localhost:8020" # Default port DEFAULT_URL_COUNT = 100 DEFAULT_MAX_CONCURRENT_REQUESTS = 1 DEFAULT_CHUNK_SIZE = 10 DEFAULT_REPORT_PATH = "reports_api" DEFAULT_STREAM_MODE = True REQUEST_TIMEOUT = 180.0 # Initialize Rich console console = Console() # --- API Health Check (Unchanged) --- async def check_server_health(client: httpx.AsyncClient, health_endpoint: str = "/health"): """Check if the API server is healthy.""" console.print(f"[bold cyan]Checking API server health at {client.base_url}{health_endpoint}...[/]", end="") try: response = await client.get(health_endpoint, timeout=10.0) response.raise_for_status() health_data = response.json() version = health_data.get('version', 'N/A') console.print(f"[bold green] Server OK! Version: {version}[/]") return True except (httpx.RequestError, httpx.HTTPStatusError) as e: console.print(f"\n[bold red]Server health check FAILED:[/]") console.print(f"Error: {e}") console.print(f"Is the server running and accessible at {client.base_url}?") return False except Exception as e: console.print(f"\n[bold red]An unexpected error occurred during health check:[/]") console.print(e) return False # --- API Stress Test Class --- class ApiStressTest: """Orchestrates the stress test by sending concurrent requests to the API.""" def __init__( self, api_url: str, url_count: int, max_concurrent_requests: int, chunk_size: int, report_path: str, stream_mode: bool, ): self.api_base_url = api_url.rstrip('/') self.url_count = url_count self.max_concurrent_requests = max_concurrent_requests self.chunk_size = chunk_size self.report_path = pathlib.Path(report_path) self.report_path.mkdir(parents=True, exist_ok=True) self.stream_mode = stream_mode # Ignore repo path and set it to current file path self.repo_path = pathlib.Path(__file__).parent.resolve() self.test_id = time.strftime("%Y%m%d_%H%M%S") self.results_summary = { "test_id": self.test_id, "api_url": api_url, "url_count": url_count, "max_concurrent_requests": max_concurrent_requests, "chunk_size": chunk_size, "stream_mode": stream_mode, "start_time": "", "end_time": "", "total_time_seconds": 0, "successful_requests": 0, "failed_requests": 0, "successful_urls": 0, "failed_urls": 0, "total_urls_processed": 0, "total_api_calls": 0, "server_memory_metrics": { # To store aggregated server memory info "batch_mode_avg_delta_mb": None, "batch_mode_max_delta_mb": None, "stream_mode_avg_max_snapshot_mb": None, "stream_mode_max_max_snapshot_mb": None, "samples": [] # Store individual request memory results } } self.http_client = httpx.AsyncClient(base_url=self.api_base_url, timeout=REQUEST_TIMEOUT, limits=httpx.Limits(max_connections=max_concurrent_requests + 5, max_keepalive_connections=max_concurrent_requests)) async def close_client(self): """Close the httpx client.""" await self.http_client.aclose() async def run(self) -> Dict: """Run the API stress test.""" # No client memory tracker needed urls_to_process = [f"https://httpbin.org/anything/{uuid.uuid4()}" for _ in range(self.url_count)] url_chunks = [urls_to_process[i:i+self.chunk_size] for i in range(0, len(urls_to_process), self.chunk_size)] self.results_summary["start_time"] = time.strftime("%Y-%m-%d %H:%M:%S") start_time = time.time() console.print(f"\n[bold cyan]Crawl4AI API Stress Test - {self.url_count} URLs, {self.max_concurrent_requests} concurrent requests[/bold cyan]") console.print(f"[bold cyan]Target API:[/bold cyan] {self.api_base_url}, [bold cyan]Mode:[/bold cyan] {'Streaming' if self.stream_mode else 'Batch'}, [bold cyan]URLs per Request:[/bold cyan] {self.chunk_size}") # Removed client memory log semaphore = asyncio.Semaphore(self.max_concurrent_requests) # Updated Batch logging header console.print("\n[bold]API Request Batch Progress:[/bold]") # Adjusted spacing and added Peak console.print("[bold] Batch | Progress | SrvMem Peak / Δ|Max (MB) | Reqs/sec | S/F URLs | Time (s) | Status [/bold]") # Adjust separator length if needed, looks okay for now console.print("─" * 95) # No client memory monitor task needed tasks = [] total_api_calls = len(url_chunks) self.results_summary["total_api_calls"] = total_api_calls try: for i, chunk in enumerate(url_chunks): task = asyncio.create_task(self._make_api_request( chunk=chunk, batch_idx=i + 1, total_batches=total_api_calls, semaphore=semaphore # No memory tracker passed )) tasks.append(task) api_results = await asyncio.gather(*tasks) # Process aggregated results including server memory total_successful_requests = sum(1 for r in api_results if r['request_success']) total_failed_requests = total_api_calls - total_successful_requests total_successful_urls = sum(r['success_urls'] for r in api_results) total_failed_urls = sum(r['failed_urls'] for r in api_results) total_urls_processed = total_successful_urls + total_failed_urls # Aggregate server memory metrics valid_samples = [r for r in api_results if r.get('server_delta_or_max_mb') is not None] # Filter results with valid mem data self.results_summary["server_memory_metrics"]["samples"] = valid_samples # Store raw samples with both peak and delta/max if valid_samples: delta_or_max_values = [r['server_delta_or_max_mb'] for r in valid_samples] if self.stream_mode: # Stream mode: delta_or_max holds max snapshot self.results_summary["server_memory_metrics"]["stream_mode_avg_max_snapshot_mb"] = sum(delta_or_max_values) / len(delta_or_max_values) self.results_summary["server_memory_metrics"]["stream_mode_max_max_snapshot_mb"] = max(delta_or_max_values) else: # Batch mode # delta_or_max holds delta self.results_summary["server_memory_metrics"]["batch_mode_avg_delta_mb"] = sum(delta_or_max_values) / len(delta_or_max_values) self.results_summary["server_memory_metrics"]["batch_mode_max_delta_mb"] = max(delta_or_max_values) # Aggregate peak values for batch mode peak_values = [r['server_peak_memory_mb'] for r in valid_samples if r.get('server_peak_memory_mb') is not None] if peak_values: self.results_summary["server_memory_metrics"]["batch_mode_avg_peak_mb"] = sum(peak_values) / len(peak_values) self.results_summary["server_memory_metrics"]["batch_mode_max_peak_mb"] = max(peak_values) self.results_summary.update({ "successful_requests": total_successful_requests, "failed_requests": total_failed_requests, "successful_urls": total_successful_urls, "failed_urls": total_failed_urls, "total_urls_processed": total_urls_processed, }) except Exception as e: console.print(f"[bold red]An error occurred during task execution: {e}[/bold red]") import traceback traceback.print_exc() # No finally block needed for monitor task end_time = time.time() self.results_summary.update({ "end_time": time.strftime("%Y-%m-%d %H:%M:%S"), "total_time_seconds": end_time - start_time, # No client memory report }) self._save_results() return self.results_summary async def _make_api_request( self, chunk: List[str], batch_idx: int, total_batches: int, semaphore: asyncio.Semaphore # No memory tracker ) -> Dict: """Makes a single API request for a chunk of URLs, handling concurrency and logging server memory.""" request_success = False success_urls = 0 failed_urls = 0 status = "Pending" status_color = "grey" server_memory_metric = None # Store delta (batch) or max snapshot (stream) api_call_start_time = time.time() async with semaphore: try: # No client memory sampling endpoint = "/crawl/stream" if self.stream_mode else "/crawl" payload = { "urls": chunk, "browser_config": {"type": "BrowserConfig", "params": {"headless": True}}, "crawler_config": { "type": "CrawlerRunConfig", "params": {"cache_mode": "BYPASS", "stream": self.stream_mode} } } if self.stream_mode: max_server_mem_snapshot = 0.0 # Track max memory seen in this stream async with self.http_client.stream("POST", endpoint, json=payload) as response: initial_status_code = response.status_code response.raise_for_status() completed_marker_received = False async for line in response.aiter_lines(): if line: try: data = json.loads(line) if data.get("status") == "completed": completed_marker_received = True break elif data.get("url"): if data.get("success"): success_urls += 1 else: failed_urls += 1 # Extract server memory snapshot per result mem_snapshot = data.get('server_memory_mb') if mem_snapshot is not None: max_server_mem_snapshot = max(max_server_mem_snapshot, float(mem_snapshot)) except json.JSONDecodeError: console.print(f"[Batch {batch_idx}] [red]Stream decode error for line:[/red] {line}") failed_urls = len(chunk) break request_success = completed_marker_received if not request_success: failed_urls = len(chunk) - success_urls server_memory_metric = max_server_mem_snapshot # Use max snapshot for stream logging else: # Batch mode response = await self.http_client.post(endpoint, json=payload) response.raise_for_status() data = response.json() # Extract server memory delta from the response server_memory_metric = data.get('server_memory_delta_mb') server_peak_mem_mb = data.get('server_peak_memory_mb') if data.get("success") and "results" in data: request_success = True results_list = data.get("results", []) for result_item in results_list: if result_item.get("success"): success_urls += 1 else: failed_urls += 1 if len(results_list) != len(chunk): console.print(f"[Batch {batch_idx}] [yellow]Warning: Result count ({len(results_list)}) doesn't match URL count ({len(chunk)})[/yellow]") failed_urls = len(chunk) - success_urls else: request_success = False failed_urls = len(chunk) # Try to get memory from error detail if available detail = data.get('detail') if isinstance(detail, str): try: detail_json = json.loads(detail) except: detail_json = {} elif isinstance(detail, dict): detail_json = detail else: detail_json = {} server_peak_mem_mb = detail_json.get('server_peak_memory_mb', None) server_memory_metric = detail_json.get('server_memory_delta_mb', None) console.print(f"[Batch {batch_idx}] [red]API request failed:[/red] {detail_json.get('error', 'No details')}") except httpx.HTTPStatusError as e: request_success = False failed_urls = len(chunk) console.print(f"[Batch {batch_idx}] [bold red]HTTP Error {e.response.status_code}:[/] {e.request.url}") try: error_detail = e.response.json() # Attempt to extract memory info even from error responses detail_content = error_detail.get('detail', {}) if isinstance(detail_content, str): # Handle if detail is stringified JSON try: detail_content = json.loads(detail_content) except: detail_content = {} server_memory_metric = detail_content.get('server_memory_delta_mb', None) server_peak_mem_mb = detail_content.get('server_peak_memory_mb', None) console.print(f"Response: {error_detail}") except Exception: console.print(f"Response Text: {e.response.text[:200]}...") except httpx.RequestError as e: request_success = False failed_urls = len(chunk) console.print(f"[Batch {batch_idx}] [bold red]Request Error:[/bold] {e.request.url} - {e}") except Exception as e: request_success = False failed_urls = len(chunk) console.print(f"[Batch {batch_idx}] [bold red]Unexpected Error:[/bold] {e}") import traceback traceback.print_exc() finally: api_call_time = time.time() - api_call_start_time total_processed_urls = success_urls + failed_urls if request_success and failed_urls == 0: status_color, status = "green", "Success" elif request_success and success_urls > 0: status_color, status = "yellow", "Partial" else: status_color, status = "red", "Failed" current_total_urls = batch_idx * self.chunk_size progress_pct = min(100.0, (current_total_urls / self.url_count) * 100) reqs_per_sec = 1.0 / api_call_time if api_call_time > 0 else float('inf') # --- New Memory Formatting --- mem_display = " N/A " # Default peak_mem_value = None delta_or_max_value = None if self.stream_mode: # server_memory_metric holds max snapshot for stream if server_memory_metric is not None: mem_display = f"{server_memory_metric:.1f} (Max)" delta_or_max_value = server_memory_metric # Store for aggregation else: # Batch mode - expect peak and delta # We need to get peak and delta from the API response peak_mem_value = locals().get('server_peak_mem_mb', None) # Get from response data if available delta_value = server_memory_metric # server_memory_metric holds delta for batch if peak_mem_value is not None and delta_value is not None: mem_display = f"{peak_mem_value:.1f} / {delta_value:+.1f}" delta_or_max_value = delta_value # Store delta for aggregation elif peak_mem_value is not None: mem_display = f"{peak_mem_value:.1f} / N/A" elif delta_value is not None: mem_display = f"N/A / {delta_value:+.1f}" delta_or_max_value = delta_value # Store delta for aggregation # --- Updated Print Statement with Adjusted Padding --- console.print( f" {batch_idx:<5} | {progress_pct:6.1f}% | {mem_display:>24} | {reqs_per_sec:8.1f} | " # Increased width for memory column f"{success_urls:^7}/{failed_urls:<6} | {api_call_time:8.2f} | [{status_color}]{status:<7}[/{status_color}] " # Added trailing space ) # --- Updated Return Dictionary --- return_data = { "batch_idx": batch_idx, "request_success": request_success, "success_urls": success_urls, "failed_urls": failed_urls, "time": api_call_time, # Return both peak (if available) and delta/max "server_peak_memory_mb": peak_mem_value, # Will be None for stream mode "server_delta_or_max_mb": delta_or_max_value # Delta for batch, Max for stream } # Add back the specific batch mode delta if needed elsewhere, but delta_or_max covers it # if not self.stream_mode: # return_data["server_memory_delta_mb"] = delta_value return return_data # No _periodic_memory_sample needed def _save_results(self) -> None: """Saves the results summary to a JSON file.""" results_path = self.report_path / f"api_test_summary_{self.test_id}.json" try: # No client memory path to convert with open(results_path, 'w', encoding='utf-8') as f: json.dump(self.results_summary, f, indent=2, default=str) except Exception as e: console.print(f"[bold red]Failed to save results summary: {e}[/bold red]") # --- run_full_test Function --- async def run_full_test(args): """Runs the full API stress test process.""" client = httpx.AsyncClient(base_url=args.api_url, timeout=REQUEST_TIMEOUT) if not await check_server_health(client): console.print("[bold red]Aborting test due to server health check failure.[/]") await client.aclose() return await client.aclose() test = ApiStressTest( api_url=args.api_url, url_count=args.urls, max_concurrent_requests=args.max_concurrent_requests, chunk_size=args.chunk_size, report_path=args.report_path, stream_mode=args.stream, ) results = {} try: results = await test.run() finally: await test.close_client() if not results: console.print("[bold red]Test did not produce results.[/bold red]") return console.print("\n" + "=" * 80) console.print("[bold green]API Stress Test Completed[/bold green]") console.print("=" * 80) success_rate_reqs = results["successful_requests"] / results["total_api_calls"] * 100 if results["total_api_calls"] > 0 else 0 success_rate_urls = results["successful_urls"] / results["url_count"] * 100 if results["url_count"] > 0 else 0 urls_per_second = results["total_urls_processed"] / results["total_time_seconds"] if results["total_time_seconds"] > 0 else 0 reqs_per_second = results["total_api_calls"] / results["total_time_seconds"] if results["total_time_seconds"] > 0 else 0 console.print(f"[bold cyan]Test ID:[/bold cyan] {results['test_id']}") console.print(f"[bold cyan]Target API:[/bold cyan] {results['api_url']}") console.print(f"[bold cyan]Configuration:[/bold cyan] {results['url_count']} URLs, {results['max_concurrent_requests']} concurrent client requests, URLs/Req: {results['chunk_size']}, Stream: {results['stream_mode']}") console.print(f"[bold cyan]API Requests:[/bold cyan] {results['successful_requests']} successful, {results['failed_requests']} failed ({results['total_api_calls']} total, {success_rate_reqs:.1f}% success)") console.print(f"[bold cyan]URL Processing:[/bold cyan] {results['successful_urls']} successful, {results['failed_urls']} failed ({results['total_urls_processed']} processed, {success_rate_urls:.1f}% success)") console.print(f"[bold cyan]Performance:[/bold cyan] {results['total_time_seconds']:.2f}s total | Avg Reqs/sec: {reqs_per_second:.2f} | Avg URLs/sec: {urls_per_second:.2f}") # Report Server Memory mem_metrics = results.get("server_memory_metrics", {}) mem_samples = mem_metrics.get("samples", []) if mem_samples: num_samples = len(mem_samples) if results['stream_mode']: avg_mem = mem_metrics.get("stream_mode_avg_max_snapshot_mb") max_mem = mem_metrics.get("stream_mode_max_max_snapshot_mb") avg_str = f"{avg_mem:.1f}" if avg_mem is not None else "N/A" max_str = f"{max_mem:.1f}" if max_mem is not None else "N/A" console.print(f"[bold cyan]Server Memory (Stream):[/bold cyan] Avg Max Snapshot: {avg_str} MB | Max Max Snapshot: {max_str} MB (across {num_samples} requests)") else: # Batch mode avg_delta = mem_metrics.get("batch_mode_avg_delta_mb") max_delta = mem_metrics.get("batch_mode_max_delta_mb") avg_peak = mem_metrics.get("batch_mode_avg_peak_mb") max_peak = mem_metrics.get("batch_mode_max_peak_mb") avg_delta_str = f"{avg_delta:.1f}" if avg_delta is not None else "N/A" max_delta_str = f"{max_delta:.1f}" if max_delta is not None else "N/A" avg_peak_str = f"{avg_peak:.1f}" if avg_peak is not None else "N/A" max_peak_str = f"{max_peak:.1f}" if max_peak is not None else "N/A" console.print(f"[bold cyan]Server Memory (Batch):[/bold cyan] Avg Peak: {avg_peak_str} MB | Max Peak: {max_peak_str} MB | Avg Delta: {avg_delta_str} MB | Max Delta: {max_delta_str} MB (across {num_samples} requests)") else: console.print("[bold cyan]Server Memory:[/bold cyan] No memory data reported by server.") # No client memory report summary_path = pathlib.Path(args.report_path) / f"api_test_summary_{results['test_id']}.json" console.print(f"[bold green]Results summary saved to {summary_path}[/bold green]") if results["failed_requests"] > 0: console.print(f"\n[bold yellow]Warning: {results['failed_requests']} API requests failed ({100-success_rate_reqs:.1f}% failure rate)[/bold yellow]") if results["failed_urls"] > 0: console.print(f"[bold yellow]Warning: {results['failed_urls']} URLs failed to process ({100-success_rate_urls:.1f}% URL failure rate)[/bold yellow]") if results["total_urls_processed"] < results["url_count"]: console.print(f"\n[bold red]Error: Only {results['total_urls_processed']} out of {results['url_count']} target URLs were processed![/bold red]") # --- main Function (Argument parsing mostly unchanged) --- def main(): """Main entry point for the script.""" parser = argparse.ArgumentParser(description="Crawl4AI API Server Stress Test") parser.add_argument("--api-url", type=str, default=DEFAULT_API_URL, help=f"Base URL of the Crawl4AI API server (default: {DEFAULT_API_URL})") parser.add_argument("--urls", type=int, default=DEFAULT_URL_COUNT, help=f"Total number of unique URLs to process via API calls (default: {DEFAULT_URL_COUNT})") parser.add_argument("--max-concurrent-requests", type=int, default=DEFAULT_MAX_CONCURRENT_REQUESTS, help=f"Maximum concurrent API requests from this client (default: {DEFAULT_MAX_CONCURRENT_REQUESTS})") parser.add_argument("--chunk-size", type=int, default=DEFAULT_CHUNK_SIZE, help=f"Number of URLs per API request payload (default: {DEFAULT_CHUNK_SIZE})") parser.add_argument("--stream", action="store_true", default=DEFAULT_STREAM_MODE, help=f"Use the /crawl/stream endpoint instead of /crawl (default: {DEFAULT_STREAM_MODE})") parser.add_argument("--report-path", type=str, default=DEFAULT_REPORT_PATH, help=f"Path to save reports and logs (default: {DEFAULT_REPORT_PATH})") parser.add_argument("--clean-reports", action="store_true", help="Clean up report directory before running") args = parser.parse_args() console.print("[bold underline]Crawl4AI API Stress Test Configuration[/bold underline]") console.print(f"API URL: {args.api_url}") console.print(f"Total URLs: {args.urls}, Concurrent Client Requests: {args.max_concurrent_requests}, URLs per Request: {args.chunk_size}") console.print(f"Mode: {'Streaming' if args.stream else 'Batch'}") console.print(f"Report Path: {args.report_path}") console.print("-" * 40) if args.clean_reports: console.print("[cyan]Option: Clean reports before test[/cyan]") console.print("-" * 40) if args.clean_reports: report_dir = pathlib.Path(args.report_path) if report_dir.exists(): console.print(f"[yellow]Cleaning up reports directory: {args.report_path}[/yellow]") shutil.rmtree(args.report_path) report_dir.mkdir(parents=True, exist_ok=True) try: asyncio.run(run_full_test(args)) except KeyboardInterrupt: console.print("\n[bold yellow]Test interrupted by user.[/bold yellow]") except Exception as e: console.print(f"\n[bold red]An unexpected error occurred:[/bold red] {e}") import traceback traceback.print_exc() if __name__ == "__main__": # No need to modify sys.path for SimpleMemoryTracker as it's removed main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/run_benchmark.py
tests/memory/run_benchmark.py
#!/usr/bin/env python3 """ Run a complete Crawl4AI benchmark test using test_stress_sdk.py and generate a report. """ import sys import os import glob import argparse import subprocess import time from datetime import datetime from rich.console import Console from rich.text import Text console = Console() # Updated TEST_CONFIGS to use max_sessions TEST_CONFIGS = { "quick": {"urls": 50, "max_sessions": 4, "chunk_size": 10, "description": "Quick test (50 URLs, 4 sessions)"}, "small": {"urls": 100, "max_sessions": 8, "chunk_size": 20, "description": "Small test (100 URLs, 8 sessions)"}, "medium": {"urls": 500, "max_sessions": 16, "chunk_size": 50, "description": "Medium test (500 URLs, 16 sessions)"}, "large": {"urls": 1000, "max_sessions": 32, "chunk_size": 100,"description": "Large test (1000 URLs, 32 sessions)"}, "extreme": {"urls": 2000, "max_sessions": 64, "chunk_size": 200,"description": "Extreme test (2000 URLs, 64 sessions)"}, } # Arguments to forward directly if present in custom_args FORWARD_ARGS = { "urls": "--urls", "max_sessions": "--max-sessions", "chunk_size": "--chunk-size", "port": "--port", "monitor_mode": "--monitor-mode", } # Boolean flags to forward if True FORWARD_FLAGS = { "stream": "--stream", "use_rate_limiter": "--use-rate-limiter", "keep_server_alive": "--keep-server-alive", "use_existing_site": "--use-existing-site", "skip_generation": "--skip-generation", "keep_site": "--keep-site", "clean_reports": "--clean-reports", # Note: clean behavior is handled here, but pass flag if needed "clean_site": "--clean-site", # Note: clean behavior is handled here, but pass flag if needed } def run_benchmark(config_name, custom_args=None, compare=True, clean=False): """Runs the stress test and optionally the report generator.""" if config_name not in TEST_CONFIGS and config_name != "custom": console.print(f"[bold red]Unknown configuration: {config_name}[/bold red]") return False # Print header title = "Crawl4AI SDK Benchmark Test" if config_name != "custom": title += f" - {TEST_CONFIGS[config_name]['description']}" else: # Safely get custom args for title urls = custom_args.get('urls', '?') if custom_args else '?' sessions = custom_args.get('max_sessions', '?') if custom_args else '?' title += f" - Custom ({urls} URLs, {sessions} sessions)" console.print(f"\n[bold blue]{title}[/bold blue]") console.print("=" * (len(title) + 4)) # Adjust underline length console.print("\n[bold white]Preparing test...[/bold white]") # --- Command Construction --- # Use the new script name cmd = ["python", "test_stress_sdk.py"] # Apply config or custom args args_to_use = {} if config_name != "custom": args_to_use = TEST_CONFIGS[config_name].copy() # If custom args are provided (e.g., boolean flags), overlay them if custom_args: args_to_use.update(custom_args) elif custom_args: # Custom config args_to_use = custom_args.copy() # Add arguments with values for key, arg_name in FORWARD_ARGS.items(): if key in args_to_use: cmd.extend([arg_name, str(args_to_use[key])]) # Add boolean flags for key, flag_name in FORWARD_FLAGS.items(): if args_to_use.get(key, False): # Check if key exists and is True # Special handling for clean flags - apply locally, don't forward? # Decide if test_stress_sdk.py also needs --clean flags or if run_benchmark handles it. # For now, let's assume run_benchmark handles cleaning based on its own --clean flag. # We'll forward other flags. if key not in ["clean_reports", "clean_site"]: cmd.append(flag_name) # Handle the top-level --clean flag for run_benchmark if clean: # Pass clean flags to the stress test script as well, if needed # This assumes test_stress_sdk.py also uses --clean-reports and --clean-site cmd.append("--clean-reports") cmd.append("--clean-site") console.print("[yellow]Applying --clean: Cleaning reports and site before test.[/yellow]") # Actual cleaning logic might reside here or be delegated entirely console.print(f"\n[bold white]Running stress test:[/bold white] {' '.join(cmd)}") start = time.time() # Execute the stress test script # Use Popen to stream output try: proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, text=True, encoding='utf-8', errors='replace') while True: line = proc.stdout.readline() if not line: break console.print(line.rstrip()) # Print line by line proc.wait() # Wait for the process to complete except FileNotFoundError: console.print(f"[bold red]Error: Script 'test_stress_sdk.py' not found. Make sure it's in the correct directory.[/bold red]") return False except Exception as e: console.print(f"[bold red]Error running stress test subprocess: {e}[/bold red]") return False if proc.returncode != 0: console.print(f"[bold red]Stress test failed with exit code {proc.returncode}[/bold red]") return False duration = time.time() - start console.print(f"[bold green]Stress test completed in {duration:.1f} seconds[/bold green]") # --- Report Generation (Optional) --- if compare: # Assuming benchmark_report.py exists and works with the generated reports report_script = "benchmark_report.py" # Keep configurable if needed report_cmd = ["python", report_script] console.print(f"\n[bold white]Generating benchmark report: {' '.join(report_cmd)}[/bold white]") # Run the report command and capture output try: report_proc = subprocess.run(report_cmd, capture_output=True, text=True, check=False, encoding='utf-8', errors='replace') # Use check=False to handle potential errors # Print the captured output from benchmark_report.py if report_proc.stdout: console.print("\n" + report_proc.stdout) if report_proc.stderr: console.print("[yellow]Report generator stderr:[/yellow]\n" + report_proc.stderr) if report_proc.returncode != 0: console.print(f"[bold yellow]Benchmark report generation script '{report_script}' failed with exit code {report_proc.returncode}[/bold yellow]") # Don't return False here, test itself succeeded else: console.print(f"[bold green]Benchmark report script '{report_script}' completed.[/bold green]") # Find and print clickable links to the reports # Assuming reports are saved in 'benchmark_reports' by benchmark_report.py report_dir = "benchmark_reports" if os.path.isdir(report_dir): report_files = glob.glob(os.path.join(report_dir, "comparison_report_*.html")) if report_files: try: latest_report = max(report_files, key=os.path.getctime) report_path = os.path.abspath(latest_report) report_url = pathlib.Path(report_path).as_uri() # Better way to create file URI console.print(f"[bold cyan]Click to open report: [link={report_url}]{report_url}[/link][/bold cyan]") except Exception as e: console.print(f"[yellow]Could not determine latest report: {e}[/yellow]") chart_files = glob.glob(os.path.join(report_dir, "memory_chart_*.png")) if chart_files: try: latest_chart = max(chart_files, key=os.path.getctime) chart_path = os.path.abspath(latest_chart) chart_url = pathlib.Path(chart_path).as_uri() console.print(f"[cyan]Memory chart: [link={chart_url}]{chart_url}[/link][/cyan]") except Exception as e: console.print(f"[yellow]Could not determine latest chart: {e}[/yellow]") else: console.print(f"[yellow]Benchmark report directory '{report_dir}' not found. Cannot link reports.[/yellow]") except FileNotFoundError: console.print(f"[bold red]Error: Report script '{report_script}' not found.[/bold red]") except Exception as e: console.print(f"[bold red]Error running report generation subprocess: {e}[/bold red]") # Prompt to exit console.print("\n[bold green]Benchmark run finished. Press Enter to exit.[/bold green]") try: input() # Wait for user input except EOFError: pass # Handle case where input is piped or unavailable return True def main(): parser = argparse.ArgumentParser(description="Run a Crawl4AI SDK benchmark test and generate a report") # --- Arguments --- parser.add_argument("config", choices=list(TEST_CONFIGS) + ["custom"], help="Test configuration: quick, small, medium, large, extreme, or custom") # Arguments for 'custom' config or to override presets parser.add_argument("--urls", type=int, help="Number of URLs") parser.add_argument("--max-sessions", type=int, help="Max concurrent sessions (replaces --workers)") parser.add_argument("--chunk-size", type=int, help="URLs per batch (for non-stream logging)") parser.add_argument("--port", type=int, help="HTTP server port") parser.add_argument("--monitor-mode", type=str, choices=["DETAILED", "AGGREGATED"], help="Monitor display mode") # Boolean flags / options parser.add_argument("--stream", action="store_true", help="Enable streaming results (disables batch logging)") parser.add_argument("--use-rate-limiter", action="store_true", help="Enable basic rate limiter") parser.add_argument("--no-report", action="store_true", help="Skip generating comparison report") parser.add_argument("--clean", action="store_true", help="Clean up reports and site before running") parser.add_argument("--keep-server-alive", action="store_true", help="Keep HTTP server running after test") parser.add_argument("--use-existing-site", action="store_true", help="Use existing site on specified port") parser.add_argument("--skip-generation", action="store_true", help="Use existing site files without regenerating") parser.add_argument("--keep-site", action="store_true", help="Keep generated site files after test") # Removed url_level_logging as it's implicitly handled by stream/batch mode now args = parser.parse_args() custom_args = {} # Populate custom_args from explicit command-line args if args.urls is not None: custom_args["urls"] = args.urls if args.max_sessions is not None: custom_args["max_sessions"] = args.max_sessions if args.chunk_size is not None: custom_args["chunk_size"] = args.chunk_size if args.port is not None: custom_args["port"] = args.port if args.monitor_mode is not None: custom_args["monitor_mode"] = args.monitor_mode if args.stream: custom_args["stream"] = True if args.use_rate_limiter: custom_args["use_rate_limiter"] = True if args.keep_server_alive: custom_args["keep_server_alive"] = True if args.use_existing_site: custom_args["use_existing_site"] = True if args.skip_generation: custom_args["skip_generation"] = True if args.keep_site: custom_args["keep_site"] = True # Clean flags are handled by the 'clean' argument passed to run_benchmark # Validate custom config requirements if args.config == "custom": required_custom = ["urls", "max_sessions", "chunk_size"] missing = [f"--{arg}" for arg in required_custom if arg not in custom_args] if missing: console.print(f"[bold red]Error: 'custom' config requires: {', '.join(missing)}[/bold red]") return 1 success = run_benchmark( config_name=args.config, custom_args=custom_args, # Pass all collected custom args compare=not args.no_report, clean=args.clean ) return 0 if success else 1 if __name__ == "__main__": sys.exit(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/test_stress_api_xs.py
tests/memory/test_stress_api_xs.py
"""Lite Crawl4AI API stress‑tester. ✔ batch or stream mode (single unified path) ✔ global stats + JSON summary ✔ rich table progress ✔ Typer CLI with presets (quick / soak) Usage examples: python api_stress_test.py # uses quick preset python api_stress_test.py soak # 5 K URLs stress run python api_stress_test.py --urls 200 --concurrent 10 --chunk 20 """ from __future__ import annotations import asyncio, json, time, uuid, pathlib, statistics from typing import List, Dict, Optional import httpx, typer from rich.console import Console from rich.table import Table # ───────────────────────── defaults / presets ────────────────────────── PRESETS = { "quick": dict(urls=1, concurrent=1, chunk=1, stream=False), "debug": dict(urls=10, concurrent=2, chunk=5, stream=False), "soak": dict(urls=5000, concurrent=20, chunk=50, stream=True), } API_HEALTH_ENDPOINT = "/health" REQUEST_TIMEOUT = 180.0 console = Console() app = typer.Typer(add_completion=False, rich_markup_mode="rich") # ───────────────────────── helpers ───────────────────────────────────── async def _check_health(client: httpx.AsyncClient) -> None: resp = await client.get(API_HEALTH_ENDPOINT, timeout=10) resp.raise_for_status() console.print(f"[green]Server healthy — version {resp.json().get('version','?')}[/]") async def _iter_results(resp: httpx.Response, stream: bool): """Yield result dicts from batch JSON or ND‑JSON stream.""" if stream: async for line in resp.aiter_lines(): if not line: continue rec = json.loads(line) if rec.get("status") == "completed": break yield rec else: data = resp.json() for rec in data.get("results", []): yield rec, data # rec + whole payload for memory delta/peak async def _consume_stream(resp: httpx.Response) -> Dict: stats = {"success_urls": 0, "failed_urls": 0, "mem_metric": 0.0} async for line in resp.aiter_lines(): if not line: continue rec = json.loads(line) if rec.get("status") == "completed": break if rec.get("success"): stats["success_urls"] += 1 else: stats["failed_urls"] += 1 mem = rec.get("server_memory_mb") if mem is not None: stats["mem_metric"] = max(stats["mem_metric"], float(mem)) return stats def _consume_batch(body: Dict) -> Dict: stats = {"success_urls": 0, "failed_urls": 0} for rec in body.get("results", []): if rec.get("success"): stats["success_urls"] += 1 else: stats["failed_urls"] += 1 stats["mem_metric"] = body.get("server_memory_delta_mb") stats["peak"] = body.get("server_peak_memory_mb") return stats async def _fetch_chunk( client: httpx.AsyncClient, urls: List[str], stream: bool, semaphore: asyncio.Semaphore, ) -> Dict: endpoint = "/crawl/stream" if stream else "/crawl" payload = { "urls": urls, "browser_config": {"type": "BrowserConfig", "params": {"headless": True}}, "crawler_config": {"type": "CrawlerRunConfig", "params": {"cache_mode": "BYPASS", "stream": stream}}, } async with semaphore: start = time.perf_counter() if stream: # ---- streaming request ---- async with client.stream("POST", endpoint, json=payload) as resp: resp.raise_for_status() stats = await _consume_stream(resp) else: # ---- batch request ---- resp = await client.post(endpoint, json=payload) resp.raise_for_status() stats = _consume_batch(resp.json()) stats["elapsed"] = time.perf_counter() - start return stats # ───────────────────────── core runner ───────────────────────────────── async def _run(api: str, urls: int, concurrent: int, chunk: int, stream: bool, report: pathlib.Path): client = httpx.AsyncClient(base_url=api, timeout=REQUEST_TIMEOUT, limits=httpx.Limits(max_connections=concurrent+5)) await _check_health(client) url_list = [f"https://httpbin.org/anything/{uuid.uuid4()}" for _ in range(urls)] chunks = [url_list[i:i+chunk] for i in range(0, len(url_list), chunk)] sem = asyncio.Semaphore(concurrent) table = Table(show_header=True, header_style="bold magenta") table.add_column("Batch", style="dim", width=6) table.add_column("Success/Fail", width=12) table.add_column("Mem", width=14) table.add_column("Time (s)") agg_success = agg_fail = 0 deltas, peaks = [], [] start = time.perf_counter() tasks = [asyncio.create_task(_fetch_chunk(client, c, stream, sem)) for c in chunks] for idx, coro in enumerate(asyncio.as_completed(tasks), 1): res = await coro agg_success += res["success_urls"] agg_fail += res["failed_urls"] if res["mem_metric"] is not None: deltas.append(res["mem_metric"]) if res["peak"] is not None: peaks.append(res["peak"]) mem_txt = f"{res['mem_metric']:.1f}" if res["mem_metric"] is not None else "‑" if res["peak"] is not None: mem_txt = f"{res['peak']:.1f}/{mem_txt}" table.add_row(str(idx), f"{res['success_urls']}/{res['failed_urls']}", mem_txt, f"{res['elapsed']:.2f}") console.print(table) total_time = time.perf_counter() - start summary = { "urls": urls, "concurrent": concurrent, "chunk": chunk, "stream": stream, "success_urls": agg_success, "failed_urls": agg_fail, "elapsed_sec": round(total_time, 2), "avg_mem": round(statistics.mean(deltas), 2) if deltas else None, "max_mem": max(deltas) if deltas else None, "avg_peak": round(statistics.mean(peaks), 2) if peaks else None, "max_peak": max(peaks) if peaks else None, } console.print("\n[bold green]Done:[/]" , summary) report.mkdir(parents=True, exist_ok=True) path = report / f"api_test_{int(time.time())}.json" path.write_text(json.dumps(summary, indent=2)) console.print(f"[green]Summary → {path}") await client.aclose() # ───────────────────────── Typer CLI ────────────────────────────────── @app.command() def main( preset: str = typer.Argument("quick", help="quick / debug / soak or custom"), api_url: str = typer.Option("http://localhost:8020", show_default=True), urls: int = typer.Option(None, help="Total URLs to crawl"), concurrent: int = typer.Option(None, help="Concurrent API requests"), chunk: int = typer.Option(None, help="URLs per request"), stream: bool = typer.Option(None, help="Use /crawl/stream"), report: pathlib.Path = typer.Option("reports_api", help="Where to save JSON summary"), ): """Run a stress test against a running Crawl4AI API server.""" if preset not in PRESETS and any(v is None for v in (urls, concurrent, chunk, stream)): console.print(f"[red]Unknown preset '{preset}' and custom params missing[/]") raise typer.Exit(1) cfg = PRESETS.get(preset, {}) urls = urls or cfg.get("urls") concurrent = concurrent or cfg.get("concurrent") chunk = chunk or cfg.get("chunk") stream = stream if stream is not None else cfg.get("stream", False) console.print(f"[cyan]API:[/] {api_url} | URLs: {urls} | Concurrency: {concurrent} | Chunk: {chunk} | Stream: {stream}") asyncio.run(_run(api_url, urls, concurrent, chunk, stream, report)) if __name__ == "__main__": app()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/test_dispatcher_stress.py
tests/memory/test_dispatcher_stress.py
import asyncio import time import psutil import logging import random from typing import List, Dict import uuid import sys import os # Import your crawler components from crawl4ai.models import DisplayMode, CrawlStatus, CrawlResult from crawl4ai.async_configs import CrawlerRunConfig, BrowserConfig, CacheMode from crawl4ai import AsyncWebCrawler from crawl4ai import MemoryAdaptiveDispatcher, CrawlerMonitor # Global configuration STREAM = False # Toggle between streaming and non-streaming modes # Configure logging to file only (to avoid breaking the rich display) os.makedirs("logs", exist_ok=True) file_handler = logging.FileHandler("logs/memory_stress_test.log") file_handler.setFormatter(logging.Formatter('%(asctime)s [%(levelname)s] %(message)s')) # Root logger - only to file, not console root_logger = logging.getLogger() root_logger.setLevel(logging.INFO) root_logger.addHandler(file_handler) # Our test logger also writes to file only logger = logging.getLogger("memory_stress_test") logger.setLevel(logging.INFO) logger.addHandler(file_handler) logger.propagate = False # Don't propagate to root logger # Create a memory restrictor to simulate limited memory environment class MemorySimulator: def __init__(self, target_percent: float = 85.0, aggressive: bool = False): """Simulates memory pressure by allocating memory""" self.target_percent = target_percent self.memory_blocks: List[bytearray] = [] self.aggressive = aggressive def apply_pressure(self, additional_percent: float = 0.0): """Fill memory until we reach target percentage""" current_percent = psutil.virtual_memory().percent target = self.target_percent + additional_percent if current_percent >= target: return # Already at target logger.info(f"Current memory: {current_percent}%, target: {target}%") # Calculate how much memory we need to allocate total_memory = psutil.virtual_memory().total target_usage = (target / 100.0) * total_memory current_usage = (current_percent / 100.0) * total_memory bytes_to_allocate = int(target_usage - current_usage) if bytes_to_allocate <= 0: return # Allocate in smaller chunks to avoid overallocation if self.aggressive: # Use larger chunks for faster allocation in aggressive mode chunk_size = min(bytes_to_allocate, 200 * 1024 * 1024) # 200MB chunks else: chunk_size = min(bytes_to_allocate, 50 * 1024 * 1024) # 50MB chunks try: logger.info(f"Allocating {chunk_size / (1024 * 1024):.1f}MB to reach target memory usage") self.memory_blocks.append(bytearray(chunk_size)) time.sleep(0.5) # Give system time to register the allocation except MemoryError: logger.warning("Unable to allocate more memory") def release_pressure(self, percent: float = None): """ Release allocated memory If percent is specified, release that percentage of blocks """ if not self.memory_blocks: return if percent is None: # Release all logger.info(f"Releasing all {len(self.memory_blocks)} memory blocks") self.memory_blocks.clear() else: # Release specified percentage blocks_to_release = int(len(self.memory_blocks) * (percent / 100.0)) if blocks_to_release > 0: logger.info(f"Releasing {blocks_to_release} of {len(self.memory_blocks)} memory blocks ({percent}%)") self.memory_blocks = self.memory_blocks[blocks_to_release:] def spike_pressure(self, duration: float = 5.0): """ Create a temporary spike in memory pressure then release Useful for forcing requeues """ logger.info(f"Creating memory pressure spike for {duration} seconds") # Save current blocks count initial_blocks = len(self.memory_blocks) # Create spike with extra 5% self.apply_pressure(additional_percent=5.0) # Schedule release after duration asyncio.create_task(self._delayed_release(duration, initial_blocks)) async def _delayed_release(self, delay: float, target_blocks: int): """Helper for spike_pressure - releases extra blocks after delay""" await asyncio.sleep(delay) # Remove blocks added since spike started if len(self.memory_blocks) > target_blocks: logger.info(f"Releasing memory spike ({len(self.memory_blocks) - target_blocks} blocks)") self.memory_blocks = self.memory_blocks[:target_blocks] # Test statistics collector class TestResults: def __init__(self): self.start_time = time.time() self.completed_urls: List[str] = [] self.failed_urls: List[str] = [] self.requeued_count = 0 self.memory_warnings = 0 self.max_memory_usage = 0.0 self.max_queue_size = 0 self.max_wait_time = 0.0 self.url_to_attempt: Dict[str, int] = {} # Track retries per URL def log_summary(self): duration = time.time() - self.start_time logger.info("===== TEST SUMMARY =====") logger.info(f"Stream mode: {'ON' if STREAM else 'OFF'}") logger.info(f"Total duration: {duration:.1f} seconds") logger.info(f"Completed URLs: {len(self.completed_urls)}") logger.info(f"Failed URLs: {len(self.failed_urls)}") logger.info(f"Requeue events: {self.requeued_count}") logger.info(f"Memory warnings: {self.memory_warnings}") logger.info(f"Max memory usage: {self.max_memory_usage:.1f}%") logger.info(f"Max queue size: {self.max_queue_size}") logger.info(f"Max wait time: {self.max_wait_time:.1f} seconds") # Log URLs with multiple attempts retried_urls = {url: count for url, count in self.url_to_attempt.items() if count > 1} if retried_urls: logger.info(f"URLs with retries: {len(retried_urls)}") # Log the top 5 most retried top_retries = sorted(retried_urls.items(), key=lambda x: x[1], reverse=True)[:5] for url, count in top_retries: logger.info(f" URL {url[-30:]} had {count} attempts") # Write summary to a separate human-readable file with open("logs/test_summary.txt", "w") as f: f.write(f"Stream mode: {'ON' if STREAM else 'OFF'}\n") f.write(f"Total duration: {duration:.1f} seconds\n") f.write(f"Completed URLs: {len(self.completed_urls)}\n") f.write(f"Failed URLs: {len(self.failed_urls)}\n") f.write(f"Requeue events: {self.requeued_count}\n") f.write(f"Memory warnings: {self.memory_warnings}\n") f.write(f"Max memory usage: {self.max_memory_usage:.1f}%\n") f.write(f"Max queue size: {self.max_queue_size}\n") f.write(f"Max wait time: {self.max_wait_time:.1f} seconds\n") # Custom monitor with stats tracking # Custom monitor that extends CrawlerMonitor with test-specific tracking class StressTestMonitor(CrawlerMonitor): def __init__(self, test_results: TestResults, **kwargs): # Initialize the parent CrawlerMonitor super().__init__(**kwargs) self.test_results = test_results def update_memory_status(self, status: str): if status != self.memory_status: logger.info(f"Memory status changed: {self.memory_status} -> {status}") if "CRITICAL" in status or "PRESSURE" in status: self.test_results.memory_warnings += 1 # Track peak memory usage in test results current_memory = psutil.virtual_memory().percent self.test_results.max_memory_usage = max(self.test_results.max_memory_usage, current_memory) # Call parent method to update the dashboard super().update_memory_status(status) def update_queue_statistics(self, total_queued: int, highest_wait_time: float, avg_wait_time: float): # Track queue metrics in test results self.test_results.max_queue_size = max(self.test_results.max_queue_size, total_queued) self.test_results.max_wait_time = max(self.test_results.max_wait_time, highest_wait_time) # Call parent method to update the dashboard super().update_queue_statistics(total_queued, highest_wait_time, avg_wait_time) def update_task(self, task_id: str, **kwargs): # Track URL status changes for test results if task_id in self.stats: old_status = self.stats[task_id].status # If this is a requeue event (requeued due to memory pressure) if 'error_message' in kwargs and 'requeued' in kwargs['error_message']: if not hasattr(self.stats[task_id], 'counted_requeue') or not self.stats[task_id].counted_requeue: self.test_results.requeued_count += 1 self.stats[task_id].counted_requeue = True # Track completion status for test results if 'status' in kwargs: new_status = kwargs['status'] if old_status != new_status: if new_status == CrawlStatus.COMPLETED: if task_id not in self.test_results.completed_urls: self.test_results.completed_urls.append(task_id) elif new_status == CrawlStatus.FAILED: if task_id not in self.test_results.failed_urls: self.test_results.failed_urls.append(task_id) # Call parent method to update the dashboard super().update_task(task_id, **kwargs) self.live.update(self._create_table()) # Generate test URLs - use example.com with unique paths to avoid browser caching def generate_test_urls(count: int) -> List[str]: urls = [] for i in range(count): # Add random path and query parameters to create unique URLs path = f"/path/{uuid.uuid4()}" query = f"?test={i}&random={random.randint(1, 100000)}" urls.append(f"https://example.com{path}{query}") return urls # Process result callback async def process_result(result, test_results: TestResults): # Track attempt counts if result.url not in test_results.url_to_attempt: test_results.url_to_attempt[result.url] = 1 else: test_results.url_to_attempt[result.url] += 1 if "requeued" in result.error_message: test_results.requeued_count += 1 logger.debug(f"Requeued due to memory pressure: {result.url}") elif result.success: test_results.completed_urls.append(result.url) logger.debug(f"Successfully processed: {result.url}") else: test_results.failed_urls.append(result.url) logger.warning(f"Failed to process: {result.url} - {result.error_message}") # Process multiple results (used in non-streaming mode) async def process_results(results, test_results: TestResults): for result in results: await process_result(result, test_results) # Main test function for extreme memory pressure simulation async def run_memory_stress_test( url_count: int = 100, target_memory_percent: float = 92.0, # Push to dangerous levels chunk_size: int = 20, # Larger chunks for more chaos aggressive: bool = False, spikes: bool = True ): test_results = TestResults() memory_simulator = MemorySimulator(target_percent=target_memory_percent, aggressive=aggressive) logger.info(f"Starting stress test with {url_count} URLs in {'STREAM' if STREAM else 'NON-STREAM'} mode") logger.info(f"Target memory usage: {target_memory_percent}%") # First, elevate memory usage to create pressure logger.info("Creating initial memory pressure...") memory_simulator.apply_pressure() # Create test URLs in chunks to simulate real-world crawling where URLs are discovered all_urls = generate_test_urls(url_count) url_chunks = [all_urls[i:i+chunk_size] for i in range(0, len(all_urls), chunk_size)] # Set up the crawler components - low memory thresholds to create more requeues browser_config = BrowserConfig(headless=True, verbose=False) run_config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, verbose=False, stream=STREAM # Use the global STREAM variable to set mode ) # Create monitor with reference to test results monitor = StressTestMonitor( test_results=test_results, display_mode=DisplayMode.DETAILED, max_visible_rows=20, total_urls=url_count # Pass total URLs count ) # Create dispatcher with EXTREME settings - pure survival mode # These settings are designed to create a memory battleground dispatcher = MemoryAdaptiveDispatcher( memory_threshold_percent=63.0, # Start throttling at just 60% memory critical_threshold_percent=70.0, # Start requeuing at 70% - incredibly aggressive recovery_threshold_percent=55.0, # Only resume normal ops when plenty of memory available check_interval=0.1, # Check extremely frequently (100ms) max_session_permit=20 if aggressive else 10, # Double the concurrent sessions - pure chaos fairness_timeout=10.0, # Extremely low timeout - rapid priority changes monitor=monitor ) # Set up spike schedule if enabled if spikes: spike_intervals = [] # Create 3-5 random spike times num_spikes = random.randint(3, 5) for _ in range(num_spikes): # Schedule spikes at random chunks chunk_index = random.randint(1, len(url_chunks) - 1) spike_intervals.append(chunk_index) logger.info(f"Scheduled memory spikes at chunks: {spike_intervals}") try: async with AsyncWebCrawler(config=browser_config) as crawler: # Process URLs in chunks to simulate discovering URLs over time for chunk_index, url_chunk in enumerate(url_chunks): logger.info(f"Processing chunk {chunk_index+1}/{len(url_chunks)} ({len(url_chunk)} URLs)") # Regular pressure increases if chunk_index % 2 == 0: logger.info("Increasing memory pressure...") memory_simulator.apply_pressure() # Memory spike if scheduled for this chunk if spikes and chunk_index in spike_intervals: logger.info(f"⚠️ CREATING MASSIVE MEMORY SPIKE at chunk {chunk_index+1} ⚠️") # Create a nightmare scenario - multiple overlapping spikes memory_simulator.spike_pressure(duration=10.0) # 10-second spike # 50% chance of double-spike (pure evil) if random.random() < 0.5: await asyncio.sleep(2.0) # Wait 2 seconds logger.info("💀 DOUBLE SPIKE - EXTREME MEMORY PRESSURE 💀") memory_simulator.spike_pressure(duration=8.0) # 8-second overlapping spike if STREAM: # Stream mode - process results as they come in async for result in dispatcher.run_urls_stream( urls=url_chunk, crawler=crawler, config=run_config ): await process_result(result, test_results) else: # Non-stream mode - get all results at once results = await dispatcher.run_urls( urls=url_chunk, crawler=crawler, config=run_config ) await process_results(results, test_results) # Simulate discovering more URLs while others are still processing await asyncio.sleep(1) # RARELY release pressure - make the system fight for resources if chunk_index % 5 == 4: # Less frequent releases release_percent = random.choice([10, 15, 20]) # Smaller, inconsistent releases logger.info(f"Releasing {release_percent}% of memory blocks - brief respite") memory_simulator.release_pressure(percent=release_percent) except Exception as e: logger.error(f"Test error: {str(e)}") raise finally: # Release memory pressure memory_simulator.release_pressure() # Log final results test_results.log_summary() # Check for success criteria if len(test_results.completed_urls) + len(test_results.failed_urls) < url_count: logger.error(f"TEST FAILED: Not all URLs were processed. {url_count - len(test_results.completed_urls) - len(test_results.failed_urls)} URLs missing.") return False logger.info("TEST PASSED: All URLs were processed without crashing.") return True # Command-line entry point if __name__ == "__main__": # Parse command line arguments url_count = int(sys.argv[1]) if len(sys.argv) > 1 else 100 target_memory = float(sys.argv[2]) if len(sys.argv) > 2 else 85.0 # Check if stream mode is specified if len(sys.argv) > 3: STREAM = sys.argv[3].lower() in ('true', 'yes', '1', 'stream') # Check if aggressive mode is specified aggressive = False if len(sys.argv) > 4: aggressive = sys.argv[4].lower() in ('true', 'yes', '1', 'aggressive') print(f"Starting test with {url_count} URLs, {target_memory}% memory target") print(f"Stream mode: {STREAM}, Aggressive: {aggressive}") print("Logs will be written to the logs directory") print("Live display starting now...") # Run the test result = asyncio.run(run_memory_stress_test( url_count=url_count, target_memory_percent=target_memory, aggressive=aggressive )) # Exit with status code sys.exit(0 if result else 1)
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/test_docker_config_gen.py
tests/memory/test_docker_config_gen.py
#!/usr/bin/env python3 """ Quick sanity‑check for /config/dump endpoint. Usage: python test_config_dump.py [http://localhost:8020] If the server isn’t running, start it first: uvicorn deploy.docker.server:app --port 8020 """ import sys, json, textwrap, requests # BASE = sys.argv[1] if len(sys.argv) > 1 else "http://localhost:8020" BASE = sys.argv[1] if len(sys.argv) > 1 else "http://localhost:11235" URL = f"{BASE.rstrip('/')}/config/dump" CASES = [ # --- CrawlRunConfig variants --- "CrawlerRunConfig()", "CrawlerRunConfig(stream=True, cache_mode=CacheMode.BYPASS)", "CrawlerRunConfig(js_only=True, wait_until='networkidle')", # --- BrowserConfig variants --- "BrowserConfig()", "BrowserConfig(headless=False, extra_args=['--disable-gpu'])", "BrowserConfig(browser_mode='builtin', proxy_config={'server': 'http://1.2.3.4:8080'})", ] for code in CASES: print("\n=== POST:", code) resp = requests.post(URL, json={"code": code}, timeout=15) if resp.ok: print(json.dumps(resp.json(), indent=2)[:400] + "...") else: print("ERROR", resp.status_code, resp.text[:200])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/cap_test.py
tests/memory/cap_test.py
#!/usr/bin/env python3 """ Hammer /crawl with many concurrent requests to prove GLOBAL_SEM works. """ import asyncio, httpx, json, uuid, argparse API = "http://localhost:8020/crawl" URLS_PER_CALL = 1 # keep it minimal so each arun() == 1 page CONCURRENT_CALLS = 20 # way above your cap payload_template = { "browser_config": {"type": "BrowserConfig", "params": {"headless": True}}, "crawler_config": { "type": "CrawlerRunConfig", "params": {"cache_mode": "BYPASS", "verbose": False}, } } async def one_call(client): payload = payload_template.copy() payload["urls"] = [f"https://httpbin.org/anything/{uuid.uuid4()}"] r = await client.post(API, json=payload) r.raise_for_status() return r.json()["server_peak_memory_mb"] async def main(): async with httpx.AsyncClient(timeout=60) as client: tasks = [asyncio.create_task(one_call(client)) for _ in range(CONCURRENT_CALLS)] mem_usages = await asyncio.gather(*tasks) print("Calls finished OK, server peaks reported:", mem_usages) if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/test_stress_docker_api.py
tests/memory/test_stress_docker_api.py
""" Crawl4AI Docker API stress tester. Examples -------- python test_stress_docker_api.py --urls 1000 --concurrency 32 python test_stress_docker_api.py --urls 1000 --concurrency 32 --stream python test_stress_docker_api.py --base-url http://10.0.0.42:11235 --http2 """ import argparse, asyncio, json, secrets, statistics, time from typing import List, Tuple import httpx from rich.console import Console from rich.progress import Progress, BarColumn, TimeElapsedColumn, TimeRemainingColumn from rich.table import Table console = Console() # ───────────────────────── helpers ───────────────────────── def make_fake_urls(n: int) -> List[str]: base = "https://httpbin.org/anything/" return [f"{base}{secrets.token_hex(8)}" for _ in range(n)] async def fire( client: httpx.AsyncClient, endpoint: str, payload: dict, sem: asyncio.Semaphore ) -> Tuple[bool, float]: async with sem: print(f"POST {endpoint} with {len(payload['urls'])} URLs") t0 = time.perf_counter() try: if endpoint.endswith("/stream"): async with client.stream("POST", endpoint, json=payload) as r: r.raise_for_status() async for _ in r.aiter_lines(): pass else: r = await client.post(endpoint, json=payload) r.raise_for_status() return True, time.perf_counter() - t0 except Exception: return False, time.perf_counter() - t0 def pct(lat: List[float], p: float) -> str: """Return percentile string even for tiny samples.""" if not lat: return "-" if len(lat) == 1: return f"{lat[0]:.2f}s" lat_sorted = sorted(lat) k = (p / 100) * (len(lat_sorted) - 1) lo = int(k) hi = min(lo + 1, len(lat_sorted) - 1) frac = k - lo val = lat_sorted[lo] * (1 - frac) + lat_sorted[hi] * frac return f"{val:.2f}s" # ───────────────────────── main ───────────────────────── def parse_args() -> argparse.Namespace: p = argparse.ArgumentParser(description="Stress test Crawl4AI Docker API") p.add_argument("--urls", type=int, default=100, help="number of URLs") p.add_argument("--concurrency", type=int, default=1, help="max POSTs in flight") p.add_argument("--chunk-size", type=int, default=50, help="URLs per request") p.add_argument("--base-url", default="http://localhost:11235", help="API root") # p.add_argument("--base-url", default="http://localhost:8020", help="API root") p.add_argument("--stream", action="store_true", help="use /crawl/stream") p.add_argument("--http2", action="store_true", help="enable HTTP/2") p.add_argument("--headless", action="store_true", default=True) return p.parse_args() async def main() -> None: args = parse_args() urls = make_fake_urls(args.urls) batches = [urls[i : i + args.chunk_size] for i in range(0, len(urls), args.chunk_size)] endpoint = "/crawl/stream" if args.stream else "/crawl" sem = asyncio.Semaphore(args.concurrency) async with httpx.AsyncClient(base_url=args.base_url, http2=args.http2, timeout=None) as client: with Progress( "[progress.description]{task.description}", BarColumn(), "[progress.percentage]{task.percentage:>3.0f}%", TimeElapsedColumn(), TimeRemainingColumn(), ) as progress: task_id = progress.add_task("[cyan]bombarding…", total=len(batches)) tasks = [] for chunk in batches: payload = { "urls": chunk, "browser_config": {"type": "BrowserConfig", "params": {"headless": args.headless}}, "crawler_config": {"type": "CrawlerRunConfig", "params": {"cache_mode": "BYPASS", "stream": args.stream}}, } tasks.append(asyncio.create_task(fire(client, endpoint, payload, sem))) progress.advance(task_id) results = await asyncio.gather(*tasks) ok_latencies = [dt for ok, dt in results if ok] err_count = sum(1 for ok, _ in results if not ok) table = Table(title="Docker API Stress‑Test Summary") table.add_column("total", justify="right") table.add_column("errors", justify="right") table.add_column("p50", justify="right") table.add_column("p95", justify="right") table.add_column("max", justify="right") table.add_row( str(len(results)), str(err_count), pct(ok_latencies, 50), pct(ok_latencies, 95), f"{max(ok_latencies):.2f}s" if ok_latencies else "-", ) console.print(table) if __name__ == "__main__": try: asyncio.run(main()) except KeyboardInterrupt: console.print("\n[yellow]aborted by user[/]")
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/memory/test_stress_sdk.py
tests/memory/test_stress_sdk.py
#!/usr/bin/env python3 """ Stress test for Crawl4AI's arun_many and dispatcher system. This version uses a local HTTP server and focuses on testing the SDK's ability to handle multiple URLs concurrently, with per-batch logging. """ import asyncio import os import time import pathlib import random import secrets import argparse import json import sys import subprocess import signal from typing import List, Dict, Optional, Union, AsyncGenerator import shutil from rich.console import Console # Crawl4AI components from crawl4ai import ( AsyncWebCrawler, CrawlerRunConfig, BrowserConfig, MemoryAdaptiveDispatcher, CrawlerMonitor, DisplayMode, CrawlResult, RateLimiter, CacheMode, ) # Constants DEFAULT_SITE_PATH = "test_site" DEFAULT_PORT = 8000 DEFAULT_MAX_SESSIONS = 16 DEFAULT_URL_COUNT = 1 DEFAULT_CHUNK_SIZE = 1 # Define chunk size for batch logging DEFAULT_REPORT_PATH = "reports" DEFAULT_STREAM_MODE = False DEFAULT_MONITOR_MODE = "DETAILED" # Initialize Rich console console = Console() # --- SiteGenerator Class (Unchanged) --- class SiteGenerator: """Generates a local test site with heavy pages for stress testing.""" def __init__(self, site_path: str = DEFAULT_SITE_PATH, page_count: int = DEFAULT_URL_COUNT): self.site_path = pathlib.Path(site_path) self.page_count = page_count self.images_dir = self.site_path / "images" self.lorem_words = " ".join("lorem ipsum dolor sit amet " * 100).split() self.html_template = """<!doctype html> <html> <head> <title>Test Page {page_num}</title> <meta charset="utf-8"> </head> <body> <h1>Test Page {page_num}</h1> {paragraphs} {images} </body> </html> """ def generate_site(self) -> None: self.site_path.mkdir(parents=True, exist_ok=True) self.images_dir.mkdir(exist_ok=True) console.print(f"Generating {self.page_count} test pages...") for i in range(self.page_count): paragraphs = "\n".join(f"<p>{' '.join(random.choices(self.lorem_words, k=200))}</p>" for _ in range(5)) images = "\n".join(f'<img src="https://picsum.photos/seed/{secrets.token_hex(8)}/300/200" loading="lazy" alt="Random image {j}"/>' for j in range(3)) page_path = self.site_path / f"page_{i}.html" page_path.write_text(self.html_template.format(page_num=i, paragraphs=paragraphs, images=images), encoding="utf-8") if (i + 1) % (self.page_count // 10 or 1) == 0 or i == self.page_count - 1: console.print(f"Generated {i+1}/{self.page_count} pages") self._create_index_page() console.print(f"[bold green]Successfully generated {self.page_count} test pages in [cyan]{self.site_path}[/cyan][/bold green]") def _create_index_page(self) -> None: index_content = """<!doctype html><html><head><title>Test Site Index</title><meta charset="utf-8"></head><body><h1>Test Site Index</h1><p>This is an automatically generated site for testing Crawl4AI.</p><div class="page-links">\n""" for i in range(self.page_count): index_content += f' <a href="page_{i}.html">Test Page {i}</a><br>\n' index_content += """ </div></body></html>""" (self.site_path / "index.html").write_text(index_content, encoding="utf-8") # --- LocalHttpServer Class (Unchanged) --- class LocalHttpServer: """Manages a local HTTP server for serving test pages.""" def __init__(self, site_path: str = DEFAULT_SITE_PATH, port: int = DEFAULT_PORT): self.site_path = pathlib.Path(site_path) self.port = port self.process = None def start(self) -> None: if not self.site_path.exists(): raise FileNotFoundError(f"Site directory {self.site_path} does not exist") console.print(f"Attempting to start HTTP server in [cyan]{self.site_path}[/cyan] on port {self.port}...") try: cmd = ["python", "-m", "http.server", str(self.port)] creationflags = 0; preexec_fn = None if sys.platform == 'win32': creationflags = subprocess.CREATE_NEW_PROCESS_GROUP self.process = subprocess.Popen(cmd, cwd=str(self.site_path), stdout=subprocess.PIPE, stderr=subprocess.PIPE, creationflags=creationflags) time.sleep(1.5) if self.is_running(): console.print(f"[bold green]HTTP server started successfully (PID: {self.process.pid})[/bold green]") else: console.print("[bold red]Failed to start HTTP server. Checking logs...[/bold red]") stdout, stderr = self.process.communicate(); print(stdout.decode(errors='ignore')); print(stderr.decode(errors='ignore')) self.stop(); raise RuntimeError("HTTP server failed to start.") except Exception as e: console.print(f"[bold red]Error starting HTTP server: {str(e)}[/bold red]"); self.stop(); raise def stop(self) -> None: if self.process and self.is_running(): console.print(f"Stopping HTTP server (PID: {self.process.pid})...") try: if sys.platform == 'win32': self.process.send_signal(signal.CTRL_BREAK_EVENT); time.sleep(0.5) self.process.terminate() try: stdout, stderr = self.process.communicate(timeout=5); console.print("[bold yellow]HTTP server stopped[/bold yellow]") except subprocess.TimeoutExpired: console.print("[bold red]Server did not terminate gracefully, killing...[/bold red]"); self.process.kill(); stdout, stderr = self.process.communicate(); console.print("[bold yellow]HTTP server killed[/bold yellow]") except Exception as e: console.print(f"[bold red]Error stopping HTTP server: {str(e)}[/bold red]"); self.process.kill() finally: self.process = None elif self.process: console.print("[dim]HTTP server process already stopped.[/dim]"); self.process = None def is_running(self) -> bool: if not self.process: return False return self.process.poll() is None # --- SimpleMemoryTracker Class (Unchanged) --- class SimpleMemoryTracker: """Basic memory tracker that doesn't rely on psutil.""" def __init__(self, report_path: str = DEFAULT_REPORT_PATH, test_id: Optional[str] = None): self.report_path = pathlib.Path(report_path); self.report_path.mkdir(parents=True, exist_ok=True) self.test_id = test_id or time.strftime("%Y%m%d_%H%M%S") self.start_time = time.time(); self.memory_samples = []; self.pid = os.getpid() self.csv_path = self.report_path / f"memory_samples_{self.test_id}.csv" with open(self.csv_path, 'w', encoding='utf-8') as f: f.write("timestamp,elapsed_seconds,memory_info_mb\n") def sample(self) -> Dict: try: memory_mb = self._get_memory_info_mb() memory_str = f"{memory_mb:.1f} MB" if memory_mb is not None else "Unknown" timestamp = time.time(); elapsed = timestamp - self.start_time sample = {"timestamp": timestamp, "elapsed_seconds": elapsed, "memory_mb": memory_mb, "memory_str": memory_str} self.memory_samples.append(sample) with open(self.csv_path, 'a', encoding='utf-8') as f: f.write(f"{timestamp},{elapsed:.2f},{memory_mb if memory_mb is not None else ''}\n") return sample except Exception as e: return {"memory_mb": None, "memory_str": "Error"} def _get_memory_info_mb(self) -> Optional[float]: pid_str = str(self.pid) try: if sys.platform == 'darwin': result = subprocess.run(["ps", "-o", "rss=", "-p", pid_str], capture_output=True, text=True, check=True, encoding='utf-8'); return int(result.stdout.strip()) / 1024.0 elif sys.platform == 'linux': with open(f"/proc/{pid_str}/status", encoding='utf-8') as f: for line in f: if line.startswith("VmRSS:"): return int(line.split()[1]) / 1024.0 return None elif sys.platform == 'win32': result = subprocess.run(["tasklist", "/fi", f"PID eq {pid_str}", "/fo", "csv", "/nh"], capture_output=True, text=True, check=True, encoding='cp850', errors='ignore'); parts = result.stdout.strip().split('","'); return int(parts[4].strip().replace('"', '').replace(' K', '').replace(',', '')) / 1024.0 if len(parts) >= 5 else None else: return None except: return None # Catch all exceptions for robustness def get_report(self) -> Dict: if not self.memory_samples: return {"error": "No memory samples collected"} total_time = time.time() - self.start_time; valid_samples = [s['memory_mb'] for s in self.memory_samples if s['memory_mb'] is not None] start_mem = valid_samples[0] if valid_samples else None; end_mem = valid_samples[-1] if valid_samples else None max_mem = max(valid_samples) if valid_samples else None; avg_mem = sum(valid_samples) / len(valid_samples) if valid_samples else None growth = (end_mem - start_mem) if start_mem is not None and end_mem is not None else None return {"test_id": self.test_id, "total_time_seconds": total_time, "sample_count": len(self.memory_samples), "valid_sample_count": len(valid_samples), "csv_path": str(self.csv_path), "platform": sys.platform, "start_memory_mb": start_mem, "end_memory_mb": end_mem, "max_memory_mb": max_mem, "average_memory_mb": avg_mem, "memory_growth_mb": growth} # --- CrawlerStressTest Class (Refactored for Per-Batch Logging) --- class CrawlerStressTest: """Orchestrates the stress test using arun_many per chunk and a dispatcher.""" def __init__( self, url_count: int = DEFAULT_URL_COUNT, port: int = DEFAULT_PORT, max_sessions: int = DEFAULT_MAX_SESSIONS, chunk_size: int = DEFAULT_CHUNK_SIZE, # Added chunk_size report_path: str = DEFAULT_REPORT_PATH, stream_mode: bool = DEFAULT_STREAM_MODE, monitor_mode: str = DEFAULT_MONITOR_MODE, use_rate_limiter: bool = False ): self.url_count = url_count self.server_port = port self.max_sessions = max_sessions self.chunk_size = chunk_size # Store chunk size self.report_path = pathlib.Path(report_path) self.report_path.mkdir(parents=True, exist_ok=True) self.stream_mode = stream_mode self.monitor_mode = DisplayMode[monitor_mode.upper()] self.use_rate_limiter = use_rate_limiter self.test_id = time.strftime("%Y%m%d_%H%M%S") self.results_summary = { "test_id": self.test_id, "url_count": url_count, "max_sessions": max_sessions, "chunk_size": chunk_size, "stream_mode": stream_mode, "monitor_mode": monitor_mode, "rate_limiter_used": use_rate_limiter, "start_time": "", "end_time": "", "total_time_seconds": 0, "successful_urls": 0, "failed_urls": 0, "urls_processed": 0, "chunks_processed": 0 } async def run(self) -> Dict: """Run the stress test and return results.""" memory_tracker = SimpleMemoryTracker(report_path=self.report_path, test_id=self.test_id) urls = [f"http://localhost:{self.server_port}/page_{i}.html" for i in range(self.url_count)] # Split URLs into chunks based on self.chunk_size url_chunks = [urls[i:i+self.chunk_size] for i in range(0, len(urls), self.chunk_size)] self.results_summary["start_time"] = time.strftime("%Y-%m-%d %H:%M:%S") start_time = time.time() config = CrawlerRunConfig( wait_for_images=False, verbose=False, stream=self.stream_mode, # Still pass stream mode, affects arun_many return type cache_mode=CacheMode.BYPASS ) total_successful_urls = 0 total_failed_urls = 0 total_urls_processed = 0 start_memory_sample = memory_tracker.sample() start_memory_str = start_memory_sample.get("memory_str", "Unknown") # monitor = CrawlerMonitor(display_mode=self.monitor_mode, total_urls=self.url_count) monitor = None rate_limiter = RateLimiter(base_delay=(0.1, 0.3)) if self.use_rate_limiter else None dispatcher = MemoryAdaptiveDispatcher(max_session_permit=self.max_sessions, monitor=monitor, rate_limiter=rate_limiter) console.print(f"\n[bold cyan]Crawl4AI Stress Test - {self.url_count} URLs, {self.max_sessions} max sessions[/bold cyan]") console.print(f"[bold cyan]Mode:[/bold cyan] {'Streaming' if self.stream_mode else 'Batch'}, [bold cyan]Monitor:[/bold cyan] {self.monitor_mode.name}, [bold cyan]Chunk Size:[/bold cyan] {self.chunk_size}") console.print(f"[bold cyan]Initial Memory:[/bold cyan] {start_memory_str}") # Print batch log header only if not streaming if not self.stream_mode: console.print("\n[bold]Batch Progress:[/bold] (Monitor below shows overall progress)") console.print("[bold] Batch | Progress | Start Mem | End Mem | URLs/sec | Success/Fail | Time (s) | Status [/bold]") console.print("─" * 90) monitor_task = asyncio.create_task(self._periodic_memory_sample(memory_tracker, 2.0)) try: async with AsyncWebCrawler( config=BrowserConfig( verbose = False) ) as crawler: # Process URLs chunk by chunk for chunk_idx, url_chunk in enumerate(url_chunks): batch_start_time = time.time() chunk_success = 0 chunk_failed = 0 # Sample memory before the chunk start_mem_sample = memory_tracker.sample() start_mem_str = start_mem_sample.get("memory_str", "Unknown") # --- Call arun_many for the current chunk --- try: # Note: dispatcher/monitor persist across calls results_gen_or_list: Union[AsyncGenerator[CrawlResult, None], List[CrawlResult]] = \ await crawler.arun_many( urls=url_chunk, config=config, dispatcher=dispatcher # Reuse the same dispatcher ) if self.stream_mode: # Process stream results if needed, but batch logging is less relevant async for result in results_gen_or_list: total_urls_processed += 1 if result.success: chunk_success += 1 else: chunk_failed += 1 # In stream mode, batch summary isn't as meaningful here # We could potentially track completion per chunk async, but it's complex else: # Batch mode # Process the list of results for this chunk for result in results_gen_or_list: total_urls_processed += 1 if result.success: chunk_success += 1 else: chunk_failed += 1 except Exception as e: console.print(f"[bold red]Error processing chunk {chunk_idx+1}: {e}[/bold red]") chunk_failed = len(url_chunk) # Assume all failed in the chunk on error total_urls_processed += len(url_chunk) # Count them as processed (failed) # --- Log batch results (only if not streaming) --- if not self.stream_mode: batch_time = time.time() - batch_start_time urls_per_sec = len(url_chunk) / batch_time if batch_time > 0 else 0 end_mem_sample = memory_tracker.sample() end_mem_str = end_mem_sample.get("memory_str", "Unknown") progress_pct = (total_urls_processed / self.url_count) * 100 if chunk_failed == 0: status_color, status = "green", "Success" elif chunk_success == 0: status_color, status = "red", "Failed" else: status_color, status = "yellow", "Partial" console.print( f" {chunk_idx+1:<5} | {progress_pct:6.1f}% | {start_mem_str:>9} | {end_mem_str:>9} | {urls_per_sec:8.1f} | " f"{chunk_success:^7}/{chunk_failed:<6} | {batch_time:8.2f} | [{status_color}]{status:<7}[/{status_color}]" ) # Accumulate totals total_successful_urls += chunk_success total_failed_urls += chunk_failed self.results_summary["chunks_processed"] += 1 # Optional small delay between starting chunks if needed # await asyncio.sleep(0.1) except Exception as e: console.print(f"[bold red]An error occurred during the main crawl loop: {e}[/bold red]") finally: if 'monitor_task' in locals() and not monitor_task.done(): monitor_task.cancel() try: await monitor_task except asyncio.CancelledError: pass end_time = time.time() self.results_summary.update({ "end_time": time.strftime("%Y-%m-%d %H:%M:%S"), "total_time_seconds": end_time - start_time, "successful_urls": total_successful_urls, "failed_urls": total_failed_urls, "urls_processed": total_urls_processed, "memory": memory_tracker.get_report() }) self._save_results() return self.results_summary async def _periodic_memory_sample(self, tracker: SimpleMemoryTracker, interval: float): """Background task to sample memory periodically.""" while True: tracker.sample() try: await asyncio.sleep(interval) except asyncio.CancelledError: break # Exit loop on cancellation def _save_results(self) -> None: results_path = self.report_path / f"test_summary_{self.test_id}.json" try: with open(results_path, 'w', encoding='utf-8') as f: json.dump(self.results_summary, f, indent=2, default=str) # console.print(f"\n[bold green]Results summary saved to {results_path}[/bold green]") # Moved summary print to run_full_test except Exception as e: console.print(f"[bold red]Failed to save results summary: {e}[/bold red]") # --- run_full_test Function (Adjusted) --- async def run_full_test(args): """Run the complete test process from site generation to crawling.""" server = None site_generated = False # --- Site Generation --- (Same as before) if not args.use_existing_site and not args.skip_generation: if os.path.exists(args.site_path): console.print(f"[yellow]Removing existing site directory: {args.site_path}[/yellow]"); shutil.rmtree(args.site_path) site_generator = SiteGenerator(site_path=args.site_path, page_count=args.urls); site_generator.generate_site(); site_generated = True elif args.use_existing_site: console.print(f"[cyan]Using existing site assumed to be running on port {args.port}[/cyan]") elif args.skip_generation: console.print(f"[cyan]Skipping site generation, using existing directory: {args.site_path}[/cyan]") if not os.path.exists(args.site_path) or not os.path.isdir(args.site_path): console.print(f"[bold red]Error: Site path '{args.site_path}' does not exist or is not a directory.[/bold red]"); return # --- Start Local Server --- (Same as before) server_started = False if not args.use_existing_site: server = LocalHttpServer(site_path=args.site_path, port=args.port) try: server.start(); server_started = True except Exception as e: console.print(f"[bold red]Failed to start local server. Aborting test.[/bold red]") if site_generated and not args.keep_site: console.print(f"[yellow]Cleaning up generated site: {args.site_path}[/yellow]"); shutil.rmtree(args.site_path) return try: # --- Run the Stress Test --- test = CrawlerStressTest( url_count=args.urls, port=args.port, max_sessions=args.max_sessions, chunk_size=args.chunk_size, # Pass chunk_size report_path=args.report_path, stream_mode=args.stream, monitor_mode=args.monitor_mode, use_rate_limiter=args.use_rate_limiter ) results = await test.run() # Run the test which now handles chunks internally # --- Print Summary --- console.print("\n" + "=" * 80) console.print("[bold green]Test Completed[/bold green]") console.print("=" * 80) # (Summary printing logic remains largely the same) success_rate = results["successful_urls"] / results["url_count"] * 100 if results["url_count"] > 0 else 0 urls_per_second = results["urls_processed"] / results["total_time_seconds"] if results["total_time_seconds"] > 0 else 0 console.print(f"[bold cyan]Test ID:[/bold cyan] {results['test_id']}") console.print(f"[bold cyan]Configuration:[/bold cyan] {results['url_count']} URLs, {results['max_sessions']} sessions, Chunk: {results['chunk_size']}, Stream: {results['stream_mode']}, Monitor: {results['monitor_mode']}") console.print(f"[bold cyan]Results:[/bold cyan] {results['successful_urls']} successful, {results['failed_urls']} failed ({results['urls_processed']} processed, {success_rate:.1f}% success)") console.print(f"[bold cyan]Performance:[/bold cyan] {results['total_time_seconds']:.2f} seconds total, {urls_per_second:.2f} URLs/second avg") mem_report = results.get("memory", {}) mem_info_str = "Memory tracking data unavailable." if mem_report and not mem_report.get("error"): start_mb = mem_report.get('start_memory_mb'); end_mb = mem_report.get('end_memory_mb'); max_mb = mem_report.get('max_memory_mb'); growth_mb = mem_report.get('memory_growth_mb') mem_parts = [] if start_mb is not None: mem_parts.append(f"Start: {start_mb:.1f} MB") if end_mb is not None: mem_parts.append(f"End: {end_mb:.1f} MB") if max_mb is not None: mem_parts.append(f"Max: {max_mb:.1f} MB") if growth_mb is not None: mem_parts.append(f"Growth: {growth_mb:.1f} MB") if mem_parts: mem_info_str = ", ".join(mem_parts) csv_path = mem_report.get('csv_path') if csv_path: console.print(f"[dim]Memory samples saved to: {csv_path}[/dim]") console.print(f"[bold cyan]Memory Usage:[/bold cyan] {mem_info_str}") console.print(f"[bold green]Results summary saved to {results['memory']['csv_path'].replace('memory_samples', 'test_summary').replace('.csv', '.json')}[/bold green]") # Infer summary path if results["failed_urls"] > 0: console.print(f"\n[bold yellow]Warning: {results['failed_urls']} URLs failed to process ({100-success_rate:.1f}% failure rate)[/bold yellow]") if results["urls_processed"] < results["url_count"]: console.print(f"\n[bold red]Error: Only {results['urls_processed']} out of {results['url_count']} URLs were processed![/bold red]") finally: # --- Stop Server / Cleanup --- (Same as before) if server_started and server and not args.keep_server_alive: server.stop() elif server_started and server and args.keep_server_alive: console.print(f"[bold cyan]Server is kept running on port {args.port}. Press Ctrl+C to stop it.[/bold cyan]") try: await asyncio.Future() # Keep running indefinitely except KeyboardInterrupt: console.print("\n[bold yellow]Stopping server due to user interrupt...[/bold yellow]"); server.stop() if site_generated and not args.keep_site: console.print(f"[yellow]Cleaning up generated site: {args.site_path}[/yellow]"); shutil.rmtree(args.site_path) elif args.clean_site and os.path.exists(args.site_path): console.print(f"[yellow]Cleaning up site directory as requested: {args.site_path}[/yellow]"); shutil.rmtree(args.site_path) # --- main Function (Added chunk_size argument) --- def main(): """Main entry point for the script.""" parser = argparse.ArgumentParser(description="Crawl4AI SDK High Volume Stress Test using arun_many") # Test parameters parser.add_argument("--urls", type=int, default=DEFAULT_URL_COUNT, help=f"Number of URLs to test (default: {DEFAULT_URL_COUNT})") parser.add_argument("--max-sessions", type=int, default=DEFAULT_MAX_SESSIONS, help=f"Maximum concurrent crawling sessions (default: {DEFAULT_MAX_SESSIONS})") parser.add_argument("--chunk-size", type=int, default=DEFAULT_CHUNK_SIZE, help=f"Number of URLs per batch for logging (default: {DEFAULT_CHUNK_SIZE})") # Added parser.add_argument("--stream", action="store_true", default=DEFAULT_STREAM_MODE, help=f"Enable streaming mode (disables batch logging) (default: {DEFAULT_STREAM_MODE})") parser.add_argument("--monitor-mode", type=str, default=DEFAULT_MONITOR_MODE, choices=["DETAILED", "AGGREGATED"], help=f"Display mode for the live monitor (default: {DEFAULT_MONITOR_MODE})") parser.add_argument("--use-rate-limiter", action="store_true", default=False, help="Enable a basic rate limiter (default: False)") # Environment parameters parser.add_argument("--site-path", type=str, default=DEFAULT_SITE_PATH, help=f"Path to generate/use the test site (default: {DEFAULT_SITE_PATH})") parser.add_argument("--port", type=int, default=DEFAULT_PORT, help=f"Port for the local HTTP server (default: {DEFAULT_PORT})") parser.add_argument("--report-path", type=str, default=DEFAULT_REPORT_PATH, help=f"Path to save reports and logs (default: {DEFAULT_REPORT_PATH})") # Site/Server management parser.add_argument("--skip-generation", action="store_true", help="Use existing test site folder without regenerating") parser.add_argument("--use-existing-site", action="store_true", help="Do not generate site or start local server; assume site exists on --port") parser.add_argument("--keep-server-alive", action="store_true", help="Keep the local HTTP server running after test") parser.add_argument("--keep-site", action="store_true", help="Keep the generated test site files after test") parser.add_argument("--clean-reports", action="store_true", help="Clean up report directory before running") parser.add_argument("--clean-site", action="store_true", help="Clean up site directory before running (if generating) or after") args = parser.parse_args() # Display config console.print("[bold underline]Crawl4AI SDK Stress Test Configuration[/bold underline]") console.print(f"URLs: {args.urls}, Max Sessions: {args.max_sessions}, Chunk Size: {args.chunk_size}") # Added chunk size console.print(f"Mode: {'Streaming' if args.stream else 'Batch'}, Monitor: {args.monitor_mode}, Rate Limit: {args.use_rate_limiter}") console.print(f"Site Path: {args.site_path}, Port: {args.port}, Report Path: {args.report_path}") console.print("-" * 40) # (Rest of config display and cleanup logic is the same) if args.use_existing_site: console.print("[cyan]Mode: Using existing external site/server[/cyan]") elif args.skip_generation: console.print("[cyan]Mode: Using existing site files, starting local server[/cyan]") else: console.print("[cyan]Mode: Generating site files, starting local server[/cyan]") if args.keep_server_alive: console.print("[cyan]Option: Keep server alive after test[/cyan]") if args.keep_site: console.print("[cyan]Option: Keep site files after test[/cyan]") if args.clean_reports: console.print("[cyan]Option: Clean reports before test[/cyan]") if args.clean_site: console.print("[cyan]Option: Clean site directory[/cyan]") console.print("-" * 40) if args.clean_reports: if os.path.exists(args.report_path): console.print(f"[yellow]Cleaning up reports directory: {args.report_path}[/yellow]"); shutil.rmtree(args.report_path) os.makedirs(args.report_path, exist_ok=True) if args.clean_site and not args.use_existing_site: if os.path.exists(args.site_path): console.print(f"[yellow]Cleaning up site directory as requested: {args.site_path}[/yellow]"); shutil.rmtree(args.site_path) # Run try: asyncio.run(run_full_test(args)) except KeyboardInterrupt: console.print("\n[bold yellow]Test interrupted by user.[/bold yellow]") except Exception as e: console.print(f"\n[bold red]An unexpected error occurred:[/bold red] {e}"); import traceback; traceback.print_exc() if __name__ == "__main__": main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_parameters_and_options.py
tests/async/test_parameters_and_options.py
import os import sys import pytest # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_word_count_threshold(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result_no_threshold = await crawler.arun( url=url, word_count_threshold=0, bypass_cache=True ) result_with_threshold = await crawler.arun( url=url, word_count_threshold=50, bypass_cache=True ) assert len(result_no_threshold.markdown) > len(result_with_threshold.markdown) @pytest.mark.asyncio async def test_css_selector(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" css_selector = "h1, h2, h3" result = await crawler.arun( url=url, css_selector=css_selector, bypass_cache=True ) assert result.success assert ( "<h1" in result.cleaned_html or "<h2" in result.cleaned_html or "<h3" in result.cleaned_html ) @pytest.mark.asyncio async def test_javascript_execution(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" # Crawl without JS result_without_more = await crawler.arun(url=url, bypass_cache=True) js_code = [ "const loadMoreButton = Array.from(document.querySelectorAll('button')).find(button => button.textContent.includes('Load More')); loadMoreButton && loadMoreButton.click();" ] result_with_more = await crawler.arun(url=url, js=js_code, bypass_cache=True) assert result_with_more.success assert len(result_with_more.markdown) > len(result_without_more.markdown) @pytest.mark.asyncio async def test_screenshot(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, screenshot=True, bypass_cache=True) assert result.success assert result.screenshot assert isinstance(result.screenshot, str) # Should be a base64 encoded string @pytest.mark.asyncio async def test_custom_user_agent(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" custom_user_agent = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36 Crawl4AI/1.0" result = await crawler.arun( url=url, user_agent=custom_user_agent, bypass_cache=True ) assert result.success # Note: We can't directly verify the user agent in the result, but we can check if the crawl was successful @pytest.mark.asyncio async def test_extract_media_and_links(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.media assert isinstance(result.media, dict) assert "images" in result.media assert result.links assert isinstance(result.links, dict) assert "internal" in result.links and "external" in result.links @pytest.mark.asyncio async def test_metadata_extraction(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.metadata assert isinstance(result.metadata, dict) # Check for common metadata fields assert any( key in result.metadata for key in ["title", "description", "keywords"] ) # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_caching.py
tests/async/test_caching.py
import os import sys import pytest import asyncio # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_caching(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" # First crawl (should not use cache) start_time = asyncio.get_event_loop().time() result1 = await crawler.arun(url=url, bypass_cache=True) end_time = asyncio.get_event_loop().time() time_taken1 = end_time - start_time assert result1.success # Second crawl (should use cache) start_time = asyncio.get_event_loop().time() result2 = await crawler.arun(url=url, bypass_cache=False) end_time = asyncio.get_event_loop().time() time_taken2 = end_time - start_time assert result2.success assert time_taken2 < time_taken1 # Cached result should be faster @pytest.mark.asyncio async def test_bypass_cache(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" # First crawl result1 = await crawler.arun(url=url, bypass_cache=False) assert result1.success # Second crawl with bypass_cache=True result2 = await crawler.arun(url=url, bypass_cache=True) assert result2.success # Content should be different (or at least, not guaranteed to be the same) assert result1.html != result2.html or result1.markdown != result2.markdown @pytest.mark.asyncio async def test_clear_cache(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" # Crawl and cache await crawler.arun(url=url, bypass_cache=False) # Clear cache await crawler.aclear_cache() # Check cache size cache_size = await crawler.aget_cache_size() assert cache_size == 0 @pytest.mark.asyncio async def test_flush_cache(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" # Crawl and cache await crawler.arun(url=url, bypass_cache=False) # Flush cache await crawler.aflush_cache() # Check cache size cache_size = await crawler.aget_cache_size() assert cache_size == 0 # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_edge_cases.py
tests/async/test_edge_cases.py
import os import re import sys import pytest from bs4 import BeautifulSoup import asyncio # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler # @pytest.mark.asyncio # async def test_large_content_page(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://en.wikipedia.org/wiki/List_of_largest_known_stars" # A page with a large table # result = await crawler.arun(url=url, bypass_cache=True) # assert result.success # assert len(result.html) > 1000000 # Expecting more than 1MB of content # @pytest.mark.asyncio # async def test_minimal_content_page(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://example.com" # A very simple page # result = await crawler.arun(url=url, bypass_cache=True) # assert result.success # assert len(result.html) < 10000 # Expecting less than 10KB of content # @pytest.mark.asyncio # async def test_single_page_application(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://reactjs.org/" # React's website is a SPA # result = await crawler.arun(url=url, bypass_cache=True) # assert result.success # assert "react" in result.html.lower() # @pytest.mark.asyncio # async def test_page_with_infinite_scroll(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://news.ycombinator.com/" # Hacker News has infinite scroll # result = await crawler.arun(url=url, bypass_cache=True) # assert result.success # assert "hacker news" in result.html.lower() # @pytest.mark.asyncio # async def test_page_with_heavy_javascript(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://www.airbnb.com/" # Airbnb uses a lot of JavaScript # result = await crawler.arun(url=url, bypass_cache=True) # assert result.success # assert "airbnb" in result.html.lower() # @pytest.mark.asyncio # async def test_page_with_mixed_content(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://github.com/" # GitHub has a mix of static and dynamic content # result = await crawler.arun(url=url, bypass_cache=True) # assert result.success # assert "github" in result.html.lower() # Add this test to your existing test file @pytest.mark.asyncio async def test_typescript_commits_multi_page(): first_commit = "" async def on_execution_started(page): nonlocal first_commit try: # Check if the page firct commit h4 text is different from the first commit (use document.querySelector('li.Box-sc-g0xbh4-0 h4')) while True: await page.wait_for_selector("li.Box-sc-g0xbh4-0 h4") commit = await page.query_selector("li.Box-sc-g0xbh4-0 h4") commit = await commit.evaluate("(element) => element.textContent") commit = re.sub(r"\s+", "", commit) if commit and commit != first_commit: first_commit = commit break await asyncio.sleep(0.5) except Exception as e: print(f"Warning: New content didn't appear after JavaScript execution: {e}") async with AsyncWebCrawler(verbose=True) as crawler: crawler.crawler_strategy.set_hook("on_execution_started", on_execution_started) url = "https://github.com/microsoft/TypeScript/commits/main" session_id = "typescript_commits_session" all_commits = [] js_next_page = """ const button = document.querySelector('a[data-testid="pagination-next-button"]'); if (button) button.click(); """ for page in range(3): # Crawl 3 pages result = await crawler.arun( url=url, # Only use URL for the first page session_id=session_id, css_selector="li.Box-sc-g0xbh4-0", js=js_next_page if page > 0 else None, # Don't click 'next' on the first page bypass_cache=True, js_only=page > 0, # Use js_only for subsequent pages ) assert result.success, f"Failed to crawl page {page + 1}" # Parse the HTML and extract commits soup = BeautifulSoup(result.cleaned_html, "html.parser") commits = soup.select("li") # Take first commit find h4 extract text first_commit = commits[0].find("h4").text first_commit = re.sub(r"\s+", "", first_commit) all_commits.extend(commits) print(f"Page {page + 1}: Found {len(commits)} commits") # Clean up the session await crawler.crawler_strategy.kill_session(session_id) # Assertions assert ( len(all_commits) >= 90 ), f"Expected at least 90 commits, but got {len(all_commits)}" print(f"Successfully crawled {len(all_commits)} commits across 3 pages") # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_crawler_strategy.py
tests/async/test_crawler_strategy.py
import os import sys import pytest # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_custom_user_agent(): async with AsyncWebCrawler(verbose=True) as crawler: custom_user_agent = "MyCustomUserAgent/1.0" crawler.crawler_strategy.update_user_agent(custom_user_agent) url = "https://httpbin.org/user-agent" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert custom_user_agent in result.html @pytest.mark.asyncio async def test_custom_headers(): async with AsyncWebCrawler(verbose=True) as crawler: custom_headers = {"X-Test-Header": "TestValue"} crawler.crawler_strategy.set_custom_headers(custom_headers) url = "https://httpbin.org/headers" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert "X-Test-Header" in result.html assert "TestValue" in result.html @pytest.mark.asyncio async def test_javascript_execution(): async with AsyncWebCrawler(verbose=True) as crawler: js_code = "document.body.innerHTML = '<h1>Modified by JS</h1>';" url = "https://www.example.com" result = await crawler.arun(url=url, bypass_cache=True, js_code=js_code) assert result.success assert "<h1>Modified by JS</h1>" in result.html @pytest.mark.asyncio async def test_hook_execution(): async with AsyncWebCrawler(verbose=True) as crawler: async def test_hook(page): await page.evaluate("document.body.style.backgroundColor = 'red';") return page crawler.crawler_strategy.set_hook("after_goto", test_hook) url = "https://www.example.com" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert "background-color: red" in result.html @pytest.mark.asyncio async def test_screenshot(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.example.com" result = await crawler.arun(url=url, bypass_cache=True, screenshot=True) assert result.success assert result.screenshot assert isinstance(result.screenshot, str) assert len(result.screenshot) > 0 # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_async_doanloader.py
tests/async/test_async_doanloader.py
import os import sys import asyncio import shutil from typing import List import tempfile # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler class TestDownloads: def __init__(self): self.temp_dir = tempfile.mkdtemp(prefix="crawl4ai_test_") self.download_dir = os.path.join(self.temp_dir, "downloads") os.makedirs(self.download_dir, exist_ok=True) self.results: List[str] = [] def cleanup(self): shutil.rmtree(self.temp_dir) def log_result(self, test_name: str, success: bool, message: str = ""): result = f"{'✅' if success else '❌'} {test_name}: {message}" self.results.append(result) print(result) async def test_basic_download(self): """Test basic file download functionality""" try: async with AsyncWebCrawler( accept_downloads=True, downloads_path=self.download_dir, verbose=True ) as crawler: # Python.org downloads page typically has stable download links result = await crawler.arun( url="https://www.python.org/downloads/", js_code=""" // Click first download link const downloadLink = document.querySelector('a[href$=".exe"]'); if (downloadLink) downloadLink.click(); """, ) success = ( result.downloaded_files is not None and len(result.downloaded_files) > 0 ) self.log_result( "Basic Download", success, f"Downloaded {len(result.downloaded_files or [])} files" if success else "No files downloaded", ) except Exception as e: self.log_result("Basic Download", False, str(e)) async def test_persistent_context_download(self): """Test downloads with persistent context""" try: user_data_dir = os.path.join(self.temp_dir, "user_data") os.makedirs(user_data_dir, exist_ok=True) async with AsyncWebCrawler( accept_downloads=True, downloads_path=self.download_dir, use_persistent_context=True, user_data_dir=user_data_dir, verbose=True, ) as crawler: result = await crawler.arun( url="https://www.python.org/downloads/", js_code=""" const downloadLink = document.querySelector('a[href$=".exe"]'); if (downloadLink) downloadLink.click(); """, ) success = ( result.downloaded_files is not None and len(result.downloaded_files) > 0 ) self.log_result( "Persistent Context Download", success, f"Downloaded {len(result.downloaded_files or [])} files" if success else "No files downloaded", ) except Exception as e: self.log_result("Persistent Context Download", False, str(e)) async def test_multiple_downloads(self): """Test multiple simultaneous downloads""" try: async with AsyncWebCrawler( accept_downloads=True, downloads_path=self.download_dir, verbose=True ) as crawler: result = await crawler.arun( url="https://www.python.org/downloads/", js_code=""" // Click multiple download links const downloadLinks = document.querySelectorAll('a[href$=".exe"]'); downloadLinks.forEach(link => link.click()); """, ) success = ( result.downloaded_files is not None and len(result.downloaded_files) > 1 ) self.log_result( "Multiple Downloads", success, f"Downloaded {len(result.downloaded_files or [])} files" if success else "Not enough files downloaded", ) except Exception as e: self.log_result("Multiple Downloads", False, str(e)) async def test_different_browsers(self): """Test downloads across different browser types""" browsers = ["chromium", "firefox", "webkit"] for browser_type in browsers: try: async with AsyncWebCrawler( accept_downloads=True, downloads_path=self.download_dir, browser_type=browser_type, verbose=True, ) as crawler: result = await crawler.arun( url="https://www.python.org/downloads/", js_code=""" const downloadLink = document.querySelector('a[href$=".exe"]'); if (downloadLink) downloadLink.click(); """, ) success = ( result.downloaded_files is not None and len(result.downloaded_files) > 0 ) self.log_result( f"{browser_type.title()} Download", success, f"Downloaded {len(result.downloaded_files or [])} files" if success else "No files downloaded", ) except Exception as e: self.log_result(f"{browser_type.title()} Download", False, str(e)) async def test_edge_cases(self): """Test various edge cases""" # Test 1: Downloads without specifying download path try: async with AsyncWebCrawler(accept_downloads=True, verbose=True) as crawler: result = await crawler.arun( url="https://www.python.org/downloads/", js_code="document.querySelector('a[href$=\".exe\"]').click()", ) self.log_result( "Default Download Path", True, f"Downloaded to default path: {result.downloaded_files[0] if result.downloaded_files else 'None'}", ) except Exception as e: self.log_result("Default Download Path", False, str(e)) # Test 2: Downloads with invalid path try: async with AsyncWebCrawler( accept_downloads=True, downloads_path="/invalid/path/that/doesnt/exist", verbose=True, ) as crawler: result = await crawler.arun( url="https://www.python.org/downloads/", js_code="document.querySelector('a[href$=\".exe\"]').click()", ) self.log_result( "Invalid Download Path", False, "Should have raised an error" ) except Exception: self.log_result( "Invalid Download Path", True, "Correctly handled invalid path" ) # Test 3: Download with accept_downloads=False try: async with AsyncWebCrawler(accept_downloads=False, verbose=True) as crawler: result = await crawler.arun( url="https://www.python.org/downloads/", js_code="document.querySelector('a[href$=\".exe\"]').click()", ) success = result.downloaded_files is None self.log_result( "Disabled Downloads", success, "Correctly ignored downloads" if success else "Unexpectedly downloaded files", ) except Exception as e: self.log_result("Disabled Downloads", False, str(e)) async def run_all_tests(self): """Run all test cases""" print("\n🧪 Running Download Tests...\n") test_methods = [ self.test_basic_download, self.test_persistent_context_download, self.test_multiple_downloads, self.test_different_browsers, self.test_edge_cases, ] for test in test_methods: print(f"\n📝 Running {test.__doc__}...") await test() await asyncio.sleep(2) # Brief pause between tests print("\n📊 Test Results Summary:") for result in self.results: print(result) successes = len([r for r in self.results if "✅" in r]) total = len(self.results) print(f"\nTotal: {successes}/{total} tests passed") self.cleanup() async def main(): tester = TestDownloads() await tester.run_all_tests() if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_content_filter_prune.py
tests/async/test_content_filter_prune.py
import os, sys import pytest parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.content_filter_strategy import PruningContentFilter @pytest.fixture def basic_html(): return """ <html> <body> <article> <h1>Main Article</h1> <p>This is a high-quality paragraph with substantial text content. It contains enough words to pass the threshold and has good text density without too many links. This kind of content should survive the pruning process.</p> <div class="sidebar">Low quality sidebar content</div> <div class="social-share">Share buttons</div> </article> </body> </html> """ @pytest.fixture def link_heavy_html(): return """ <html> <body> <div class="content"> <p>Good content paragraph that should remain.</p> <div class="links"> <a href="#">Link 1</a> <a href="#">Link 2</a> <a href="#">Link 3</a> <a href="#">Link 4</a> </div> </div> </body> </html> """ @pytest.fixture def mixed_content_html(): return """ <html> <body> <article> <h1>Article Title</h1> <p class="summary">Short summary.</p> <div class="content"> <p>Long high-quality paragraph with substantial content that should definitely survive the pruning process. This content has good text density and proper formatting which makes it valuable for retention.</p> </div> <div class="comments"> <p>Short comment 1</p> <p>Short comment 2</p> </div> </article> </body> </html> """ class TestPruningContentFilter: def test_basic_pruning(self, basic_html): """Test basic content pruning functionality""" filter = PruningContentFilter(min_word_threshold=5) contents = filter.filter_content(basic_html) combined_content = " ".join(contents).lower() assert "high-quality paragraph" in combined_content assert "sidebar content" not in combined_content assert "share buttons" not in combined_content def test_min_word_threshold(self, mixed_content_html): """Test minimum word threshold filtering""" filter = PruningContentFilter(min_word_threshold=10) contents = filter.filter_content(mixed_content_html) combined_content = " ".join(contents).lower() assert "short summary" not in combined_content assert "long high-quality paragraph" in combined_content assert "short comment" not in combined_content def test_threshold_types(self, basic_html): """Test fixed vs dynamic thresholds""" fixed_filter = PruningContentFilter(threshold_type="fixed", threshold=0.48) dynamic_filter = PruningContentFilter(threshold_type="dynamic", threshold=0.45) fixed_contents = fixed_filter.filter_content(basic_html) dynamic_contents = dynamic_filter.filter_content(basic_html) assert len(fixed_contents) != len( dynamic_contents ), "Fixed and dynamic thresholds should yield different results" def test_link_density_impact(self, link_heavy_html): """Test handling of link-heavy content""" filter = PruningContentFilter(threshold_type="dynamic") contents = filter.filter_content(link_heavy_html) combined_content = " ".join(contents).lower() assert "good content paragraph" in combined_content assert ( len([c for c in contents if "href" in c]) < 2 ), "Should prune link-heavy sections" def test_tag_importance(self, mixed_content_html): """Test tag importance in scoring""" filter = PruningContentFilter(threshold_type="dynamic") contents = filter.filter_content(mixed_content_html) has_article = any("article" in c.lower() for c in contents) has_h1 = any("h1" in c.lower() for c in contents) assert has_article or has_h1, "Should retain important tags" def test_empty_input(self): """Test handling of empty input""" filter = PruningContentFilter() assert filter.filter_content("") == [] assert filter.filter_content(None) == [] def test_malformed_html(self): """Test handling of malformed HTML""" malformed_html = "<div>Unclosed div<p>Nested<span>content</div>" filter = PruningContentFilter() contents = filter.filter_content(malformed_html) assert isinstance(contents, list) def test_performance(self, basic_html): """Test performance with timer""" filter = PruningContentFilter() import time start = time.perf_counter() filter.filter_content(basic_html) duration = time.perf_counter() - start # Extra strict on performance since you mentioned milliseconds matter assert duration < 0.1, f"Processing took too long: {duration:.3f} seconds" @pytest.mark.parametrize( "threshold,expected_count", [ (0.3, 4), # Very lenient (0.48, 2), # Default (0.7, 1), # Very strict ], ) def test_threshold_levels(self, mixed_content_html, threshold, expected_count): """Test different threshold levels""" filter = PruningContentFilter(threshold_type="fixed", threshold=threshold) contents = filter.filter_content(mixed_content_html) assert ( len(contents) <= expected_count ), f"Expected {expected_count} or fewer elements with threshold {threshold}" def test_consistent_output(self, basic_html): """Test output consistency across multiple runs""" filter = PruningContentFilter() first_run = filter.filter_content(basic_html) second_run = filter.filter_content(basic_html) assert first_run == second_run, "Output should be consistent" if __name__ == "__main__": pytest.main([__file__])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_content_extraction.py
tests/async/test_content_extraction.py
import os import sys import pytest # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_extract_markdown(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.markdown assert isinstance(result.markdown, str) assert len(result.markdown) > 0 @pytest.mark.asyncio async def test_extract_cleaned_html(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.cleaned_html assert isinstance(result.cleaned_html, str) assert len(result.cleaned_html) > 0 @pytest.mark.asyncio async def test_extract_media(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.media media = result.media assert isinstance(media, dict) assert "images" in media assert isinstance(media["images"], list) for image in media["images"]: assert "src" in image assert "alt" in image assert "type" in image @pytest.mark.asyncio async def test_extract_links(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.links links = result.links assert isinstance(links, dict) assert "internal" in links assert "external" in links assert isinstance(links["internal"], list) assert isinstance(links["external"], list) for link in links["internal"] + links["external"]: assert "href" in link assert "text" in link @pytest.mark.asyncio async def test_extract_metadata(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.metadata metadata = result.metadata assert isinstance(metadata, dict) assert "title" in metadata assert isinstance(metadata["title"], str) @pytest.mark.asyncio async def test_css_selector_extraction(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" css_selector = "h1, h2, h3" result = await crawler.arun( url=url, bypass_cache=True, css_selector=css_selector ) assert result.success assert result.markdown assert all(heading in result.markdown for heading in ["#", "##", "###"]) @pytest.mark.asyncio async def test_base_tag_link_extraction(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://sohamkukreti.github.io/portfolio" result = await crawler.arun(url=url) assert result.success assert result.links assert isinstance(result.links, dict) assert "internal" in result.links assert "external" in result.links assert any("github.com" in x["href"] for x in result.links["external"]) # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_evaluation_scraping_methods_performance.configs.py
tests/async/test_evaluation_scraping_methods_performance.configs.py
import json import time from bs4 import BeautifulSoup from crawl4ai.content_scraping_strategy import ( WebScrapingStrategy, LXMLWebScrapingStrategy, ) from typing import Dict, List, Tuple import difflib from lxml import html as lhtml, etree def normalize_dom(element): """ Recursively normalizes an lxml HTML element: - Removes comment nodes - Sorts attributes on each node - Removes <head> if you want (optional) Returns the same element (mutated). """ # Remove comment nodes comments = element.xpath("//comment()") for c in comments: p = c.getparent() if p is not None: p.remove(c) # If you'd like to remove <head>, or unify <html>/<body>, you could do so here. # For example, remove <head> entirely: # heads = element.xpath('//head') # for h in heads: # parent = h.getparent() # if parent is not None: # parent.remove(h) # Sort attributes (to avoid false positives due to attr order) for el in element.iter(): if el.attrib: # Convert to a sorted list of (k, v), then reassign sorted_attribs = sorted(el.attrib.items()) el.attrib.clear() for k, v in sorted_attribs: el.set(k, v) return element def strip_html_body(root): """ If 'root' is <html>, find its <body> child and move all of <body>'s children into a new <div>. Return that <div>. If 'root' is <body>, similarly move all of its children into a new <div> and return it. Otherwise, return 'root' as-is. """ tag_name = (root.tag or "").lower() # Case 1: The root is <html> if tag_name == "html": bodies = root.xpath("./body") if bodies: body = bodies[0] new_div = lhtml.Element("div") for child in body: new_div.append(child) return new_div else: # No <body> found; just return the <html> root return root # Case 2: The root is <body> elif tag_name == "body": new_div = lhtml.Element("div") for child in root: new_div.append(child) return new_div # Case 3: Neither <html> nor <body> else: return root def compare_nodes(node1, node2, differences, path="/"): """ Recursively compare two lxml nodes, appending textual differences to `differences`. `path` is used to indicate the location in the tree (like an XPath). """ # 1) Compare tag names if node1.tag != node2.tag: differences.append(f"Tag mismatch at {path}: '{node1.tag}' vs. '{node2.tag}'") return # 2) Compare attributes # By now, they are sorted in normalize_dom() attrs1 = list(node1.attrib.items()) attrs2 = list(node2.attrib.items()) if attrs1 != attrs2: differences.append( f"Attribute mismatch at {path}/{node1.tag}: {attrs1} vs. {attrs2}" ) # 3) Compare text (trim or unify whitespace as needed) text1 = (node1.text or "").strip() text2 = (node2.text or "").strip() # Normalize whitespace text1 = " ".join(text1.split()) text2 = " ".join(text2.split()) if text1 != text2: # If you prefer ignoring newlines or multiple whitespace, do a more robust cleanup differences.append( f"Text mismatch at {path}/{node1.tag}: '{text1}' vs. '{text2}'" ) # 4) Compare number of children children1 = list(node1) children2 = list(node2) if len(children1) != len(children2): differences.append( f"Child count mismatch at {path}/{node1.tag}: {len(children1)} vs. {len(children2)}" ) return # If counts differ, no point comparing child by child # 5) Recursively compare each child for i, (c1, c2) in enumerate(zip(children1, children2)): # Build a path for child child_path = f"{path}/{node1.tag}[{i}]" compare_nodes(c1, c2, differences, child_path) # 6) Compare tail text tail1 = (node1.tail or "").strip() tail2 = (node2.tail or "").strip() if tail1 != tail2: differences.append( f"Tail mismatch after {path}/{node1.tag}: '{tail1}' vs. '{tail2}'" ) def compare_html_structurally(html1, html2): """ Compare two HTML strings using a structural approach with lxml. Returns a list of differences (if any). If empty, they're effectively the same. """ # 1) Parse both try: tree1 = lhtml.fromstring(html1) except etree.ParserError: return ["Error parsing HTML1"] try: tree2 = lhtml.fromstring(html2) except etree.ParserError: return ["Error parsing HTML2"] # 2) Normalize both DOMs (remove comments, sort attributes, etc.) tree1 = normalize_dom(tree1) tree2 = normalize_dom(tree2) # 3) Possibly strip <html>/<body> wrappers for better apples-to-apples comparison tree1 = strip_html_body(tree1) tree2 = strip_html_body(tree2) # 4) Compare recursively differences = [] compare_nodes(tree1, tree2, differences, path="") return differences def generate_large_html(n_elements=1000): html = ["<!DOCTYPE html><html><head></head><body>"] for i in range(n_elements): html.append( f""" <div class="article"> <h2>Heading {i}</h2> <p>This is paragraph {i} with some content and a <a href="http://example.com/{i}">link</a></p> <img src="image{i}.jpg" alt="Image {i}"> <ul> <li>List item {i}.1</li> <li>List item {i}.2</li> </ul> </div> """ ) html.append("</body></html>") return "".join(html) def generate_complicated_html(): """ HTML with multiple domains, forms, data attributes, various images, comments, style, and noscript to test all parameter toggles. """ return """ <!DOCTYPE html> <html> <head> <title>Complicated Test Page</title> <meta name="description" content="A very complicated page for testing."> <style> .hidden { display: none; } .highlight { color: red; } </style> </head> <body> <!-- This is a comment that we may remove if remove_comments=True --> <header> <h1>Main Title of the Page</h1> <nav> <a href="http://example.com/home">Home</a> <a href="http://social.com/profile">Social Profile</a> <a href="javascript:void(0)">JS Void Link</a> </nav> </header> <noscript> <p>JavaScript is disabled or not supported.</p> </noscript> <form action="submit.php" method="post"> <input type="text" name="username" /> <button type="submit">Submit</button> </form> <section> <article> <h2>Article Title</h2> <p> This paragraph has a good amount of text to exceed word_count_threshold if it's set to something small. But it might not exceed a very high threshold. </p> <img src="http://images.example.com/photo.jpg" alt="Descriptive alt text" style="width:200px;height:150px;" data-lazy="true"> <img src="icon.png" alt="Icon" style="display:none;"> <p>Another short text. <a href="/local-link">Local Link</a></p> </article> </section> <section id="promo-section"> <p>Promo text <a href="http://ads.example.com/ad">Ad Link</a></p> </section> <aside class="sidebar"> <img src="data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAA..." alt="Base64 Image"> <div data-info="secret" class="social-widget"> <p>Follow us on <a href="http://facebook.com/brand">Facebook</a></p> </div> </aside> <!-- Another comment below this line --> <script>console.log("script that might be removed");</script> <div style="display:none;"> <p>This is hidden</p> </div> <footer> <small>Footer Info &copy; 2025</small> </footer> </body> </html> """ def get_test_scenarios(): """ Returns a dictionary of parameter sets (test scenarios) for the scraper. Each scenario name maps to a dictionary of keyword arguments that will be passed into scrap() for testing various features. """ TEST_SCENARIOS = { # "default": {}, # "exclude_domains": { # "exclude_domains": {"images.example.com", "ads.example.com"} # }, # "exclude_social_media_links": { # "exclude_social_media_links": True # }, # "high_word_threshold": { # "word_count_threshold": 100 # }, # "keep_data_attrs": { # "keep_data_attributes": True # }, # "remove_forms_and_comments": { # "remove_forms": True, # "remove_comments": True # }, # "exclude_tags_and_selector": { # "excluded_tags": ["aside", "script"], # "excluded_selector": ".social-widget" # }, # "only_text_mode": { # "only_text": True # }, # "combo_mode": { # "exclude_domains": {"images.example.com", "ads.example.com"}, # "exclude_social_media_links": True, # "remove_forms": True, # "remove_comments": True, # "excluded_tags": ["aside"], # "excluded_selector": "#promo-section", # "only_text": False, # "keep_data_attributes": True, # "word_count_threshold": 20 # }, # "exclude_external_images": { # "exclude_external_images": True, # "exclude_social_media_links": True # }, # "strict_image_scoring": { # "image_score_threshold": 3, # "image_description_min_word_threshold": 10 # }, # "custom_css_selector": { # "css_selector": "section#promo-section" # }, # "remove_noscript": { # "excluded_tags": ["noscript"] # }, # "exclude_external_links": { # "exclude_external_links": True # }, # "large_word_count": { # "word_count_threshold": 500 # }, # "super_strict_images": { # "image_score_threshold": 5, # "image_description_min_word_threshold": 15 # }, # "exclude_style_and_script": { # "excluded_tags": ["style", "script"] # }, # "keep_data_and_remove_forms": { # "keep_data_attributes": True, # "remove_forms": True # }, # "only_text_high_word_count": { # "only_text": True, # "word_count_threshold": 40 # }, # "reduce_to_selector": { # "css_selector": "section > article" # }, # "exclude_all_links": { # # Removes all external links and also excludes example.com & social.com # "exclude_domains": {"example.com", "social.com", "facebook.com"}, # "exclude_external_links": True # }, # "comprehensive_removal": { # # Exclude multiple tags, remove forms & comments, # # and also remove targeted selectors # "excluded_tags": ["aside", "noscript", "script"], # "excluded_selector": "#promo-section, .social-widget", # "remove_comments": True, # "remove_forms": True # } } return TEST_SCENARIOS class ScraperEquivalenceTester: def __init__(self): self.test_cases = { "basic": self.generate_basic_html(), "complex": self.generate_complex_html(), "malformed": self.generate_malformed_html(), # 'real_world': self.load_real_samples() } def generate_basic_html(self): return generate_large_html(1000) # Your existing function def generate_complex_html(self): return """ <html><body> <div class="nested-content"> <article> <h1>Main Title</h1> <img src="test.jpg" srcset="test-1x.jpg 1x, test-2x.jpg 2x" data-src="lazy.jpg"> <p>Text with <a href="http://test.com">mixed <b>formatting</b></a></p> <iframe src="embedded.html"></iframe> </article> <nav> <ul> <li><a href="/page1">Link 1</a></li> <li><a href="javascript:void(0)">JS Link</a></li> </ul> </nav> </div> </body></html> """ def generate_malformed_html(self): return """ <div>Unclosed div <p>Unclosed paragraph <a href="test.com">Link</a> <img src=no-quotes> <script>document.write("<div>Dynamic</div>");</script> <!-- Malformed comment -- > --> <![CDATA[Test CDATA]]> """ def load_real_samples(self): # Load some real-world HTML samples you've collected samples = { "article": open("tests/samples/article.html").read(), "product": open("tests/samples/product.html").read(), "blog": open("tests/samples/blog.html").read(), } return samples def deep_compare_links(self, old_links: Dict, new_links: Dict) -> List[str]: """Detailed comparison of link structures""" differences = [] for category in ["internal", "external"]: old_urls = {link["href"] for link in old_links[category]} new_urls = {link["href"] for link in new_links[category]} missing = old_urls - new_urls extra = new_urls - old_urls if missing: differences.append(f"Missing {category} links: {missing}") if extra: differences.append(f"Extra {category} links: {extra}") # Compare link attributes for common URLs common = old_urls & new_urls for url in common: old_link = next(l for l in old_links[category] if l["href"] == url) new_link = next(l for l in new_links[category] if l["href"] == url) for attr in ["text", "title"]: if old_link[attr] != new_link[attr]: differences.append( f"Link attribute mismatch for {url} - {attr}:" f" old='{old_link[attr]}' vs new='{new_link[attr]}'" ) return differences def deep_compare_media(self, old_media: Dict, new_media: Dict) -> List[str]: """Detailed comparison of media elements""" differences = [] for media_type in ["images", "videos", "audios"]: old_srcs = {item["src"] for item in old_media[media_type]} new_srcs = {item["src"] for item in new_media[media_type]} missing = old_srcs - new_srcs extra = new_srcs - old_srcs if missing: differences.append(f"Missing {media_type}: {missing}") if extra: differences.append(f"Extra {media_type}: {extra}") # Compare media attributes for common sources common = old_srcs & new_srcs for src in common: old_item = next(m for m in old_media[media_type] if m["src"] == src) new_item = next(m for m in new_media[media_type] if m["src"] == src) for attr in ["alt", "description"]: if old_item.get(attr) != new_item.get(attr): differences.append( f"{media_type} attribute mismatch for {src} - {attr}:" f" old='{old_item.get(attr)}' vs new='{new_item.get(attr)}'" ) return differences def compare_html_content(self, old_html: str, new_html: str) -> List[str]: """Compare HTML content structure and text""" # return compare_html_structurally(old_html, new_html) differences = [] def normalize_html(html: str) -> Tuple[str, str]: soup = BeautifulSoup(html, "lxml") # Get both structure and text structure = " ".join(tag.name for tag in soup.find_all()) text = " ".join(soup.get_text().split()) return structure, text old_structure, old_text = normalize_html(old_html) new_structure, new_text = normalize_html(new_html) # Compare structure if abs(len(old_structure) - len(new_structure)) > 100: # if old_structure != new_structure: diff = difflib.unified_diff( old_structure.split(), new_structure.split(), lineterm="" ) differences.append("HTML structure differences:\n" + "\n".join(diff)) # Compare text content if abs(len(old_text) - len(new_text)) > 100: # if old_text != new_text: # Show detailed text differences text_diff = difflib.unified_diff( old_text.split(), new_text.split(), lineterm="" ) differences.append("Text content differences:\n" + "\n".join(text_diff)) return differences def compare_results( self, old_result: Dict, new_result: Dict ) -> Dict[str, List[str]]: """Comprehensive comparison of scraper outputs""" differences = {} # Compare links link_differences = self.deep_compare_links( old_result["links"], new_result["links"] ) if link_differences: differences["links"] = link_differences # Compare media media_differences = self.deep_compare_media( old_result["media"], new_result["media"] ) if media_differences: differences["media"] = media_differences # Compare HTML html_differences = self.compare_html_content( old_result["cleaned_html"], new_result["cleaned_html"] ) if html_differences: differences["html"] = html_differences return differences def run_tests(self) -> Dict: """Run comparison tests using the complicated HTML with multiple parameter scenarios.""" # We'll still keep some "test_cases" logic from above (basic, complex, malformed). # But we add a new section for the complicated HTML scenarios. results = {"tests": [], "summary": {"passed": 0, "failed": 0}} # 1) First, run the existing 3 built-in test cases (basic, complex, malformed). # for case_name, html in self.test_cases.items(): # print(f"\nTesting built-in case: {case_name}...") # original = WebScrapingStrategy() # lxml = LXMLWebScrapingStrategy() # start = time.time() # orig_result = original.scrap("http://test.com", html) # orig_time = time.time() - start # print("\nOriginal Mode:") # print(f"Cleaned HTML size: {len(orig_result['cleaned_html'])/1024:.2f} KB") # print(f"Images: {len(orig_result['media']['images'])}") # print(f"External links: {len(orig_result['links']['external'])}") # print(f"Times - Original: {orig_time:.3f}s") # start = time.time() # lxml_result = lxml.scrap("http://test.com", html) # lxml_time = time.time() - start # print("\nLXML Mode:") # print(f"Cleaned HTML size: {len(lxml_result['cleaned_html'])/1024:.2f} KB") # print(f"Images: {len(lxml_result['media']['images'])}") # print(f"External links: {len(lxml_result['links']['external'])}") # print(f"Times - LXML: {lxml_time:.3f}s") # # Compare # diffs = {} # link_diff = self.deep_compare_links(orig_result['links'], lxml_result['links']) # if link_diff: # diffs['links'] = link_diff # media_diff = self.deep_compare_media(orig_result['media'], lxml_result['media']) # if media_diff: # diffs['media'] = media_diff # html_diff = self.compare_html_content(orig_result['cleaned_html'], lxml_result['cleaned_html']) # if html_diff: # diffs['html'] = html_diff # test_result = { # 'case': case_name, # 'lxml_mode': { # 'differences': diffs, # 'execution_time': lxml_time # }, # 'original_time': orig_time # } # results['tests'].append(test_result) # if not diffs: # results['summary']['passed'] += 1 # else: # results['summary']['failed'] += 1 # 2) Now, run the complicated HTML with multiple parameter scenarios. complicated_html = generate_complicated_html() print("\n=== Testing complicated HTML with multiple parameter scenarios ===") # Create the scrapers once (or you can re-create if needed) original = WebScrapingStrategy() lxml = LXMLWebScrapingStrategy() for scenario_name, params in get_test_scenarios().items(): print(f"\nScenario: {scenario_name}") start = time.time() orig_result = original.scrap("http://test.com", complicated_html, **params) orig_time = time.time() - start start = time.time() lxml_result = lxml.scrap("http://test.com", complicated_html, **params) lxml_time = time.time() - start diffs = {} link_diff = self.deep_compare_links( orig_result["links"], lxml_result["links"] ) if link_diff: diffs["links"] = link_diff media_diff = self.deep_compare_media( orig_result["media"], lxml_result["media"] ) if media_diff: diffs["media"] = media_diff html_diff = self.compare_html_content( orig_result["cleaned_html"], lxml_result["cleaned_html"] ) if html_diff: diffs["html"] = html_diff test_result = { "case": f"complicated_{scenario_name}", "lxml_mode": {"differences": diffs, "execution_time": lxml_time}, "original_time": orig_time, } results["tests"].append(test_result) if not diffs: results["summary"]["passed"] += 1 print( f"✅ [OK] No differences found. Time(Orig: {orig_time:.3f}s, LXML: {lxml_time:.3f}s)" ) else: results["summary"]["failed"] += 1 print("❌ Differences found:") for category, dlist in diffs.items(): print(f" {category}:") for d in dlist: print(f" - {d}") return results def print_report(self, results: Dict): """Generate detailed equivalence report""" print("\n=== Scraper Equivalence Test Report ===\n") print(f"Total Cases: {len(results['tests'])}") print(f"Passed: {results['summary']['passed']}") print(f"Failed: {results['summary']['failed']}") for test in results["tests"]: print(f"\nTest Case: {test['case']}") if not test["lxml_mode"]["differences"]: print("✅ All implementations produced identical results") print( f"Times - Original: {test['original_time']:.3f}s, " f"LXML: {test['lxml_mode']['execution_time']:.3f}s" ) else: print("❌ Differences found:") if test["lxml_mode"]["differences"]: print("\nLXML Mode Differences:") for category, diffs in test["lxml_mode"]["differences"].items(): print(f"\n{category}:") for diff in diffs: print(f" - {diff}") def main(): tester = ScraperEquivalenceTester() results = tester.run_tests() tester.print_report(results) # Save detailed results for debugging with open("scraper_equivalence_results.json", "w") as f: json.dump(results, f, indent=2) if __name__ == "__main__": main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_content_scraper_strategy.py
tests/async/test_content_scraper_strategy.py
import os import sys import time import csv from tabulate import tabulate from dataclasses import dataclass from typing import List parent_dir = os.path.dirname( os.path.dirname(os.path.dirname(os.path.abspath(__file__))) ) sys.path.append(parent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) from crawl4ai.content_scraping_strategy import LXMLWebScrapingStrategy # This test compares the same strategy with itself now since WebScrapingStrategy is deprecated @dataclass class TestResult: name: str success: bool images: int internal_links: int external_links: int markdown_length: int execution_time: float class StrategyTester: def __init__(self): self.new_scraper = LXMLWebScrapingStrategy() self.current_scraper = LXMLWebScrapingStrategy() # Same strategy now with open(__location__ + "/sample_wikipedia.html", "r", encoding="utf-8") as f: self.WIKI_HTML = f.read() self.results = {"new": [], "current": []} def run_test(self, name: str, **kwargs) -> tuple[TestResult, TestResult]: results = [] for scraper in [self.new_scraper, self.current_scraper]: start_time = time.time() result = scraper._get_content_of_website_optimized( url="https://en.wikipedia.org/wiki/Test", html=self.WIKI_HTML, **kwargs ) execution_time = time.time() - start_time test_result = TestResult( name=name, success=result["success"], images=len(result["media"]["images"]), internal_links=len(result["links"]["internal"]), external_links=len(result["links"]["external"]), markdown_length=len(result["markdown"]), execution_time=execution_time, ) results.append(test_result) return results[0], results[1] # new, current def run_all_tests(self): test_cases = [ ("Basic Extraction", {}), ("Exclude Tags", {"excluded_tags": ["table", "div.infobox", "div.navbox"]}), ("Word Threshold", {"word_count_threshold": 50}), ("CSS Selector", {"css_selector": "div.mw-parser-output > p"}), ( "Link Exclusions", { "exclude_external_links": True, "exclude_social_media_links": True, "exclude_domains": ["facebook.com", "twitter.com"], }, ), ( "Media Handling", { "exclude_external_images": True, "image_description_min_word_threshold": 20, }, ), ("Text Only", {"only_text": True, "remove_forms": True}), ("HTML Cleaning", {"clean_html": True, "keep_data_attributes": True}), ( "HTML2Text Options", { "html2text": { "skip_internal_links": True, "single_line_break": True, "mark_code": True, "preserve_tags": ["pre", "code"], } }, ), ] all_results = [] for name, kwargs in test_cases: try: new_result, current_result = self.run_test(name, **kwargs) all_results.append((name, new_result, current_result)) except Exception as e: print(f"Error in {name}: {str(e)}") self.save_results_to_csv(all_results) self.print_comparison_table(all_results) def save_results_to_csv(self, all_results: List[tuple]): csv_file = os.path.join(__location__, "strategy_comparison_results.csv") with open(csv_file, "w", newline="") as f: writer = csv.writer(f) writer.writerow( [ "Test Name", "Strategy", "Success", "Images", "Internal Links", "External Links", "Markdown Length", "Execution Time", ] ) for name, new_result, current_result in all_results: writer.writerow( [ name, "New", new_result.success, new_result.images, new_result.internal_links, new_result.external_links, new_result.markdown_length, f"{new_result.execution_time:.3f}", ] ) writer.writerow( [ name, "Current", current_result.success, current_result.images, current_result.internal_links, current_result.external_links, current_result.markdown_length, f"{current_result.execution_time:.3f}", ] ) def print_comparison_table(self, all_results: List[tuple]): table_data = [] headers = [ "Test Name", "Strategy", "Success", "Images", "Internal Links", "External Links", "Markdown Length", "Time (s)", ] for name, new_result, current_result in all_results: # Check for differences differences = [] if new_result.images != current_result.images: differences.append("images") if new_result.internal_links != current_result.internal_links: differences.append("internal_links") if new_result.external_links != current_result.external_links: differences.append("external_links") if new_result.markdown_length != current_result.markdown_length: differences.append("markdown") # Add row for new strategy new_row = [ name, "New", new_result.success, new_result.images, new_result.internal_links, new_result.external_links, new_result.markdown_length, f"{new_result.execution_time:.3f}", ] table_data.append(new_row) # Add row for current strategy current_row = [ "", "Current", current_result.success, current_result.images, current_result.internal_links, current_result.external_links, current_result.markdown_length, f"{current_result.execution_time:.3f}", ] table_data.append(current_row) # Add difference summary if any if differences: table_data.append( ["", "⚠️ Differences", ", ".join(differences), "", "", "", "", ""] ) # Add empty row for better readability table_data.append([""] * len(headers)) print("\nStrategy Comparison Results:") print(tabulate(table_data, headers=headers, tablefmt="grid")) if __name__ == "__main__": tester = StrategyTester() tester.run_all_tests()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_dispatchers.py
tests/async/test_dispatchers.py
import pytest import time from crawl4ai import ( AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, MemoryAdaptiveDispatcher, SemaphoreDispatcher, RateLimiter, CrawlerMonitor, DisplayMode, CacheMode, ) @pytest.fixture def browser_config(): return BrowserConfig(headless=True, verbose=False) @pytest.fixture def run_config(): return CrawlerRunConfig(cache_mode=CacheMode.BYPASS, verbose=False) @pytest.fixture def test_urls(): return [ "http://example.com", "http://example.com/page1", "http://example.com/page2", ] @pytest.mark.asyncio class TestDispatchStrategies: async def test_memory_adaptive_basic(self, browser_config, run_config, test_urls): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher( memory_threshold_percent=70.0, max_session_permit=2, check_interval=0.1 ) results = await crawler.arun_many( test_urls, config=run_config, dispatcher=dispatcher ) assert len(results) == len(test_urls) assert all(r.success for r in results) async def test_memory_adaptive_with_rate_limit( self, browser_config, run_config, test_urls ): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher( memory_threshold_percent=70.0, max_session_permit=2, check_interval=0.1, rate_limiter=RateLimiter( base_delay=(0.1, 0.2), max_delay=1.0, max_retries=2 ), ) results = await crawler.arun_many( test_urls, config=run_config, dispatcher=dispatcher ) assert len(results) == len(test_urls) assert all(r.success for r in results) async def test_semaphore_basic(self, browser_config, run_config, test_urls): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = SemaphoreDispatcher(semaphore_count=2) results = await crawler.arun_many( test_urls, config=run_config, dispatcher=dispatcher ) assert len(results) == len(test_urls) assert all(r.success for r in results) async def test_semaphore_with_rate_limit( self, browser_config, run_config, test_urls ): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = SemaphoreDispatcher( semaphore_count=2, rate_limiter=RateLimiter( base_delay=(0.1, 0.2), max_delay=1.0, max_retries=2 ), ) results = await crawler.arun_many( test_urls, config=run_config, dispatcher=dispatcher ) assert len(results) == len(test_urls) assert all(r.success for r in results) async def test_memory_adaptive_memory_error( self, browser_config, run_config, test_urls ): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher( memory_threshold_percent=1.0, # Set unrealistically low threshold max_session_permit=2, check_interval=0.1, memory_wait_timeout=1.0, # Short timeout for testing ) with pytest.raises(MemoryError): await crawler.arun_many( test_urls, config=run_config, dispatcher=dispatcher ) async def test_empty_urls(self, browser_config, run_config): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher(max_session_permit=2) results = await crawler.arun_many( [], config=run_config, dispatcher=dispatcher ) assert len(results) == 0 async def test_single_url(self, browser_config, run_config): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher(max_session_permit=2) results = await crawler.arun_many( ["http://example.com"], config=run_config, dispatcher=dispatcher ) assert len(results) == 1 assert results[0].success async def test_invalid_urls(self, browser_config, run_config): async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher(max_session_permit=2) results = await crawler.arun_many( ["http://invalid.url.that.doesnt.exist"], config=run_config, dispatcher=dispatcher, ) assert len(results) == 1 assert not results[0].success async def test_rate_limit_backoff(self, browser_config, run_config): urls = ["http://example.com"] * 5 # Multiple requests to same domain async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher( max_session_permit=2, rate_limiter=RateLimiter( base_delay=(0.1, 0.2), max_delay=1.0, max_retries=2, rate_limit_codes=[200], # Force rate limiting for testing ), ) start_time = time.time() results = await crawler.arun_many( urls, config=run_config, dispatcher=dispatcher ) duration = time.time() - start_time assert len(results) == len(urls) assert duration > 1.0 # Ensure rate limiting caused delays async def test_monitor_integration(self, browser_config, run_config, test_urls): async with AsyncWebCrawler(config=browser_config) as crawler: monitor = CrawlerMonitor( max_visible_rows=5, display_mode=DisplayMode.DETAILED ) dispatcher = MemoryAdaptiveDispatcher(max_session_permit=2, monitor=monitor) results = await crawler.arun_many( test_urls, config=run_config, dispatcher=dispatcher ) assert len(results) == len(test_urls) # Check monitor stats assert len(monitor.stats) == len(test_urls) assert all(stat.end_time is not None for stat in monitor.stats.values()) if __name__ == "__main__": pytest.main([__file__, "-v", "--asyncio-mode=auto"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_markdown_genertor.py
tests/async/test_markdown_genertor.py
# ## Issue #236 # - **Last Updated:** 2024-11-11 01:42:14 # - **Title:** [user data crawling opens two windows, unable to control correct user browser](https://github.com/unclecode/crawl4ai/issues/236) # - **State:** open import os, sys, time parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) import os import time from typing import Dict, Any from crawl4ai.markdown_generation_strategy import DefaultMarkdownGenerator # Get current directory __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) def print_test_result(name: str, result: Dict[str, Any], execution_time: float): """Helper function to print test results.""" print(f"\n{'='*20} {name} {'='*20}") print(f"Execution time: {execution_time:.4f} seconds") # Save markdown to files for key, content in result.items(): if isinstance(content, str): with open(__location__ + f"/output/{name.lower()}_{key}.md", "w") as f: f.write(content) # # Print first few lines of each markdown version # for key, content in result.items(): # if isinstance(content, str): # preview = '\n'.join(content.split('\n')[:3]) # print(f"\n{key} (first 3 lines):") # print(preview) # print(f"Total length: {len(content)} characters") def test_basic_markdown_conversion(): """Test basic markdown conversion with links.""" with open(__location__ + "/data/wikipedia.html", "r") as f: cleaned_html = f.read() generator = DefaultMarkdownGenerator() start_time = time.perf_counter() result = generator.generate_markdown( cleaned_html=cleaned_html, base_url="https://en.wikipedia.org" ) execution_time = time.perf_counter() - start_time print_test_result( "Basic Markdown Conversion", { "raw": result.raw_markdown, "with_citations": result.markdown_with_citations, "references": result.references_markdown, }, execution_time, ) # Basic assertions assert result.raw_markdown, "Raw markdown should not be empty" assert result.markdown_with_citations, "Markdown with citations should not be empty" assert result.references_markdown, "References should not be empty" assert "⟨" in result.markdown_with_citations, "Citations should use ⟨⟩ brackets" assert ( "## References" in result.references_markdown ), "Should contain references section" def test_relative_links(): """Test handling of relative links with base URL.""" markdown = """ Here's a [relative link](/wiki/Apple) and an [absolute link](https://example.com). Also an [image](/images/test.png) and another [page](/wiki/Banana). """ generator = DefaultMarkdownGenerator() result = generator.generate_markdown( cleaned_html=markdown, base_url="https://en.wikipedia.org" ) assert "https://en.wikipedia.org/wiki/Apple" in result.references_markdown assert "https://example.com" in result.references_markdown assert "https://en.wikipedia.org/images/test.png" in result.references_markdown def test_duplicate_links(): """Test handling of duplicate links.""" markdown = """ Here's a [link](/test) and another [link](/test) and a [different link](/other). """ generator = DefaultMarkdownGenerator() result = generator.generate_markdown( cleaned_html=markdown, base_url="https://example.com" ) # Count citations in markdown citations = result.markdown_with_citations.count("⟨1⟩") assert citations == 2, "Same link should use same citation number" def test_link_descriptions(): """Test handling of link titles and descriptions.""" markdown = """ Here's a [link with title](/test "Test Title") and a [link with description](/other) to test. """ generator = DefaultMarkdownGenerator() result = generator.generate_markdown( cleaned_html=markdown, base_url="https://example.com" ) assert ( "Test Title" in result.references_markdown ), "Link title should be in references" assert ( "link with description" in result.references_markdown ), "Link text should be in references" def test_performance_large_document(): """Test performance with large document.""" with open(__location__ + "/data/wikipedia.md", "r") as f: markdown = f.read() # Test with multiple iterations iterations = 5 times = [] generator = DefaultMarkdownGenerator() for i in range(iterations): start_time = time.perf_counter() result = generator.generate_markdown( cleaned_html=markdown, base_url="https://en.wikipedia.org" ) end_time = time.perf_counter() times.append(end_time - start_time) avg_time = sum(times) / len(times) print(f"\n{'='*20} Performance Test {'='*20}") print( f"Average execution time over {iterations} iterations: {avg_time:.4f} seconds" ) print(f"Min time: {min(times):.4f} seconds") print(f"Max time: {max(times):.4f} seconds") def test_image_links(): """Test handling of image links.""" markdown = """ Here's an ![image](/image.png "Image Title") and another ![image](/other.jpg). And a regular [link](/page). """ generator = DefaultMarkdownGenerator() result = generator.generate_markdown( cleaned_html=markdown, base_url="https://example.com" ) assert ( "![" in result.markdown_with_citations ), "Image markdown syntax should be preserved" assert ( "Image Title" in result.references_markdown ), "Image title should be in references" if __name__ == "__main__": print("Running markdown generation strategy tests...") test_basic_markdown_conversion() test_relative_links() test_duplicate_links() test_link_descriptions() test_performance_large_document() test_image_links()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_basic_crawling.py
tests/async/test_basic_crawling.py
import os import sys import pytest import time # Add the parent directory to the Python path parent_dir = os.path.dirname( os.path.dirname(os.path.dirname(os.path.abspath(__file__))) ) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_successful_crawl(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" result = await crawler.arun(url=url, bypass_cache=True) assert result.success assert result.url == url assert result.html assert result.markdown assert result.cleaned_html @pytest.mark.asyncio async def test_invalid_url(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.invalidurl12345.com" result = await crawler.arun(url=url, bypass_cache=True) assert not result.success assert result.error_message @pytest.mark.asyncio async def test_multiple_urls(): async with AsyncWebCrawler(verbose=True) as crawler: urls = [ "https://www.nbcnews.com/business", "https://www.example.com", "https://www.python.org", ] results = await crawler.arun_many(urls=urls, bypass_cache=True) assert len(results) == len(urls) assert all(result.success for result in results) assert all(result.html for result in results) @pytest.mark.asyncio async def test_javascript_execution(): async with AsyncWebCrawler(verbose=True) as crawler: js_code = "document.body.innerHTML = '<h1>Modified by JS</h1>';" url = "https://www.example.com" result = await crawler.arun(url=url, bypass_cache=True, js_code=js_code) assert result.success assert "<h1>Modified by JS</h1>" in result.html @pytest.mark.asyncio async def test_concurrent_crawling_performance(): async with AsyncWebCrawler(verbose=True) as crawler: urls = [ "https://www.nbcnews.com/business", "https://www.example.com", "https://www.python.org", "https://www.github.com", "https://www.stackoverflow.com", ] start_time = time.time() results = await crawler.arun_many(urls=urls, bypass_cache=True) end_time = time.time() total_time = end_time - start_time print(f"Total time for concurrent crawling: {total_time:.2f} seconds") assert all(result.success for result in results) assert len(results) == len(urls) # Assert that concurrent crawling is faster than sequential # This multiplier may need adjustment based on the number of URLs and their complexity assert ( total_time < len(urls) * 5 ), f"Concurrent crawling not significantly faster: {total_time:.2f} seconds" # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_chunking_and_extraction_strategies.py
tests/async/test_chunking_and_extraction_strategies.py
import os import sys import pytest import json # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai import LLMConfig from crawl4ai.async_webcrawler import AsyncWebCrawler from crawl4ai.chunking_strategy import RegexChunking from crawl4ai import LLMExtractionStrategy @pytest.mark.asyncio async def test_regex_chunking(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" chunking_strategy = RegexChunking(patterns=["\n\n"]) result = await crawler.arun( url=url, chunking_strategy=chunking_strategy, bypass_cache=True ) assert result.success assert result.extracted_content chunks = json.loads(result.extracted_content) assert len(chunks) > 1 # Ensure multiple chunks were created # @pytest.mark.asyncio # async def test_cosine_strategy(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://www.nbcnews.com/business" # extraction_strategy = CosineStrategy(word_count_threshold=10, max_dist=0.2, linkage_method="ward", top_k=3, sim_threshold=0.3) # result = await crawler.arun( # url=url, # extraction_strategy=extraction_strategy, # bypass_cache=True # ) # assert result.success # assert result.extracted_content # extracted_data = json.loads(result.extracted_content) # assert len(extracted_data) > 0 # assert all('tags' in item for item in extracted_data) @pytest.mark.asyncio async def test_llm_extraction_strategy(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" extraction_strategy = LLMExtractionStrategy( llm_config=LLMConfig(provider="openai/gpt-4o-mini",api_token=os.getenv("OPENAI_API_KEY")), instruction="Extract only content related to technology", ) result = await crawler.arun( url=url, extraction_strategy=extraction_strategy, bypass_cache=True ) assert result.success assert result.extracted_content extracted_data = json.loads(result.extracted_content) assert len(extracted_data) > 0 assert all("content" in item for item in extracted_data) # @pytest.mark.asyncio # async def test_combined_chunking_and_extraction(): # async with AsyncWebCrawler(verbose=True) as crawler: # url = "https://www.nbcnews.com/business" # chunking_strategy = RegexChunking(patterns=["\n\n"]) # extraction_strategy = CosineStrategy(word_count_threshold=10, max_dist=0.2, linkage_method="ward", top_k=3, sim_threshold=0.3) # result = await crawler.arun( # url=url, # chunking_strategy=chunking_strategy, # extraction_strategy=extraction_strategy, # bypass_cache=True # ) # assert result.success # assert result.extracted_content # extracted_data = json.loads(result.extracted_content) # assert len(extracted_data) > 0 # assert all('tags' in item for item in extracted_data) # assert all('content' in item for item in extracted_data) # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_content_filter_bm25.py
tests/async/test_content_filter_bm25.py
import os, sys import pytest from bs4 import BeautifulSoup # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.content_filter_strategy import BM25ContentFilter @pytest.fixture def basic_html(): return """ <html> <head> <title>Test Article</title> <meta name="description" content="Test description"> <meta name="keywords" content="test, keywords"> </head> <body> <h1>Main Heading</h1> <article> <p>This is a long paragraph with more than fifty words. It continues with more text to ensure we meet the minimum word count threshold. We need to make sure this paragraph is substantial enough to be considered for extraction according to our filtering rules. This should be enough words now.</p> <div class="navigation">Skip this nav content</div> </article> </body> </html> """ @pytest.fixture def wiki_html(): return """ <html> <head> <title>Wikipedia Article</title> </head> <body> <h1>Article Title</h1> <h2>Section 1</h2> <p>Short but important section header description.</p> <div class="content"> <p>Long paragraph with sufficient words to meet the minimum threshold. This paragraph continues with more text to ensure we have enough content for proper testing. We need to make sure this has enough words to pass our filters and be considered valid content for extraction purposes.</p> </div> </body> </html> """ @pytest.fixture def no_meta_html(): return """ <html> <body> <h1>Simple Page</h1> <p>First paragraph that should be used as fallback for query when no meta tags exist. This text needs to be long enough to serve as a meaningful fallback for our content extraction process.</p> </body> </html> """ class TestBM25ContentFilter: def test_basic_extraction(self, basic_html): """Test basic content extraction functionality""" filter = BM25ContentFilter() contents = filter.filter_content(basic_html) assert contents, "Should extract content" assert len(contents) >= 1, "Should extract at least one content block" assert "long paragraph" in " ".join(contents).lower() assert "navigation" not in " ".join(contents).lower() def test_user_query_override(self, basic_html): """Test that user query overrides metadata extraction""" user_query = "specific test query" filter = BM25ContentFilter(user_query=user_query) # Access internal state to verify query usage soup = BeautifulSoup(basic_html, "lxml") extracted_query = filter.extract_page_query(soup.find("head")) assert extracted_query == user_query assert "Test description" not in extracted_query def test_header_extraction(self, wiki_html): """Test that headers are properly extracted despite length""" filter = BM25ContentFilter() contents = filter.filter_content(wiki_html) combined_content = " ".join(contents).lower() assert "section 1" in combined_content, "Should include section header" assert "article title" in combined_content, "Should include main title" def test_no_metadata_fallback(self, no_meta_html): """Test fallback behavior when no metadata is present""" filter = BM25ContentFilter() contents = filter.filter_content(no_meta_html) assert contents, "Should extract content even without metadata" assert "First paragraph" in " ".join( contents ), "Should use first paragraph content" def test_empty_input(self): """Test handling of empty input""" filter = BM25ContentFilter() assert filter.filter_content("") == [] assert filter.filter_content(None) == [] def test_malformed_html(self): """Test handling of malformed HTML""" malformed_html = "<p>Unclosed paragraph<div>Nested content</p></div>" filter = BM25ContentFilter() contents = filter.filter_content(malformed_html) assert isinstance(contents, list), "Should return list even with malformed HTML" def test_threshold_behavior(self, basic_html): """Test different BM25 threshold values""" strict_filter = BM25ContentFilter(bm25_threshold=2.0) lenient_filter = BM25ContentFilter(bm25_threshold=0.5) strict_contents = strict_filter.filter_content(basic_html) lenient_contents = lenient_filter.filter_content(basic_html) assert len(strict_contents) <= len( lenient_contents ), "Strict threshold should extract fewer elements" def test_html_cleaning(self, basic_html): """Test HTML cleaning functionality""" filter = BM25ContentFilter() contents = filter.filter_content(basic_html) cleaned_content = " ".join(contents) assert "class=" not in cleaned_content, "Should remove class attributes" assert "style=" not in cleaned_content, "Should remove style attributes" assert "<script" not in cleaned_content, "Should remove script tags" def test_large_content(self): """Test handling of large content blocks""" large_html = f""" <html><body> <article>{'<p>Test content. ' * 1000}</article> </body></html> """ filter = BM25ContentFilter() contents = filter.filter_content(large_html) assert contents, "Should handle large content blocks" @pytest.mark.parametrize( "unwanted_tag", ["script", "style", "nav", "footer", "header"] ) def test_excluded_tags(self, unwanted_tag): """Test that specific tags are properly excluded""" html = f""" <html><body> <{unwanted_tag}>Should not appear</{unwanted_tag}> <p>Should appear</p> </body></html> """ filter = BM25ContentFilter() contents = filter.filter_content(html) combined_content = " ".join(contents).lower() assert "should not appear" not in combined_content def test_performance(self, basic_html): """Test performance with timer""" filter = BM25ContentFilter() import time start = time.perf_counter() filter.filter_content(basic_html) duration = time.perf_counter() - start assert duration < 1.0, f"Processing took too long: {duration:.2f} seconds" if __name__ == "__main__": pytest.main([__file__])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_screenshot.py
tests/async/test_screenshot.py
import os import sys import pytest import base64 from PIL import Image import io # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_basic_screenshot(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://example.com" # A static website result = await crawler.arun(url=url, bypass_cache=True, screenshot=True) assert result.success assert result.screenshot is not None # Verify the screenshot is a valid image image_data = base64.b64decode(result.screenshot) image = Image.open(io.BytesIO(image_data)) assert image.format == "PNG" @pytest.mark.asyncio async def test_screenshot_with_wait_for(): async with AsyncWebCrawler(verbose=True) as crawler: # Using a website with dynamic content url = "https://www.youtube.com" wait_for = "css:#content" # Wait for the main content to load result = await crawler.arun( url=url, bypass_cache=True, screenshot=True, wait_for=wait_for ) assert result.success assert result.screenshot is not None # Verify the screenshot is a valid image image_data = base64.b64decode(result.screenshot) image = Image.open(io.BytesIO(image_data)) assert image.format == "PNG" # You might want to add more specific checks here, like image dimensions # or even use image recognition to verify certain elements are present @pytest.mark.asyncio async def test_screenshot_with_js_wait_for(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.amazon.com" wait_for = "js:() => document.querySelector('#nav-logo-sprites') !== null" result = await crawler.arun( url=url, bypass_cache=True, screenshot=True, wait_for=wait_for ) assert result.success assert result.screenshot is not None image_data = base64.b64decode(result.screenshot) image = Image.open(io.BytesIO(image_data)) assert image.format == "PNG" @pytest.mark.asyncio async def test_screenshot_without_wait_for(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nytimes.com" # A website with lots of dynamic content result = await crawler.arun(url=url, bypass_cache=True, screenshot=True) assert result.success assert result.screenshot is not None image_data = base64.b64decode(result.screenshot) image = Image.open(io.BytesIO(image_data)) assert image.format == "PNG" @pytest.mark.asyncio async def test_screenshot_comparison(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.reddit.com" wait_for = "css:#SHORTCUT_FOCUSABLE_DIV" # Take screenshot without wait_for result_without_wait = await crawler.arun( url=url, bypass_cache=True, screenshot=True ) # Take screenshot with wait_for result_with_wait = await crawler.arun( url=url, bypass_cache=True, screenshot=True, wait_for=wait_for ) assert result_without_wait.success and result_with_wait.success assert result_without_wait.screenshot is not None assert result_with_wait.screenshot is not None # Compare the two screenshots image_without_wait = Image.open( io.BytesIO(base64.b64decode(result_without_wait.screenshot)) ) image_with_wait = Image.open( io.BytesIO(base64.b64decode(result_with_wait.screenshot)) ) # This is a simple size comparison. In a real-world scenario, you might want to use # more sophisticated image comparison techniques. assert image_with_wait.size[0] >= image_without_wait.size[0] assert image_with_wait.size[1] >= image_without_wait.size[1] # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_error_handling.py
tests/async/test_error_handling.py
# import os # import sys # import pytest # import asyncio # # Add the parent directory to the Python path # parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # sys.path.append(parent_dir) # from crawl4ai.async_webcrawler import AsyncWebCrawler # from crawl4ai.utils import InvalidCSSSelectorError # class AsyncCrawlerWrapper: # def __init__(self): # self.crawler = None # async def setup(self): # self.crawler = AsyncWebCrawler(verbose=True) # await self.crawler.awarmup() # async def cleanup(self): # if self.crawler: # await self.crawler.aclear_cache() # @pytest.fixture(scope="module") # def crawler_wrapper(): # wrapper = AsyncCrawlerWrapper() # asyncio.get_event_loop().run_until_complete(wrapper.setup()) # yield wrapper # asyncio.get_event_loop().run_until_complete(wrapper.cleanup()) # @pytest.mark.asyncio # async def test_network_error(crawler_wrapper): # url = "https://www.nonexistentwebsite123456789.com" # result = await crawler_wrapper.crawler.arun(url=url, bypass_cache=True) # assert not result.success # assert "Failed to crawl" in result.error_message # # @pytest.mark.asyncio # # async def test_timeout_error(crawler_wrapper): # # # Simulating a timeout by using a very short timeout value # # url = "https://www.nbcnews.com/business" # # result = await crawler_wrapper.crawler.arun(url=url, bypass_cache=True, timeout=0.001) # # assert not result.success # # assert "timeout" in result.error_message.lower() # # @pytest.mark.asyncio # # async def test_invalid_css_selector(crawler_wrapper): # # url = "https://www.nbcnews.com/business" # # with pytest.raises(InvalidCSSSelectorError): # # await crawler_wrapper.crawler.arun(url=url, bypass_cache=True, css_selector="invalid>>selector") # # @pytest.mark.asyncio # # async def test_js_execution_error(crawler_wrapper): # # url = "https://www.nbcnews.com/business" # # invalid_js = "This is not valid JavaScript code;" # # result = await crawler_wrapper.crawler.arun(url=url, bypass_cache=True, js=invalid_js) # # assert not result.success # # assert "JavaScript" in result.error_message # # @pytest.mark.asyncio # # async def test_empty_page(crawler_wrapper): # # # Use a URL that typically returns an empty page # # url = "http://example.com/empty" # # result = await crawler_wrapper.crawler.arun(url=url, bypass_cache=True) # # assert result.success # The crawl itself should succeed # # assert not result.markdown.strip() # The markdown content should be empty or just whitespace # # @pytest.mark.asyncio # # async def test_rate_limiting(crawler_wrapper): # # # Simulate rate limiting by making multiple rapid requests # # url = "https://www.nbcnews.com/business" # # results = await asyncio.gather(*[crawler_wrapper.crawler.arun(url=url, bypass_cache=True) for _ in range(10)]) # # assert any(not result.success and "rate limit" in result.error_message.lower() for result in results) # # Entry point for debugging # if __name__ == "__main__": # pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_performance.py
tests/async/test_performance.py
import os import sys import pytest import time # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_crawl_speed(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" start_time = time.time() result = await crawler.arun(url=url, bypass_cache=True) end_time = time.time() assert result.success crawl_time = end_time - start_time print(f"Crawl time: {crawl_time:.2f} seconds") assert crawl_time < 10, f"Crawl took too long: {crawl_time:.2f} seconds" @pytest.mark.asyncio async def test_concurrent_crawling_performance(): async with AsyncWebCrawler(verbose=True) as crawler: urls = [ "https://www.nbcnews.com/business", "https://www.example.com", "https://www.python.org", "https://www.github.com", "https://www.stackoverflow.com", ] start_time = time.time() results = await crawler.arun_many(urls=urls, bypass_cache=True) end_time = time.time() total_time = end_time - start_time print(f"Total time for concurrent crawling: {total_time:.2f} seconds") assert all(result.success for result in results) assert len(results) == len(urls) assert ( total_time < len(urls) * 5 ), f"Concurrent crawling not significantly faster: {total_time:.2f} seconds" @pytest.mark.asyncio async def test_crawl_speed_with_caching(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.nbcnews.com/business" start_time = time.time() result1 = await crawler.arun(url=url, bypass_cache=True) end_time = time.time() first_crawl_time = end_time - start_time start_time = time.time() result2 = await crawler.arun(url=url, bypass_cache=False) end_time = time.time() second_crawl_time = end_time - start_time assert result1.success and result2.success print(f"First crawl time: {first_crawl_time:.2f} seconds") print(f"Second crawl time (cached): {second_crawl_time:.2f} seconds") assert ( second_crawl_time < first_crawl_time / 2 ), "Cached crawl not significantly faster" if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_0.4.2_browser_manager.py
tests/async/test_0.4.2_browser_manager.py
import os import sys import asyncio from crawl4ai import AsyncWebCrawler, CacheMode from crawl4ai.markdown_generation_strategy import DefaultMarkdownGenerator parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) # Assuming that the changes made allow different configurations # for managed browser, persistent context, and so forth. async def test_default_headless(): async with AsyncWebCrawler( headless=True, verbose=True, user_agent_mode="random", user_agent_generator_config={"device_type": "mobile", "os_type": "android"}, use_managed_browser=False, use_persistent_context=False, ignore_https_errors=True, # Testing normal ephemeral context ) as crawler: result = await crawler.arun( url="https://www.kidocode.com/degrees/technology", cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_default_headless] success:", result.success) print("HTML length:", len(result.html if result.html else "")) async def test_managed_browser_persistent(): # Treating use_persistent_context=True as managed_browser scenario. async with AsyncWebCrawler( headless=False, verbose=True, user_agent_mode="random", user_agent_generator_config={"device_type": "desktop", "os_type": "mac"}, use_managed_browser=True, use_persistent_context=True, # now should behave same as managed browser user_data_dir="./outpu/test_profile", # This should store and reuse profile data across runs ) as crawler: result = await crawler.arun( url="https://www.google.com", cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_managed_browser_persistent] success:", result.success) print("HTML length:", len(result.html if result.html else "")) async def test_session_reuse(): # Test creating a session, using it for multiple calls session_id = "my_session" async with AsyncWebCrawler( headless=False, verbose=True, user_agent="Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36", # Fixed user-agent for consistency use_managed_browser=False, use_persistent_context=False, ) as crawler: # First call: create session result1 = await crawler.arun( url="https://www.example.com", cache_mode=CacheMode.BYPASS, session_id=session_id, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_session_reuse first call] success:", result1.success) # Second call: same session, possibly cookie retained result2 = await crawler.arun( url="https://www.example.com/about", cache_mode=CacheMode.BYPASS, session_id=session_id, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_session_reuse second call] success:", result2.success) async def test_magic_mode(): # Test magic mode with override_navigator and simulate_user async with AsyncWebCrawler( headless=False, verbose=True, user_agent_mode="random", user_agent_generator_config={"device_type": "desktop", "os_type": "windows"}, use_managed_browser=False, use_persistent_context=False, magic=True, override_navigator=True, simulate_user=True, ) as crawler: result = await crawler.arun( url="https://www.kidocode.com/degrees/business", cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_magic_mode] success:", result.success) print("HTML length:", len(result.html if result.html else "")) async def test_proxy_settings(): # Test with a proxy (if available) to ensure code runs with proxy async with AsyncWebCrawler( headless=True, verbose=False, user_agent="Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36", proxy_config={"server": "http://127.0.0.1:8080"}, # Assuming local proxy server for test use_managed_browser=False, use_persistent_context=False, ) as crawler: result = await crawler.arun( url="https://httpbin.org/ip", cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_proxy_settings] success:", result.success) if result.success: print("HTML preview:", result.html[:200] if result.html else "") async def test_ignore_https_errors(): # Test ignore HTTPS errors with a self-signed or invalid cert domain # This is just conceptual, the domain should be one that triggers SSL error. # Using a hypothetical URL that fails SSL: async with AsyncWebCrawler( headless=True, verbose=True, user_agent="Mozilla/5.0", ignore_https_errors=True, use_managed_browser=False, use_persistent_context=False, ) as crawler: result = await crawler.arun( url="https://self-signed.badssl.com/", cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator(options={"ignore_links": True}), ) print("[test_ignore_https_errors] success:", result.success) async def main(): print("Running tests...") # await test_default_headless() # await test_managed_browser_persistent() # await test_session_reuse() # await test_magic_mode() # await test_proxy_settings() await test_ignore_https_errors() if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_database_operations.py
tests/async/test_database_operations.py
import os import sys import pytest # Add the parent directory to the Python path parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) from crawl4ai.async_webcrawler import AsyncWebCrawler @pytest.mark.asyncio async def test_cache_url(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.example.com" # First run to cache the URL result1 = await crawler.arun(url=url, bypass_cache=True) assert result1.success # Second run to retrieve from cache result2 = await crawler.arun(url=url, bypass_cache=False) assert result2.success assert result2.html == result1.html @pytest.mark.asyncio async def test_bypass_cache(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.python.org" # First run to cache the URL result1 = await crawler.arun(url=url, bypass_cache=True) assert result1.success # Second run bypassing cache result2 = await crawler.arun(url=url, bypass_cache=True) assert result2.success assert ( result2.html != result1.html ) # Content might be different due to dynamic nature of websites @pytest.mark.asyncio async def test_cache_size(): async with AsyncWebCrawler(verbose=True) as crawler: initial_size = await crawler.aget_cache_size() url = "https://www.nbcnews.com/business" await crawler.arun(url=url, bypass_cache=True) new_size = await crawler.aget_cache_size() assert new_size == initial_size + 1 @pytest.mark.asyncio async def test_clear_cache(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.example.org" await crawler.arun(url=url, bypass_cache=True) initial_size = await crawler.aget_cache_size() assert initial_size > 0 await crawler.aclear_cache() new_size = await crawler.aget_cache_size() assert new_size == 0 @pytest.mark.asyncio async def test_flush_cache(): async with AsyncWebCrawler(verbose=True) as crawler: url = "https://www.example.net" await crawler.arun(url=url, bypass_cache=True) initial_size = await crawler.aget_cache_size() assert initial_size > 0 await crawler.aflush_cache() new_size = await crawler.aget_cache_size() assert new_size == 0 # Try to retrieve the previously cached URL result = await crawler.arun(url=url, bypass_cache=False) assert ( result.success ) # The crawler should still succeed, but it will fetch the content anew # Entry point for debugging if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_redirect_url_resolution.py
tests/async/test_redirect_url_resolution.py
"""Test delayed redirect WITH wait_for - does link resolution use correct URL?""" import asyncio import threading from http.server import HTTPServer, SimpleHTTPRequestHandler class RedirectTestHandler(SimpleHTTPRequestHandler): def log_message(self, format, *args): pass def do_GET(self): if self.path == "/page-a": self.send_response(200) self.send_header("Content-type", "text/html") self.end_headers() content = """ <!DOCTYPE html> <html> <head><title>Page A</title></head> <body> <h1>Page A - Will redirect after 200ms</h1> <script> setTimeout(function() { window.location.href = '/redirect-target/'; }, 200); </script> </body> </html> """ self.wfile.write(content.encode()) elif self.path.startswith("/redirect-target"): self.send_response(200) self.send_header("Content-type", "text/html") self.end_headers() content = """ <!DOCTYPE html> <html> <head><title>Redirect Target</title></head> <body> <h1>Redirect Target</h1> <nav id="target-nav"> <a href="subpage-1">Subpage 1</a> <a href="subpage-2">Subpage 2</a> </nav> </body> </html> """ self.wfile.write(content.encode()) else: self.send_response(404) self.end_headers() async def main(): import socket class ReuseAddrHTTPServer(HTTPServer): allow_reuse_address = True server = ReuseAddrHTTPServer(("localhost", 8769), RedirectTestHandler) thread = threading.Thread(target=server.serve_forever) thread.daemon = True thread.start() try: import sys sys.path.insert(0, '/Users/nasrin/vscode/c4ai-uc/develop') from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig print("=" * 60) print("TEST: Delayed JS redirect WITH wait_for='css:#target-nav'") print("This waits for the redirect to complete") print("=" * 60) browser_config = BrowserConfig(headless=True, verbose=False) crawl_config = CrawlerRunConfig( cache_mode="bypass", wait_for="css:#target-nav" # Wait for element on redirect target ) async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun( url="http://localhost:8769/page-a", config=crawl_config ) print(f"Original URL: http://localhost:8769/page-a") print(f"Redirected URL returned: {result.redirected_url}") print(f"HTML contains 'Redirect Target': {'Redirect Target' in result.html}") print() if "/redirect-target" in (result.redirected_url or ""): print("✓ redirected_url is CORRECT") else: print("✗ BUG #1: redirected_url is WRONG - still shows original URL!") # Check links all_links = [] if isinstance(result.links, dict): all_links = result.links.get("internal", []) + result.links.get("external", []) print(f"\nLinks found ({len(all_links)} total):") bug_found = False for link in all_links: href = link.get("href", "") if isinstance(link, dict) else getattr(link, 'href', "") if "subpage" in href: print(f" {href}") if "/page-a/" in href: print(" ^^^ BUG #2: Link resolved with WRONG base URL!") bug_found = True elif "/redirect-target/" in href: print(" ^^^ CORRECT") if not bug_found and all_links: print("\n✓ Link resolution is CORRECT") finally: server.shutdown() if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/async/test_0.4.2_config_params.py
tests/async/test_0.4.2_config_params.py
import os, sys parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) import asyncio from crawl4ai import AsyncWebCrawler, CacheMode from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.content_filter_strategy import PruningContentFilter from crawl4ai import JsonCssExtractionStrategy from crawl4ai.chunking_strategy import RegexChunking # Category 1: Browser Configuration Tests async def test_browser_config_object(): """Test the new BrowserConfig object with various browser settings""" browser_config = BrowserConfig( browser_type="chromium", headless=False, viewport_width=1920, viewport_height=1080, use_managed_browser=True, user_agent_mode="random", user_agent_generator_config={"device_type": "desktop", "os_type": "windows"}, ) async with AsyncWebCrawler(config=browser_config, verbose=True) as crawler: result = await crawler.arun("https://example.com", cache_mode=CacheMode.BYPASS) assert result.success, "Browser config crawl failed" assert len(result.html) > 0, "No HTML content retrieved" async def test_browser_performance_config(): """Test browser configurations focused on performance""" browser_config = BrowserConfig( text_mode=True, light_mode=True, extra_args=["--disable-gpu", "--disable-software-rasterizer"], ignore_https_errors=True, java_script_enabled=False, ) async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun("https://example.com") assert result.success, "Performance optimized crawl failed" assert result.status_code == 200, "Unexpected status code" # Category 2: Content Processing Tests async def test_content_extraction_config(): """Test content extraction with various strategies""" crawler_config = CrawlerRunConfig( word_count_threshold=300, extraction_strategy=JsonCssExtractionStrategy( schema={ "name": "article", "baseSelector": "div", "fields": [{"name": "title", "selector": "h1", "type": "text"}], } ), chunking_strategy=RegexChunking(), content_filter=PruningContentFilter(), ) async with AsyncWebCrawler() as crawler: result = await crawler.arun( "https://example.com/article", config=crawler_config ) assert result.extracted_content is not None, "Content extraction failed" assert "title" in result.extracted_content, "Missing expected content field" # Category 3: Cache and Session Management Tests async def test_cache_and_session_management(): """Test different cache modes and session handling""" browser_config = BrowserConfig(use_persistent_context=True) crawler_config = CrawlerRunConfig( cache_mode=CacheMode.WRITE_ONLY, process_iframes=True, remove_overlay_elements=True, ) async with AsyncWebCrawler(config=browser_config) as crawler: # First request - should write to cache result1 = await crawler.arun("https://example.com", config=crawler_config) # Second request - should use fresh fetch due to WRITE_ONLY mode result2 = await crawler.arun("https://example.com", config=crawler_config) assert result1.success and result2.success, "Cache mode crawl failed" assert result1.html == result2.html, "Inconsistent results between requests" # Category 4: Media Handling Tests async def test_media_handling_config(): """Test configurations related to media handling""" # Get the base path for home directroy ~/.crawl4ai/downloads, make sure it exists os.makedirs(os.path.expanduser("~/.crawl4ai/downloads"), exist_ok=True) browser_config = BrowserConfig( viewport_width=1920, viewport_height=1080, accept_downloads=True, downloads_path=os.path.expanduser("~/.crawl4ai/downloads"), ) crawler_config = CrawlerRunConfig( screenshot=True, pdf=True, adjust_viewport_to_content=True, wait_for_images=True, screenshot_height_threshold=20000, ) async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun("https://example.com", config=crawler_config) assert result.screenshot is not None, "Screenshot capture failed" assert result.pdf is not None, "PDF generation failed" # Category 5: Anti-Bot and Site Interaction Tests async def test_antibot_config(): """Test configurations for handling anti-bot measures""" crawler_config = CrawlerRunConfig( simulate_user=True, override_navigator=True, magic=True, wait_for="js:()=>document.querySelector('body')", delay_before_return_html=1.0, log_console=True, cache_mode=CacheMode.BYPASS, ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://example.com", config=crawler_config) assert result.success, "Anti-bot measure handling failed" # Category 6: Parallel Processing Tests async def test_parallel_processing(): """Test parallel processing capabilities""" crawler_config = CrawlerRunConfig(mean_delay=0.5, max_range=1.0, semaphore_count=5) urls = ["https://example.com/1", "https://example.com/2", "https://example.com/3"] async with AsyncWebCrawler() as crawler: results = await crawler.arun_many(urls, config=crawler_config) assert len(results) == len(urls), "Not all URLs were processed" assert all(r.success for r in results), "Some parallel requests failed" # Category 7: Backwards Compatibility Tests async def test_legacy_parameter_support(): """Test that legacy parameters still work""" async with AsyncWebCrawler( headless=True, browser_type="chromium", viewport_width=1024, viewport_height=768 ) as crawler: result = await crawler.arun( "https://example.com", screenshot=True, word_count_threshold=200, bypass_cache=True, css_selector=".main-content", ) assert result.success, "Legacy parameter support failed" # Category 8: Mixed Configuration Tests async def test_mixed_config_usage(): """Test mixing new config objects with legacy parameters""" browser_config = BrowserConfig(headless=True) crawler_config = CrawlerRunConfig(screenshot=True) async with AsyncWebCrawler( config=browser_config, verbose=True, # legacy parameter ) as crawler: result = await crawler.arun( "https://example.com", config=crawler_config, cache_mode=CacheMode.BYPASS, # legacy parameter css_selector="body", # legacy parameter ) assert result.success, "Mixed configuration usage failed" if __name__ == "__main__": async def run_tests(): test_functions = [ test_browser_config_object, # test_browser_performance_config, # test_content_extraction_config, # test_cache_and_session_management, # test_media_handling_config, # test_antibot_config, # test_parallel_processing, # test_legacy_parameter_support, # test_mixed_config_usage ] for test in test_functions: print(f"\nRunning {test.__name__}...") try: await test() print(f"✓ {test.__name__} passed") except AssertionError as e: print(f"✗ {test.__name__} failed: {str(e)}") except Exception as e: print(f"✗ {test.__name__} error: {str(e)}") asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_url_pattern.py
tests/general/test_url_pattern.py
import sys import os # Get the grandparent directory grandparent_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(grandparent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) import asyncio from crawl4ai.deep_crawling.filters import URLPatternFilter def test_prefix_boundary_matching(): """Test that prefix patterns respect path boundaries""" print("=== Testing URLPatternFilter Prefix Boundary Fix ===") filter_obj = URLPatternFilter(patterns=['https://langchain-ai.github.io/langgraph/*']) test_cases = [ ('https://langchain-ai.github.io/langgraph/', True), ('https://langchain-ai.github.io/langgraph/concepts/', True), ('https://langchain-ai.github.io/langgraph/tutorials/', True), ('https://langchain-ai.github.io/langgraph?param=1', True), ('https://langchain-ai.github.io/langgraph#section', True), ('https://langchain-ai.github.io/langgraphjs/', False), ('https://langchain-ai.github.io/langgraphjs/concepts/', False), ('https://other-site.com/langgraph/', False), ] all_passed = True for url, expected in test_cases: result = filter_obj.apply(url) status = "PASS" if result == expected else "FAIL" if result != expected: all_passed = False print(f"{status:4} | Expected: {expected:5} | Got: {result:5} | {url}") return all_passed def test_edge_cases(): """Test edge cases for path boundary matching""" print("\n=== Testing Edge Cases ===") test_patterns = [ ('/api/*', [ ('/api/', True), ('/api/v1', True), ('/api?param=1', True), ('/apiv2/', False), ('/api_old/', False), ]), ('*/docs/*', [ ('example.com/docs/', True), ('example.com/docs/guide', True), ('example.com/documentation/', False), ('example.com/docs_old/', False), ]), ] all_passed = True for pattern, test_cases in test_patterns: print(f"\nPattern: {pattern}") filter_obj = URLPatternFilter(patterns=[pattern]) for url, expected in test_cases: result = filter_obj.apply(url) status = "PASS" if result == expected else "FAIL" if result != expected: all_passed = False print(f" {status:4} | Expected: {expected:5} | Got: {result:5} | {url}") return all_passed if __name__ == "__main__": test1_passed = test_prefix_boundary_matching() test2_passed = test_edge_cases() if test1_passed and test2_passed: print("\n✅ All tests passed!") sys.exit(0) else: print("\n❌ Some tests failed!") sys.exit(1)
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_async_markdown_generator.py
tests/general/test_async_markdown_generator.py
import asyncio from typing import Dict from crawl4ai.content_filter_strategy import BM25ContentFilter, PruningContentFilter from crawl4ai.markdown_generation_strategy import DefaultMarkdownGenerator import time # Test HTML samples TEST_HTML_SAMPLES = { "basic": """ <body> <h1>Test Title</h1> <p>This is a test paragraph with <a href="http://example.com">a link</a>.</p> <div class="content"> <h2>Section 1</h2> <p>More content here with <b>bold text</b>.</p> </div> </body> """, "complex": """ <body> <nav>Navigation menu that should be removed</nav> <header>Header content to remove</header> <main> <article> <h1>Main Article</h1> <p>Important content paragraph with <a href="http://test.com">useful link</a>.</p> <section> <h2>Key Section</h2> <p>Detailed explanation with multiple sentences. This should be kept in the final output. Very important information here.</p> </section> </article> <aside>Sidebar content to remove</aside> </main> <footer>Footer content to remove</footer> </body> """, "edge_cases": """ <body> <div> <p></p> <p> </p> <script>alert('remove me');</script> <div class="advertisement">Ad content to remove</div> <p class="social-share">Share buttons to remove</p> <h1>!!Special>> Characters## Title!!</h1> <pre><code>def test(): pass</code></pre> </div> </body> """, "links_citations": """ <body> <h1>Document with Links</h1> <p>First link to <a href="http://example.com/1">Example 1</a></p> <p>Second link to <a href="http://example.com/2" title="Example 2">Test 2</a></p> <p>Image link: <img src="test.jpg" alt="test image"></p> <p>Repeated link to <a href="http://example.com/1">Example 1 again</a></p> </body> """, } def test_content_filters() -> Dict[str, Dict[str, int]]: """Test various content filtering strategies and return length comparisons.""" results = {} # Initialize filters pruning_filter = PruningContentFilter( threshold=0.48, threshold_type="fixed", min_word_threshold=2 ) bm25_filter = BM25ContentFilter( bm25_threshold=1.0, user_query="test article content important" ) # Test each HTML sample for test_name, html in TEST_HTML_SAMPLES.items(): # Store results for this test case results[test_name] = {} # Test PruningContentFilter start_time = time.time() pruned_content = pruning_filter.filter_content(html) pruning_time = time.time() - start_time # Test BM25ContentFilter start_time = time.time() bm25_content = bm25_filter.filter_content(html) bm25_time = time.time() - start_time # Store results results[test_name] = { "original_length": len(html), "pruned_length": sum(len(c) for c in pruned_content), "bm25_length": sum(len(c) for c in bm25_content), "pruning_time": pruning_time, "bm25_time": bm25_time } return results def test_markdown_generation(): """Test markdown generation with different configurations.""" results = [] # Initialize generators with different configurations generators = { "no_filter": DefaultMarkdownGenerator(), "pruning": DefaultMarkdownGenerator( content_filter=PruningContentFilter(threshold=0.48) ), "bm25": DefaultMarkdownGenerator( content_filter=BM25ContentFilter( user_query="test article content important" ) ) } # Test each generator with each HTML sample for test_name, html in TEST_HTML_SAMPLES.items(): for gen_name, generator in generators.items(): start_time = time.time() result = generator.generate_markdown( html, base_url="http://example.com", citations=True ) results.append({ "test_case": test_name, "generator": gen_name, "time": time.time() - start_time, "raw_length": len(result.raw_markdown), "fit_length": len(result.fit_markdown) if result.fit_markdown else 0, "citations": len(result.references_markdown) }) return results def main(): """Run all tests and print results.""" print("Starting content filter tests...") filter_results = test_content_filters() print("\nContent Filter Results:") print("-" * 50) for test_name, metrics in filter_results.items(): print(f"\nTest case: {test_name}") print(f"Original length: {metrics['original_length']}") print(f"Pruned length: {metrics['pruned_length']} ({metrics['pruning_time']:.3f}s)") print(f"BM25 length: {metrics['bm25_length']} ({metrics['bm25_time']:.3f}s)") print("\nStarting markdown generation tests...") markdown_results = test_markdown_generation() print("\nMarkdown Generation Results:") print("-" * 50) for result in markdown_results: print(f"\nTest: {result['test_case']} - Generator: {result['generator']}") print(f"Time: {result['time']:.3f}s") print(f"Raw length: {result['raw_length']}") print(f"Fit length: {result['fit_length']}") print(f"Citations: {result['citations']}") if __name__ == "__main__": main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_crawlers.py
tests/general/test_crawlers.py
# example_usageexample_usageexample_usage# example_usage.py import asyncio from crawl4ai.crawlers import get_crawler async def main(): # Get the registered crawler example_crawler = get_crawler("example_site.content") # Crawl example.com result = await example_crawler(url="https://example.com") print(result) if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_deep_crawl_scorers.py
tests/general/test_deep_crawl_scorers.py
from crawl4ai.deep_crawling.scorers import CompositeScorer, ContentTypeScorer, DomainAuthorityScorer, FreshnessScorer, KeywordRelevanceScorer, PathDepthScorer def test_scorers(): test_cases = [ # Keyword Scorer Tests { "scorer_type": "keyword", "config": { "keywords": ["python", "blog"], "weight": 1.0, "case_sensitive": False }, "urls": { "https://example.com/python-blog": 1.0, "https://example.com/PYTHON-BLOG": 1.0, "https://example.com/python-only": 0.5, "https://example.com/other": 0.0 } }, # Path Depth Scorer Tests { "scorer_type": "path_depth", "config": { "optimal_depth": 2, "weight": 1.0 }, "urls": { "https://example.com/a/b": 1.0, "https://example.com/a": 0.5, "https://example.com/a/b/c": 0.5, "https://example.com": 0.33333333 } }, # Content Type Scorer Tests { "scorer_type": "content_type", "config": { "type_weights": { ".html$": 1.0, ".pdf$": 0.8, ".jpg$": 0.6 }, "weight": 1.0 }, "urls": { "https://example.com/doc.html": 1.0, "https://example.com/doc.pdf": 0.8, "https://example.com/img.jpg": 0.6, "https://example.com/other.txt": 0.0 } }, # Freshness Scorer Tests { "scorer_type": "freshness", "config": { "weight": 1.0, # Remove current_year since original doesn't support it }, "urls": { "https://example.com/2024/01/post": 1.0, "https://example.com/2023/12/post": 0.9, "https://example.com/2022/post": 0.8, "https://example.com/no-date": 0.5 } }, # Domain Authority Scorer Tests { "scorer_type": "domain", "config": { "domain_weights": { "python.org": 1.0, "github.com": 0.8, "medium.com": 0.6 }, "default_weight": 0.3, "weight": 1.0 }, "urls": { "https://python.org/about": 1.0, "https://github.com/repo": 0.8, "https://medium.com/post": 0.6, "https://unknown.com": 0.3 } } ] def create_scorer(scorer_type, config): if scorer_type == "keyword": return KeywordRelevanceScorer(**config) elif scorer_type == "path_depth": return PathDepthScorer(**config) elif scorer_type == "content_type": return ContentTypeScorer(**config) elif scorer_type == "freshness": return FreshnessScorer(**config,current_year=2024) elif scorer_type == "domain": return DomainAuthorityScorer(**config) def run_accuracy_test(): print("\nAccuracy Tests:") print("-" * 50) all_passed = True for test_case in test_cases: print(f"\nTesting {test_case['scorer_type']} scorer:") scorer = create_scorer( test_case['scorer_type'], test_case['config'] ) for url, expected in test_case['urls'].items(): score = round(scorer.score(url), 8) expected = round(expected, 8) if abs(score - expected) > 0.00001: print(f"❌ Scorer Failed: URL '{url}'") print(f" Expected: {expected}, Got: {score}") all_passed = False else: print(f"✅ Scorer Passed: URL '{url}'") return all_passed def run_composite_test(): print("\nTesting Composite Scorer:") print("-" * 50) # Create test data test_urls = { "https://python.org/blog/2024/01/new-release.html":0.86666667, "https://github.com/repo/old-code.pdf": 0.62, "https://unknown.com/random": 0.26 } # Create composite scorers with all types scorers = [] for test_case in test_cases: scorer = create_scorer( test_case['scorer_type'], test_case['config'] ) scorers.append(scorer) composite = CompositeScorer(scorers, normalize=True) all_passed = True for url, expected in test_urls.items(): score = round(composite.score(url), 8) if abs(score - expected) > 0.00001: print(f"❌ Composite Failed: URL '{url}'") print(f" Expected: {expected}, Got: {score}") all_passed = False else: print(f"✅ Composite Passed: URL '{url}'") return all_passed # Run tests print("Running Scorer Tests...") accuracy_passed = run_accuracy_test() composite_passed = run_composite_test() if accuracy_passed and composite_passed: print("\n✨ All tests passed!") # Note: Already have performance tests in run_scorer_performance_test() else: print("\n❌ Some tests failed!") if __name__ == "__main__": test_scorers()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/tets_robot.py
tests/general/tets_robot.py
import asyncio from crawl4ai import * async def test_real_websites(): print("\n=== Testing Real Website Robots.txt Compliance ===\n") browser_config = BrowserConfig(headless=True, verbose=True) async with AsyncWebCrawler(config=browser_config) as crawler: # Test cases with URLs test_cases = [ # Public sites that should be allowed ("https://example.com", True), # Simple public site ("https://httpbin.org/get", True), # API endpoint # Sites with known strict robots.txt ("https://www.facebook.com/robots.txt", False), # Social media ("https://www.google.com/search", False), # Search pages # Edge cases ("https://api.github.com", True), # API service ("https://raw.githubusercontent.com", True), # Content delivery # Non-existent/error cases ("https://thisisnotarealwebsite.com", True), # Non-existent domain ("https://localhost:12345", True), # Invalid port ] for url, expected in test_cases: print(f"\nTesting: {url}") try: config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, check_robots_txt=True, # Enable robots.txt checking verbose=True ) result = await crawler.arun(url=url, config=config) allowed = result.success and not result.error_message print(f"Expected: {'allowed' if expected else 'denied'}") print(f"Actual: {'allowed' if allowed else 'denied'}") print(f"Status Code: {result.status_code}") if result.error_message: print(f"Error: {result.error_message}") # Optional: Print robots.txt content if available if result.metadata and 'robots_txt' in result.metadata: print(f"Robots.txt rules:\n{result.metadata['robots_txt']}") except Exception as e: print(f"Test failed with error: {str(e)}") async def main(): try: await test_real_websites() except Exception as e: print(f"Test suite failed: {str(e)}") raise if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_content_source_parameter.py
tests/general/test_content_source_parameter.py
""" Tests for the content_source parameter in markdown generation. """ import unittest import asyncio from unittest.mock import patch, MagicMock from crawl4ai.markdown_generation_strategy import DefaultMarkdownGenerator, MarkdownGenerationStrategy from crawl4ai.async_webcrawler import AsyncWebCrawler from crawl4ai.async_configs import CrawlerRunConfig from crawl4ai.models import MarkdownGenerationResult HTML_SAMPLE = """ <html> <head><title>Test Page</title></head> <body> <h1>Test Content</h1> <p>This is a test paragraph.</p> <div class="container"> <p>This is content within a container.</p> </div> </body> </html> """ class TestContentSourceParameter(unittest.TestCase): """Test cases for the content_source parameter in markdown generation.""" def setUp(self): """Set up test fixtures.""" self.loop = asyncio.new_event_loop() asyncio.set_event_loop(self.loop) def tearDown(self): """Tear down test fixtures.""" self.loop.close() def test_default_content_source(self): """Test that the default content_source is 'cleaned_html'.""" # Can't directly instantiate abstract class, so just test DefaultMarkdownGenerator generator = DefaultMarkdownGenerator() self.assertEqual(generator.content_source, "cleaned_html") def test_custom_content_source(self): """Test that content_source can be customized.""" generator = DefaultMarkdownGenerator(content_source="fit_html") self.assertEqual(generator.content_source, "fit_html") @patch('crawl4ai.markdown_generation_strategy.CustomHTML2Text') def test_html_processing_using_input_html(self, mock_html2text): """Test that generate_markdown uses input_html parameter.""" # Setup mock mock_instance = MagicMock() mock_instance.handle.return_value = "# Test Content\n\nThis is a test paragraph." mock_html2text.return_value = mock_instance # Create generator and call generate_markdown generator = DefaultMarkdownGenerator() result = generator.generate_markdown(input_html="<h1>Test Content</h1><p>This is a test paragraph.</p>") # Verify input_html was passed to HTML2Text handler mock_instance.handle.assert_called_once() # Get the first positional argument args, _ = mock_instance.handle.call_args self.assertEqual(args[0], "<h1>Test Content</h1><p>This is a test paragraph.</p>") # Check result self.assertIsInstance(result, MarkdownGenerationResult) self.assertEqual(result.raw_markdown, "# Test Content\n\nThis is a test paragraph.") def test_html_source_selection_logic(self): """Test that the HTML source selection logic works correctly.""" # We'll test the dispatch pattern directly to avoid async complexities # Create test data raw_html = "<html><body><h1>Raw HTML</h1></body></html>" cleaned_html = "<html><body><h1>Cleaned HTML</h1></body></html>" fit_html = "<html><body><h1>Preprocessed HTML</h1></body></html>" # Test the dispatch pattern html_source_selector = { "raw_html": lambda: raw_html, "cleaned_html": lambda: cleaned_html, "fit_html": lambda: fit_html, } # Test Case 1: content_source="cleaned_html" source_lambda = html_source_selector.get("cleaned_html") self.assertEqual(source_lambda(), cleaned_html) # Test Case 2: content_source="raw_html" source_lambda = html_source_selector.get("raw_html") self.assertEqual(source_lambda(), raw_html) # Test Case 3: content_source="fit_html" source_lambda = html_source_selector.get("fit_html") self.assertEqual(source_lambda(), fit_html) # Test Case 4: Invalid content_source falls back to cleaned_html source_lambda = html_source_selector.get("invalid_source", lambda: cleaned_html) self.assertEqual(source_lambda(), cleaned_html) if __name__ == '__main__': unittest.main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_advanced_deep_crawl.py
tests/general/test_advanced_deep_crawl.py
import asyncio import time from crawl4ai import CrawlerRunConfig, AsyncWebCrawler, CacheMode from crawl4ai.content_scraping_strategy import LXMLWebScrapingStrategy from crawl4ai.deep_crawling import BFSDeepCrawlStrategy, BestFirstCrawlingStrategy from crawl4ai.deep_crawling.filters import FilterChain, URLPatternFilter, DomainFilter, ContentTypeFilter, ContentRelevanceFilter from crawl4ai.deep_crawling.scorers import KeywordRelevanceScorer # from crawl4ai.deep_crawling import BFSDeepCrawlStrategy, BestFirstCrawlingStrategy async def main(): """Example deep crawl of documentation site.""" filter_chain = FilterChain([ URLPatternFilter(patterns=["*2025*"]), DomainFilter(allowed_domains=["techcrunch.com"]), ContentRelevanceFilter(query="Use of artificial intelligence in Defence applications", threshold=1), ContentTypeFilter(allowed_types=["text/html","application/javascript"]) ]) config = CrawlerRunConfig( deep_crawl_strategy = BestFirstCrawlingStrategy( max_depth=2, include_external=False, filter_chain=filter_chain, url_scorer=KeywordRelevanceScorer(keywords=["anduril", "defence", "AI"]), ), stream=False, verbose=True, cache_mode=CacheMode.BYPASS, scraping_strategy=LXMLWebScrapingStrategy() ) async with AsyncWebCrawler() as crawler: print("Starting deep crawl in streaming mode:") config.stream = True start_time = time.perf_counter() async for result in await crawler.arun( url="https://techcrunch.com", config=config ): print(f"→ {result.url} (Depth: {result.metadata.get('depth', 0)})") print(f"Duration: {time.perf_counter() - start_time:.2f} seconds") if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_schema_builder.py
tests/general/test_schema_builder.py
# https://claude.ai/chat/c4bbe93d-fb54-44ce-92af-76b4c8086c6b # https://claude.ai/chat/c24a768c-d8b2-478a-acc7-d76d42a308da import os, sys parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) import asyncio from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, CacheMode from crawl4ai.markdown_generation_strategy import DefaultMarkdownGenerator from crawl4ai import JsonCssExtractionStrategy, JsonXPathExtractionStrategy from crawl4ai.utils import preprocess_html_for_schema, JsonXPathExtractionStrategy import json # Test HTML - A complex job board with companies, departments, and positions test_html = """ <div class="company-listings"> <div class="company" data-company-id="123"> <div class="company-header"> <img class="company-logo" src="google.png" alt="Google"> <h1 class="company-name">Google</h1> <div class="company-meta"> <span class="company-size">10,000+ employees</span> <span class="company-industry">Technology</span> <a href="https://google.careers" class="careers-link">Careers Page</a> </div> </div> <div class="departments"> <div class="department"> <h2 class="department-name">Engineering</h2> <div class="positions"> <div class="position-card" data-position-id="eng-1"> <h3 class="position-title">Senior Software Engineer</h3> <span class="salary-range">$150,000 - $250,000</span> <div class="position-meta"> <span class="location">Mountain View, CA</span> <span class="job-type">Full-time</span> <span class="experience">5+ years</span> </div> <div class="skills-required"> <span class="skill">Python</span> <span class="skill">Kubernetes</span> <span class="skill">Machine Learning</span> </div> <p class="position-description">Join our core engineering team...</p> <div class="application-info"> <span class="posting-date">Posted: 2024-03-15</span> <button class="apply-btn" data-req-id="REQ12345">Apply Now</button> </div> </div> <!-- More positions --> </div> </div> <div class="department"> <h2 class="department-name">Marketing</h2> <div class="positions"> <div class="position-card" data-position-id="mkt-1"> <h3 class="position-title">Growth Marketing Manager</h3> <span class="salary-range">$120,000 - $180,000</span> <div class="position-meta"> <span class="location">New York, NY</span> <span class="job-type">Full-time</span> <span class="experience">3+ years</span> </div> <div class="skills-required"> <span class="skill">SEO</span> <span class="skill">Analytics</span> <span class="skill">Content Strategy</span> </div> <p class="position-description">Drive our growth initiatives...</p> <div class="application-info"> <span class="posting-date">Posted: 2024-03-14</span> <button class="apply-btn" data-req-id="REQ12346">Apply Now</button> </div> </div> </div> </div> </div> </div> </div> """ # Test cases def test_schema_generation(): # Test 1: No query (should extract everything) print("\nTest 1: No Query (Full Schema)") schema1 = JsonCssExtractionStrategy.generate_schema(test_html) print(json.dumps(schema1, indent=2)) # Test 2: Query for just basic job info print("\nTest 2: Basic Job Info Query") query2 = "I only need job titles, salaries, and locations" schema2 = JsonCssExtractionStrategy.generate_schema(test_html, query2) print(json.dumps(schema2, indent=2)) # Test 3: Query for company and department structure print("\nTest 3: Organizational Structure Query") query3 = "Extract company details and department names, without position details" schema3 = JsonCssExtractionStrategy.generate_schema(test_html, query3) print(json.dumps(schema3, indent=2)) # Test 4: Query for specific skills tracking print("\nTest 4: Skills Analysis Query") query4 = "I want to analyze required skills across all positions" schema4 = JsonCssExtractionStrategy.generate_schema(test_html, query4) print(json.dumps(schema4, indent=2)) if __name__ == "__main__": test_schema_generation()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/generate_dummy_site.py
tests/general/generate_dummy_site.py
# ==== File: build_dummy_site.py ==== import os import random import argparse from pathlib import Path from urllib.parse import quote # --- Configuration --- NUM_CATEGORIES = 3 NUM_SUBCATEGORIES_PER_CAT = 2 # Results in NUM_CATEGORIES * NUM_SUBCATEGORIES_PER_CAT total L2 categories NUM_PRODUCTS_PER_SUBCAT = 5 # Products listed on L3 pages MAX_DEPTH_TARGET = 5 # Explicitly set target depth # --- Helper Functions --- def generate_lorem(words=20): """Generates simple placeholder text.""" lorem_words = ["lorem", "ipsum", "dolor", "sit", "amet", "consectetur", "adipiscing", "elit", "sed", "do", "eiusmod", "tempor", "incididunt", "ut", "labore", "et", "dolore", "magna", "aliqua"] return " ".join(random.choice(lorem_words) for _ in range(words)).capitalize() + "." def create_html_page(filepath: Path, title: str, body_content: str, breadcrumbs: list = [], head_extras: str = ""): """Creates an HTML file with basic structure and inline CSS.""" os.makedirs(filepath.parent, exist_ok=True) # Generate breadcrumb HTML using the 'link' provided in the breadcrumbs list breadcrumb_html = "" if breadcrumbs: links_html = " » ".join(f'<a href="{bc["link"]}">{bc["name"]}</a>' for bc in breadcrumbs) breadcrumb_html = f"<nav class='breadcrumbs'>{links_html} » {title}</nav>" # Basic CSS for structure identification (kept the same) css = """ <style> body { font-family: sans-serif; padding: 20px; background-color: #1e1e1e; color: #d1d1d1; } .container { max-width: 960px; margin: auto; background: #2c2c2c; padding: 20px; border-radius: 5px; box-shadow: 0 2px 5px rgba(0, 0, 0, 0.5); } h1, h2 { color: #ccc; } a { color: #9bcdff; text-decoration: none; } a:hover { text-decoration: underline; } ul { list-style: none; padding-left: 0; } li { margin-bottom: 10px; } .category-link, .subcategory-link, .product-link, .details-link, .reviews-link { display: block; padding: 8px; background-color: #3a3a3a; border-radius: 3px; } .product-preview { border: 1px solid #444; padding: 10px; margin-bottom: 10px; border-radius: 4px; background-color: #2a2a2a; } .product-title { color: #d1d1d1; } .product-price { font-weight: bold; color: #85e085; } .product-description, .product-specs, .product-reviews { margin-top: 15px; line-height: 1.6; } .product-specs li { margin-bottom: 5px; font-size: 0.9em; } .spec-name { font-weight: bold; } .breadcrumbs { margin-bottom: 20px; font-size: 0.9em; color: #888; } .breadcrumbs a { color: #9bcdff; } </style> """ html_content = f"""<!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <meta name="viewport" content="width=device-width, initial-scale=1.0"> <title>{title} - FakeShop</title> {head_extras} {css} </head> <body> <div class="container"> {breadcrumb_html} <h1>{title}</h1> {body_content} </div> </body> </html>""" with open(filepath, "w", encoding="utf-8") as f: f.write(html_content) # Keep print statement concise for clarity # print(f"Created: {filepath}") def generate_site(base_dir: Path, site_name: str = "FakeShop", base_path: str = ""): """Generates the dummy website structure.""" base_dir.mkdir(parents=True, exist_ok=True) # --- Clean and prepare the base path for URL construction --- # Ensure it starts with '/' if not empty, and remove any trailing '/' if base_path: full_base_path = "/" + base_path.strip('/') else: full_base_path = "" # Represents the root print(f"Using base path for links: '{full_base_path}'") # --- Level 0: Homepage --- home_body = "<h2>Welcome to FakeShop!</h2><p>Your one-stop shop for imaginary items.</p><h3>Categories:</h3>\n<ul>" # Define the *actual* link path for the homepage breadcrumb home_link_path = f"{full_base_path}/index.html" breadcrumbs_home = [{"name": "Home", "link": home_link_path}] # Base breadcrumb # Links *within* the page content should remain relative for i in range(NUM_CATEGORIES): cat_name = f"Category-{i+1}" cat_folder_name = quote(cat_name.lower().replace(" ", "-")) # This path is relative to the current directory (index.html) cat_relative_page_path = f"{cat_folder_name}/index.html" home_body += f'<li><a class="category-link" href="{cat_relative_page_path}">{cat_name}</a> - {generate_lorem(10)}</li>' home_body += "</ul>" create_html_page(base_dir / "index.html", "Homepage", home_body, []) # No breadcrumbs *on* the homepage itself # --- Levels 1-5 --- for i in range(NUM_CATEGORIES): cat_name = f"Category-{i+1}" cat_folder_name = quote(cat_name.lower().replace(" ", "-")) cat_dir = base_dir / cat_folder_name # This is the *absolute* path for the breadcrumb link cat_link_path = f"{full_base_path}/{cat_folder_name}/index.html" # Update breadcrumbs list for this level breadcrumbs_cat = breadcrumbs_home + [{"name": cat_name, "link": cat_link_path}] # --- Level 1: Category Page --- cat_body = f"<p>{generate_lorem(15)} for {cat_name}.</p><h3>Sub-Categories:</h3>\n<ul>" for j in range(NUM_SUBCATEGORIES_PER_CAT): subcat_name = f"{cat_name}-Sub-{j+1}" subcat_folder_name = quote(subcat_name.lower().replace(" ", "-")) # Path relative to the category page subcat_relative_page_path = f"{subcat_folder_name}/index.html" cat_body += f'<li><a class="subcategory-link" href="{subcat_relative_page_path}">{subcat_name}</a> - {generate_lorem(8)}</li>' cat_body += "</ul>" # Pass the updated breadcrumbs list create_html_page(cat_dir / "index.html", cat_name, cat_body, breadcrumbs_home) # Parent breadcrumb needed here for j in range(NUM_SUBCATEGORIES_PER_CAT): subcat_name = f"{cat_name}-Sub-{j+1}" subcat_folder_name = quote(subcat_name.lower().replace(" ", "-")) subcat_dir = cat_dir / subcat_folder_name # Absolute path for the breadcrumb link subcat_link_path = f"{full_base_path}/{cat_folder_name}/{subcat_folder_name}/index.html" # Update breadcrumbs list for this level breadcrumbs_subcat = breadcrumbs_cat + [{"name": subcat_name, "link": subcat_link_path}] # --- Level 2: Sub-Category Page (Product List) --- subcat_body = f"<p>Explore products in {subcat_name}. {generate_lorem(12)}</p><h3>Products:</h3>\n<ul class='product-list'>" for k in range(NUM_PRODUCTS_PER_SUBCAT): prod_id = f"P{i+1}{j+1}{k+1:03d}" # e.g., P11001 prod_name = f"{subcat_name} Product {k+1} ({prod_id})" # Filename relative to the subcategory page prod_filename = f"product_{prod_id}.html" # Absolute path for the breadcrumb link prod_link_path = f"{full_base_path}/{cat_folder_name}/{subcat_folder_name}/{prod_filename}" # Preview on list page (link remains relative) subcat_body += f""" <li> <div class="product-preview"> <a class="product-link" href="{prod_filename}"><strong>{prod_name}</strong></a> <p>{generate_lorem(10)}</p> <span class="product-price">£{random.uniform(10, 500):.2f}</span> </div> </li>""" # --- Level 3: Product Page --- prod_price = random.uniform(10, 500) prod_desc = generate_lorem(40) prod_specs = {f"Spec {s+1}": generate_lorem(3) for s in range(random.randint(3,6))} prod_reviews_count = random.randint(0, 150) # Relative filenames for links on this page details_filename_relative = f"product_{prod_id}_details.html" reviews_filename_relative = f"product_{prod_id}_reviews.html" prod_body = f""" <p class="product-price">Price: £{prod_price:.2f}</p> <div class="product-description"> <h2>Description</h2> <p>{prod_desc}</p> </div> <div class="product-specs"> <h2>Specifications</h2> <ul> {''.join(f'<li><span class="spec-name">{name}</span>: <span class="spec-value">{value}</span></li>' for name, value in prod_specs.items())} </ul> </div> <div class="product-reviews"> <h2>Reviews</h2> <p>Total Reviews: <span class="review-count">{prod_reviews_count}</span></p> </div> <hr> <p> <a class="details-link" href="{details_filename_relative}">View More Details</a> | <a class="reviews-link" href="{reviews_filename_relative}">See All Reviews</a> </p> """ # Update breadcrumbs list for this level breadcrumbs_prod = breadcrumbs_subcat + [{"name": prod_name, "link": prod_link_path}] # Pass the updated breadcrumbs list create_html_page(subcat_dir / prod_filename, prod_name, prod_body, breadcrumbs_subcat) # Parent breadcrumb needed here # --- Level 4: Product Details Page --- details_filename = f"product_{prod_id}_details.html" # Actual filename # Absolute path for the breadcrumb link details_link_path = f"{full_base_path}/{cat_folder_name}/{subcat_folder_name}/{details_filename}" details_body = f"<p>This page contains extremely detailed information about {prod_name}.</p>{generate_lorem(100)}" # Update breadcrumbs list for this level breadcrumbs_details = breadcrumbs_prod + [{"name": "Details", "link": details_link_path}] # Pass the updated breadcrumbs list create_html_page(subcat_dir / details_filename, f"{prod_name} - Details", details_body, breadcrumbs_prod) # Parent breadcrumb needed here # --- Level 5: Product Reviews Page --- reviews_filename = f"product_{prod_id}_reviews.html" # Actual filename # Absolute path for the breadcrumb link reviews_link_path = f"{full_base_path}/{cat_folder_name}/{subcat_folder_name}/{reviews_filename}" reviews_body = f"<p>All {prod_reviews_count} reviews for {prod_name} are listed here.</p><ul>" for r in range(prod_reviews_count): reviews_body += f"<li>Review {r+1}: {generate_lorem(random.randint(15, 50))}</li>" reviews_body += "</ul>" # Update breadcrumbs list for this level breadcrumbs_reviews = breadcrumbs_prod + [{"name": "Reviews", "link": reviews_link_path}] # Pass the updated breadcrumbs list create_html_page(subcat_dir / reviews_filename, f"{prod_name} - Reviews", reviews_body, breadcrumbs_prod) # Parent breadcrumb needed here subcat_body += "</ul>" # Close product-list ul # Pass the correct breadcrumbs list for the subcategory index page create_html_page(subcat_dir / "index.html", subcat_name, subcat_body, breadcrumbs_cat) # Parent breadcrumb needed here # --- Main Execution --- if __name__ == "__main__": parser = argparse.ArgumentParser(description="Generate a dummy multi-level retail website.") parser.add_argument( "-o", "--output-dir", type=str, default="dummy_retail_site", help="Directory to generate the website in." ) parser.add_argument( "-n", "--site-name", type=str, default="FakeShop", help="Name of the fake shop." ) parser.add_argument( "-b", "--base-path", type=str, default="", help="Base path for hosting the site (e.g., 'samples/deepcrawl'). Leave empty if hosted at the root." ) # Optional: Add more args to configure counts if needed args = parser.parse_args() output_directory = Path(args.output_dir) site_name = args.site_name base_path = args.base_path print(f"Generating dummy site '{site_name}' in '{output_directory}'...") # Pass the base_path to the generation function generate_site(output_directory, site_name, base_path) print(f"\nCreated {sum(1 for _ in output_directory.rglob('*.html'))} HTML pages.") print("Dummy site generation complete.") print(f"To serve locally (example): python -m http.server --directory {output_directory} 8000") if base_path: print(f"Access the site at: http://localhost:8000/{base_path.strip('/')}/index.html") else: print(f"Access the site at: http://localhost:8000/index.html")
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_deep_crawl.py
tests/general/test_deep_crawl.py
import asyncio import time from crawl4ai import CrawlerRunConfig, AsyncWebCrawler, CacheMode from crawl4ai.content_scraping_strategy import LXMLWebScrapingStrategy from crawl4ai.deep_crawling import BFSDeepCrawlStrategy # from crawl4ai.deep_crawling import BFSDeepCrawlStrategy, BestFirstCrawlingStrategy async def main(): """Example deep crawl of documentation site.""" config = CrawlerRunConfig( deep_crawl_strategy = BFSDeepCrawlStrategy( max_depth=2, include_external=False ), stream=False, verbose=True, cache_mode=CacheMode.BYPASS, scraping_strategy=LXMLWebScrapingStrategy() ) async with AsyncWebCrawler() as crawler: start_time = time.perf_counter() print("\nStarting deep crawl in batch mode:") results = await crawler.arun( url="https://docs.crawl4ai.com", config=config ) print(f"Crawled {len(results)} pages") print(f"Example page: {results[0].url}") print(f"Duration: {time.perf_counter() - start_time:.2f} seconds\n") print("Starting deep crawl in streaming mode:") config.stream = True start_time = time.perf_counter() async for result in await crawler.arun( url="https://docs.crawl4ai.com", config=config ): print(f"→ {result.url} (Depth: {result.metadata.get('depth', 0)})") print(f"Duration: {time.perf_counter() - start_time:.2f} seconds") if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_deep_crawl_filters.py
tests/general/test_deep_crawl_filters.py
from crawl4ai.deep_crawling.filters import ContentRelevanceFilter, URLPatternFilter, DomainFilter, ContentTypeFilter, SEOFilter async def test_pattern_filter(): # Test cases as list of tuples instead of dict for multiple patterns test_cases = [ # Simple suffix patterns (*.html) ("*.html", { "https://example.com/page.html": True, "https://example.com/path/doc.html": True, "https://example.com/page.htm": False, "https://example.com/page.html?param=1": True, }), # Path prefix patterns (/foo/*) ("*/article/*", { "https://example.com/article/123": True, "https://example.com/blog/article/456": True, "https://example.com/articles/789": False, "https://example.com/article": False, }), # Complex patterns ("blog-*-[0-9]", { "https://example.com/blog-post-1": True, "https://example.com/blog-test-9": True, "https://example.com/blog-post": False, "https://example.com/blog-post-x": False, }), # Multiple patterns case (["*.pdf", "*/download/*"], { "https://example.com/doc.pdf": True, "https://example.com/download/file.txt": True, "https://example.com/path/download/doc": True, "https://example.com/uploads/file.txt": False, }), # Edge cases ("*", { "https://example.com": True, "": True, "http://test.com/path": True, }), # Complex regex (r"^https?://.*\.example\.com/\d+", { "https://sub.example.com/123": True, "http://test.example.com/456": True, "https://example.com/789": False, "https://sub.example.com/abc": False, }) ] def run_accuracy_test(): print("\nAccuracy Tests:") print("-" * 50) all_passed = True for patterns, test_urls in test_cases: filter_obj = URLPatternFilter(patterns) for url, expected in test_urls.items(): result = filter_obj.apply(url) if result != expected: print(f"❌ Failed: Pattern '{patterns}' with URL '{url}'") print(f" Expected: {expected}, Got: {result}") all_passed = False else: print(f"✅ Passed: Pattern '{patterns}' with URL '{url}'") return all_passed # Run tests print("Running Pattern Filter Tests...") accuracy_passed = run_accuracy_test() if accuracy_passed: print("\n✨ All accuracy tests passed!") else: print("\n❌ Some accuracy tests failed!") async def test_domain_filter(): from itertools import chain # Test cases test_cases = [ # Allowed domains ({"allowed": "example.com"}, { "https://example.com/page": True, "http://example.com": True, "https://sub.example.com": False, "https://other.com": False, }), ({"allowed": ["example.com", "test.com"]}, { "https://example.com/page": True, "https://test.com/home": True, "https://other.com": False, }), # Blocked domains ({"blocked": "malicious.com"}, { "https://malicious.com": False, "https://safe.com": True, "http://malicious.com/login": False, }), ({"blocked": ["spam.com", "ads.com"]}, { "https://spam.com": False, "https://ads.com/banner": False, "https://example.com": True, }), # Allowed and Blocked combination ({"allowed": "example.com", "blocked": "sub.example.com"}, { "https://example.com": True, "https://sub.example.com": False, "https://other.com": False, }), ] def run_accuracy_test(): print("\nAccuracy Tests:") print("-" * 50) all_passed = True for params, test_urls in test_cases: filter_obj = DomainFilter( allowed_domains=params.get("allowed"), blocked_domains=params.get("blocked"), ) for url, expected in test_urls.items(): result = filter_obj.apply(url) if result != expected: print(f"\u274C Failed: Params {params} with URL '{url}'") print(f" Expected: {expected}, Got: {result}") all_passed = False else: print(f"\u2705 Passed: Params {params} with URL '{url}'") return all_passed # Run tests print("Running Domain Filter Tests...") accuracy_passed = run_accuracy_test() if accuracy_passed: print("\n\u2728 All accuracy tests passed!") else: print("\n\u274C Some accuracy tests failed!") async def test_content_relevance_filter(): relevance_filter = ContentRelevanceFilter( query="What was the cause of american civil war?", threshold=1 ) test_cases = { "https://en.wikipedia.org/wiki/Cricket": False, "https://en.wikipedia.org/wiki/American_Civil_War": True, } print("\nRunning Content Relevance Filter Tests...") print("-" * 50) all_passed = True for url, expected in test_cases.items(): result = await relevance_filter.apply(url) if result != expected: print(f"\u274C Failed: URL '{url}'") print(f" Expected: {expected}, Got: {result}") all_passed = False else: print(f"\u2705 Passed: URL '{url}'") if all_passed: print("\n\u2728 All content relevance tests passed!") else: print("\n\u274C Some content relevance tests failed!") async def test_content_type_filter(): from itertools import chain # Test cases test_cases = [ # Allowed single type ({"allowed": "image/png"}, { "https://example.com/image.png": True, "https://example.com/photo.jpg": False, "https://example.com/document.pdf": False, }), # Multiple allowed types ({"allowed": ["image/jpeg", "application/pdf"]}, { "https://example.com/photo.jpg": True, "https://example.com/document.pdf": True, "https://example.com/script.js": False, }), # No extension should be allowed ({"allowed": "application/json"}, { "https://example.com/api/data": True, "https://example.com/data.json": True, "https://example.com/page.html": False, }), # Unknown extensions should not be allowed ({"allowed": "application/octet-stream"}, { "https://example.com/file.unknown": True, "https://example.com/archive.zip": False, "https://example.com/software.exe": False, }), ] def run_accuracy_test(): print("\nAccuracy Tests:") print("-" * 50) all_passed = True for params, test_urls in test_cases: filter_obj = ContentTypeFilter( allowed_types=params.get("allowed"), ) for url, expected in test_urls.items(): result = filter_obj.apply(url) if result != expected: print(f"\u274C Failed: Params {params} with URL '{url}'") print(f" Expected: {expected}, Got: {result}") all_passed = False else: print(f"\u2705 Passed: Params {params} with URL '{url}'") return all_passed # Run tests print("Running Content Type Filter Tests...") accuracy_passed = run_accuracy_test() if accuracy_passed: print("\n\u2728 All accuracy tests passed!") else: print("\n\u274C Some accuracy tests failed!") async def test_seo_filter(): seo_filter = SEOFilter(threshold=0.5, keywords=["SEO", "search engines", "Optimization"]) test_cases = { "https://en.wikipedia.org/wiki/Search_engine_optimization": True, "https://en.wikipedia.org/wiki/Randomness": False, } print("\nRunning SEO Filter Tests...") print("-" * 50) all_passed = True for url, expected in test_cases.items(): result = await seo_filter.apply(url) if result != expected: print(f"\u274C Failed: URL '{url}'") print(f" Expected: {expected}, Got: {result}") all_passed = False else: print(f"\u2705 Passed: URL '{url}'") if all_passed: print("\n\u2728 All SEO filter tests passed!") else: print("\n\u274C Some SEO filter tests failed!") import asyncio if __name__ == "__main__": asyncio.run(test_pattern_filter()) asyncio.run(test_domain_filter()) asyncio.run(test_content_type_filter()) asyncio.run(test_content_relevance_filter()) asyncio.run(test_seo_filter())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_llm_filter.py
tests/general/test_llm_filter.py
import os import asyncio from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, CacheMode from crawl4ai import LLMConfig from crawl4ai.content_filter_strategy import LLMContentFilter async def test_llm_filter(): # Create an HTML source that needs intelligent filtering url = "https://docs.python.org/3/tutorial/classes.html" browser_config = BrowserConfig( headless=True, verbose=True ) # run_config = CrawlerRunConfig(cache_mode=CacheMode.BYPASS) run_config = CrawlerRunConfig(cache_mode=CacheMode.ENABLED) async with AsyncWebCrawler(config=browser_config) as crawler: # First get the raw HTML result = await crawler.arun(url, config=run_config) html = result.cleaned_html # Initialize LLM filter with focused instruction filter = LLMContentFilter( llm_config=LLMConfig(provider="openai/gpt-4o",api_token=os.getenv('OPENAI_API_KEY')), instruction=""" Focus on extracting the core educational content about Python classes. Include: - Key concepts and their explanations - Important code examples - Essential technical details Exclude: - Navigation elements - Sidebars - Footer content - Version information - Any non-essential UI elements Format the output as clean markdown with proper code blocks and headers. """, verbose=True ) filter = LLMContentFilter( llm_config = LLMConfig(provider="openai/gpt-4o",api_token=os.getenv('OPENAI_API_KEY')), chunk_token_threshold=2 ** 12 * 2, # 2048 * 2 instruction=""" Extract the main educational content while preserving its original wording and substance completely. Your task is to: 1. Maintain the exact language and terminology used in the main content 2. Keep all technical explanations, examples, and educational content intact 3. Preserve the original flow and structure of the core content 4. Remove only clearly irrelevant elements like: - Navigation menus - Advertisement sections - Cookie notices - Footers with site information - Sidebars with external links - Any UI elements that don't contribute to learning The goal is to create a clean markdown version that reads exactly like the original article, keeping all valuable content but free from distracting elements. Imagine you're creating a perfect reading experience where nothing valuable is lost, but all noise is removed. """, verbose=True ) # Apply filtering filtered_content = filter.filter_content(html, ignore_cache = True) # Show results print("\nFiltered Content Length:", len(filtered_content)) print("\nFirst 500 chars of filtered content:") if filtered_content: print(filtered_content[0][:500]) # Save on disc the markdown version with open("filtered_content.md", "w", encoding="utf-8") as f: f.write("\n".join(filtered_content)) # Show token usage filter.show_usage() if __name__ == "__main__": asyncio.run(test_llm_filter())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_network_console_capture.py
tests/general/test_network_console_capture.py
from crawl4ai.async_webcrawler import AsyncWebCrawler from crawl4ai.async_configs import CrawlerRunConfig, BrowserConfig import asyncio import aiohttp from aiohttp import web import tempfile import shutil import os, sys, time, json async def start_test_server(): app = web.Application() async def basic_page(request): return web.Response(text=""" <!DOCTYPE html> <html> <head> <title>Network Request Test</title> </head> <body> <h1>Test Page for Network Capture</h1> <p>This page performs network requests and console logging.</p> <img src="/image.png" alt="Test Image"> <script> console.log("Basic console log"); console.error("Error message"); console.warn("Warning message"); // Make some XHR requests const xhr = new XMLHttpRequest(); xhr.open('GET', '/api/data', true); xhr.send(); // Make a fetch request fetch('/api/json') .then(response => response.json()) .catch(error => console.error('Fetch error:', error)); // Trigger an error setTimeout(() => { try { nonExistentFunction(); } catch (e) { console.error("Caught error:", e); } }, 100); </script> </body> </html> """, content_type="text/html") async def image(request): # Return a small 1x1 transparent PNG return web.Response(body=bytes.fromhex('89504E470D0A1A0A0000000D49484452000000010000000108060000001F15C4890000000D4944415478DA63FAFFFF3F030079DB00018D959DE70000000049454E44AE426082'), content_type="image/png") async def api_data(request): return web.Response(text="sample data") async def api_json(request): return web.json_response({"status": "success", "message": "JSON data"}) # Register routes app.router.add_get('/', basic_page) app.router.add_get('/image.png', image) app.router.add_get('/api/data', api_data) app.router.add_get('/api/json', api_json) runner = web.AppRunner(app) await runner.setup() site = web.TCPSite(runner, 'localhost', 8080) await site.start() return runner async def test_network_console_capture(): print("\n=== Testing Network and Console Capture ===\n") # Start test server runner = await start_test_server() try: browser_config = BrowserConfig(headless=True) # Test with capture disabled (default) print("\n1. Testing with capture disabled (default)...") async with AsyncWebCrawler(config=browser_config) as crawler: config = CrawlerRunConfig( wait_until="networkidle", # Wait for network to be idle ) result = await crawler.arun(url="http://localhost:8080/", config=config) assert result.network_requests is None, "Network requests should be None when capture is disabled" assert result.console_messages is None, "Console messages should be None when capture is disabled" print("✓ Default config correctly returns None for network_requests and console_messages") # Test with network capture enabled print("\n2. Testing with network capture enabled...") async with AsyncWebCrawler(config=browser_config) as crawler: config = CrawlerRunConfig( wait_until="networkidle", # Wait for network to be idle capture_network_requests=True ) result = await crawler.arun(url="http://localhost:8080/", config=config) assert result.network_requests is not None, "Network requests should be captured" print(f"✓ Captured {len(result.network_requests)} network requests") # Check if we have both requests and responses request_count = len([r for r in result.network_requests if r.get("event_type") == "request"]) response_count = len([r for r in result.network_requests if r.get("event_type") == "response"]) print(f" - {request_count} requests, {response_count} responses") # Check if we captured specific resources urls = [r.get("url") for r in result.network_requests] has_image = any("/image.png" in url for url in urls) has_api_data = any("/api/data" in url for url in urls) has_api_json = any("/api/json" in url for url in urls) assert has_image, "Should have captured image request" assert has_api_data, "Should have captured API data request" assert has_api_json, "Should have captured API JSON request" print("✓ Captured expected network requests (image, API endpoints)") # Test with console capture enabled print("\n3. Testing with console capture enabled...") async with AsyncWebCrawler(config=browser_config) as crawler: config = CrawlerRunConfig( wait_until="networkidle", # Wait for network to be idle capture_console_messages=True ) result = await crawler.arun(url="http://localhost:8080/", config=config) assert result.console_messages is not None, "Console messages should be captured" print(f"✓ Captured {len(result.console_messages)} console messages") # Check if we have different types of console messages message_types = set(msg.get("type") for msg in result.console_messages if "type" in msg) print(f" - Message types: {', '.join(message_types)}") # Print all captured messages for debugging print(" - Captured messages:") for msg in result.console_messages: print(f" * Type: {msg.get('type', 'N/A')}, Text: {msg.get('text', 'N/A')}") # Look for specific messages messages = [msg.get("text") for msg in result.console_messages if "text" in msg] has_basic_log = any("Basic console log" in msg for msg in messages) has_error_msg = any("Error message" in msg for msg in messages) has_warning_msg = any("Warning message" in msg for msg in messages) assert has_basic_log, "Should have captured basic console.log message" assert has_error_msg, "Should have captured console.error message" assert has_warning_msg, "Should have captured console.warn message" print("✓ Captured expected console messages (log, error, warning)") # Test with both captures enabled print("\n4. Testing with both network and console capture enabled...") async with AsyncWebCrawler(config=browser_config) as crawler: config = CrawlerRunConfig( wait_until="networkidle", # Wait for network to be idle capture_network_requests=True, capture_console_messages=True ) result = await crawler.arun(url="http://localhost:8080/", config=config) assert result.network_requests is not None, "Network requests should be captured" assert result.console_messages is not None, "Console messages should be captured" print(f"✓ Successfully captured both {len(result.network_requests)} network requests and {len(result.console_messages)} console messages") finally: await runner.cleanup() print("\nTest server shutdown") async def main(): try: await test_network_console_capture() print("\n✅ All tests passed successfully!") except Exception as e: print(f"\n❌ Test failed: {str(e)}") raise if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_async_webcrawler.py
tests/general/test_async_webcrawler.py
import asyncio import pytest from typing import List from crawl4ai import ( AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, MemoryAdaptiveDispatcher, RateLimiter, CacheMode ) from crawl4ai.extraction_strategy import ExtractionStrategy class MockExtractionStrategy(ExtractionStrategy): """Mock extraction strategy for testing URL parameter handling""" def __init__(self): super().__init__() self.run_calls = [] def extract(self, url: str, html: str, *args, **kwargs): return [{"test": "data"}] def run(self, url: str, sections: List[str], *args, **kwargs): self.run_calls.append(url) return super().run(url, sections, *args, **kwargs) @pytest.mark.asyncio @pytest.mark.parametrize("viewport", [ (800, 600), (1024, 768), (1920, 1080) ]) async def test_viewport_config(viewport): """Test different viewport configurations""" width, height = viewport browser_config = BrowserConfig( browser_type="chromium", headless=True, viewport_width=width, viewport_height=height ) async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun( url="https://example.com", config=CrawlerRunConfig( # cache_mode=CacheMode.BYPASS, page_timeout=30000 # 30 seconds ) ) assert result.success @pytest.mark.asyncio async def test_memory_management(): """Test memory-adaptive dispatching""" browser_config = BrowserConfig( browser_type="chromium", headless=True, viewport_width=1024, viewport_height=768 ) dispatcher = MemoryAdaptiveDispatcher( memory_threshold_percent=70.0, check_interval=1.0, max_session_permit=5 ) urls = ["https://example.com"] * 3 # Test with multiple identical URLs async with AsyncWebCrawler(config=browser_config) as crawler: results = await crawler.arun_many( urls=urls, config=CrawlerRunConfig(page_timeout=30000), dispatcher=dispatcher ) assert len(results) == len(urls) @pytest.mark.asyncio async def test_rate_limiting(): """Test rate limiting functionality""" browser_config = BrowserConfig( browser_type="chromium", headless=True ) dispatcher = MemoryAdaptiveDispatcher( rate_limiter=RateLimiter( base_delay=(1.0, 2.0), max_delay=5.0, max_retries=2 ), memory_threshold_percent=70.0 ) urls = [ "https://example.com", "https://example.org", "https://example.net" ] async with AsyncWebCrawler(config=browser_config) as crawler: results = await crawler.arun_many( urls=urls, config=CrawlerRunConfig(page_timeout=30000), dispatcher=dispatcher ) assert len(results) == len(urls) @pytest.mark.asyncio async def test_javascript_execution(): """Test JavaScript execution capabilities""" browser_config = BrowserConfig( browser_type="chromium", headless=True, java_script_enabled=True ) js_code = """ document.body.style.backgroundColor = 'red'; return document.body.style.backgroundColor; """ async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun( url="https://example.com", config=CrawlerRunConfig( js_code=js_code, page_timeout=30000 ) ) assert result.success @pytest.mark.asyncio @pytest.mark.parametrize("error_url", [ "https://invalid.domain.test", "https://httpbin.org/status/404", "https://httpbin.org/status/503", "https://httpbin.org/status/403" ]) async def test_error_handling(error_url): """Test error handling for various failure scenarios""" browser_config = BrowserConfig( browser_type="chromium", headless=True ) async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun( url=error_url, config=CrawlerRunConfig( page_timeout=10000, # Short timeout for error cases cache_mode=CacheMode.BYPASS ) ) assert not result.success assert result.error_message is not None @pytest.mark.asyncio async def test_extraction_strategy_run_with_regular_url(): """ Regression test for extraction_strategy.run URL parameter handling with regular URLs. This test verifies that when is_raw_html=False (regular URL), extraction_strategy.run is called with the actual URL. """ browser_config = BrowserConfig( browser_type="chromium", headless=True ) async with AsyncWebCrawler(config=browser_config) as crawler: mock_strategy = MockExtractionStrategy() # Test regular URL (is_raw_html=False) regular_url = "https://example.com" result = await crawler.arun( url=regular_url, config=CrawlerRunConfig( page_timeout=30000, extraction_strategy=mock_strategy, cache_mode=CacheMode.BYPASS ) ) assert result.success assert len(mock_strategy.run_calls) == 1 assert mock_strategy.run_calls[0] == regular_url, f"Expected '{regular_url}', got '{mock_strategy.run_calls[0]}'" @pytest.mark.asyncio async def test_extraction_strategy_run_with_raw_html(): """ Regression test for extraction_strategy.run URL parameter handling with raw HTML. This test verifies that when is_raw_html=True (URL starts with "raw:"), extraction_strategy.run is called with "Raw HTML" instead of the actual URL. """ browser_config = BrowserConfig( browser_type="chromium", headless=True ) async with AsyncWebCrawler(config=browser_config) as crawler: mock_strategy = MockExtractionStrategy() # Test raw HTML URL (is_raw_html=True automatically set) raw_html_url = "raw:<html><body><h1>Test HTML</h1><p>This is a test.</p></body></html>" result = await crawler.arun( url=raw_html_url, config=CrawlerRunConfig( page_timeout=30000, extraction_strategy=mock_strategy, cache_mode=CacheMode.BYPASS ) ) assert result.success assert len(mock_strategy.run_calls) == 1 assert mock_strategy.run_calls[0] == "Raw HTML", f"Expected 'Raw HTML', got '{mock_strategy.run_calls[0]}'" if __name__ == "__main__": asyncio.run(test_viewport_config((1024, 768))) asyncio.run(test_memory_management()) asyncio.run(test_rate_limiting()) asyncio.run(test_javascript_execution()) asyncio.run(test_extraction_strategy_run_with_regular_url()) asyncio.run(test_extraction_strategy_run_with_raw_html())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_bff_scoring.py
tests/general/test_bff_scoring.py
#!/usr/bin/env python3 """ Simple test to verify BestFirstCrawlingStrategy fixes. This test crawls a real website and shows that: 1. Higher-scoring pages are crawled first (priority queue fix) 2. Links are scored before truncation (link discovery fix) """ import asyncio from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig from crawl4ai.deep_crawling import BestFirstCrawlingStrategy from crawl4ai.deep_crawling.scorers import KeywordRelevanceScorer async def test_best_first_strategy(): """Test BestFirstCrawlingStrategy with keyword scoring""" print("=" * 70) print("Testing BestFirstCrawlingStrategy with Real URL") print("=" * 70) print("\nThis test will:") print("1. Crawl Python.org documentation") print("2. Score pages based on keywords: 'tutorial', 'guide', 'reference'") print("3. Show that higher-scoring pages are crawled first") print("-" * 70) # Create a keyword scorer that prioritizes tutorial/guide pages scorer = KeywordRelevanceScorer( keywords=["tutorial", "guide", "reference", "documentation"], weight=1.0, case_sensitive=False ) # Create the strategy with scoring strategy = BestFirstCrawlingStrategy( max_depth=2, # Crawl 2 levels deep max_pages=10, # Limit to 10 pages total url_scorer=scorer, # Use keyword scoring include_external=False # Only internal links ) # Configure browser and crawler browser_config = BrowserConfig( headless=True, # Run in background verbose=False # Reduce output noise ) crawler_config = CrawlerRunConfig( deep_crawl_strategy=strategy, verbose=False ) print("\nStarting crawl of https://docs.python.org/3/") print("Looking for pages with keywords: tutorial, guide, reference, documentation") print("-" * 70) crawled_urls = [] async with AsyncWebCrawler(config=browser_config) as crawler: # Crawl and collect results results = await crawler.arun( url="https://docs.python.org/3/", config=crawler_config ) # Process results if isinstance(results, list): for result in results: score = result.metadata.get('score', 0) if result.metadata else 0 depth = result.metadata.get('depth', 0) if result.metadata else 0 crawled_urls.append({ 'url': result.url, 'score': score, 'depth': depth, 'success': result.success }) print("\n" + "=" * 70) print("CRAWL RESULTS (in order of crawling)") print("=" * 70) for i, item in enumerate(crawled_urls, 1): status = "✓" if item['success'] else "✗" # Highlight high-scoring pages if item['score'] > 0.5: print(f"{i:2}. [{status}] Score: {item['score']:.2f} | Depth: {item['depth']} | {item['url']}") print(f" ^ HIGH SCORE - Contains keywords!") else: print(f"{i:2}. [{status}] Score: {item['score']:.2f} | Depth: {item['depth']} | {item['url']}") print("\n" + "=" * 70) print("ANALYSIS") print("=" * 70) # Check if higher scores appear early in the crawl scores = [item['score'] for item in crawled_urls[1:]] # Skip initial URL high_score_indices = [i for i, s in enumerate(scores) if s > 0.3] if high_score_indices and high_score_indices[0] < len(scores) / 2: print("✅ SUCCESS: Higher-scoring pages (with keywords) were crawled early!") print(" This confirms the priority queue fix is working.") else: print("⚠️ Check the crawl order above - higher scores should appear early") # Show score distribution print(f"\nScore Statistics:") print(f" - Total pages crawled: {len(crawled_urls)}") print(f" - Average score: {sum(item['score'] for item in crawled_urls) / len(crawled_urls):.2f}") print(f" - Max score: {max(item['score'] for item in crawled_urls):.2f}") print(f" - Pages with keywords: {sum(1 for item in crawled_urls if item['score'] > 0.3)}") print("\n" + "=" * 70) print("TEST COMPLETE") print("=" * 70) if __name__ == "__main__": print("\n🔍 BestFirstCrawlingStrategy Simple Test\n") asyncio.run(test_best_first_strategy())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_stream_dispatch.py
tests/general/test_stream_dispatch.py
import os, sys # append 2 parent directories to sys.path to import crawl4ai parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) parent_parent_dir = os.path.dirname(parent_dir) sys.path.append(parent_parent_dir) import asyncio from typing import List from crawl4ai import * from crawl4ai.async_dispatcher import MemoryAdaptiveDispatcher async def test_streaming(): browser_config = BrowserConfig(headless=True, verbose=True) crawler_config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator( # content_filter=PruningContentFilter( # threshold=0.48, # threshold_type="fixed", # min_word_threshold=0 # ) ), ) urls = ["http://example.com"] * 10 async with AsyncWebCrawler(config=browser_config) as crawler: dispatcher = MemoryAdaptiveDispatcher( max_session_permit=5, check_interval=0.5 ) async for result in dispatcher.run_urls_stream(urls, crawler, crawler_config): print(f"Got result for {result.url} - Success: {result.result.success}") if __name__ == "__main__": asyncio.run(test_streaming())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_acyn_crawl_wuth_http_crawler_strategy.py
tests/general/test_acyn_crawl_wuth_http_crawler_strategy.py
import asyncio from crawl4ai import ( AsyncWebCrawler, CrawlerRunConfig, HTTPCrawlerConfig, CacheMode, DefaultMarkdownGenerator, PruningContentFilter ) from crawl4ai.async_crawler_strategy import AsyncHTTPCrawlerStrategy from crawl4ai.async_logger import AsyncLogger async def main(): # Initialize HTTP crawler strategy http_strategy = AsyncHTTPCrawlerStrategy( browser_config=HTTPCrawlerConfig( method="GET", verify_ssl=True, follow_redirects=True ), logger=AsyncLogger(verbose=True) ) # Initialize web crawler with HTTP strategy async with AsyncWebCrawler(crawler_strategy=http_strategy) as crawler: crawler_config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator( content_filter=PruningContentFilter( threshold=0.48, threshold_type="fixed", min_word_threshold=0 ) ) ) # Test different URLs urls = [ "https://example.com", "https://httpbin.org/get", "raw://<html><body>Test content</body></html>" ] for url in urls: print(f"\n=== Testing {url} ===") try: result = await crawler.arun(url=url, config=crawler_config) print(f"Status: {result.status_code}") print(f"Raw HTML length: {len(result.html)}") if hasattr(result, 'markdown'): print(f"Markdown length: {len(result.markdown.raw_markdown)}") except Exception as e: print(f"Error: {e}") if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_persistent_context.py
tests/general/test_persistent_context.py
import asyncio import os from crawl4ai.async_webcrawler import AsyncWebCrawler from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig, CacheMode # Simple concurrency test for persistent context page creation # Usage: python scripts/test_persistent_context.py URLS = [ # "https://example.com", "https://httpbin.org/html", "https://www.python.org/", "https://www.rust-lang.org/", ] async def main(): profile_dir = os.path.join(os.path.expanduser("~"), ".crawl4ai", "profiles", "test-persistent-profile") os.makedirs(profile_dir, exist_ok=True) browser_config = BrowserConfig( browser_type="chromium", headless=True, use_persistent_context=True, user_data_dir=profile_dir, use_managed_browser=True, verbose=True, ) run_cfg = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, stream=False, verbose=True, ) async with AsyncWebCrawler(config=browser_config) as crawler: results = await crawler.arun_many(URLS, config=run_cfg) for r in results: print(r.url, r.success, len(r.markdown.raw_markdown) if r.markdown else 0) # r = await crawler.arun(url=URLS[0], config=run_cfg) # print(r.url, r.success, len(r.markdown.raw_markdown) if r.markdown else 0) if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_robot_parser.py
tests/general/test_robot_parser.py
from crawl4ai.utils import RobotsParser import asyncio import aiohttp from aiohttp import web import tempfile import shutil import os, sys, time, json async def test_robots_parser(): print("\n=== Testing RobotsParser ===\n") # Setup temporary directory for testing temp_dir = tempfile.mkdtemp() try: # 1. Basic setup test print("1. Testing basic initialization...") parser = RobotsParser(cache_dir=temp_dir) assert os.path.exists(parser.db_path), "Database file not created" print("✓ Basic initialization passed") # 2. Test common cases print("\n2. Testing common cases...") allowed = await parser.can_fetch("https://www.example.com", "MyBot/1.0") print(f"✓ Regular website fetch: {'allowed' if allowed else 'denied'}") # Test caching print("Testing cache...") start = time.time() await parser.can_fetch("https://www.example.com", "MyBot/1.0") duration = time.time() - start print(f"✓ Cached lookup took: {duration*1000:.2f}ms") assert duration < 0.03, "Cache lookup too slow" # 3. Edge cases print("\n3. Testing edge cases...") # Empty URL result = await parser.can_fetch("", "MyBot/1.0") print(f"✓ Empty URL handled: {'allowed' if result else 'denied'}") # Invalid URL result = await parser.can_fetch("not_a_url", "MyBot/1.0") print(f"✓ Invalid URL handled: {'allowed' if result else 'denied'}") # URL without scheme result = await parser.can_fetch("example.com/page", "MyBot/1.0") print(f"✓ URL without scheme handled: {'allowed' if result else 'denied'}") # 4. Test with local server async def start_test_server(): app = web.Application() async def robots_txt(request): return web.Response(text="""User-agent: * Disallow: /private/ Allow: /public/ """) async def malformed_robots(request): return web.Response(text="<<<malformed>>>") async def timeout_robots(request): await asyncio.sleep(5) return web.Response(text="Should timeout") async def empty_robots(request): return web.Response(text="") async def giant_robots(request): return web.Response(text="User-agent: *\nDisallow: /\n" * 10000) # Mount all handlers at root level app.router.add_get('/robots.txt', robots_txt) app.router.add_get('/malformed/robots.txt', malformed_robots) app.router.add_get('/timeout/robots.txt', timeout_robots) app.router.add_get('/empty/robots.txt', empty_robots) app.router.add_get('/giant/robots.txt', giant_robots) runner = web.AppRunner(app) await runner.setup() site = web.TCPSite(runner, 'localhost', 8080) await site.start() return runner runner = await start_test_server() try: print("\n4. Testing robots.txt rules...") base_url = "http://localhost:8080" # Test public access result = await parser.can_fetch(f"{base_url}/public/page", "bot") print(f"Public access (/public/page): {'allowed' if result else 'denied'}") assert result, "Public path should be allowed" # Test private access result = await parser.can_fetch(f"{base_url}/private/secret", "bot") print(f"Private access (/private/secret): {'allowed' if result else 'denied'}") assert not result, "Private path should be denied" # Test malformed result = await parser.can_fetch("http://localhost:8080/malformed/page", "bot") print(f"✓ Malformed robots.txt handled: {'allowed' if result else 'denied'}") # Test timeout start = time.time() result = await parser.can_fetch("http://localhost:8080/timeout/page", "bot") duration = time.time() - start print(f"✓ Timeout handled (took {duration:.2f}s): {'allowed' if result else 'denied'}") assert duration < 3, "Timeout not working" # Test empty result = await parser.can_fetch("http://localhost:8080/empty/page", "bot") print(f"✓ Empty robots.txt handled: {'allowed' if result else 'denied'}") # Test giant file start = time.time() result = await parser.can_fetch("http://localhost:8080/giant/page", "bot") duration = time.time() - start print(f"✓ Giant robots.txt handled (took {duration:.2f}s): {'allowed' if result else 'denied'}") finally: await runner.cleanup() # 5. Cache manipulation print("\n5. Testing cache manipulation...") # Clear expired parser.clear_expired() print("✓ Clear expired entries completed") # Clear all parser.clear_cache() print("✓ Clear all cache completed") # Test with custom TTL custom_parser = RobotsParser(cache_dir=temp_dir, cache_ttl=1) # 1 second TTL await custom_parser.can_fetch("https://www.example.com", "bot") print("✓ Custom TTL fetch completed") await asyncio.sleep(1.1) start = time.time() await custom_parser.can_fetch("https://www.example.com", "bot") print(f"✓ TTL expiry working (refetched after {time.time() - start:.2f}s)") finally: # Cleanup shutil.rmtree(temp_dir) print("\nTest cleanup completed") async def main(): try: await test_robots_parser() except Exception as e: print(f"Test failed: {str(e)}") raise if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_max_scroll.py
tests/general/test_max_scroll.py
""" Sample script to test the max_scroll_steps parameter implementation """ import asyncio import os import sys # Get the grandparent directory grandparent_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(grandparent_dir) __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) from crawl4ai import AsyncWebCrawler from crawl4ai.async_configs import CrawlerRunConfig async def test_max_scroll_steps(): """ Test the max_scroll_steps parameter with different configurations """ print("🚀 Testing max_scroll_steps parameter implementation") print("=" * 60) async with AsyncWebCrawler(verbose=True) as crawler: # Test 1: Without max_scroll_steps (unlimited scrolling) print("\\n📋 Test 1: Unlimited scrolling (max_scroll_steps=None)") config1 = CrawlerRunConfig( scan_full_page=True, scroll_delay=0.1, max_scroll_steps=None, # Default behavior verbose=True ) print(f"Config: scan_full_page={config1.scan_full_page}, max_scroll_steps={config1.max_scroll_steps}") try: result1 = await crawler.arun( url="https://example.com", # Simple page for testing config=config1 ) print(f"✅ Test 1 Success: Crawled {len(result1.markdown)} characters") except Exception as e: print(f"❌ Test 1 Failed: {e}") # Test 2: With limited scroll steps print("\\n📋 Test 2: Limited scrolling (max_scroll_steps=3)") config2 = CrawlerRunConfig( scan_full_page=True, scroll_delay=0.1, max_scroll_steps=3, # Limit to 3 scroll steps verbose=True ) print(f"Config: scan_full_page={config2.scan_full_page}, max_scroll_steps={config2.max_scroll_steps}") try: result2 = await crawler.arun( url="https://techcrunch.com/", # Another test page config=config2 ) print(f"✅ Test 2 Success: Crawled {len(result2.markdown)} characters") except Exception as e: print(f"❌ Test 2 Failed: {e}") # Test 3: Test serialization/deserialization print("\\n📋 Test 3: Configuration serialization test") config3 = CrawlerRunConfig( scan_full_page=True, max_scroll_steps=5, scroll_delay=0.2 ) # Test to_dict config_dict = config3.to_dict() print(f"Serialized max_scroll_steps: {config_dict.get('max_scroll_steps')}") # Test from_kwargs config4 = CrawlerRunConfig.from_kwargs({ 'scan_full_page': True, 'max_scroll_steps': 7, 'scroll_delay': 0.3 }) print(f"Deserialized max_scroll_steps: {config4.max_scroll_steps}") print("✅ Test 3 Success: Serialization works correctly") # Test 4: Edge case - max_scroll_steps = 0 print("\\n📋 Test 4: Edge case (max_scroll_steps=0)") config5 = CrawlerRunConfig( scan_full_page=True, max_scroll_steps=0, # Should not scroll at all verbose=True ) try: result5 = await crawler.arun( url="https://techcrunch.com/", config=config5 ) print(f"✅ Test 4 Success: No scrolling performed, crawled {len(result5.markdown)} characters") except Exception as e: print(f"❌ Test 4 Failed: {e}") print("\\n" + "=" * 60) print("🎉 All tests completed!") print("\\nThe max_scroll_steps parameter is working correctly:") print("- None: Unlimited scrolling (default behavior)") print("- Positive integer: Limits scroll steps to that number") print("- 0: No scrolling performed") print("- Properly serializes/deserializes in config") if __name__ == "__main__": print("Starting max_scroll_steps test...") asyncio.run(test_max_scroll_steps())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_async_url_seeder_bm25.py
tests/general/test_async_url_seeder_bm25.py
""" Comprehensive test cases for AsyncUrlSeeder with BM25 scoring functionality. Tests cover all features including query-based scoring, metadata extraction, edge cases, and integration scenarios. """ import asyncio import pytest from typing import List, Dict, Any from crawl4ai import AsyncUrlSeeder, SeedingConfig, AsyncLogger import json from datetime import datetime # Test domain - using docs.crawl4ai.com as it has the actual documentation TEST_DOMAIN = "kidocode.com" TEST_DOMAIN = "docs.crawl4ai.com" TEST_DOMAIN = "www.bbc.com/sport" class TestAsyncUrlSeederBM25: """Comprehensive test suite for AsyncUrlSeeder with BM25 scoring.""" async def create_seeder(self): """Create an AsyncUrlSeeder instance for testing.""" logger = AsyncLogger() return AsyncUrlSeeder(logger=logger) # ============================================ # Basic BM25 Scoring Tests # ============================================ @pytest.mark.asyncio async def test_basic_bm25_scoring(self, seeder): """Test basic BM25 scoring with a simple query.""" config = SeedingConfig( source="sitemap", extract_head=True, query="premier league highlights", scoring_method="bm25", max_urls=200, verbose=True, force=True # Force fresh fetch ) results = await seeder.urls(TEST_DOMAIN, config) # Verify results have relevance scores assert all("relevance_score" in r for r in results) # Verify scores are normalized between 0 and 1 scores = [r["relevance_score"] for r in results] assert all(0.0 <= s <= 1.0 for s in scores) # Verify results are sorted by relevance (descending) assert scores == sorted(scores, reverse=True) # Print top 5 results for manual verification print("\nTop 5 results for 'web crawling tutorial':") for i, r in enumerate(results[:5]): print(f"{i+1}. Score: {r['relevance_score']:.3f} - {r['url']}") @pytest.mark.asyncio async def test_query_variations(self, seeder): """Test BM25 scoring with different query variations.""" queries = [ "VAR controversy", "player ratings", "live score update", "transfer rumours", "post match analysis", "injury news" ] for query in queries: config = SeedingConfig( source="sitemap", extract_head=True, query=query, scoring_method="bm25", max_urls=100, # force=True ) results = await seeder.urls(TEST_DOMAIN, config) # Verify each query produces scored results assert len(results) > 0 assert all("relevance_score" in r for r in results) print(f"\nTop result for '{query}':") if results: top = results[0] print(f" Score: {top['relevance_score']:.3f} - {top['url']}") # ============================================ # Score Threshold Tests # ============================================ @pytest.mark.asyncio async def test_score_threshold_filtering(self, seeder): """Test filtering results by minimum relevance score.""" thresholds = [0.1, 0.3, 0.5, 0.7] for threshold in thresholds: config = SeedingConfig( source="sitemap", extract_head=True, query="league standings", score_threshold=threshold, scoring_method="bm25", max_urls=50 ) results = await seeder.urls(TEST_DOMAIN, config) # Verify all results meet threshold if results: assert all(r["relevance_score"] >= threshold for r in results) print(f"\nThreshold {threshold}: {len(results)} URLs passed") @pytest.mark.asyncio async def test_extreme_thresholds(self, seeder): """Test edge cases with extreme threshold values.""" # Very low threshold - should return many results config_low = SeedingConfig( source="sitemap", extract_head=True, query="match", score_threshold=0.001, scoring_method="bm25" ) results_low = await seeder.urls(TEST_DOMAIN, config_low) # Very high threshold - might return few or no results config_high = SeedingConfig( source="sitemap", extract_head=True, query="match", score_threshold=0.99, scoring_method="bm25" ) results_high = await seeder.urls(TEST_DOMAIN, config_high) # Low threshold should return more results than high assert len(results_low) >= len(results_high) print(f"\nLow threshold (0.001): {len(results_low)} results") print(f"High threshold (0.99): {len(results_high)} results") # ============================================ # Metadata Extraction Tests # ============================================ @pytest.mark.asyncio async def test_comprehensive_metadata_extraction(self, seeder): """Test extraction of all metadata types including JSON-LD.""" config = SeedingConfig( source="sitemap", extract_head=True, query="match report", scoring_method="bm25", max_urls=5, verbose=True ) results = await seeder.urls(TEST_DOMAIN, config) for result in results: head_data = result.get("head_data", {}) # Check for various metadata fields print(f"\nMetadata for {result['url']}:") print(f" Title: {head_data.get('title', 'N/A')}") print(f" Charset: {head_data.get('charset', 'N/A')}") print(f" Lang: {head_data.get('lang', 'N/A')}") # Check meta tags meta = head_data.get("meta", {}) if meta: print(" Meta tags found:") for key in ["description", "keywords", "author", "viewport"]: if key in meta: print(f" {key}: {meta[key][:50]}...") # Check for Open Graph tags og_tags = {k: v for k, v in meta.items() if k.startswith("og:")} if og_tags: print(" Open Graph tags found:") for k, v in list(og_tags.items())[:3]: print(f" {k}: {v[:50]}...") # Check JSON-LD if head_data.get("jsonld"): print(f" JSON-LD schemas found: {len(head_data['jsonld'])}") @pytest.mark.asyncio async def test_jsonld_extraction_scoring(self, seeder): """Test that JSON-LD data contributes to BM25 scoring.""" config = SeedingConfig( source="sitemap", extract_head=True, query="Premier League match report highlights", scoring_method="bm25", max_urls=20 ) results = await seeder.urls(TEST_DOMAIN, config) # Find results with JSON-LD data jsonld_results = [r for r in results if r.get("head_data", {}).get("jsonld")] if jsonld_results: print(f"\nFound {len(jsonld_results)} URLs with JSON-LD data") for r in jsonld_results[:3]: print(f" Score: {r['relevance_score']:.3f} - {r['url']}") jsonld_data = r["head_data"]["jsonld"] print(f" JSON-LD types: {[item.get('@type', 'Unknown') for item in jsonld_data if isinstance(item, dict)]}") # ============================================ # Edge Cases and Error Handling # ============================================ @pytest.mark.asyncio async def test_empty_query(self, seeder): """Test behavior with empty query string.""" config = SeedingConfig( source="sitemap", extract_head=True, query="", scoring_method="bm25", max_urls=10 ) results = await seeder.urls(TEST_DOMAIN, config) # Should return results but all with zero scores assert len(results) > 0 assert all(r.get("relevance_score", 0) == 0 for r in results) @pytest.mark.asyncio async def test_query_without_extract_head(self, seeder): """Test query scoring when extract_head is False.""" config = SeedingConfig( source="sitemap", extract_head=False, # This should trigger a warning query="Premier League match report highlights", scoring_method="bm25", max_urls=10 ) results = await seeder.urls(TEST_DOMAIN, config) # Results should not have relevance scores assert all("relevance_score" not in r for r in results) print("\nVerified: No scores added when extract_head=False") @pytest.mark.asyncio async def test_special_characters_in_query(self, seeder): """Test queries with special characters and symbols.""" special_queries = [ "premier league + analytics", "injury/rehab routines", "AI-powered scouting", "match stats & xG", "tactical@breakdown", "transfer-window.yml" ] for query in special_queries: config = SeedingConfig( source="sitemap", extract_head=True, query=query, scoring_method="bm25", max_urls=5 ) try: results = await seeder.urls(TEST_DOMAIN, config) assert isinstance(results, list) print(f"\n✓ Query '{query}' processed successfully") except Exception as e: pytest.fail(f"Failed on query '{query}': {str(e)}") @pytest.mark.asyncio async def test_unicode_query(self, seeder): """Test queries with Unicode characters.""" unicode_queries = [ "网页爬虫", # Chinese "веб-краулер", # Russian "🚀 crawl4ai", # Emoji "naïve implementation", # Accented characters ] for query in unicode_queries: config = SeedingConfig( source="sitemap", extract_head=True, query=query, scoring_method="bm25", max_urls=5 ) try: results = await seeder.urls(TEST_DOMAIN, config) assert isinstance(results, list) print(f"\n✓ Unicode query '{query}' processed successfully") except Exception as e: print(f"\n✗ Unicode query '{query}' failed: {str(e)}") # ============================================ # Performance and Scalability Tests # ============================================ @pytest.mark.asyncio async def test_large_scale_scoring(self, seeder): """Test BM25 scoring with many URLs.""" config = SeedingConfig( source="cc+sitemap", # Use both sources for more URLs extract_head=True, query="world cup group standings", scoring_method="bm25", max_urls=100, concurrency=20, hits_per_sec=10 ) start_time = asyncio.get_event_loop().time() results = await seeder.urls(TEST_DOMAIN, config) elapsed = asyncio.get_event_loop().time() - start_time print(f"\nProcessed {len(results)} URLs in {elapsed:.2f} seconds") print(f"Average time per URL: {elapsed/len(results)*1000:.1f}ms") # Verify scoring worked at scale assert all("relevance_score" in r for r in results) # Check score distribution scores = [r["relevance_score"] for r in results] print(f"Score distribution:") print(f" Min: {min(scores):.3f}") print(f" Max: {max(scores):.3f}") print(f" Avg: {sum(scores)/len(scores):.3f}") @pytest.mark.asyncio async def test_concurrent_scoring_consistency(self, seeder): """Test that concurrent requests produce consistent scores.""" config = SeedingConfig( source="sitemap", extract_head=True, query="live score update", scoring_method="bm25", max_urls=20, concurrency=10 ) # Run the same query multiple times results_list = [] for _ in range(3): results = await seeder.urls(TEST_DOMAIN, config) results_list.append(results) # Compare scores across runs (they should be identical for same URLs) url_scores = {} for results in results_list: for r in results: url = r["url"] score = r["relevance_score"] if url in url_scores: # Scores should be very close (allowing for tiny float differences) assert abs(url_scores[url] - score) < 0.001 else: url_scores[url] = score print(f"\n✓ Consistent scores across {len(results_list)} runs") # ============================================ # Multi-Domain Tests # ============================================ @pytest.mark.asyncio async def test_many_urls_with_scoring(self, seeder): """Test many_urls method with BM25 scoring.""" domains = [TEST_DOMAIN, "docs.crawl4ai.com", "example.com"] config = SeedingConfig( source="sitemap", extract_head=True, # live_check=True, query="fixture list", scoring_method="bm25", score_threshold=0.2, max_urls=10, force=True, # Force fresh fetch ) results_dict = await seeder.many_urls(domains, config) for domain, results in results_dict.items(): print(f"\nDomain: {domain}") print(f" Found {len(results)} URLs above threshold") if results: top = results[0] print(f" Top result: {top['relevance_score']:.3f} - {top['url']}") # ============================================ # Complex Query Tests # ============================================ @pytest.mark.asyncio async def test_multi_word_complex_queries(self, seeder): """Test complex multi-word queries.""" complex_queries = [ "how to follow live match commentary", "extract expected goals stats from match data", "premier league match report analysis", "transfer rumours and confirmed signings tracker", "tactical breakdown of high press strategy" ] for query in complex_queries: config = SeedingConfig( source="sitemap", extract_head=True, query=query, scoring_method="bm25", max_urls=5 ) results = await seeder.urls(TEST_DOMAIN, config) if results: print(f"\nQuery: '{query}'") print(f"Top match: {results[0]['relevance_score']:.3f} - {results[0]['url']}") # Extract matched terms from metadata head_data = results[0].get("head_data", {}) title = head_data.get("title", "") description = head_data.get("meta", {}).get("description", "") # Simple term matching for verification query_terms = set(query.lower().split()) title_terms = set(title.lower().split()) desc_terms = set(description.lower().split()) matched_terms = query_terms & (title_terms | desc_terms) if matched_terms: print(f"Matched terms: {', '.join(matched_terms)}") # ============================================ # Cache and Force Tests # ============================================ @pytest.mark.asyncio async def test_scoring_with_cache(self, seeder): """Test that scoring works correctly with cached results.""" config = SeedingConfig( source="sitemap", extract_head=True, query="injury update timeline", scoring_method="bm25", max_urls=10, force=False # Use cache ) # First run - populate cache results1 = await seeder.urls(TEST_DOMAIN, config) # Second run - should use cache results2 = await seeder.urls(TEST_DOMAIN, config) # Results should be identical assert len(results1) == len(results2) for r1, r2 in zip(results1, results2): assert r1["url"] == r2["url"] assert abs(r1["relevance_score"] - r2["relevance_score"]) < 0.001 print("\n✓ Cache produces consistent scores") @pytest.mark.asyncio async def test_force_refresh_scoring(self, seeder): """Test force=True bypasses cache for fresh scoring.""" config_cached = SeedingConfig( source="sitemap", extract_head=True, query="transfer window", scoring_method="bm25", max_urls=5, force=False ) config_forced = SeedingConfig( source="sitemap", extract_head=True, query="transfer window", scoring_method="bm25", max_urls=5, force=True ) # Run with cache start1 = asyncio.get_event_loop().time() results1 = await seeder.urls(TEST_DOMAIN, config_cached) time1 = asyncio.get_event_loop().time() - start1 # Run with force (should be slower due to fresh fetch) start2 = asyncio.get_event_loop().time() results2 = await seeder.urls(TEST_DOMAIN, config_forced) time2 = asyncio.get_event_loop().time() - start2 print(f"\nCached run: {time1:.2f}s") print(f"Forced run: {time2:.2f}s") # Both should produce scored results assert all("relevance_score" in r for r in results1) assert all("relevance_score" in r for r in results2) # ============================================ # Source Combination Tests # ============================================ @pytest.mark.asyncio async def test_scoring_with_multiple_sources(self, seeder): """Test BM25 scoring with combined sources (cc+sitemap).""" config = SeedingConfig( source="cc+sitemap", extract_head=True, query="match highlights video", scoring_method="bm25", score_threshold=0.3, max_urls=30, concurrency=15 ) results = await seeder.urls(TEST_DOMAIN, config) # Verify we got results from both sources print(f"\nCombined sources returned {len(results)} URLs above threshold") # Check URL diversity unique_paths = set() for r in results: path = r["url"].replace("https://", "").replace("http://", "").split("/", 1)[-1] unique_paths.add(path.split("?")[0]) # Remove query params print(f"Unique paths found: {len(unique_paths)}") # All should be scored and above threshold assert all(r["relevance_score"] >= 0.3 for r in results) # ============================================ # Integration Tests # ============================================ @pytest.mark.asyncio async def test_full_workflow_integration(self, seeder): """Test complete workflow: discover -> score -> filter -> use.""" # Step 1: Discover and score URLs config = SeedingConfig( source="sitemap", extract_head=True, query="premier league opening fixtures", scoring_method="bm25", score_threshold=0.4, max_urls=10, verbose=True ) results = await seeder.urls(TEST_DOMAIN, config) print(f"\nStep 1: Found {len(results)} relevant URLs") # Step 2: Analyze top results if results: top_urls = results[:3] print("\nStep 2: Top 3 URLs for crawling:") for i, r in enumerate(top_urls): print(f"{i+1}. Score: {r['relevance_score']:.3f}") print(f" URL: {r['url']}") print(f" Title: {r['head_data'].get('title', 'N/A')}") # Check metadata quality meta = r['head_data'].get('meta', {}) if 'description' in meta: print(f" Description: {meta['description'][:80]}...") # Step 3: Verify these URLs would be good for actual crawling assert all(r["status"] == "valid" for r in results[:3]) print("\nStep 3: All top URLs are valid for crawling ✓") # ============================================ # Report Generation # ============================================ @pytest.mark.asyncio async def test_generate_scoring_report(self, seeder): """Generate a comprehensive report of BM25 scoring effectiveness.""" queries = { "beginner": "match schedule", "advanced": "tactical analysis pressing", "api": "VAR decision explanation", "deployment": "fixture changes due to weather", "extraction": "expected goals statistics" } report = { "timestamp": datetime.now().isoformat(), "domain": TEST_DOMAIN, "results": {} } for category, query in queries.items(): config = SeedingConfig( source="sitemap", extract_head=True, query=query, scoring_method="bm25", max_urls=10 ) results = await seeder.urls(TEST_DOMAIN, config) report["results"][category] = { "query": query, "total_results": len(results), "top_results": [ { "url": r["url"], "score": r["relevance_score"], "title": r["head_data"].get("title", "") } for r in results[:3] ], "score_distribution": { "min": min(r["relevance_score"] for r in results) if results else 0, "max": max(r["relevance_score"] for r in results) if results else 0, "avg": sum(r["relevance_score"] for r in results) / len(results) if results else 0 } } # Print report print("\n" + "="*60) print("BM25 SCORING EFFECTIVENESS REPORT") print("="*60) print(f"Domain: {report['domain']}") print(f"Timestamp: {report['timestamp']}") print("\nResults by Category:") for category, data in report["results"].items(): print(f"\n{category.upper()}: '{data['query']}'") print(f" Total results: {data['total_results']}") print(f" Score range: {data['score_distribution']['min']:.3f} - {data['score_distribution']['max']:.3f}") print(f" Average score: {data['score_distribution']['avg']:.3f}") print(" Top matches:") for i, result in enumerate(data['top_results']): print(f" {i+1}. [{result['score']:.3f}] {result['title']}") # ============================================ # Standalone test runner # ============================================ async def run_all_tests(): """Run all tests standalone (without pytest).""" print("Running AsyncUrlSeeder BM25 Tests...") print("="*60) test_instance = TestAsyncUrlSeederBM25() seeder = await test_instance.create_seeder() # Run each test method test_methods = [ # test_instance.test_basic_bm25_scoring, # test_instance.test_query_variations, # test_instance.test_score_threshold_filtering, # test_instance.test_extreme_thresholds, # test_instance.test_comprehensive_metadata_extraction, # test_instance.test_jsonld_extraction_scoring, # test_instance.test_empty_query, # test_instance.test_query_without_extract_head, # test_instance.test_special_characters_in_query, # test_instance.test_unicode_query, # test_instance.test_large_scale_scoring, # test_instance.test_concurrent_scoring_consistency, # test_instance.test_many_urls_with_scoring, test_instance.test_multi_word_complex_queries, test_instance.test_scoring_with_cache, test_instance.test_force_refresh_scoring, test_instance.test_scoring_with_multiple_sources, test_instance.test_full_workflow_integration, test_instance.test_generate_scoring_report ] for test_method in test_methods: try: print(f"\nRunning {test_method.__name__}...") await test_method(seeder) print(f"✓ {test_method.__name__} passed") except Exception as e: import traceback print(f"✗ {test_method.__name__} failed: {str(e)}") print(f" Error type: {type(e).__name__}") traceback.print_exc() print("\n" + "="*60) print("Test suite completed!") if __name__ == "__main__": # Run tests directly asyncio.run(run_all_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_download_file.py
tests/general/test_download_file.py
import asyncio from crawl4ai import CrawlerRunConfig, AsyncWebCrawler, BrowserConfig from pathlib import Path import os async def test_basic_download(): # Custom folder (otherwise defaults to ~/.crawl4ai/downloads) downloads_path = os.path.join(Path.home(), ".crawl4ai", "downloads") os.makedirs(downloads_path, exist_ok=True) browser_config = BrowserConfig( accept_downloads=True, downloads_path=downloads_path ) async with AsyncWebCrawler(config=browser_config) as crawler: run_config = CrawlerRunConfig( js_code=""" const link = document.querySelector('a[href$=".exe"]'); if (link) { link.click(); } """, delay_before_return_html=5 ) result = await crawler.arun("https://www.python.org/downloads/", config=run_config) if result.downloaded_files: print("Downloaded files:") for file_path in result.downloaded_files: print("•", file_path) else: print("No files downloaded.") if __name__ == "__main__": asyncio.run(test_basic_download())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_mhtml.py
tests/general/test_mhtml.py
# test_mhtml_capture.py import pytest import asyncio import re # For more robust MHTML checks # Assuming these can be imported directly from the crawl4ai library from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, CrawlResult # A reliable, simple static HTML page for testing # Using httpbin as it's designed for testing clients TEST_URL_SIMPLE = "https://httpbin.org/html" EXPECTED_CONTENT_SIMPLE = "Herman Melville - Moby-Dick" # A slightly more complex page that might involve JS (good secondary test) TEST_URL_JS = "https://quotes.toscrape.com/js/" EXPECTED_CONTENT_JS = "Quotes to Scrape" # Title of the page, which should be present in MHTML # Removed the custom event_loop fixture as pytest-asyncio provides a default one. @pytest.mark.asyncio async def test_mhtml_capture_when_enabled(): """ Verify that when CrawlerRunConfig has capture_mhtml=True, the CrawlResult contains valid MHTML content. """ # Create a fresh browser config and crawler instance for this test browser_config = BrowserConfig(headless=True) # Use headless for testing CI/CD # --- Key: Enable MHTML capture in the run config --- run_config = CrawlerRunConfig(capture_mhtml=True) # Create a fresh crawler instance crawler = AsyncWebCrawler(config=browser_config) try: # Start the browser await crawler.start() # Perform the crawl with the MHTML-enabled config result: CrawlResult = await crawler.arun(TEST_URL_SIMPLE, config=run_config) # --- Assertions --- assert result is not None, "Crawler should return a result object" assert result.success is True, f"Crawling {TEST_URL_SIMPLE} should succeed. Error: {result.error_message}" # 1. Check if the mhtml attribute exists (will fail if CrawlResult not updated) assert hasattr(result, 'mhtml'), "CrawlResult object must have an 'mhtml' attribute" # 2. Check if mhtml is populated assert result.mhtml is not None, "MHTML content should be captured when enabled" assert isinstance(result.mhtml, str), "MHTML content should be a string" assert len(result.mhtml) > 500, "MHTML content seems too short, likely invalid" # Basic sanity check # 3. Check for MHTML structure indicators (more robust than simple string contains) # MHTML files are multipart MIME messages assert re.search(r"Content-Type: multipart/related;", result.mhtml, re.IGNORECASE), \ "MHTML should contain 'Content-Type: multipart/related;'" # Should contain a boundary definition assert re.search(r"boundary=\"----MultipartBoundary", result.mhtml), \ "MHTML should contain a multipart boundary" # Should contain the main HTML part assert re.search(r"Content-Type: text/html", result.mhtml, re.IGNORECASE), \ "MHTML should contain a 'Content-Type: text/html' part" # 4. Check if the *actual page content* is within the MHTML string # This confirms the snapshot captured the rendered page assert EXPECTED_CONTENT_SIMPLE in result.mhtml, \ f"Expected content '{EXPECTED_CONTENT_SIMPLE}' not found within the captured MHTML" # 5. Ensure standard HTML is still present and correct assert result.html is not None, "Standard HTML should still be present" assert isinstance(result.html, str), "Standard HTML should be a string" assert EXPECTED_CONTENT_SIMPLE in result.html, \ f"Expected content '{EXPECTED_CONTENT_SIMPLE}' not found within the standard HTML" finally: # Important: Ensure browser is completely closed even if assertions fail await crawler.close() # Help the garbage collector clean up crawler = None @pytest.mark.asyncio async def test_mhtml_capture_when_disabled_explicitly(): """ Verify that when CrawlerRunConfig explicitly has capture_mhtml=False, the CrawlResult.mhtml attribute is None. """ # Create a fresh browser config and crawler instance for this test browser_config = BrowserConfig(headless=True) # --- Key: Explicitly disable MHTML capture --- run_config = CrawlerRunConfig(capture_mhtml=False) # Create a fresh crawler instance crawler = AsyncWebCrawler(config=browser_config) try: # Start the browser await crawler.start() result: CrawlResult = await crawler.arun(TEST_URL_SIMPLE, config=run_config) assert result is not None assert result.success is True, f"Crawling {TEST_URL_SIMPLE} should succeed. Error: {result.error_message}" # 1. Check attribute existence (important for TDD start) assert hasattr(result, 'mhtml'), "CrawlResult object must have an 'mhtml' attribute" # 2. Check mhtml is None assert result.mhtml is None, "MHTML content should be None when explicitly disabled" # 3. Ensure standard HTML is still present assert result.html is not None assert EXPECTED_CONTENT_SIMPLE in result.html finally: # Important: Ensure browser is completely closed even if assertions fail await crawler.close() # Help the garbage collector clean up crawler = None @pytest.mark.asyncio async def test_mhtml_capture_when_disabled_by_default(): """ Verify that if capture_mhtml is not specified (using its default), the CrawlResult.mhtml attribute is None. (This assumes the default value for capture_mhtml in CrawlerRunConfig is False) """ # Create a fresh browser config and crawler instance for this test browser_config = BrowserConfig(headless=True) # --- Key: Use default run config --- run_config = CrawlerRunConfig() # Do not specify capture_mhtml # Create a fresh crawler instance crawler = AsyncWebCrawler(config=browser_config) try: # Start the browser await crawler.start() result: CrawlResult = await crawler.arun(TEST_URL_SIMPLE, config=run_config) assert result is not None assert result.success is True, f"Crawling {TEST_URL_SIMPLE} should succeed. Error: {result.error_message}" # 1. Check attribute existence assert hasattr(result, 'mhtml'), "CrawlResult object must have an 'mhtml' attribute" # 2. Check mhtml is None (assuming default is False) assert result.mhtml is None, "MHTML content should be None when using default config (assuming default=False)" # 3. Ensure standard HTML is still present assert result.html is not None assert EXPECTED_CONTENT_SIMPLE in result.html finally: # Important: Ensure browser is completely closed even if assertions fail await crawler.close() # Help the garbage collector clean up crawler = None # Optional: Add a test for a JS-heavy page if needed @pytest.mark.asyncio async def test_mhtml_capture_on_js_page_when_enabled(): """ Verify MHTML capture works on a page requiring JavaScript execution. """ # Create a fresh browser config and crawler instance for this test browser_config = BrowserConfig(headless=True) run_config = CrawlerRunConfig( capture_mhtml=True, # Add a small wait or JS execution if needed for the JS page to fully render # For quotes.toscrape.com/js/, it renders quickly, but a wait might be safer # wait_for_timeout=2000 # Example: wait up to 2 seconds js_code="await new Promise(r => setTimeout(r, 500));" # Small delay after potential load ) # Create a fresh crawler instance crawler = AsyncWebCrawler(config=browser_config) try: # Start the browser await crawler.start() result: CrawlResult = await crawler.arun(TEST_URL_JS, config=run_config) assert result is not None assert result.success is True, f"Crawling {TEST_URL_JS} should succeed. Error: {result.error_message}" assert hasattr(result, 'mhtml'), "CrawlResult object must have an 'mhtml' attribute" assert result.mhtml is not None, "MHTML content should be captured on JS page when enabled" assert isinstance(result.mhtml, str), "MHTML content should be a string" assert len(result.mhtml) > 500, "MHTML content from JS page seems too short" # Check for MHTML structure assert re.search(r"Content-Type: multipart/related;", result.mhtml, re.IGNORECASE) assert re.search(r"Content-Type: text/html", result.mhtml, re.IGNORECASE) # Check for content rendered by JS within the MHTML assert EXPECTED_CONTENT_JS in result.mhtml, \ f"Expected JS-rendered content '{EXPECTED_CONTENT_JS}' not found within the captured MHTML" # Check standard HTML too assert result.html is not None assert EXPECTED_CONTENT_JS in result.html, \ f"Expected JS-rendered content '{EXPECTED_CONTENT_JS}' not found within the standard HTML" finally: # Important: Ensure browser is completely closed even if assertions fail await crawler.close() # Help the garbage collector clean up crawler = None if __name__ == "__main__": # Use pytest for async tests pytest.main(["-xvs", __file__])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_cache_context.py
tests/general/test_cache_context.py
import asyncio from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, CacheMode from playwright.async_api import Page, BrowserContext async def test_reuse_context_by_config(): # We will store each context ID in these maps to confirm reuse context_ids_for_A = [] context_ids_for_B = [] # Create a small hook to track context creation async def on_page_context_created(page: Page, context: BrowserContext, config: CrawlerRunConfig, **kwargs): c_id = id(context) print(f"[HOOK] on_page_context_created - Context ID: {c_id}") # Distinguish which config we used by checking a custom hook param config_label = config.shared_data.get("config_label", "unknown") if config_label == "A": context_ids_for_A.append(c_id) elif config_label == "B": context_ids_for_B.append(c_id) return page # Browser config - Headless, verbose so we see logs browser_config = BrowserConfig(headless=True, verbose=True) # Two crawler run configs that differ (for example, text_mode): configA = CrawlerRunConfig( only_text=True, cache_mode=CacheMode.BYPASS, wait_until="domcontentloaded", shared_data = { "config_label" : "A" } ) configB = CrawlerRunConfig( only_text=False, cache_mode=CacheMode.BYPASS, wait_until="domcontentloaded", shared_data = { "config_label" : "B" } ) # Create the crawler crawler = AsyncWebCrawler(config=browser_config) # Attach our custom hook # Note: "on_page_context_created" will be called each time a new context+page is generated crawler.crawler_strategy.set_hook("on_page_context_created", on_page_context_created) # Start the crawler (launches the browser) await crawler.start() # For demonstration, we’ll crawl a benign site multiple times with each config test_url = "https://example.com" print("\n--- Crawling with config A (text_mode=True) ---") for _ in range(2): # Pass an extra kwarg to the hook so we know which config is being used await crawler.arun(test_url, config=configA) print("\n--- Crawling with config B (text_mode=False) ---") for _ in range(2): await crawler.arun(test_url, config=configB) # Close the crawler (shuts down the browser, closes contexts) await crawler.close() # Validate and show the results print("\n=== RESULTS ===") print(f"Config A context IDs: {context_ids_for_A}") print(f"Config B context IDs: {context_ids_for_B}") if len(set(context_ids_for_A)) == 1: print("✅ All config A crawls used the SAME BrowserContext.") else: print("❌ Config A crawls created multiple contexts unexpectedly.") if len(set(context_ids_for_B)) == 1: print("✅ All config B crawls used the SAME BrowserContext.") else: print("❌ Config B crawls created multiple contexts unexpectedly.") if set(context_ids_for_A).isdisjoint(context_ids_for_B): print("✅ Config A context is different from Config B context.") else: print("❌ A and B ended up sharing the same context somehow!") if __name__ == "__main__": asyncio.run(test_reuse_context_by_config())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_stream.py
tests/general/test_stream.py
import os, sys # append 2 parent directories to sys.path to import crawl4ai parent_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.append(parent_dir) parent_parent_dir = os.path.dirname(parent_dir) sys.path.append(parent_parent_dir) import asyncio from crawl4ai import * async def test_crawler(): # Setup configurations browser_config = BrowserConfig(headless=True, verbose=False) crawler_config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, markdown_generator=DefaultMarkdownGenerator( content_filter=PruningContentFilter( threshold=0.48, threshold_type="fixed", min_word_threshold=0 ) ), ) # Test URLs - mix of different sites urls = [ "http://example.com", "http://example.org", "http://example.net", ] * 10 # 15 total URLs async with AsyncWebCrawler(config=browser_config) as crawler: print("\n=== Testing Streaming Mode ===") async for result in await crawler.arun_many( urls=urls, config=crawler_config.clone(stream=True), ): print(f"Received result for: {result.url} - Success: {result.success}") print("\n=== Testing Batch Mode ===") results = await crawler.arun_many( urls=urls, config=crawler_config, ) print(f"Received all {len(results)} results at once") for result in results: print(f"Batch result for: {result.url} - Success: {result.success}") if __name__ == "__main__": asyncio.run(test_crawler())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_http_crawler_strategy.py
tests/general/test_http_crawler_strategy.py
from tkinter import N from crawl4ai.async_crawler_strategy import AsyncHTTPCrawlerStrategy from crawl4ai.async_logger import AsyncLogger from crawl4ai import CrawlerRunConfig, HTTPCrawlerConfig from crawl4ai.async_crawler_strategy import ConnectionTimeoutError import asyncio import os async def main(): """Test the AsyncHTTPCrawlerStrategy with various scenarios""" logger = AsyncLogger(verbose=True) # Initialize the strategy with default HTTPCrawlerConfig crawler = AsyncHTTPCrawlerStrategy( browser_config=HTTPCrawlerConfig(), logger=logger ) # Test 1: Basic HTTP GET print("\n=== Test 1: Basic HTTP GET ===") result = await crawler.crawl("https://example.com") print(f"Status: {result.status_code}") print(f"Content length: {len(result.html)}") print(f"Headers: {dict(result.response_headers)}") # Test 2: POST request with JSON print("\n=== Test 2: POST with JSON ===") crawler.browser_config = crawler.browser_config.clone( method="POST", json={"test": "data"}, headers={"Content-Type": "application/json"} ) try: result = await crawler.crawl( "https://httpbin.org/post", ) print(f"Status: {result.status_code}") print(f"Response: {result.html[:200]}...") except Exception as e: print(f"Error: {e}") # Test 3: File handling crawler.browser_config = HTTPCrawlerConfig() print("\n=== Test 3: Local file handling ===") # Create a tmp file with test content from tempfile import NamedTemporaryFile with NamedTemporaryFile(delete=False) as f: f.write(b"<html><body>Test content</body></html>") f.close() result = await crawler.crawl(f"file://{f.name}") print(f"File content: {result.html}") # Test 4: Raw content print("\n=== Test 4: Raw content handling ===") raw_html = "raw://<html><body>Raw test content</body></html>" result = await crawler.crawl(raw_html) print(f"Raw content: {result.html}") # Test 5: Custom hooks print("\n=== Test 5: Custom hooks ===") async def before_request(url, kwargs): print(f"Before request to {url}") kwargs['headers']['X-Custom'] = 'test' async def after_request(response): print(f"After request, status: {response.status_code}") crawler.set_hook('before_request', before_request) crawler.set_hook('after_request', after_request) result = await crawler.crawl("https://example.com") # Test 6: Error handling print("\n=== Test 6: Error handling ===") try: await crawler.crawl("https://nonexistent.domain.test") except Exception as e: print(f"Expected error: {e}") # Test 7: Redirects print("\n=== Test 7: Redirect handling ===") crawler.browser_config = HTTPCrawlerConfig(follow_redirects=True) result = await crawler.crawl("http://httpbin.org/redirect/1") print(f"Final URL: {result.redirected_url}") # Test 8: Custom timeout print("\n=== Test 8: Custom timeout ===") try: await crawler.crawl( "https://httpbin.org/delay/5", config=CrawlerRunConfig(page_timeout=2) ) except ConnectionTimeoutError as e: print(f"Expected timeout: {e}") # Test 9: SSL verification print("\n=== Test 9: SSL verification ===") crawler.browser_config = HTTPCrawlerConfig(verify_ssl=False) try: await crawler.crawl("https://expired.badssl.com/") print("Connected to invalid SSL site with verification disabled") except Exception as e: print(f"SSL error: {e}") # Test 10: Large file streaming print("\n=== Test 10: Large file streaming ===") from tempfile import NamedTemporaryFile with NamedTemporaryFile(delete=False) as f: f.write(b"<html><body>" + b"X" * 1024 * 1024 * 10 + b"</body></html>") f.close() result = await crawler.crawl("file://" + f.name) print(f"Large file content length: {len(result.html)}") os.remove(f.name) crawler.close() if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/general/test_async_crawler_strategy.py
tests/general/test_async_crawler_strategy.py
import pytest import pytest_asyncio import asyncio from typing import Dict, Any from pathlib import Path from unittest.mock import MagicMock, patch import os from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_crawler_strategy import AsyncPlaywrightCrawlerStrategy from crawl4ai.models import AsyncCrawlResponse from crawl4ai.async_logger import AsyncLogger, LogLevel CRAWL4AI_HOME_DIR = Path(os.path.expanduser("~")).joinpath(".crawl4ai") if not CRAWL4AI_HOME_DIR.joinpath("profiles", "test_profile").exists(): CRAWL4AI_HOME_DIR.joinpath("profiles", "test_profile").mkdir(parents=True) @pytest.fixture def basic_html(): return """ <html lang="en"> <head> <title>Basic HTML</title> </head> <body> <h1>Main Heading</h1> <main> <div class="container"> <p>Basic HTML document for testing purposes.</p> </div> </main> </body> </html> """ # Test Config Files @pytest.fixture def basic_browser_config(): return BrowserConfig( browser_type="chromium", headless=True, verbose=True ) @pytest.fixture def advanced_browser_config(): return BrowserConfig( browser_type="chromium", headless=True, use_managed_browser=True, user_data_dir=CRAWL4AI_HOME_DIR.joinpath("profiles", "test_profile"), # proxy="http://localhost:8080", viewport_width=1920, viewport_height=1080, user_agent_mode="random" ) @pytest.fixture def basic_crawler_config(): return CrawlerRunConfig( word_count_threshold=100, wait_until="domcontentloaded", page_timeout=30000 ) @pytest.fixture def logger(): return AsyncLogger(verbose=True, log_level=LogLevel.DEBUG) @pytest_asyncio.fixture async def crawler_strategy(basic_browser_config, logger): strategy = AsyncPlaywrightCrawlerStrategy(browser_config=basic_browser_config, logger=logger) await strategy.start() yield strategy await strategy.close() # Browser Configuration Tests @pytest.mark.asyncio async def test_browser_config_initialization(): config = BrowserConfig( browser_type="chromium", user_agent_mode="random" ) assert config.browser_type == "chromium" assert config.user_agent is not None assert config.headless is True @pytest.mark.asyncio async def test_persistent_browser_config(): config = BrowserConfig( use_persistent_context=True, user_data_dir="/tmp/test_dir" ) assert config.use_managed_browser is True assert config.user_data_dir == "/tmp/test_dir" # Crawler Strategy Tests @pytest.mark.asyncio async def test_basic_page_load(crawler_strategy): response = await crawler_strategy.crawl( "https://example.com", CrawlerRunConfig() ) assert response.status_code == 200 assert len(response.html) > 0 assert "Example Domain" in response.html @pytest.mark.asyncio async def test_screenshot_capture(crawler_strategy): config = CrawlerRunConfig(screenshot=True) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.screenshot is not None assert len(response.screenshot) > 0 @pytest.mark.asyncio async def test_pdf_generation(crawler_strategy): config = CrawlerRunConfig(pdf=True) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.pdf_data is not None assert len(response.pdf_data) > 0 @pytest.mark.asyncio async def test_handle_js_execution(crawler_strategy): config = CrawlerRunConfig( js_code="document.body.style.backgroundColor = 'red';" ) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 assert 'background-color: red' in response.html.lower() @pytest.mark.asyncio async def test_multiple_js_commands(crawler_strategy): js_commands = [ "document.body.style.backgroundColor = 'blue';", "document.title = 'Modified Title';", "const div = document.createElement('div'); div.id = 'test'; div.textContent = 'Test Content'; document.body.appendChild(div);" ] config = CrawlerRunConfig(js_code=js_commands) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 assert 'background-color: blue' in response.html.lower() assert 'id="test"' in response.html assert '>Test Content<' in response.html assert '<title>Modified Title</title>' in response.html @pytest.mark.asyncio async def test_complex_dom_manipulation(crawler_strategy): js_code = """ // Create a complex structure const container = document.createElement('div'); container.className = 'test-container'; const list = document.createElement('ul'); list.className = 'test-list'; for (let i = 1; i <= 3; i++) { const item = document.createElement('li'); item.textContent = `Item ${i}`; item.className = `item-${i}`; list.appendChild(item); } container.appendChild(list); document.body.appendChild(container); """ config = CrawlerRunConfig(js_code=js_code) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 assert 'class="test-container"' in response.html assert 'class="test-list"' in response.html assert 'class="item-1"' in response.html assert '>Item 1<' in response.html assert '>Item 2<' in response.html assert '>Item 3<' in response.html @pytest.mark.asyncio async def test_style_modifications(crawler_strategy): js_code = """ const testDiv = document.createElement('div'); testDiv.id = 'style-test'; testDiv.style.cssText = 'color: green; font-size: 20px; margin: 10px;'; testDiv.textContent = 'Styled Content'; document.body.appendChild(testDiv); """ config = CrawlerRunConfig(js_code=js_code) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 assert 'id="style-test"' in response.html assert 'color: green' in response.html.lower() assert 'font-size: 20px' in response.html.lower() assert 'margin: 10px' in response.html.lower() assert '>Styled Content<' in response.html @pytest.mark.asyncio async def test_dynamic_content_loading(crawler_strategy): js_code = """ // Simulate dynamic content loading setTimeout(() => { const dynamic = document.createElement('div'); dynamic.id = 'dynamic-content'; dynamic.textContent = 'Dynamically Loaded'; document.body.appendChild(dynamic); }, 1000); // Add a loading indicator immediately const loading = document.createElement('div'); loading.id = 'loading'; loading.textContent = 'Loading...'; document.body.appendChild(loading); """ config = CrawlerRunConfig(js_code=js_code, delay_before_return_html=2.0) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 assert 'id="loading"' in response.html assert '>Loading...</' in response.html assert 'dynamic-content' in response.html assert '>Dynamically Loaded<' in response.html # @pytest.mark.asyncio # async def test_js_return_values(crawler_strategy): # js_code = """ # return { # title: document.title, # metaCount: document.getElementsByTagName('meta').length, # bodyClass: document.body.className # }; # """ # config = CrawlerRunConfig(js_code=js_code) # response = await crawler_strategy.crawl( # "https://example.com", # config # ) # assert response.status_code == 200 # assert 'Example Domain' in response.html # assert 'meta name="viewport"' in response.html # assert 'class="main"' in response.html @pytest.mark.asyncio async def test_async_js_execution(crawler_strategy): js_code = """ await new Promise(resolve => setTimeout(resolve, 1000)); document.body.style.color = 'green'; const computedStyle = window.getComputedStyle(document.body); return computedStyle.color; """ config = CrawlerRunConfig(js_code=js_code) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 assert 'color: green' in response.html.lower() # @pytest.mark.asyncio # async def test_js_error_handling(crawler_strategy): # js_code = """ # // Intentionally cause different types of errors # const results = []; # try { # nonExistentFunction(); # } catch (e) { # results.push(e.name); # } # try { # JSON.parse('{invalid}'); # } catch (e) { # results.push(e.name); # } # return results; # """ # config = CrawlerRunConfig(js_code=js_code) # response = await crawler_strategy.crawl( # "https://example.com", # config # ) # assert response.status_code == 200 # assert 'ReferenceError' in response.html # assert 'SyntaxError' in response.html @pytest.mark.asyncio async def test_handle_navigation_timeout(): config = CrawlerRunConfig(page_timeout=1) # 1ms timeout with pytest.raises(Exception): async with AsyncPlaywrightCrawlerStrategy() as strategy: await strategy.crawl("https://example.com", config) @pytest.mark.asyncio async def test_session_management(crawler_strategy): config = CrawlerRunConfig(session_id="test_session") response1 = await crawler_strategy.crawl( "https://example.com", config ) response2 = await crawler_strategy.crawl( "https://example.com", config ) assert response1.status_code == 200 assert response2.status_code == 200 @pytest.mark.asyncio async def test_process_iframes(crawler_strategy): config = CrawlerRunConfig( process_iframes=True, wait_for_images=True ) response = await crawler_strategy.crawl( "https://example.com", config ) assert response.status_code == 200 @pytest.mark.asyncio async def test_stealth_mode(crawler_strategy): config = CrawlerRunConfig( simulate_user=True, override_navigator=True ) response = await crawler_strategy.crawl( "https://bot.sannysoft.com", config ) assert response.status_code == 200 @pytest.mark.asyncio @pytest.mark.parametrize("prefix", ("raw:", "raw://")) async def test_raw_urls(crawler_strategy, basic_html, prefix): url = f"{prefix}{basic_html}" response = await crawler_strategy.crawl(url, CrawlerRunConfig()) assert response.html == basic_html # Error Handling Tests @pytest.mark.asyncio async def test_invalid_url(): with pytest.raises(ValueError): async with AsyncPlaywrightCrawlerStrategy() as strategy: await strategy.crawl("not_a_url", CrawlerRunConfig()) @pytest.mark.asyncio async def test_network_error_handling(): config = CrawlerRunConfig() with pytest.raises(Exception): async with AsyncPlaywrightCrawlerStrategy() as strategy: await strategy.crawl("https://invalid.example.com", config) @pytest.mark.asyncio async def test_remove_overlay_elements(crawler_strategy): config = CrawlerRunConfig( remove_overlay_elements=True, delay_before_return_html=5, ) response = await crawler_strategy.crawl( "https://www2.hm.com/en_us/index.html", config ) assert response.status_code == 200 assert "Accept all cookies" not in response.html if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/hub/test_simple.py
tests/hub/test_simple.py
# test.py from crawl4ai import CrawlerHub import json async def amazon_example(): if (crawler_cls := CrawlerHub.get("amazon_product")) : crawler = crawler_cls() print(f"Crawler version: {crawler_cls.meta['version']}") print(f"Rate limits: {crawler_cls.meta.get('rate_limit', 'Unlimited')}") print(await crawler.run("https://amazon.com/test")) else: print("Crawler not found!") async def google_example(): # Get crawler dynamically crawler_cls = CrawlerHub.get("google_search") crawler = crawler_cls() # Text search text_results = await crawler.run( query="apple inc", search_type="text", schema_cache_path="/Users/unclecode/.crawl4ai" ) print(json.dumps(json.loads(text_results), indent=4)) # Image search # image_results = await crawler.run(query="apple inc", search_type="image") # print(image_results) if __name__ == "__main__": import asyncio # asyncio.run(amazon_example()) asyncio.run(google_example())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/deep_crwaling/test_filter.py
tests/deep_crwaling/test_filter.py
# // File: tests/deep_crawling/test_filters.py import pytest from urllib.parse import urlparse from crawl4ai import ContentTypeFilter, URLFilter # Minimal URLFilter base class stub if not already importable directly for tests # In a real scenario, this would be imported from the library if not hasattr(URLFilter, '_update_stats'): # Check if it's a basic stub class URLFilter: # Basic stub for testing if needed def __init__(self, name=None): self.name = name def apply(self, url: str) -> bool: raise NotImplementedError def _update_stats(self, passed: bool): pass # Mock implementation # Assume ContentTypeFilter is structured as discussed. If its definition is not fully # available for direct import in the test environment, a more elaborate stub or direct # instantiation of the real class (if possible) would be needed. # For this example, we assume ContentTypeFilter can be imported and used. class TestContentTypeFilter: @pytest.mark.parametrize( "url, allowed_types, expected", [ # Existing tests (examples) ("http://example.com/page.html", ["text/html"], True), ("http://example.com/page.json", ["application/json"], True), ("http://example.com/image.png", ["text/html"], False), ("http://example.com/document.pdf", ["application/pdf"], True), ("http://example.com/page", ["text/html"], True), # No extension, allowed ("http://example.com/page", ["text/html"], False), # No extension, disallowed ("http://example.com/page.unknown", ["text/html"], False), # Unknown extension # Tests for PHP extensions ("http://example.com/index.php", ["application/x-httpd-php"], True), ("http://example.com/script.php3", ["application/x-httpd-php"], True), ("http://example.com/legacy.php4", ["application/x-httpd-php"], True), ("http://example.com/main.php5", ["application/x-httpd-php"], True), ("http://example.com/api.php7", ["application/x-httpd-php"], True), ("http://example.com/index.phtml", ["application/x-httpd-php"], True), ("http://example.com/source.phps", ["application/x-httpd-php-source"], True), # Test rejection of PHP extensions ("http://example.com/index.php", ["text/html"], False), ("http://example.com/script.php3", ["text/plain"], False), ("http://example.com/source.phps", ["application/x-httpd-php"], False), # Mismatch MIME ("http://example.com/source.php", ["application/x-httpd-php-source"], False), # Mismatch MIME for .php # Test case-insensitivity of extensions in URL ("http://example.com/PAGE.HTML", ["text/html"], True), ("http://example.com/INDEX.PHP", ["application/x-httpd-php"], True), ("http://example.com/SOURCE.PHPS", ["application/x-httpd-php-source"], True), # Test case-insensitivity of allowed_types ("http://example.com/index.php", ["APPLICATION/X-HTTPD-PHP"], True), ], ) def test_apply(self, url, allowed_types, expected): content_filter = ContentTypeFilter( allowed_types=allowed_types ) assert content_filter.apply(url) == expected @pytest.mark.parametrize( "url, expected_extension", [ ("http://example.com/file.html", "html"), ("http://example.com/file.tar.gz", "gz"), ("http://example.com/path/", ""), ("http://example.com/nodot", ""), ("http://example.com/.config", "config"), # hidden file with extension ("http://example.com/path/to/archive.BIG.zip", "zip"), # Case test ] ) def test_extract_extension(self, url, expected_extension): # Test the static method directly assert ContentTypeFilter._extract_extension(url) == expected_extension
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/unit/test_sitemap_namespace_parsing.py
tests/unit/test_sitemap_namespace_parsing.py
import sys from types import SimpleNamespace import pytest # Provide a lightweight stub for rank_bm25 before importing the seeder to avoid # optional dependency issues (e.g., incompatible wheels in CI). class _FakeBM25: def __init__(self, corpus): self._scores = [1.0] * len(corpus) def get_scores(self, tokens): return self._scores sys.modules.setdefault("rank_bm25", SimpleNamespace(BM25Okapi=_FakeBM25)) from crawl4ai.async_url_seeder import AsyncUrlSeeder class DummyResponse: def __init__(self, request_url: str, text: str): self.status_code = 200 self._content = text.encode("utf-8") self.url = request_url def raise_for_status(self): return None @property def content(self): return self._content @property def text(self): return self._content.decode("utf-8") class DummyAsyncClient: def __init__(self, response_map): self._responses = response_map async def get(self, url, **kwargs): payload = self._responses[url] if callable(payload): payload = payload() return DummyResponse(url, payload) @pytest.mark.asyncio async def test_iter_sitemap_handles_namespace_less_sitemaps(): xml = """<?xml version="1.0"?> <urlset> <url><loc>https://example.com/a</loc></url> <url><loc>https://example.com/b</loc></url> </urlset> """ seeder = AsyncUrlSeeder(client=DummyAsyncClient({"https://example.com/sitemap.xml": xml})) urls = [] async for u in seeder._iter_sitemap("https://example.com/sitemap.xml"): urls.append(u) assert urls == ["https://example.com/a", "https://example.com/b"] @pytest.mark.asyncio async def test_iter_sitemap_handles_custom_namespace(): xml = """<?xml version="1.0"?> <urlset xmlns="https://custom.namespace/schema"> <url><loc>https://example.com/ns</loc></url> </urlset> """ seeder = AsyncUrlSeeder(client=DummyAsyncClient({"https://example.com/ns-sitemap.xml": xml})) urls = [] async for u in seeder._iter_sitemap("https://example.com/ns-sitemap.xml"): urls.append(u) assert urls == ["https://example.com/ns"] @pytest.mark.asyncio async def test_iter_sitemap_handles_namespace_index_and_children(): index_xml = """<?xml version="1.0"?> <sitemapindex xmlns="http://another.example/ns"> <sitemap> <loc>https://example.com/child-1.xml</loc> </sitemap> <sitemap> <loc>https://example.com/child-2.xml</loc> </sitemap> </sitemapindex> """ child_xml = """<?xml version="1.0"?> <urlset xmlns="http://irrelevant"> <url><loc>https://example.com/page-{n}</loc></url> </urlset> """ responses = { "https://example.com/index.xml": index_xml, "https://example.com/child-1.xml": child_xml.format(n=1), "https://example.com/child-2.xml": child_xml.format(n=2), } seeder = AsyncUrlSeeder(client=DummyAsyncClient(responses)) urls = [] async for u in seeder._iter_sitemap("https://example.com/index.xml"): urls.append(u) assert sorted(urls) == [ "https://example.com/page-1", "https://example.com/page-2", ] @pytest.mark.asyncio async def test_iter_sitemap_normalizes_relative_locations(): xml = """<?xml version="1.0"?> <urlset> <url><loc>/relative-path</loc></url> <url><loc>https://example.com/absolute</loc></url> </urlset> """ seeder = AsyncUrlSeeder(client=DummyAsyncClient({"https://example.com/sitemap.xml": xml})) urls = [] async for u in seeder._iter_sitemap("https://example.com/sitemap.xml"): urls.append(u) assert urls == [ "https://example.com/relative-path", "https://example.com/absolute", ]
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/cli/test_cli.py
tests/cli/test_cli.py
import pytest from click.testing import CliRunner from pathlib import Path import json import yaml from crawl4ai.cli import cli, load_config_file, parse_key_values import tempfile import os import click @pytest.fixture def runner(): return CliRunner() @pytest.fixture def temp_config_dir(): with tempfile.TemporaryDirectory() as tmpdir: old_home = os.environ.get('HOME') os.environ['HOME'] = tmpdir yield Path(tmpdir) if old_home: os.environ['HOME'] = old_home @pytest.fixture def sample_configs(temp_config_dir): configs = { 'browser.yml': { 'headless': True, 'viewport_width': 1280, 'user_agent_mode': 'random' }, 'crawler.yml': { 'cache_mode': 'bypass', 'wait_until': 'networkidle', 'scan_full_page': True }, 'extract_css.yml': { 'type': 'json-css', 'params': {'verbose': True} }, 'css_schema.json': { 'name': 'ArticleExtractor', 'baseSelector': '.article', 'fields': [ {'name': 'title', 'selector': 'h1.title', 'type': 'text'}, {'name': 'link', 'selector': 'a.read-more', 'type': 'attribute', 'attribute': 'href'} ] } } for filename, content in configs.items(): path = temp_config_dir / filename with open(path, 'w') as f: if filename.endswith('.yml'): yaml.dump(content, f) else: json.dump(content, f) return {name: str(temp_config_dir / name) for name in configs} class TestCLIBasics: def test_help(self, runner): result = runner.invoke(cli, ['--help']) assert result.exit_code == 0 assert 'Crawl4AI CLI' in result.output def test_examples(self, runner): result = runner.invoke(cli, ['--example']) assert result.exit_code == 0 assert 'Examples' in result.output def test_missing_url(self, runner): result = runner.invoke(cli) assert result.exit_code != 0 assert 'URL argument is required' in result.output class TestConfigParsing: def test_parse_key_values_basic(self): result = parse_key_values(None, None, "key1=value1,key2=true") assert result == {'key1': 'value1', 'key2': True} def test_parse_key_values_invalid(self): with pytest.raises(click.BadParameter): parse_key_values(None, None, "invalid_format") class TestConfigLoading: def test_load_yaml_config(self, sample_configs): config = load_config_file(sample_configs['browser.yml']) assert config['headless'] is True assert config['viewport_width'] == 1280 def test_load_json_config(self, sample_configs): config = load_config_file(sample_configs['css_schema.json']) assert config['name'] == 'ArticleExtractor' assert len(config['fields']) == 2 def test_load_nonexistent_config(self): with pytest.raises(click.BadParameter): load_config_file('nonexistent.yml') class TestLLMConfig: def test_llm_config_creation(self, temp_config_dir, runner): def input_simulation(inputs): return runner.invoke(cli, ['https://example.com', '-q', 'test question'], input='\n'.join(inputs)) class TestCrawlingFeatures: def test_basic_crawl(self, runner): result = runner.invoke(cli, ['https://example.com']) assert result.exit_code == 0 class TestErrorHandling: def test_invalid_config_file(self, runner): result = runner.invoke(cli, [ 'https://example.com', '--browser-config', 'nonexistent.yml' ]) assert result.exit_code != 0 def test_invalid_schema(self, runner, temp_config_dir): invalid_schema = temp_config_dir / 'invalid_schema.json' with open(invalid_schema, 'w') as f: f.write('invalid json') result = runner.invoke(cli, [ 'https://example.com', '--schema', str(invalid_schema) ]) assert result.exit_code != 0 if __name__ == '__main__': pytest.main(['-v', '-s', '--tb=native', __file__])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/releases/test_release_0.6.4.py
tests/releases/test_release_0.6.4.py
import pytest import asyncio import time from crawl4ai import AsyncWebCrawler, CrawlerRunConfig, BrowserConfig, CacheMode @pytest.mark.asyncio async def test_wait_for_timeout_separate_from_page_timeout(): """Test that wait_for has its own timeout separate from page_timeout""" browser_config = BrowserConfig(headless=True) # Test with short wait_for_timeout but longer page_timeout config = CrawlerRunConfig( wait_for="css:.nonexistent-element", wait_for_timeout=2000, # 2 seconds page_timeout=10000, # 10 seconds cache_mode=CacheMode.BYPASS ) async with AsyncWebCrawler(config=browser_config) as crawler: start_time = time.time() result = await crawler.arun("https://example.com", config=config) elapsed = time.time() - start_time # Should timeout after ~2 seconds (wait_for_timeout), not 10 seconds assert elapsed < 5, f"Expected timeout around 2s, but took {elapsed:.2f}s" assert result.success, "Crawl should still succeed even if wait_for times out" @pytest.mark.asyncio async def test_wait_for_timeout_with_existing_element(): """Test that wait_for_timeout works correctly when element exists""" browser_config = BrowserConfig(headless=True) config = CrawlerRunConfig( wait_for="css:body", # This should exist quickly wait_for_timeout=5000, cache_mode=CacheMode.BYPASS ) async with AsyncWebCrawler(config=browser_config) as crawler: start_time = time.time() result = await crawler.arun("https://example.com", config=config) elapsed = time.time() - start_time # Should complete quickly since body element exists assert elapsed < 3, f"Expected quick completion, but took {elapsed:.2f}s" assert result.success assert "<body" in result.html @pytest.mark.asyncio async def test_javascript_wait_for_timeout(): """Test wait_for_timeout with JavaScript condition""" browser_config = BrowserConfig(headless=True) config = CrawlerRunConfig( wait_for="js:() => window.nonExistentVariable === true", wait_for_timeout=2000, cache_mode=CacheMode.BYPASS ) async with AsyncWebCrawler(config=browser_config) as crawler: start_time = time.time() result = await crawler.arun("https://example.com", config=config) elapsed = time.time() - start_time # Should timeout after ~2 seconds assert elapsed < 4, f"Expected timeout around 2s, but took {elapsed:.2f}s" assert result.success @pytest.mark.asyncio async def test_google_analytics_integration(): """Test that Google Analytics scripts are properly integrated""" browser_config = BrowserConfig(headless=True) config = CrawlerRunConfig(cache_mode=CacheMode.BYPASS) # Test with a simple HTML page that we can control html_content = """ <!DOCTYPE html> <html> <head> <title>Test GA Integration</title> <!-- Google tag (gtag.js) --> <script async src="https://www.googletagmanager.com/gtag/js?id=GA_MEASUREMENT_ID"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('config', 'GA_MEASUREMENT_ID'); </script> </head> <body> <h1>Test Page</h1> <p>Testing Google Analytics integration</p> </body> </html> """ async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun(f"raw://{html_content}", config=config) assert result.success # Check that GA scripts are preserved in the HTML assert "googletagmanager.com/gtag/js" in result.html assert "dataLayer" in result.html assert "gtag('config'" in result.html @pytest.mark.asyncio async def test_mkdocs_no_duplicate_gtag(): """Test that there are no duplicate gtag.js entries in documentation""" browser_config = BrowserConfig(headless=True) config = CrawlerRunConfig(cache_mode=CacheMode.BYPASS) # Simulate MkDocs-like HTML structure html_content = """ <!DOCTYPE html> <html> <head> <title>Crawl4AI Documentation</title> <script async src="https://www.googletagmanager.com/gtag/js?id=G-XXXXXXXXXX"></script> <script> window.dataLayer = window.dataLayer || []; function gtag(){dataLayer.push(arguments);} gtag('config', 'G-XXXXXXXXXX'); </script> </head> <body> <h1>Crawl4AI Documentation</h1> <p>Welcome to the documentation</p> </body> </html> """ async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun(f"raw://{html_content}", config=config) assert result.success # Count occurrences of gtag.js to ensure no duplicates gtag_count = result.html.count("googletagmanager.com/gtag/js") assert gtag_count <= 1, f"Found {gtag_count} gtag.js scripts, expected at most 1" # Ensure the analytics functionality is still there if gtag_count == 1: assert "dataLayer" in result.html assert "gtag('config'" in result.html if __name__ == "__main__": pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/releases/test_release_0.7.0.py
tests/releases/test_release_0.7.0.py
#!/usr/bin/env python3 import asyncio import pytest import os import json import tempfile from pathlib import Path from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, CacheMode from crawl4ai import JsonCssExtractionStrategy, LLMExtractionStrategy, LLMConfig from crawl4ai.content_filter_strategy import BM25ContentFilter from crawl4ai.markdown_generation_strategy import DefaultMarkdownGenerator from crawl4ai.async_url_seeder import AsyncUrlSeeder from crawl4ai.utils import RobotsParser class TestCrawl4AIv070: """Test suite for Crawl4AI v0.7.0 changes""" @pytest.mark.asyncio async def test_raw_url_parsing(self): """Test raw:// URL parsing logic fix""" html_content = "<html><body><h1>Test Content</h1><p>This is a test paragraph.</p></body></html>" async with AsyncWebCrawler() as crawler: # Test raw:// prefix result1 = await crawler.arun(f"raw://{html_content}") assert result1.success assert "Test Content" in result1.markdown # Test raw: prefix result2 = await crawler.arun(f"raw:{html_content}") assert result2.success assert "Test Content" in result2.markdown @pytest.mark.asyncio async def test_max_pages_limit_batch_processing(self): """Test max_pages limit is respected during batch processing""" urls = [ "https://httpbin.org/html", "https://httpbin.org/json", "https://httpbin.org/xml" ] config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, max_pages=2 ) async with AsyncWebCrawler() as crawler: results = await crawler.arun_many(urls, config=config) # Should only process 2 pages due to max_pages limit successful_results = [r for r in results if r.success] assert len(successful_results) <= 2 @pytest.mark.asyncio async def test_navigation_abort_handling(self): """Test handling of navigation aborts during file downloads""" async with AsyncWebCrawler() as crawler: # Test with a URL that might cause navigation issues result = await crawler.arun( "https://httpbin.org/status/404", config=CrawlerRunConfig(cache_mode=CacheMode.BYPASS) ) # Should not crash even with navigation issues assert result is not None @pytest.mark.asyncio async def test_screenshot_capture_fix(self): """Test screenshot capture improvements""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, screenshot=True ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success assert result.screenshot is not None assert len(result.screenshot) > 0 @pytest.mark.asyncio async def test_redirect_status_codes(self): """Test that real redirect status codes are surfaced""" async with AsyncWebCrawler() as crawler: # Test with a redirect URL result = await crawler.arun( "https://httpbin.org/redirect/1", config=CrawlerRunConfig(cache_mode=CacheMode.BYPASS) ) assert result.success # Should have redirect information assert result.status_code in [200, 301, 302, 303, 307, 308] @pytest.mark.asyncio async def test_local_file_processing(self): """Test local file processing with captured_console initialization""" with tempfile.NamedTemporaryFile(mode='w', suffix='.html', delete=False) as f: f.write("<html><body><h1>Local File Test</h1></body></html>") temp_file = f.name try: async with AsyncWebCrawler() as crawler: result = await crawler.arun(f"file://{temp_file}") assert result.success assert "Local File Test" in result.markdown finally: os.unlink(temp_file) @pytest.mark.asyncio async def test_robots_txt_wildcard_support(self): """Test robots.txt wildcard rules support""" parser = RobotsParser() # Test wildcard patterns robots_content = "User-agent: *\nDisallow: /admin/*\nDisallow: *.pdf" # This should work without throwing exceptions assert parser is not None @pytest.mark.asyncio async def test_exclude_external_images(self): """Test exclude_external_images flag""" html_with_images = ''' <html><body> <img src="/local-image.jpg" alt="Local"> <img src="https://external.com/image.jpg" alt="External"> </body></html> ''' config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, exclude_external_images=True ) async with AsyncWebCrawler() as crawler: result = await crawler.arun(f"raw://{html_with_images}", config=config) assert result.success # External images should be excluded assert "external.com" not in result.cleaned_html @pytest.mark.asyncio async def test_llm_extraction_strategy_fix(self): """Test LLM extraction strategy choices error fix""" if not os.getenv("OPENAI_API_KEY"): pytest.skip("OpenAI API key not available") llm_config = LLMConfig( provider="openai/gpt-4o-mini", api_token=os.getenv("OPENAI_API_KEY") ) strategy = LLMExtractionStrategy( llm_config=llm_config, instruction="Extract the main heading", extraction_type="block" ) config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, extraction_strategy=strategy ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success # Should not throw 'str' object has no attribute 'choices' error assert result.extracted_content is not None @pytest.mark.asyncio async def test_wait_for_timeout(self): """Test separate timeout for wait_for condition""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, wait_for="css:non-existent-element", wait_for_timeout=1000 # 1 second timeout ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) # Should timeout gracefully and still return result assert result is not None @pytest.mark.asyncio async def test_bm25_content_filter_language_parameter(self): """Test BM25 filter with language parameter for stemming""" content_filter = BM25ContentFilter( user_query="test content", language="english", use_stemming=True ) markdown_generator = DefaultMarkdownGenerator( content_filter=content_filter ) config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, markdown_generator=markdown_generator ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success assert result.markdown is not None @pytest.mark.asyncio async def test_url_normalization(self): """Test URL normalization for invalid schemes and trailing slashes""" async with AsyncWebCrawler() as crawler: # Test with trailing slash result = await crawler.arun( "https://httpbin.org/html/", config=CrawlerRunConfig(cache_mode=CacheMode.BYPASS) ) assert result.success @pytest.mark.asyncio async def test_max_scroll_steps(self): """Test max_scroll_steps parameter for full page scanning""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, scan_full_page=True, max_scroll_steps=3 ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success @pytest.mark.asyncio async def test_async_url_seeder(self): """Test AsyncUrlSeeder functionality""" seeder = AsyncUrlSeeder( base_url="https://httpbin.org", max_depth=1, max_urls=5 ) async with AsyncWebCrawler() as crawler: urls = await seeder.seed(crawler) assert isinstance(urls, list) assert len(urls) <= 5 @pytest.mark.asyncio async def test_pdf_processing_timeout(self): """Test PDF processing with timeout""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, pdf=True, pdf_timeout=10000 # 10 seconds ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success # PDF might be None for HTML pages, but should not hang assert result.pdf is not None or result.pdf is None @pytest.mark.asyncio async def test_browser_session_management(self): """Test improved browser session management""" browser_config = BrowserConfig( headless=True, use_persistent_context=True ) async with AsyncWebCrawler(config=browser_config) as crawler: result = await crawler.arun( "https://httpbin.org/html", config=CrawlerRunConfig(cache_mode=CacheMode.BYPASS) ) assert result.success @pytest.mark.asyncio async def test_memory_management(self): """Test memory management features""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, memory_threshold_percent=80.0, check_interval=1.0, memory_wait_timeout=600 # 10 minutes default ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success @pytest.mark.asyncio async def test_virtual_scroll_support(self): """Test virtual scroll support for modern web scraping""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, scan_full_page=True, virtual_scroll=True ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success @pytest.mark.asyncio async def test_adaptive_crawling(self): """Test adaptive crawling feature""" config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, adaptive_crawling=True ) async with AsyncWebCrawler() as crawler: result = await crawler.arun("https://httpbin.org/html", config=config) assert result.success if __name__ == "__main__": # Run the tests pytest.main([__file__, "-v"])
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_cdp_strategy.py
tests/browser/test_cdp_strategy.py
"""Test examples for CDPBrowserStrategy. These examples demonstrate the functionality of CDPBrowserStrategy and serve as functional tests. """ import asyncio import os import sys import time # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.browser_manager import BrowserManager from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def test_cdp_launch_connect(): """Test launching a browser and connecting via CDP.""" logger.info("Testing launch and connect via CDP", tag="TEST") browser_config = BrowserConfig( browser_mode="cdp", use_managed_browser=True, headless=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() logger.info("Browser launched and connected via CDP", tag="TEST") # Test with multiple pages pages = [] for i in range(3): crawler_config = CrawlerRunConfig() page, context = await manager.get_page(crawler_config) await page.goto(f"https://example.com?test={i}") pages.append(page) logger.info(f"Created page {i+1}", tag="TEST") # Verify all pages are working for i, page in enumerate(pages): title = await page.title() logger.info(f"Page {i+1} title: {title}", tag="TEST") await manager.close() logger.info("Browser closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def test_cdp_with_user_data_dir(): """Test CDP browser with a user data directory and storage state.""" logger.info("Testing CDP browser with user data directory", tag="TEST") # Create a temporary user data directory import tempfile user_data_dir = tempfile.mkdtemp(prefix="crawl4ai-test-") storage_state_file = os.path.join(user_data_dir, "storage_state.json") logger.info(f"Created temporary user data directory: {user_data_dir}", tag="TEST") browser_config = BrowserConfig( headless=True, use_managed_browser=True, user_data_dir=user_data_dir ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() logger.info("Browser launched with user data directory", tag="TEST") # Navigate to a page and store some data crawler_config = CrawlerRunConfig() page, context = await manager.get_page(crawler_config) # Visit the site first await page.goto("https://example.com", wait_until="domcontentloaded") # Set a cookie via JavaScript (more reliable for persistence) await page.evaluate(""" document.cookie = 'test_cookie=test_value; path=/; max-age=86400'; """) # Also set via context API for double coverage await context.add_cookies([{ "name": "test_cookie_api", "value": "test_value_api", "domain": "example.com", "path": "/" }]) # Verify cookies were set cookies = await context.cookies(["https://example.com"]) has_test_cookie = any(cookie["name"] in ["test_cookie", "test_cookie_api"] for cookie in cookies) logger.info(f"Cookie set successfully: {has_test_cookie}", tag="TEST") # Save storage state before closing await context.storage_state(path=storage_state_file) logger.info(f"Storage state saved to: {storage_state_file}", tag="TEST") # Close the browser await manager.close() logger.info("First browser session closed", tag="TEST") # Wait a moment for clean shutdown await asyncio.sleep(1.0) # Start a new browser with the same user data directory and storage state logger.info("Starting second browser session with same user data directory", tag="TEST") browser_config2 = BrowserConfig( headless=True, use_managed_browser=True, user_data_dir=user_data_dir, storage_state=storage_state_file ) manager2 = BrowserManager(browser_config=browser_config2, logger=logger) await manager2.start() # Get a new page and check if the cookie persists page2, context2 = await manager2.get_page(crawler_config) await page2.goto("https://example.com", wait_until="domcontentloaded") # Verify cookie persisted cookies2 = await context2.cookies(["https://example.com"]) has_test_cookie2 = any(cookie["name"] in ["test_cookie", "test_cookie_api"] for cookie in cookies2) logger.info(f"Cookie persisted across sessions: {has_test_cookie2}", tag="TEST") logger.info(f"Cookies found: {[c['name'] for c in cookies2]}", tag="TEST") # Clean up await manager2.close() # Remove temporary directory import shutil shutil.rmtree(user_data_dir, ignore_errors=True) logger.info(f"Removed temporary user data directory", tag="TEST") return has_test_cookie and has_test_cookie2 except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass try: await manager2.close() except: pass # Clean up temporary directory try: import shutil shutil.rmtree(user_data_dir, ignore_errors=True) except: pass return False async def test_cdp_session_management(): """Test session management with CDP browser - focused on session tracking.""" logger.info("Testing session management with CDP browser", tag="TEST") browser_config = BrowserConfig( use_managed_browser=True, headless=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() logger.info("Browser launched successfully", tag="TEST") # Test session tracking and lifecycle management session1_id = "test_session_1" session2_id = "test_session_2" # Set up first session crawler_config1 = CrawlerRunConfig(session_id=session1_id) page1, context1 = await manager.get_page(crawler_config1) await page1.goto("https://example.com", wait_until="domcontentloaded") # Get page URL and title for verification page1_url = page1.url page1_title = await page1.title() logger.info(f"Session 1 setup - URL: {page1_url}, Title: {page1_title}", tag="TEST") # Set up second session crawler_config2 = CrawlerRunConfig(session_id=session2_id) page2, context2 = await manager.get_page(crawler_config2) await page2.goto("https://httpbin.org/html", wait_until="domcontentloaded") page2_url = page2.url page2_title = await page2.title() logger.info(f"Session 2 setup - URL: {page2_url}, Title: {page2_title}", tag="TEST") # Verify sessions exist in manager session1_exists = session1_id in manager.sessions session2_exists = session2_id in manager.sessions logger.info(f"Sessions in manager - S1: {session1_exists}, S2: {session2_exists}", tag="TEST") # Test session reuse page1_again, context1_again = await manager.get_page(crawler_config1) is_same_page = page1 == page1_again is_same_context = context1 == context1_again logger.info(f"Session 1 reuse - Same page: {is_same_page}, Same context: {is_same_context}", tag="TEST") # Test that sessions are properly tracked with timestamps session1_info = manager.sessions.get(session1_id) session2_info = manager.sessions.get(session2_id) session1_has_timestamp = session1_info and len(session1_info) == 3 session2_has_timestamp = session2_info and len(session2_info) == 3 logger.info(f"Session tracking - S1 complete: {session1_has_timestamp}, S2 complete: {session2_has_timestamp}", tag="TEST") # In managed browser mode, pages might be shared. Let's test what actually happens pages_same_or_different = page1 == page2 logger.info(f"Pages same object: {pages_same_or_different}", tag="TEST") # Test that we can distinguish sessions by their stored info session1_context, session1_page, session1_time = session1_info session2_context, session2_page, session2_time = session2_info sessions_have_different_timestamps = session1_time != session2_time logger.info(f"Sessions have different timestamps: {sessions_have_different_timestamps}", tag="TEST") # Test session killing await manager.kill_session(session1_id) logger.info(f"Killed session 1", tag="TEST") # Verify session was removed session1_removed = session1_id not in manager.sessions session2_still_exists = session2_id in manager.sessions logger.info(f"After kill - S1 removed: {session1_removed}, S2 exists: {session2_still_exists}", tag="TEST") # Test page state after killing session page1_closed = page1.is_closed() logger.info(f"Page1 closed after kill: {page1_closed}", tag="TEST") # Clean up remaining session try: await manager.kill_session(session2_id) logger.info("Killed session 2", tag="TEST") session2_removed = session2_id not in manager.sessions except Exception as e: logger.info(f"Session 2 cleanup: {e}", tag="TEST") session2_removed = False # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") # Success criteria for managed browser sessions: # 1. Sessions can be created and tracked with proper info # 2. Same page/context returned for same session ID # 3. Sessions have proper timestamp tracking # 4. Sessions can be killed and removed from tracking # 5. Session cleanup works properly success = (session1_exists and session2_exists and is_same_page and session1_has_timestamp and session2_has_timestamp and sessions_have_different_timestamps and session1_removed and session2_removed) logger.info(f"Test success: {success}", tag="TEST") return success except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def test_cdp_timing_fix_fast_startup(): """ Test that the CDP timing fix handles fast browser startup correctly. This should work without any delays or retries. """ logger.info("Testing CDP timing fix with fast startup", tag="TEST") browser_config = BrowserConfig( use_managed_browser=True, browser_mode="cdp", headless=True, debugging_port=9223, # Use different port to avoid conflicts verbose=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: start_time = time.time() await manager.start() startup_time = time.time() - start_time logger.info(f"Browser started successfully in {startup_time:.2f}s", tag="TEST") # Test basic functionality crawler_config = CrawlerRunConfig(url="https://example.com") page, context = await manager.get_page(crawler_config) await page.goto("https://example.com", wait_until="domcontentloaded") title = await page.title() logger.info(f"Successfully navigated to page: {title}", tag="TEST") await manager.close() logger.success("test_cdp_timing_fix_fast_startup completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_cdp_timing_fix_fast_startup failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def test_cdp_timing_fix_delayed_browser_start(): """ Test CDP timing fix by actually delaying the browser startup process. This simulates a real scenario where the browser takes time to expose CDP. """ logger.info("Testing CDP timing fix with delayed browser startup", tag="TEST") browser_config = BrowserConfig( use_managed_browser=True, browser_mode="cdp", headless=True, debugging_port=9224, verbose=True ) # Start the managed browser separately to control timing from crawl4ai.browser_manager import ManagedBrowser managed_browser = ManagedBrowser(browser_config=browser_config, logger=logger) try: # Start browser process but it will take time for CDP to be ready cdp_url = await managed_browser.start() logger.info(f"Managed browser started at {cdp_url}", tag="TEST") # Small delay to simulate the browser needing time to fully initialize CDP await asyncio.sleep(1.0) # Now create BrowserManager and connect - this should use the CDP verification fix manager = BrowserManager(browser_config=browser_config, logger=logger) manager.config.cdp_url = cdp_url # Use the CDP URL from managed browser start_time = time.time() await manager.start() startup_time = time.time() - start_time logger.info(f"BrowserManager connected successfully in {startup_time:.2f}s", tag="TEST") # Test basic functionality crawler_config = CrawlerRunConfig(url="https://example.com") page, context = await manager.get_page(crawler_config) await page.goto("https://example.com", wait_until="domcontentloaded") title = await page.title() logger.info(f"Successfully navigated to page: {title}", tag="TEST") # Clean up await manager.close() await managed_browser.cleanup() logger.success("test_cdp_timing_fix_delayed_browser_start completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_cdp_timing_fix_delayed_browser_start failed: {str(e)}", tag="TEST") try: await manager.close() await managed_browser.cleanup() except: pass return False async def test_cdp_verification_backoff_behavior(): """ Test the exponential backoff behavior of CDP verification in isolation. """ logger.info("Testing CDP verification exponential backoff behavior", tag="TEST") browser_config = BrowserConfig( use_managed_browser=True, debugging_port=9225, # Use different port verbose=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: # Test with a non-existent CDP URL to trigger retries fake_cdp_url = "http://localhost:19999" # This should not exist start_time = time.time() result = await manager._verify_cdp_ready(fake_cdp_url) elapsed_time = time.time() - start_time # Should return False after all retries assert result is False, "Expected CDP verification to fail with non-existent endpoint" # Should take some time due to retries and backoff assert elapsed_time > 2.0, f"Expected backoff delays, but took only {elapsed_time:.2f}s" logger.info(f"CDP verification correctly failed after {elapsed_time:.2f}s with exponential backoff", tag="TEST") logger.success("test_cdp_verification_backoff_behavior completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_cdp_verification_backoff_behavior failed: {str(e)}", tag="TEST") return False async def run_tests(): """Run all tests sequentially.""" import time results = [] # Original CDP strategy tests logger.info("Running original CDP strategy tests", tag="SUITE") # results.append(await test_cdp_launch_connect()) results.append(await test_cdp_with_user_data_dir()) results.append(await test_cdp_session_management()) # CDP timing fix tests logger.info("Running CDP timing fix tests", tag="SUITE") results.append(await test_cdp_timing_fix_fast_startup()) results.append(await test_cdp_timing_fix_delayed_browser_start()) results.append(await test_cdp_verification_backoff_behavior()) # Print summary total = len(results) passed = sum(results) logger.info(f"Tests complete: {passed}/{total} passed", tag="SUMMARY") if passed == total: logger.success("All tests passed!", tag="SUMMARY") else: logger.error(f"{total - passed} tests failed", tag="SUMMARY") if __name__ == "__main__": asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_builtin_strategy.py
tests/browser/test_builtin_strategy.py
"""Test examples for BuiltinBrowserStrategy. These examples demonstrate the functionality of BuiltinBrowserStrategy and serve as functional tests. """ import asyncio import os import sys # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.browser import BrowserManager from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def test_builtin_browser(): """Test using a builtin browser that persists between sessions.""" logger.info("Testing builtin browser", tag="TEST") browser_config = BrowserConfig( browser_mode="builtin", headless=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: # Start should connect to existing builtin browser or create one await manager.start() logger.info("Connected to builtin browser", tag="TEST") # Test page creation crawler_config = CrawlerRunConfig() page, context = await manager.get_page(crawler_config) # Test navigation await page.goto("https://example.com") title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Close manager (should not close the builtin browser) await manager.close() logger.info("First session closed", tag="TEST") # Create a second manager to verify browser persistence logger.info("Creating second session to verify persistence", tag="TEST") manager2 = BrowserManager(browser_config=browser_config, logger=logger) await manager2.start() logger.info("Connected to existing builtin browser", tag="TEST") page2, context2 = await manager2.get_page(crawler_config) await page2.goto("https://example.org") title2 = await page2.title() logger.info(f"Second session page title: {title2}", tag="TEST") await manager2.close() logger.info("Second session closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def test_builtin_browser_status(): """Test getting status of the builtin browser.""" logger.info("Testing builtin browser status", tag="TEST") from crawl4ai.browser.strategies import BuiltinBrowserStrategy browser_config = BrowserConfig( browser_mode="builtin", headless=True ) # Create strategy directly to access its status methods strategy = BuiltinBrowserStrategy(browser_config, logger) try: # Get status before starting (should be not running) status_before = await strategy.get_builtin_browser_status() logger.info(f"Initial status: {status_before}", tag="TEST") # Start the browser await strategy.start() logger.info("Browser started successfully", tag="TEST") # Get status after starting status_after = await strategy.get_builtin_browser_status() logger.info(f"Status after start: {status_after}", tag="TEST") # Create a page to verify functionality crawler_config = CrawlerRunConfig() page, context = await strategy.get_page(crawler_config) await page.goto("https://example.com") title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Close strategy (should not kill the builtin browser) await strategy.close() logger.info("Strategy closed successfully", tag="TEST") # Create a new strategy object strategy2 = BuiltinBrowserStrategy(browser_config, logger) # Get status again (should still be running) status_final = await strategy2.get_builtin_browser_status() logger.info(f"Final status: {status_final}", tag="TEST") # Verify that the status shows the browser is running is_running = status_final.get('running', False) logger.info(f"Builtin browser persistence confirmed: {is_running}", tag="TEST") # Kill the builtin browser to clean up logger.info("Killing builtin browser", tag="TEST") success = await strategy2.kill_builtin_browser() logger.info(f"Killed builtin browser successfully: {success}", tag="TEST") return is_running and success except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await strategy.close() # Try to kill the builtin browser to clean up strategy2 = BuiltinBrowserStrategy(browser_config, logger) await strategy2.kill_builtin_browser() except: pass return False async def run_tests(): """Run all tests sequentially.""" results = [] results.append(await test_builtin_browser()) results.append(await test_builtin_browser_status()) # Print summary total = len(results) passed = sum(results) logger.info(f"Tests complete: {passed}/{total} passed", tag="SUMMARY") if passed == total: logger.success("All tests passed!", tag="SUMMARY") else: logger.error(f"{total - passed} tests failed", tag="SUMMARY") if __name__ == "__main__": asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_playwright_strategy.py
tests/browser/test_playwright_strategy.py
"""Test examples for PlaywrightBrowserStrategy. These examples demonstrate the functionality of PlaywrightBrowserStrategy and serve as functional tests. """ import asyncio import os import re import sys # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.browser import BrowserManager from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def test_start_close(): # Create browser config for standard Playwright browser_config = BrowserConfig( headless=True, viewport_width=1280, viewport_height=800 ) # Create browser manager with the config manager = BrowserManager(browser_config=browser_config, logger=logger) try: for _ in range(4): # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Get a page page, context = await manager.get_page(CrawlerRunConfig()) logger.info("Got page successfully", tag="TEST") # Navigate to a website await page.goto("https://example.com") logger.info("Navigated to example.com", tag="TEST") # Get page title title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") await asyncio.sleep(1) # Wait for a moment before restarting except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False return True async def test_playwright_basic(): """Test basic Playwright browser functionality.""" logger.info("Testing standard Playwright browser", tag="TEST") # Create browser config for standard Playwright browser_config = BrowserConfig( headless=True, viewport_width=1280, viewport_height=800 ) # Create browser manager with the config manager = BrowserManager(browser_config=browser_config, logger=logger) try: # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create crawler config crawler_config = CrawlerRunConfig(url="https://example.com") # Get a page page, context = await manager.get_page(crawler_config) logger.info("Got page successfully", tag="TEST") # Navigate to a website await page.goto("https://example.com") logger.info("Navigated to example.com", tag="TEST") # Get page title title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False async def test_playwright_text_mode(): """Test Playwright browser in text-only mode.""" logger.info("Testing Playwright text mode", tag="TEST") # Create browser config with text mode enabled browser_config = BrowserConfig( headless=True, text_mode=True # Enable text-only mode ) # Create browser manager with the config manager = BrowserManager(browser_config=browser_config, logger=logger) try: # Start the browser await manager.start() logger.info("Browser started successfully in text mode", tag="TEST") # Get a page crawler_config = CrawlerRunConfig(url="https://example.com") page, context = await manager.get_page(crawler_config) # Navigate to a website await page.goto("https://example.com") logger.info("Navigated to example.com", tag="TEST") # Get page title title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Check if images are blocked in text mode # We'll check if any image requests were made has_images = False async with page.expect_request("**/*.{png,jpg,jpeg,gif,webp,svg}", timeout=1000) as request_info: try: # Try to load a page with images await page.goto("https://picsum.photos/", wait_until="domcontentloaded") request = await request_info.value has_images = True except: # Timeout without image requests means text mode is working has_images = False logger.info(f"Text mode image blocking working: {not has_images}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False async def test_playwright_context_reuse(): """Test context caching and reuse with identical configurations.""" logger.info("Testing context reuse with identical configurations", tag="TEST") # Create browser config browser_config = BrowserConfig(headless=True) # Create browser manager manager = BrowserManager(browser_config=browser_config, logger=logger) try: # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create identical crawler configs crawler_config1 = CrawlerRunConfig( css_selector="body", ) crawler_config2 = CrawlerRunConfig( css_selector="body", ) # Get pages with these configs page1, context1 = await manager.get_page(crawler_config1) page2, context2 = await manager.get_page(crawler_config2) # Check if contexts are reused is_same_context = context1 == context2 logger.info(f"Contexts reused: {is_same_context}", tag="TEST") # Now try with a different config crawler_config3 = CrawlerRunConfig() page3, context3 = await manager.get_page(crawler_config3) # This should be a different context is_different_context = context1 != context3 logger.info(f"Different contexts for different configs: {is_different_context}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") # Both tests should pass for success return is_same_context and is_different_context except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False async def test_playwright_session_management(): """Test session management with Playwright browser.""" logger.info("Testing session management with Playwright browser", tag="TEST") browser_config = BrowserConfig( headless=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() logger.info("Browser launched successfully", tag="TEST") # Create two sessions session1_id = "playwright_session_1" session2_id = "playwright_session_2" # Set up first session crawler_config1 = CrawlerRunConfig(session_id=session1_id, url="https://example.com") page1, context1 = await manager.get_page(crawler_config1) await page1.goto("https://example.com") await page1.evaluate("localStorage.setItem('playwright_session1_data', 'test_value1')") logger.info(f"Set up session 1 with ID: {session1_id}", tag="TEST") # Set up second session crawler_config2 = CrawlerRunConfig(session_id=session2_id, url="https://example.org") page2, context2 = await manager.get_page(crawler_config2) await page2.goto("https://example.org") await page2.evaluate("localStorage.setItem('playwright_session2_data', 'test_value2')") logger.info(f"Set up session 2 with ID: {session2_id}", tag="TEST") # Get first session again page1_again, context1_again = await manager.get_page(crawler_config1) # Verify it's the same page and data persists is_same_page = page1 == page1_again is_same_context = context1 == context1_again data1 = await page1_again.evaluate("localStorage.getItem('playwright_session1_data')") logger.info(f"Session 1 reuse successful: {is_same_page}, data: {data1}", tag="TEST") # Kill first session await manager.kill_session(session1_id) logger.info(f"Killed session 1", tag="TEST") # Verify second session still works data2 = await page2.evaluate("localStorage.getItem('playwright_session2_data')") logger.info(f"Session 2 still functional after killing session 1, data: {data2}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") return is_same_page and is_same_context and data1 == "test_value1" and data2 == "test_value2" except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def run_tests(): """Run all tests sequentially.""" results = [] # results.append(await test_start_close()) # results.append(await test_playwright_basic()) # results.append(await test_playwright_text_mode()) # results.append(await test_playwright_context_reuse()) results.append(await test_playwright_session_management()) # Print summary total = len(results) passed = sum(results) logger.info(f"Tests complete: {passed}/{total} passed", tag="SUMMARY") if passed == total: logger.success("All tests passed!", tag="SUMMARY") else: logger.error(f"{total - passed} tests failed", tag="SUMMARY") if __name__ == "__main__": asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_combined.py
tests/browser/test_combined.py
"""Combined test runner for all browser module tests. This script runs all the browser module tests in sequence and provides a comprehensive summary. """ import asyncio import os import sys import time # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def run_test_module(module_name, header): """Run all tests in a module and return results.""" logger.info(f"\n{'-'*30}", tag="TEST") logger.info(f"RUNNING: {header}", tag="TEST") logger.info(f"{'-'*30}", tag="TEST") # Import the module dynamically module = __import__(f"tests.browser.{module_name}", fromlist=["run_tests"]) # Track time for performance measurement start_time = time.time() # Run the tests await module.run_tests() # Calculate time taken time_taken = time.time() - start_time logger.info(f"Time taken: {time_taken:.2f} seconds", tag="TIMING") return time_taken async def main(): """Run all test modules.""" logger.info("STARTING COMPREHENSIVE BROWSER MODULE TESTS", tag="MAIN") # List of test modules to run test_modules = [ ("test_browser_manager", "Browser Manager Tests"), ("test_playwright_strategy", "Playwright Strategy Tests"), ("test_cdp_strategy", "CDP Strategy Tests"), ("test_builtin_strategy", "Builtin Browser Strategy Tests"), ("test_profiles", "Profile Management Tests") ] # Run each test module timings = {} for module_name, header in test_modules: try: time_taken = await run_test_module(module_name, header) timings[module_name] = time_taken except Exception as e: logger.error(f"Error running {module_name}: {str(e)}", tag="ERROR") # Print summary logger.info("\n\nTEST SUMMARY:", tag="SUMMARY") logger.info(f"{'-'*50}", tag="SUMMARY") for module_name, header in test_modules: if module_name in timings: logger.info(f"{header}: {timings[module_name]:.2f} seconds", tag="SUMMARY") else: logger.error(f"{header}: FAILED TO RUN", tag="SUMMARY") logger.info(f"{'-'*50}", tag="SUMMARY") total_time = sum(timings.values()) logger.info(f"Total time: {total_time:.2f} seconds", tag="SUMMARY") if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_builtin_browser.py
tests/browser/test_builtin_browser.py
""" Test script for builtin browser functionality in the browser module. This script tests: 1. Creating a builtin browser 2. Getting browser information 3. Killing the browser 4. Restarting the browser 5. Testing operations with different browser strategies 6. Testing edge cases """ import asyncio import os import sys import time from typing import List, Dict, Any from colorama import Fore, Style, init # Add the project root to the path for imports sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), "../.."))) from rich.console import Console from rich.table import Table from rich.panel import Panel from rich.text import Text from rich.box import Box, SIMPLE from crawl4ai.browser import BrowserManager from crawl4ai.browser.strategies import BuiltinBrowserStrategy from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Initialize colorama for cross-platform colored terminal output init() # Define colors for pretty output SUCCESS = Fore.GREEN WARNING = Fore.YELLOW ERROR = Fore.RED INFO = Fore.CYAN RESET = Fore.RESET # Create logger logger = AsyncLogger(verbose=True) async def test_builtin_browser_creation(): """Test creating a builtin browser using the BrowserManager with BuiltinBrowserStrategy""" print(f"\n{INFO}========== Testing Builtin Browser Creation =========={RESET}") # Step 1: Create a BrowserManager with builtin mode print(f"\n{INFO}1. Creating BrowserManager with builtin mode{RESET}") browser_config = BrowserConfig(browser_mode="builtin", headless=True, verbose=True) manager = BrowserManager(browser_config=browser_config, logger=logger) # Step 2: Check if we have a BuiltinBrowserStrategy print(f"\n{INFO}2. Checking if we have a BuiltinBrowserStrategy{RESET}") if isinstance(manager.strategy, BuiltinBrowserStrategy): print( f"{SUCCESS}Correct strategy type: {manager.strategy.__class__.__name__}{RESET}" ) else: print( f"{ERROR}Wrong strategy type: {manager.strategy.__class__.__name__}{RESET}" ) return None # Step 3: Start the manager to launch or connect to builtin browser print(f"\n{INFO}3. Starting the browser manager{RESET}") try: await manager.start() print(f"{SUCCESS}Browser manager started successfully{RESET}") except Exception as e: print(f"{ERROR}Failed to start browser manager: {str(e)}{RESET}") return None # Step 4: Get browser info from the strategy print(f"\n{INFO}4. Getting browser information{RESET}") browser_info = manager.strategy.get_browser_info() if browser_info: print(f"{SUCCESS}Browser info retrieved:{RESET}") for key, value in browser_info.items(): if key != "config": # Skip the verbose config section print(f" {key}: {value}") cdp_url = browser_info.get("cdp_url") print(f"{SUCCESS}CDP URL: {cdp_url}{RESET}") else: print(f"{ERROR}Failed to get browser information{RESET}") cdp_url = None # Save manager for later tests return manager, cdp_url async def test_page_operations(manager: BrowserManager): """Test page operations with the builtin browser""" print( f"\n{INFO}========== Testing Page Operations with Builtin Browser =========={RESET}" ) # Step 1: Get a single page print(f"\n{INFO}1. Getting a single page{RESET}") try: crawler_config = CrawlerRunConfig() page, context = await manager.get_page(crawler_config) print(f"{SUCCESS}Got page successfully{RESET}") # Navigate to a test URL await page.goto("https://example.com") title = await page.title() print(f"{SUCCESS}Page title: {title}{RESET}") # Close the page await page.close() print(f"{SUCCESS}Page closed successfully{RESET}") except Exception as e: print(f"{ERROR}Page operation failed: {str(e)}{RESET}") return False # Step 2: Get multiple pages print(f"\n{INFO}2. Getting multiple pages with get_pages(){RESET}") try: # Request 3 pages crawler_config = CrawlerRunConfig() pages = await manager.get_pages(crawler_config, count=3) print(f"{SUCCESS}Got {len(pages)} pages{RESET}") # Test each page for i, (page, context) in enumerate(pages): await page.goto(f"https://example.com?test={i}") title = await page.title() print(f"{SUCCESS}Page {i + 1} title: {title}{RESET}") await page.close() print(f"{SUCCESS}All pages tested and closed successfully{RESET}") except Exception as e: print(f"{ERROR}Multiple page operation failed: {str(e)}{RESET}") return False return True async def test_browser_status_management(manager: BrowserManager): """Test browser status and management operations""" print(f"\n{INFO}========== Testing Browser Status and Management =========={RESET}") # Step 1: Get browser status print(f"\n{INFO}1. Getting browser status{RESET}") try: status = await manager.strategy.get_builtin_browser_status() print(f"{SUCCESS}Browser status:{RESET}") print(f" Running: {status['running']}") print(f" CDP URL: {status['cdp_url']}") except Exception as e: print(f"{ERROR}Failed to get browser status: {str(e)}{RESET}") return False # Step 2: Test killing the browser print(f"\n{INFO}2. Testing killing the browser{RESET}") try: result = await manager.strategy.kill_builtin_browser() if result: print(f"{SUCCESS}Browser killed successfully{RESET}") else: print(f"{ERROR}Failed to kill browser{RESET}") except Exception as e: print(f"{ERROR}Browser kill operation failed: {str(e)}{RESET}") return False # Step 3: Check status after kill print(f"\n{INFO}3. Checking status after kill{RESET}") try: status = await manager.strategy.get_builtin_browser_status() if not status["running"]: print(f"{SUCCESS}Browser is correctly reported as not running{RESET}") else: print(f"{ERROR}Browser is incorrectly reported as still running{RESET}") except Exception as e: print(f"{ERROR}Failed to get browser status: {str(e)}{RESET}") return False # Step 4: Launch a new browser print(f"\n{INFO}4. Launching a new browser{RESET}") try: cdp_url = await manager.strategy.launch_builtin_browser( browser_type="chromium", headless=True ) if cdp_url: print(f"{SUCCESS}New browser launched at: {cdp_url}{RESET}") else: print(f"{ERROR}Failed to launch new browser{RESET}") return False except Exception as e: print(f"{ERROR}Browser launch failed: {str(e)}{RESET}") return False return True async def test_multiple_managers(): """Test creating multiple BrowserManagers that use the same builtin browser""" print(f"\n{INFO}========== Testing Multiple Browser Managers =========={RESET}") # Step 1: Create first manager print(f"\n{INFO}1. Creating first browser manager{RESET}") browser_config1 = BrowserConfig(browser_mode="builtin", headless=True) manager1 = BrowserManager(browser_config=browser_config1, logger=logger) # Step 2: Create second manager print(f"\n{INFO}2. Creating second browser manager{RESET}") browser_config2 = BrowserConfig(browser_mode="builtin", headless=True) manager2 = BrowserManager(browser_config=browser_config2, logger=logger) # Step 3: Start both managers (should connect to the same builtin browser) print(f"\n{INFO}3. Starting both managers{RESET}") try: await manager1.start() print(f"{SUCCESS}First manager started{RESET}") await manager2.start() print(f"{SUCCESS}Second manager started{RESET}") # Check if they got the same CDP URL cdp_url1 = manager1.strategy.config.cdp_url cdp_url2 = manager2.strategy.config.cdp_url if cdp_url1 == cdp_url2: print( f"{SUCCESS}Both managers connected to the same browser: {cdp_url1}{RESET}" ) else: print( f"{WARNING}Managers connected to different browsers: {cdp_url1} and {cdp_url2}{RESET}" ) except Exception as e: print(f"{ERROR}Failed to start managers: {str(e)}{RESET}") return False # Step 4: Test using both managers print(f"\n{INFO}4. Testing operations with both managers{RESET}") try: # First manager creates a page page1, ctx1 = await manager1.get_page(CrawlerRunConfig()) await page1.goto("https://example.com") title1 = await page1.title() print(f"{SUCCESS}Manager 1 page title: {title1}{RESET}") # Second manager creates a page page2, ctx2 = await manager2.get_page(CrawlerRunConfig()) await page2.goto("https://example.org") title2 = await page2.title() print(f"{SUCCESS}Manager 2 page title: {title2}{RESET}") # Clean up await page1.close() await page2.close() except Exception as e: print(f"{ERROR}Failed to use both managers: {str(e)}{RESET}") return False # Step 5: Close both managers print(f"\n{INFO}5. Closing both managers{RESET}") try: await manager1.close() print(f"{SUCCESS}First manager closed{RESET}") await manager2.close() print(f"{SUCCESS}Second manager closed{RESET}") except Exception as e: print(f"{ERROR}Failed to close managers: {str(e)}{RESET}") return False return True async def test_edge_cases(): """Test edge cases like multiple starts, killing browser during operations, etc.""" print(f"\n{INFO}========== Testing Edge Cases =========={RESET}") # Step 1: Test multiple starts with the same manager print(f"\n{INFO}1. Testing multiple starts with the same manager{RESET}") browser_config = BrowserConfig(browser_mode="builtin", headless=True) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() print(f"{SUCCESS}First start successful{RESET}") # Try to start again await manager.start() print(f"{SUCCESS}Second start completed without errors{RESET}") # Test if it's still functional page, context = await manager.get_page(CrawlerRunConfig()) await page.goto("https://example.com") title = await page.title() print( f"{SUCCESS}Page operations work after multiple starts. Title: {title}{RESET}" ) await page.close() except Exception as e: print(f"{ERROR}Multiple starts test failed: {str(e)}{RESET}") return False finally: await manager.close() # Step 2: Test killing the browser while manager is active print(f"\n{INFO}2. Testing killing the browser while manager is active{RESET}") manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() print(f"{SUCCESS}Manager started{RESET}") # Kill the browser directly print(f"{INFO}Killing the browser...{RESET}") await manager.strategy.kill_builtin_browser() print(f"{SUCCESS}Browser killed{RESET}") # Try to get a page (should fail or launch a new browser) try: page, context = await manager.get_page(CrawlerRunConfig()) print( f"{WARNING}Page request succeeded despite killed browser (might have auto-restarted){RESET}" ) title = await page.title() print(f"{SUCCESS}Got page title: {title}{RESET}") await page.close() except Exception as e: print( f"{SUCCESS}Page request failed as expected after browser was killed: {str(e)}{RESET}" ) except Exception as e: print(f"{ERROR}Kill during operation test failed: {str(e)}{RESET}") return False finally: await manager.close() return True async def cleanup_browsers(): """Clean up any remaining builtin browsers""" print(f"\n{INFO}========== Cleaning Up Builtin Browsers =========={RESET}") browser_config = BrowserConfig(browser_mode="builtin", headless=True) manager = BrowserManager(browser_config=browser_config, logger=logger) try: # No need to start, just access the strategy directly strategy = manager.strategy if isinstance(strategy, BuiltinBrowserStrategy): result = await strategy.kill_builtin_browser() if result: print(f"{SUCCESS}Successfully killed all builtin browsers{RESET}") else: print(f"{WARNING}No builtin browsers found to kill{RESET}") else: print(f"{ERROR}Wrong strategy type: {strategy.__class__.__name__}{RESET}") except Exception as e: print(f"{ERROR}Cleanup failed: {str(e)}{RESET}") finally: # Just to be safe try: await manager.close() except: pass async def test_performance_scaling(): """Test performance with multiple browsers and pages. This test creates multiple browsers on different ports, spawns multiple pages per browser, and measures performance metrics. """ print(f"\n{INFO}========== Testing Performance Scaling =========={RESET}") # Configuration parameters num_browsers = 10 pages_per_browser = 10 total_pages = num_browsers * pages_per_browser base_port = 9222 # Set up a measuring mechanism for memory import psutil import gc # Force garbage collection before starting gc.collect() process = psutil.Process() initial_memory = process.memory_info().rss / 1024 / 1024 # in MB peak_memory = initial_memory # Report initial configuration print( f"{INFO}Test configuration: {num_browsers} browsers × {pages_per_browser} pages = {total_pages} total crawls{RESET}" ) # List to track managers managers: List[BrowserManager] = [] all_pages = [] # Get crawl4ai home directory crawl4ai_home = os.path.expanduser("~/.crawl4ai") temp_dir = os.path.join(crawl4ai_home, "temp") os.makedirs(temp_dir, exist_ok=True) # Create all managers but don't start them yet manager_configs = [] for i in range(num_browsers): port = base_port + i browser_config = BrowserConfig( browser_mode="builtin", headless=True, debugging_port=port, user_data_dir=os.path.join(temp_dir, f"browser_profile_{i}"), ) manager = BrowserManager(browser_config=browser_config, logger=logger) manager.strategy.shutting_down = True manager_configs.append((manager, i, port)) # Define async function to start a single manager async def start_manager(manager, index, port): try: await manager.start() return manager except Exception as e: print( f"{ERROR}Failed to start browser {index + 1} on port {port}: {str(e)}{RESET}" ) return None # Start all managers in parallel start_tasks = [ start_manager(manager, i, port) for manager, i, port in manager_configs ] started_managers = await asyncio.gather(*start_tasks) # Filter out None values (failed starts) and add to managers list managers = [m for m in started_managers if m is not None] if len(managers) == 0: print(f"{ERROR}All browser managers failed to start. Aborting test.{RESET}") return False if len(managers) < num_browsers: print( f"{WARNING}Only {len(managers)} out of {num_browsers} browser managers started successfully{RESET}" ) # Create pages for each browser for i, manager in enumerate(managers): try: pages = await manager.get_pages(CrawlerRunConfig(), count=pages_per_browser) all_pages.extend(pages) except Exception as e: print(f"{ERROR}Failed to create pages for browser {i + 1}: {str(e)}{RESET}") # Check memory after page creation gc.collect() current_memory = process.memory_info().rss / 1024 / 1024 peak_memory = max(peak_memory, current_memory) # Ask for confirmation before loading confirmation = input( f"{WARNING}Do you want to proceed with loading pages? (y/n): {RESET}" ) # Step 1: Create and start multiple browser managers in parallel start_time = time.time() if confirmation.lower() == "y": load_start_time = time.time() # Function to load a single page async def load_page(page_ctx, index): page, _ = page_ctx try: await page.goto(f"https://example.com/page{index}", timeout=30000) title = await page.title() return title except Exception as e: return f"Error: {str(e)}" # Load all pages concurrently load_tasks = [load_page(page_ctx, i) for i, page_ctx in enumerate(all_pages)] load_results = await asyncio.gather(*load_tasks, return_exceptions=True) # Count successes and failures successes = sum( 1 for r in load_results if isinstance(r, str) and not r.startswith("Error") ) failures = len(load_results) - successes load_time = time.time() - load_start_time total_test_time = time.time() - start_time # Check memory after loading (peak memory) gc.collect() current_memory = process.memory_info().rss / 1024 / 1024 peak_memory = max(peak_memory, current_memory) # Calculate key metrics memory_per_page = peak_memory / successes if successes > 0 else 0 time_per_crawl = total_test_time / successes if successes > 0 else 0 crawls_per_second = successes / total_test_time if total_test_time > 0 else 0 crawls_per_minute = crawls_per_second * 60 crawls_per_hour = crawls_per_minute * 60 # Print simplified performance summary from rich.console import Console from rich.table import Table console = Console() # Create a simple summary table table = Table(title="CRAWL4AI PERFORMANCE SUMMARY") table.add_column("Metric", style="cyan") table.add_column("Value", style="green") table.add_row("Total Crawls Completed", f"{successes}") table.add_row("Total Time", f"{total_test_time:.2f} seconds") table.add_row("Time Per Crawl", f"{time_per_crawl:.2f} seconds") table.add_row("Crawling Speed", f"{crawls_per_second:.2f} crawls/second") table.add_row("Projected Rate (1 minute)", f"{crawls_per_minute:.0f} crawls") table.add_row("Projected Rate (1 hour)", f"{crawls_per_hour:.0f} crawls") table.add_row("Peak Memory Usage", f"{peak_memory:.2f} MB") table.add_row("Memory Per Crawl", f"{memory_per_page:.2f} MB") # Display the table console.print(table) # Ask confirmation before cleanup confirmation = input( f"{WARNING}Do you want to proceed with cleanup? (y/n): {RESET}" ) if confirmation.lower() != "y": print(f"{WARNING}Cleanup aborted by user{RESET}") return False # Close all pages for page, _ in all_pages: try: await page.close() except: pass # Close all managers for manager in managers: try: await manager.close() except: pass # Remove the temp directory import shutil if os.path.exists(temp_dir): shutil.rmtree(temp_dir) return True async def test_performance_scaling_lab( num_browsers: int = 10, pages_per_browser: int = 10): """Test performance with multiple browsers and pages. This test creates multiple browsers on different ports, spawns multiple pages per browser, and measures performance metrics. """ print(f"\n{INFO}========== Testing Performance Scaling =========={RESET}") # Configuration parameters num_browsers = num_browsers pages_per_browser = pages_per_browser total_pages = num_browsers * pages_per_browser base_port = 9222 # Set up a measuring mechanism for memory import psutil import gc # Force garbage collection before starting gc.collect() process = psutil.Process() initial_memory = process.memory_info().rss / 1024 / 1024 # in MB peak_memory = initial_memory # Report initial configuration print( f"{INFO}Test configuration: {num_browsers} browsers × {pages_per_browser} pages = {total_pages} total crawls{RESET}" ) # List to track managers managers: List[BrowserManager] = [] all_pages = [] # Get crawl4ai home directory crawl4ai_home = os.path.expanduser("~/.crawl4ai") temp_dir = os.path.join(crawl4ai_home, "temp") os.makedirs(temp_dir, exist_ok=True) # Create all managers but don't start them yet manager_configs = [] for i in range(num_browsers): port = base_port + i browser_config = BrowserConfig( browser_mode="builtin", headless=True, debugging_port=port, user_data_dir=os.path.join(temp_dir, f"browser_profile_{i}"), ) manager = BrowserManager(browser_config=browser_config, logger=logger) manager.strategy.shutting_down = True manager_configs.append((manager, i, port)) # Define async function to start a single manager async def start_manager(manager, index, port): try: await manager.start() return manager except Exception as e: print( f"{ERROR}Failed to start browser {index + 1} on port {port}: {str(e)}{RESET}" ) return None # Start all managers in parallel start_tasks = [ start_manager(manager, i, port) for manager, i, port in manager_configs ] started_managers = await asyncio.gather(*start_tasks) # Filter out None values (failed starts) and add to managers list managers = [m for m in started_managers if m is not None] if len(managers) == 0: print(f"{ERROR}All browser managers failed to start. Aborting test.{RESET}") return False if len(managers) < num_browsers: print( f"{WARNING}Only {len(managers)} out of {num_browsers} browser managers started successfully{RESET}" ) # Create pages for each browser for i, manager in enumerate(managers): try: pages = await manager.get_pages(CrawlerRunConfig(), count=pages_per_browser) all_pages.extend(pages) except Exception as e: print(f"{ERROR}Failed to create pages for browser {i + 1}: {str(e)}{RESET}") # Check memory after page creation gc.collect() current_memory = process.memory_info().rss / 1024 / 1024 peak_memory = max(peak_memory, current_memory) # Ask for confirmation before loading confirmation = input( f"{WARNING}Do you want to proceed with loading pages? (y/n): {RESET}" ) # Step 1: Create and start multiple browser managers in parallel start_time = time.time() if confirmation.lower() == "y": load_start_time = time.time() # Function to load a single page async def load_page(page_ctx, index): page, _ = page_ctx try: await page.goto(f"https://example.com/page{index}", timeout=30000) title = await page.title() return title except Exception as e: return f"Error: {str(e)}" # Load all pages concurrently load_tasks = [load_page(page_ctx, i) for i, page_ctx in enumerate(all_pages)] load_results = await asyncio.gather(*load_tasks, return_exceptions=True) # Count successes and failures successes = sum( 1 for r in load_results if isinstance(r, str) and not r.startswith("Error") ) failures = len(load_results) - successes load_time = time.time() - load_start_time total_test_time = time.time() - start_time # Check memory after loading (peak memory) gc.collect() current_memory = process.memory_info().rss / 1024 / 1024 peak_memory = max(peak_memory, current_memory) # Calculate key metrics memory_per_page = peak_memory / successes if successes > 0 else 0 time_per_crawl = total_test_time / successes if successes > 0 else 0 crawls_per_second = successes / total_test_time if total_test_time > 0 else 0 crawls_per_minute = crawls_per_second * 60 crawls_per_hour = crawls_per_minute * 60 # Print simplified performance summary from rich.console import Console from rich.table import Table console = Console() # Create a simple summary table table = Table(title="CRAWL4AI PERFORMANCE SUMMARY") table.add_column("Metric", style="cyan") table.add_column("Value", style="green") table.add_row("Total Crawls Completed", f"{successes}") table.add_row("Total Time", f"{total_test_time:.2f} seconds") table.add_row("Time Per Crawl", f"{time_per_crawl:.2f} seconds") table.add_row("Crawling Speed", f"{crawls_per_second:.2f} crawls/second") table.add_row("Projected Rate (1 minute)", f"{crawls_per_minute:.0f} crawls") table.add_row("Projected Rate (1 hour)", f"{crawls_per_hour:.0f} crawls") table.add_row("Peak Memory Usage", f"{peak_memory:.2f} MB") table.add_row("Memory Per Crawl", f"{memory_per_page:.2f} MB") # Display the table console.print(table) # Ask confirmation before cleanup confirmation = input( f"{WARNING}Do you want to proceed with cleanup? (y/n): {RESET}" ) if confirmation.lower() != "y": print(f"{WARNING}Cleanup aborted by user{RESET}") return False # Close all pages for page, _ in all_pages: try: await page.close() except: pass # Close all managers for manager in managers: try: await manager.close() except: pass # Remove the temp directory import shutil if os.path.exists(temp_dir): shutil.rmtree(temp_dir) return True async def main(): """Run all tests""" try: print(f"{INFO}Starting builtin browser tests with browser module{RESET}") # # Run browser creation test # manager, cdp_url = await test_builtin_browser_creation() # if not manager: # print(f"{ERROR}Browser creation failed, cannot continue tests{RESET}") # return # # Run page operations test # await test_page_operations(manager) # # Run browser status and management test # await test_browser_status_management(manager) # # Close manager before multiple manager test # await manager.close() # Run multiple managers test await test_multiple_managers() # Run performance scaling test await test_performance_scaling() # Run cleanup test await cleanup_browsers() # Run edge cases test await test_edge_cases() print(f"\n{SUCCESS}All tests completed!{RESET}") except Exception as e: print(f"\n{ERROR}Test failed with error: {str(e)}{RESET}") import traceback traceback.print_exc() finally: # Clean up: kill any remaining builtin browsers await cleanup_browsers() print(f"{SUCCESS}Test cleanup complete{RESET}") if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_parallel_crawling.py
tests/browser/test_parallel_crawling.py
""" Test examples for parallel crawling with the browser module. These examples demonstrate the functionality of parallel page creation and serve as functional tests for multi-page crawling performance. """ import asyncio import os import sys import time from typing import List # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.browser import BrowserManager from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def test_get_pages_basic(): """Test basic functionality of get_pages method.""" logger.info("Testing basic get_pages functionality", tag="TEST") browser_config = BrowserConfig(headless=True) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() # Request 3 pages crawler_config = CrawlerRunConfig() pages = await manager.get_pages(crawler_config, count=3) # Verify we got the correct number of pages assert len(pages) == 3, f"Expected 3 pages, got {len(pages)}" # Verify each page is valid for i, (page, context) in enumerate(pages): await page.goto("https://example.com") title = await page.title() logger.info(f"Page {i+1} title: {title}", tag="TEST") assert title, f"Page {i+1} has no title" await manager.close() logger.success("Basic get_pages test completed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def test_parallel_approaches_comparison(): """Compare two parallel crawling approaches: 1. Create a page for each URL on-demand (get_page + gather) 2. Get all pages upfront with get_pages, then use them (get_pages + gather) """ logger.info("Comparing different parallel crawling approaches", tag="TEST") urls = [ "https://example.com/page1", "https://crawl4ai.com", "https://kidocode.com", "https://bbc.com", # "https://example.com/page1", # "https://example.com/page2", # "https://example.com/page3", # "https://example.com/page4", ] browser_config = BrowserConfig(headless=False) manager = BrowserManager(browser_config=browser_config, logger=logger) try: await manager.start() # Approach 1: Create a page for each URL on-demand and run in parallel logger.info("Testing approach 1: get_page for each URL + gather", tag="TEST") start_time = time.time() async def fetch_title_approach1(url): """Create a new page for each URL, go to the URL, and get title""" crawler_config = CrawlerRunConfig(url=url) page, context = await manager.get_page(crawler_config) try: await page.goto(url) title = await page.title() return title finally: await page.close() # Run fetch_title_approach1 for each URL in parallel tasks = [fetch_title_approach1(url) for url in urls] approach1_results = await asyncio.gather(*tasks) approach1_time = time.time() - start_time logger.info(f"Approach 1 time (get_page + gather): {approach1_time:.2f}s", tag="TEST") # Approach 2: Get all pages upfront with get_pages, then use them in parallel logger.info("Testing approach 2: get_pages upfront + gather", tag="TEST") start_time = time.time() # Get all pages upfront crawler_config = CrawlerRunConfig() pages = await manager.get_pages(crawler_config, count=len(urls)) async def fetch_title_approach2(page_ctx, url): """Use a pre-created page to go to URL and get title""" page, _ = page_ctx try: await page.goto(url) title = await page.title() return title finally: await page.close() # Use the pre-created pages to fetch titles in parallel tasks = [fetch_title_approach2(page_ctx, url) for page_ctx, url in zip(pages, urls)] approach2_results = await asyncio.gather(*tasks) approach2_time = time.time() - start_time logger.info(f"Approach 2 time (get_pages + gather): {approach2_time:.2f}s", tag="TEST") # Compare results and performance speedup = approach1_time / approach2_time if approach2_time > 0 else 0 if speedup > 1: logger.success(f"Approach 2 (get_pages upfront) was {speedup:.2f}x faster", tag="TEST") else: logger.info(f"Approach 1 (get_page + gather) was {1/speedup:.2f}x faster", tag="TEST") # Verify same content was retrieved in both approaches assert len(approach1_results) == len(approach2_results), "Result count mismatch" # Sort results for comparison since parallel execution might complete in different order assert sorted(approach1_results) == sorted(approach2_results), "Results content mismatch" await manager.close() return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") try: await manager.close() except: pass return False async def test_multi_browser_scaling(num_browsers=3, pages_per_browser=5): """Test performance with multiple browsers and pages per browser. Compares two approaches: 1. On-demand page creation (get_page + gather) 2. Pre-created pages (get_pages + gather) """ logger.info(f"Testing multi-browser scaling with {num_browsers} browsers × {pages_per_browser} pages", tag="TEST") # Generate test URLs total_pages = num_browsers * pages_per_browser urls = [f"https://example.com/page_{i}" for i in range(total_pages)] # Create browser managers managers = [] base_port = 9222 try: # Start all browsers in parallel start_tasks = [] for i in range(num_browsers): browser_config = BrowserConfig( headless=True # Using default browser mode like in test_parallel_approaches_comparison ) manager = BrowserManager(browser_config=browser_config, logger=logger) start_tasks.append(manager.start()) managers.append(manager) await asyncio.gather(*start_tasks) # Distribute URLs among managers urls_per_manager = {} for i, manager in enumerate(managers): start_idx = i * pages_per_browser end_idx = min(start_idx + pages_per_browser, len(urls)) urls_per_manager[manager] = urls[start_idx:end_idx] # Approach 1: Create a page for each URL on-demand and run in parallel logger.info("Testing approach 1: get_page for each URL + gather", tag="TEST") start_time = time.time() async def fetch_title_approach1(manager, url): """Create a new page for the URL, go to the URL, and get title""" crawler_config = CrawlerRunConfig(url=url) page, context = await manager.get_page(crawler_config) try: await page.goto(url) title = await page.title() return title finally: await page.close() # Run fetch_title_approach1 for each URL in parallel tasks = [] for manager, manager_urls in urls_per_manager.items(): for url in manager_urls: tasks.append(fetch_title_approach1(manager, url)) approach1_results = await asyncio.gather(*tasks) approach1_time = time.time() - start_time logger.info(f"Approach 1 time (get_page + gather): {approach1_time:.2f}s", tag="TEST") # Approach 2: Get all pages upfront with get_pages, then use them in parallel logger.info("Testing approach 2: get_pages upfront + gather", tag="TEST") start_time = time.time() # Get all pages upfront for each manager all_pages = [] for manager, manager_urls in urls_per_manager.items(): crawler_config = CrawlerRunConfig() pages = await manager.get_pages(crawler_config, count=len(manager_urls)) all_pages.extend(zip(pages, manager_urls)) async def fetch_title_approach2(page_ctx, url): """Use a pre-created page to go to URL and get title""" page, _ = page_ctx try: await page.goto(url) title = await page.title() return title finally: await page.close() # Use the pre-created pages to fetch titles in parallel tasks = [fetch_title_approach2(page_ctx, url) for page_ctx, url in all_pages] approach2_results = await asyncio.gather(*tasks) approach2_time = time.time() - start_time logger.info(f"Approach 2 time (get_pages + gather): {approach2_time:.2f}s", tag="TEST") # Compare results and performance speedup = approach1_time / approach2_time if approach2_time > 0 else 0 pages_per_second = total_pages / approach2_time # Show a simple summary logger.info(f"📊 Summary: {num_browsers} browsers × {pages_per_browser} pages = {total_pages} total crawls", tag="TEST") logger.info(f"⚡ Performance: {pages_per_second:.1f} pages/second ({pages_per_second*60:.0f} pages/minute)", tag="TEST") logger.info(f"🚀 Total crawl time: {approach2_time:.2f} seconds", tag="TEST") if speedup > 1: logger.success(f"✅ Approach 2 (get_pages upfront) was {speedup:.2f}x faster", tag="TEST") else: logger.info(f"✅ Approach 1 (get_page + gather) was {1/speedup:.2f}x faster", tag="TEST") # Close all managers for manager in managers: await manager.close() return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Clean up for manager in managers: try: await manager.close() except: pass return False async def grid_search_optimal_configuration(total_urls=50): """Perform a grid search to find the optimal balance between number of browsers and pages per browser. This function tests different combinations of browser count and pages per browser, while keeping the total number of URLs constant. It measures performance metrics for each configuration to find the "sweet spot" that provides the best speed with reasonable memory usage. Args: total_urls: Total number of URLs to crawl (default: 50) """ logger.info(f"=== GRID SEARCH FOR OPTIMAL CRAWLING CONFIGURATION ({total_urls} URLs) ===", tag="TEST") # Generate test URLs once urls = [f"https://example.com/page_{i}" for i in range(total_urls)] # Define grid search configurations # We'll use more flexible approach: test all browser counts from 1 to min(20, total_urls) # and distribute pages evenly (some browsers may have 1 more page than others) configurations = [] # Maximum number of browsers to test max_browsers_to_test = min(20, total_urls) # Try configurations with 1 to max_browsers_to_test browsers for num_browsers in range(1, max_browsers_to_test + 1): base_pages_per_browser = total_urls // num_browsers remainder = total_urls % num_browsers # Generate exact page distribution array if remainder > 0: # First 'remainder' browsers get one more page page_distribution = [base_pages_per_browser + 1] * remainder + [base_pages_per_browser] * (num_browsers - remainder) pages_distribution = f"{base_pages_per_browser+1} pages × {remainder} browsers, {base_pages_per_browser} pages × {num_browsers - remainder} browsers" else: # All browsers get the same number of pages page_distribution = [base_pages_per_browser] * num_browsers pages_distribution = f"{base_pages_per_browser} pages × {num_browsers} browsers" # Format the distribution as a tuple string like (4, 4, 3, 3) distribution_str = str(tuple(page_distribution)) configurations.append((num_browsers, base_pages_per_browser, pages_distribution, page_distribution, distribution_str)) # Track results results = [] # Test each configuration for num_browsers, pages_per_browser, pages_distribution, page_distribution, distribution_str in configurations: logger.info("-" * 80, tag="TEST") logger.info(f"Testing configuration: {num_browsers} browsers with distribution: {distribution_str}", tag="TEST") logger.info(f"Details: {pages_distribution}", tag="TEST") # Sleep a bit for randomness await asyncio.sleep(0.5) try: # Import psutil for memory tracking try: import psutil process = psutil.Process() initial_memory = process.memory_info().rss / (1024 * 1024) # MB except ImportError: logger.warning("psutil not available, memory metrics will not be tracked", tag="TEST") initial_memory = 0 # Create and start browser managers managers = [] start_time = time.time() # Start all browsers in parallel start_tasks = [] for i in range(num_browsers): browser_config = BrowserConfig( headless=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) start_tasks.append(manager.start()) managers.append(manager) await asyncio.gather(*start_tasks) browser_startup_time = time.time() - start_time # Measure memory after browser startup if initial_memory > 0: browser_memory = process.memory_info().rss / (1024 * 1024) - initial_memory else: browser_memory = 0 # Distribute URLs among managers using the exact page distribution urls_per_manager = {} total_assigned = 0 for i, manager in enumerate(managers): if i < len(page_distribution): # Get the exact number of pages for this browser from our distribution manager_pages = page_distribution[i] # Get the URL slice for this manager start_idx = total_assigned end_idx = start_idx + manager_pages urls_per_manager[manager] = urls[start_idx:end_idx] total_assigned += manager_pages else: # If we have more managers than our distribution (should never happen) urls_per_manager[manager] = [] # Use the more efficient approach (pre-created pages) logger.info("Running page crawling test...", tag="TEST") crawl_start_time = time.time() # Get all pages upfront for each manager all_pages = [] for manager, manager_urls in urls_per_manager.items(): if not manager_urls: # Skip managers with no URLs continue crawler_config = CrawlerRunConfig() pages = await manager.get_pages(crawler_config, count=len(manager_urls)) all_pages.extend(zip(pages, manager_urls)) # Measure memory after page creation if initial_memory > 0: pages_memory = process.memory_info().rss / (1024 * 1024) - browser_memory - initial_memory else: pages_memory = 0 # Function to crawl a URL with a pre-created page async def fetch_title(page_ctx, url): page, _ = page_ctx try: await page.goto(url) title = await page.title() return title finally: await page.close() # Use the pre-created pages to fetch titles in parallel tasks = [fetch_title(page_ctx, url) for page_ctx, url in all_pages] crawl_results = await asyncio.gather(*tasks) crawl_time = time.time() - crawl_start_time total_time = time.time() - start_time # Final memory measurement if initial_memory > 0: peak_memory = max(browser_memory + pages_memory, process.memory_info().rss / (1024 * 1024) - initial_memory) else: peak_memory = 0 # Close all managers for manager in managers: await manager.close() # Calculate metrics pages_per_second = total_urls / crawl_time # Store result metrics result = { "num_browsers": num_browsers, "pages_per_browser": pages_per_browser, "page_distribution": page_distribution, "distribution_str": distribution_str, "total_urls": total_urls, "browser_startup_time": browser_startup_time, "crawl_time": crawl_time, "total_time": total_time, "browser_memory": browser_memory, "pages_memory": pages_memory, "peak_memory": peak_memory, "pages_per_second": pages_per_second, # Calculate efficiency score (higher is better) # This balances speed vs memory usage "efficiency_score": pages_per_second / (peak_memory + 1) if peak_memory > 0 else pages_per_second, } results.append(result) # Log the results logger.info(f"Browser startup: {browser_startup_time:.2f}s", tag="TEST") logger.info(f"Crawl time: {crawl_time:.2f}s", tag="TEST") logger.info(f"Total time: {total_time:.2f}s", tag="TEST") logger.info(f"Performance: {pages_per_second:.1f} pages/second", tag="TEST") if peak_memory > 0: logger.info(f"Browser memory: {browser_memory:.1f}MB", tag="TEST") logger.info(f"Pages memory: {pages_memory:.1f}MB", tag="TEST") logger.info(f"Peak memory: {peak_memory:.1f}MB", tag="TEST") logger.info(f"Efficiency score: {result['efficiency_score']:.6f}", tag="TEST") except Exception as e: logger.error(f"Error testing configuration: {str(e)}", tag="TEST") import traceback traceback.print_exc() # Clean up for manager in managers: try: await manager.close() except: pass # Print summary of all configurations logger.info("=" * 100, tag="TEST") logger.info("GRID SEARCH RESULTS SUMMARY", tag="TEST") logger.info("=" * 100, tag="TEST") # Rank configurations by efficiency score ranked_results = sorted(results, key=lambda x: x["efficiency_score"], reverse=True) # Also determine rankings by different metrics fastest = sorted(results, key=lambda x: x["crawl_time"])[0] lowest_memory = sorted(results, key=lambda x: x["peak_memory"] if x["peak_memory"] > 0 else float('inf'))[0] most_efficient = ranked_results[0] # Print top performers by category logger.info("🏆 TOP PERFORMERS BY CATEGORY:", tag="TEST") logger.info(f"⚡ Fastest: {fastest['num_browsers']} browsers × ~{fastest['pages_per_browser']} pages " + f"({fastest['crawl_time']:.2f}s, {fastest['pages_per_second']:.1f} pages/s)", tag="TEST") if lowest_memory["peak_memory"] > 0: logger.info(f"💾 Lowest memory: {lowest_memory['num_browsers']} browsers × ~{lowest_memory['pages_per_browser']} pages " + f"({lowest_memory['peak_memory']:.1f}MB)", tag="TEST") logger.info(f"🌟 Most efficient: {most_efficient['num_browsers']} browsers × ~{most_efficient['pages_per_browser']} pages " + f"(score: {most_efficient['efficiency_score']:.6f})", tag="TEST") # Print result table header logger.info("\n📊 COMPLETE RANKING TABLE (SORTED BY EFFICIENCY SCORE):", tag="TEST") logger.info("-" * 120, tag="TEST") # Define table header header = f"{'Rank':<5} | {'Browsers':<8} | {'Distribution':<55} | {'Total Time(s)':<12} | {'Speed(p/s)':<12} | {'Memory(MB)':<12} | {'Efficiency':<10} | {'Notes'}" logger.info(header, tag="TEST") logger.info("-" * 120, tag="TEST") # Print each configuration in ranked order for rank, result in enumerate(ranked_results, 1): # Add special notes for top performers notes = [] if result == fastest: notes.append("⚡ Fastest") if result == lowest_memory: notes.append("💾 Lowest Memory") if result == most_efficient: notes.append("🌟 Most Efficient") notes_str = " | ".join(notes) if notes else "" # Format memory if available memory_str = f"{result['peak_memory']:.1f}" if result['peak_memory'] > 0 else "N/A" # Get the distribution string dist_str = result.get('distribution_str', str(tuple([result['pages_per_browser']] * result['num_browsers']))) # Build the row row = f"{rank:<5} | {result['num_browsers']:<8} | {dist_str:<55} | {result['total_time']:.2f}s{' ':<7} | " row += f"{result['pages_per_second']:.2f}{' ':<6} | {memory_str}{' ':<6} | {result['efficiency_score']:.4f}{' ':<4} | {notes_str}" logger.info(row, tag="TEST") logger.info("-" * 120, tag="TEST") # Generate visualization if matplotlib is available try: import matplotlib.pyplot as plt import numpy as np # Extract data for plotting from ranked results browser_counts = [r["num_browsers"] for r in ranked_results] efficiency_scores = [r["efficiency_score"] for r in ranked_results] crawl_times = [r["crawl_time"] for r in ranked_results] total_times = [r["total_time"] for r in ranked_results] # Filter results with memory data memory_results = [r for r in ranked_results if r["peak_memory"] > 0] memory_browser_counts = [r["num_browsers"] for r in memory_results] peak_memories = [r["peak_memory"] for r in memory_results] # Create figure with clean design plt.figure(figsize=(14, 12), facecolor='white') plt.style.use('ggplot') # Create grid for subplots gs = plt.GridSpec(3, 1, height_ratios=[1, 1, 1], hspace=0.3) # Plot 1: Efficiency Score (higher is better) ax1 = plt.subplot(gs[0]) bar_colors = ['#3498db'] * len(browser_counts) # Highlight the most efficient most_efficient_idx = browser_counts.index(most_efficient["num_browsers"]) bar_colors[most_efficient_idx] = '#e74c3c' # Red for most efficient bars = ax1.bar(range(len(browser_counts)), efficiency_scores, color=bar_colors) ax1.set_xticks(range(len(browser_counts))) ax1.set_xticklabels([f"{bc}" for bc in browser_counts], rotation=45) ax1.set_xlabel('Number of Browsers') ax1.set_ylabel('Efficiency Score (higher is better)') ax1.set_title('Browser Configuration Efficiency (higher is better)') # Add value labels on top of bars for bar, score in zip(bars, efficiency_scores): height = bar.get_height() ax1.text(bar.get_x() + bar.get_width()/2., height + 0.02*max(efficiency_scores), f'{score:.3f}', ha='center', va='bottom', rotation=90, fontsize=8) # Highlight best configuration ax1.text(0.02, 0.90, f"🌟 Most Efficient: {most_efficient['num_browsers']} browsers with ~{most_efficient['pages_per_browser']} pages", transform=ax1.transAxes, fontsize=12, verticalalignment='top', bbox=dict(boxstyle='round,pad=0.5', facecolor='yellow', alpha=0.3)) # Plot 2: Time Performance ax2 = plt.subplot(gs[1]) # Plot both total time and crawl time ax2.plot(browser_counts, crawl_times, 'bo-', label='Crawl Time (s)', linewidth=2) ax2.plot(browser_counts, total_times, 'go--', label='Total Time (s)', linewidth=2, alpha=0.6) # Mark the fastest configuration fastest_idx = browser_counts.index(fastest["num_browsers"]) ax2.plot(browser_counts[fastest_idx], crawl_times[fastest_idx], 'ro', ms=10, label=f'Fastest: {fastest["num_browsers"]} browsers') ax2.set_xlabel('Number of Browsers') ax2.set_ylabel('Time (seconds)') ax2.set_title(f'Time Performance for {total_urls} URLs by Browser Count') ax2.grid(True, linestyle='--', alpha=0.7) ax2.legend(loc='upper right') # Plot pages per second on second y-axis pages_per_second = [total_urls/t for t in crawl_times] ax2_twin = ax2.twinx() ax2_twin.plot(browser_counts, pages_per_second, 'r^--', label='Pages/second', alpha=0.5) ax2_twin.set_ylabel('Pages per second') # Add note about the fastest configuration ax2.text(0.02, 0.90, f"⚡ Fastest: {fastest['num_browsers']} browsers with ~{fastest['pages_per_browser']} pages" + f"\n {fastest['crawl_time']:.2f}s ({fastest['pages_per_second']:.1f} pages/s)", transform=ax2.transAxes, fontsize=12, verticalalignment='top', bbox=dict(boxstyle='round,pad=0.5', facecolor='lightblue', alpha=0.3)) # Plot 3: Memory Usage (if available) if memory_results: ax3 = plt.subplot(gs[2]) # Prepare data for grouped bar chart memory_per_browser = [m/n for m, n in zip(peak_memories, memory_browser_counts)] memory_per_page = [m/(n*p) for m, n, p in zip( [r["peak_memory"] for r in memory_results], [r["num_browsers"] for r in memory_results], [r["pages_per_browser"] for r in memory_results])] x = np.arange(len(memory_browser_counts)) width = 0.35 # Create grouped bars ax3.bar(x - width/2, peak_memories, width, label='Total Memory (MB)', color='#9b59b6') ax3.bar(x + width/2, memory_per_browser, width, label='Memory per Browser (MB)', color='#3498db') # Configure axis ax3.set_xticks(x) ax3.set_xticklabels([f"{bc}" for bc in memory_browser_counts], rotation=45) ax3.set_xlabel('Number of Browsers') ax3.set_ylabel('Memory (MB)') ax3.set_title('Memory Usage by Browser Configuration') ax3.legend(loc='upper left') ax3.grid(True, linestyle='--', alpha=0.7) # Add second y-axis for memory per page ax3_twin = ax3.twinx() ax3_twin.plot(x, memory_per_page, 'ro-', label='Memory per Page (MB)') ax3_twin.set_ylabel('Memory per Page (MB)') # Get lowest memory configuration lowest_memory_idx = memory_browser_counts.index(lowest_memory["num_browsers"]) # Add note about lowest memory configuration ax3.text(0.02, 0.90, f"💾 Lowest Memory: {lowest_memory['num_browsers']} browsers with ~{lowest_memory['pages_per_browser']} pages" + f"\n {lowest_memory['peak_memory']:.1f}MB ({lowest_memory['peak_memory']/total_urls:.2f}MB per page)", transform=ax3.transAxes, fontsize=12, verticalalignment='top', bbox=dict(boxstyle='round,pad=0.5', facecolor='lightgreen', alpha=0.3)) # Add overall title plt.suptitle(f'Browser Scaling Grid Search Results for {total_urls} URLs', fontsize=16, y=0.98) # Add timestamp and info at the bottom plt.figtext(0.5, 0.01, f"Generated by Crawl4AI at {time.strftime('%Y-%m-%d %H:%M:%S')}", ha="center", fontsize=10, style='italic') # Get current directory and save the figure there import os __current_file = os.path.abspath(__file__) current_dir = os.path.dirname(__current_file) output_file = os.path.join(current_dir, 'browser_scaling_grid_search.png') # Adjust layout and save figure with high DPI plt.tight_layout(rect=[0, 0.03, 1, 0.97]) plt.savefig(output_file, dpi=200, bbox_inches='tight') logger.success(f"Visualization saved to {output_file}", tag="TEST") except ImportError: logger.warning("matplotlib not available, skipping visualization", tag="TEST") return most_efficient["num_browsers"], most_efficient["pages_per_browser"] async def find_optimal_browser_config(total_urls=50, verbose=True, rate_limit_delay=0.2): """Find optimal browser configuration for crawling a specific number of URLs. Args: total_urls: Number of URLs to crawl verbose: Whether to print progress rate_limit_delay: Delay between page loads to avoid rate limiting Returns: dict: Contains fastest, lowest_memory, and optimal configurations """ if verbose: print(f"\n=== Finding optimal configuration for crawling {total_urls} URLs ===\n") # Generate test URLs with timestamp to avoid caching timestamp = int(time.time()) urls = [f"https://example.com/page_{i}?t={timestamp}" for i in range(total_urls)] # Limit browser configurations to test (1 browser to max 10) max_browsers = min(10, total_urls) configs_to_test = [] # Generate configurations (browser count, pages distribution) for num_browsers in range(1, max_browsers + 1): base_pages = total_urls // num_browsers remainder = total_urls % num_browsers # Create distribution array like [3, 3, 2, 2] (some browsers get one more page) if remainder > 0: distribution = [base_pages + 1] * remainder + [base_pages] * (num_browsers - remainder) else: distribution = [base_pages] * num_browsers configs_to_test.append((num_browsers, distribution)) results = [] # Test each configuration for browser_count, page_distribution in configs_to_test: if verbose: print(f"Testing {browser_count} browsers with distribution {tuple(page_distribution)}") try: # Track memory if possible try: import psutil process = psutil.Process() start_memory = process.memory_info().rss / (1024 * 1024) # MB except ImportError: if verbose: print("Memory tracking not available (psutil not installed)") start_memory = 0 # Start browsers in parallel managers = [] start_tasks = [] start_time = time.time() for i in range(browser_count): config = BrowserConfig(headless=True)
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_browser_manager.py
tests/browser/test_browser_manager.py
"""Test examples for BrowserManager. These examples demonstrate the functionality of BrowserManager and serve as functional tests. """ import asyncio import os import sys from typing import List # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.browser import BrowserManager from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def test_basic_browser_manager(): """Test basic BrowserManager functionality with default configuration.""" logger.info("Starting test_basic_browser_manager", tag="TEST") try: # Create a browser manager with default config manager = BrowserManager(logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Get a page crawler_config = CrawlerRunConfig(url="https://example.com") page, context = await manager.get_page(crawler_config) logger.info("Page created successfully", tag="TEST") # Navigate to a website await page.goto("https://example.com") title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Clean up await manager.close() logger.success("test_basic_browser_manager completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_basic_browser_manager failed: {str(e)}", tag="TEST") return False async def test_custom_browser_config(): """Test BrowserManager with custom browser configuration.""" logger.info("Starting test_custom_browser_config", tag="TEST") try: # Create a custom browser config browser_config = BrowserConfig( browser_type="chromium", headless=True, viewport_width=1280, viewport_height=800, light_mode=True ) # Create browser manager with the config manager = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully with custom config", tag="TEST") # Get a page crawler_config = CrawlerRunConfig(url="https://example.com") page, context = await manager.get_page(crawler_config) # Navigate to a website await page.goto("https://example.com") title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Verify viewport size viewport_size = await page.evaluate("() => ({ width: window.innerWidth, height: window.innerHeight })") logger.info(f"Viewport size: {viewport_size}", tag="TEST") # Clean up await manager.close() logger.success("test_custom_browser_config completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_custom_browser_config failed: {str(e)}", tag="TEST") return False async def test_multiple_pages(): """Test BrowserManager with multiple pages.""" logger.info("Starting test_multiple_pages", tag="TEST") try: # Create browser manager manager = BrowserManager(logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create multiple pages pages = [] urls = ["https://example.com", "https://example.org", "https://mozilla.org"] for i, url in enumerate(urls): crawler_config = CrawlerRunConfig(url=url) page, context = await manager.get_page(crawler_config) await page.goto(url) pages.append((page, url)) logger.info(f"Created page {i+1} for {url}", tag="TEST") # Verify all pages are loaded correctly for i, (page, url) in enumerate(pages): title = await page.title() logger.info(f"Page {i+1} title: {title}", tag="TEST") # Clean up await manager.close() logger.success("test_multiple_pages completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_multiple_pages failed: {str(e)}", tag="TEST") return False async def test_session_management(): """Test session management in BrowserManager.""" logger.info("Starting test_session_management", tag="TEST") try: # Create browser manager manager = BrowserManager(logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create a session session_id = "test_session_1" crawler_config = CrawlerRunConfig(url="https://example.com", session_id=session_id) page1, context1 = await manager.get_page(crawler_config) await page1.goto("https://example.com") logger.info(f"Created session with ID: {session_id}", tag="TEST") # Get the same session again page2, context2 = await manager.get_page(crawler_config) # Verify it's the same page/context is_same_page = page1 == page2 is_same_context = context1 == context2 logger.info(f"Same page: {is_same_page}, Same context: {is_same_context}", tag="TEST") # Kill the session await manager.kill_session(session_id) logger.info(f"Killed session with ID: {session_id}", tag="TEST") # Clean up await manager.close() logger.success("test_session_management completed successfully", tag="TEST") return True except Exception as e: logger.error(f"test_session_management failed: {str(e)}", tag="TEST") return False async def run_tests(): """Run all tests sequentially.""" results = [] results.append(await test_basic_browser_manager()) results.append(await test_custom_browser_config()) results.append(await test_multiple_pages()) results.append(await test_session_management()) # Print summary total = len(results) passed = sum(results) logger.info(f"Tests complete: {passed}/{total} passed", tag="SUMMARY") if passed == total: logger.success("All tests passed!", tag="SUMMARY") else: logger.error(f"{total - passed} tests failed", tag="SUMMARY") if __name__ == "__main__": asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_profiles.py
tests/browser/test_profiles.py
"""Test examples for BrowserProfileManager. These examples demonstrate the functionality of BrowserProfileManager and serve as functional tests. """ import asyncio import os import sys import uuid import shutil from crawl4ai import BrowserProfiler from crawl4ai.browser_manager import BrowserManager # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../..'))) from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) async def test_profile_creation(): """Test creating and managing browser profiles.""" logger.info("Testing profile creation and management", tag="TEST") profile_manager = BrowserProfiler(logger=logger) try: # List existing profiles profiles = profile_manager.list_profiles() logger.info(f"Found {len(profiles)} existing profiles", tag="TEST") # Generate a unique profile name for testing test_profile_name = f"test-profile-{uuid.uuid4().hex[:8]}" # Create a test profile directory profile_path = os.path.join(profile_manager.profiles_dir, test_profile_name) os.makedirs(os.path.join(profile_path, "Default"), exist_ok=True) # Create a dummy Preferences file to simulate a Chrome profile with open(os.path.join(profile_path, "Default", "Preferences"), "w") as f: f.write("{\"test\": true}") logger.info(f"Created test profile at: {profile_path}", tag="TEST") # Verify the profile is now in the list profiles = profile_manager.list_profiles() profile_found = any(p["name"] == test_profile_name for p in profiles) logger.info(f"Profile found in list: {profile_found}", tag="TEST") # Try to get the profile path retrieved_path = profile_manager.get_profile_path(test_profile_name) path_match = retrieved_path == profile_path logger.info(f"Retrieved correct profile path: {path_match}", tag="TEST") # Delete the profile success = profile_manager.delete_profile(test_profile_name) logger.info(f"Profile deletion successful: {success}", tag="TEST") # Verify it's gone profiles_after = profile_manager.list_profiles() profile_removed = not any(p["name"] == test_profile_name for p in profiles_after) logger.info(f"Profile removed from list: {profile_removed}", tag="TEST") # Clean up just in case if os.path.exists(profile_path): shutil.rmtree(profile_path, ignore_errors=True) return profile_found and path_match and success and profile_removed except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Clean up test directory try: if os.path.exists(profile_path): shutil.rmtree(profile_path, ignore_errors=True) except: pass return False async def test_profile_with_browser(): """Test using a profile with a browser.""" logger.info("Testing using a profile with a browser", tag="TEST") profile_manager = BrowserProfiler(logger=logger) test_profile_name = f"test-browser-profile-{uuid.uuid4().hex[:8]}" profile_path = None try: # Create a test profile directory profile_path = os.path.join(profile_manager.profiles_dir, test_profile_name) os.makedirs(os.path.join(profile_path, "Default"), exist_ok=True) # Create a dummy Preferences file to simulate a Chrome profile with open(os.path.join(profile_path, "Default", "Preferences"), "w") as f: f.write("{\"test\": true}") logger.info(f"Created test profile at: {profile_path}", tag="TEST") # Now use this profile with a browser browser_config = BrowserConfig( user_data_dir=profile_path, use_managed_browser=True, use_persistent_context=True, headless=True ) manager = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser with the profile await manager.start() logger.info("Browser started with profile", tag="TEST") # Create a page crawler_config = CrawlerRunConfig() page, context = await manager.get_page(crawler_config) # Navigate and set some data to verify profile works await page.goto("https://example.com") await page.evaluate("localStorage.setItem('test_data', 'profile_value')") # Close browser await manager.close() logger.info("First browser session closed", tag="TEST") # Create a new browser with the same profile manager2 = BrowserManager(browser_config=browser_config, logger=logger) await manager2.start() logger.info("Second browser session started with same profile", tag="TEST") # Get a page and check if the data persists page2, context2 = await manager2.get_page(crawler_config) await page2.goto("https://example.com") data = await page2.evaluate("localStorage.getItem('test_data')") # Verify data persisted data_persisted = data == "profile_value" logger.info(f"Data persisted across sessions: {data_persisted}", tag="TEST") # Clean up await manager2.close() logger.info("Second browser session closed", tag="TEST") # Delete the test profile success = profile_manager.delete_profile(test_profile_name) logger.info(f"Test profile deleted: {success}", tag="TEST") return data_persisted and success except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Clean up try: if profile_path and os.path.exists(profile_path): shutil.rmtree(profile_path, ignore_errors=True) except: pass return False async def run_tests(): """Run all tests sequentially.""" results = [] results.append(await test_profile_creation()) results.append(await test_profile_with_browser()) # Print summary total = len(results) passed = sum(results) logger.info(f"Tests complete: {passed}/{total} passed", tag="SUMMARY") if passed == total: logger.success("All tests passed!", tag="SUMMARY") else: logger.error(f"{total - passed} tests failed", tag="SUMMARY") if __name__ == "__main__": asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/test_launch_standalone.py
tests/browser/test_launch_standalone.py
from crawl4ai.browser_profiler import BrowserProfiler import asyncio if __name__ == "__main__": # Test launching a standalone browser async def test_standalone_browser(): profiler = BrowserProfiler() cdp_url = await profiler.launch_standalone_browser( browser_type="chromium", user_data_dir="~/.crawl4ai/browser_profile/test-browser-data", debugging_port=9222, headless=False ) print(f"CDP URL: {cdp_url}") asyncio.run(test_standalone_browser())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/docker/__init__.py
tests/browser/docker/__init__.py
"""Docker browser strategy tests. This package contains tests for the Docker browser strategy implementation. """
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/docker/test_docker_browser.py
tests/browser/docker/test_docker_browser.py
"""Test examples for Docker Browser Strategy. These examples demonstrate the functionality of Docker Browser Strategy and serve as functional tests. """ import asyncio import os import sys import shutil import uuid # Add the project root to Python path if running directly if __name__ == "__main__": sys.path.insert(0, os.path.abspath(os.path.join(os.path.dirname(__file__), '../../..'))) from crawl4ai.browser import BrowserManager from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger from crawl4ai.browser import DockerConfig from crawl4ai.browser import DockerRegistry from crawl4ai.browser import DockerUtils # Create a logger for clear terminal output logger = AsyncLogger(verbose=True, log_file=None) # Global Docker utils instance docker_utils = DockerUtils(logger) async def test_docker_components(): """Test Docker utilities, registry, and image building. This function tests the core Docker components before running the browser tests. It validates DockerRegistry, DockerUtils, and builds test images to ensure everything is functioning correctly. """ logger.info("Testing Docker components", tag="SETUP") # Create a test registry directory registry_dir = os.path.join(os.path.dirname(__file__), "test_registry") registry_file = os.path.join(registry_dir, "test_registry.json") os.makedirs(registry_dir, exist_ok=True) try: # 1. Test DockerRegistry logger.info("Testing DockerRegistry...", tag="SETUP") registry = DockerRegistry(registry_file) # Test saving and loading registry test_container_id = "test-container-123" registry.register_container(test_container_id, 9876, "test-hash-123") registry.save() # Create a new registry instance that loads from the file registry2 = DockerRegistry(registry_file) port = registry2.get_container_host_port(test_container_id) hash_value = registry2.get_container_config_hash(test_container_id) if port != 9876 or hash_value != "test-hash-123": logger.error("DockerRegistry persistence failed", tag="SETUP") return False # Clean up test container from registry registry2.unregister_container(test_container_id) logger.success("DockerRegistry works correctly", tag="SETUP") # 2. Test DockerUtils logger.info("Testing DockerUtils...", tag="SETUP") # Test port detection in_use = docker_utils.is_port_in_use(22) # SSH port is usually in use logger.info(f"Port 22 in use: {in_use}", tag="SETUP") # Get next available port available_port = docker_utils.get_next_available_port(9000) logger.info(f"Next available port: {available_port}", tag="SETUP") # Test config hash generation config_dict = {"mode": "connect", "headless": True} config_hash = docker_utils.generate_config_hash(config_dict) logger.info(f"Generated config hash: {config_hash[:8]}...", tag="SETUP") # 3. Test Docker is available logger.info("Checking Docker availability...", tag="SETUP") if not await check_docker_available(): logger.error("Docker is not available - cannot continue tests", tag="SETUP") return False # 4. Test building connect image logger.info("Building connect mode Docker image...", tag="SETUP") connect_image = await docker_utils.ensure_docker_image_exists(None, "connect") if not connect_image: logger.error("Failed to build connect mode image", tag="SETUP") return False logger.success(f"Successfully built connect image: {connect_image}", tag="SETUP") # 5. Test building launch image logger.info("Building launch mode Docker image...", tag="SETUP") launch_image = await docker_utils.ensure_docker_image_exists(None, "launch") if not launch_image: logger.error("Failed to build launch mode image", tag="SETUP") return False logger.success(f"Successfully built launch image: {launch_image}", tag="SETUP") # 6. Test creating and removing container logger.info("Testing container creation and removal...", tag="SETUP") container_id = await docker_utils.create_container( image_name=launch_image, host_port=available_port, container_name="crawl4ai-test-container" ) if not container_id: logger.error("Failed to create test container", tag="SETUP") return False logger.info(f"Created test container: {container_id[:12]}", tag="SETUP") # Verify container is running running = await docker_utils.is_container_running(container_id) if not running: logger.error("Test container is not running", tag="SETUP") await docker_utils.remove_container(container_id) return False # Test commands in container logger.info("Testing command execution in container...", tag="SETUP") returncode, stdout, stderr = await docker_utils.exec_in_container( container_id, ["ls", "-la", "/"] ) if returncode != 0: logger.error(f"Command execution failed: {stderr}", tag="SETUP") await docker_utils.remove_container(container_id) return False # Verify Chrome is installed in the container returncode, stdout, stderr = await docker_utils.exec_in_container( container_id, ["which", "chromium"] ) if returncode != 0: logger.error("Chrome not found in container", tag="SETUP") await docker_utils.remove_container(container_id) return False chrome_path = stdout.strip() logger.info(f"Chrome found at: {chrome_path}", tag="SETUP") # Test Chrome version returncode, stdout, stderr = await docker_utils.exec_in_container( container_id, ["chromium", "--version"] ) if returncode != 0: logger.error(f"Failed to get Chrome version: {stderr}", tag="SETUP") await docker_utils.remove_container(container_id) return False logger.info(f"Chrome version: {stdout.strip()}", tag="SETUP") # Remove test container removed = await docker_utils.remove_container(container_id) if not removed: logger.error("Failed to remove test container", tag="SETUP") return False logger.success("Test container removed successfully", tag="SETUP") # All components tested successfully logger.success("All Docker components tested successfully", tag="SETUP") return True except Exception as e: logger.error(f"Docker component tests failed: {str(e)}", tag="SETUP") return False finally: # Clean up registry test directory if os.path.exists(registry_dir): shutil.rmtree(registry_dir) async def test_docker_connect_mode(): """Test Docker browser in connect mode. This tests the basic functionality of creating a browser in Docker connect mode and using it for navigation. """ logger.info("Testing Docker browser in connect mode", tag="TEST") # Create temp directory for user data temp_dir = os.path.join(os.path.dirname(__file__), "tmp_user_data") os.makedirs(temp_dir, exist_ok=True) try: # Create Docker configuration docker_config = DockerConfig( mode="connect", persistent=False, remove_on_exit=True, user_data_dir=temp_dir ) # Create browser configuration browser_config = BrowserConfig( browser_mode="docker", headless=True, docker_config=docker_config ) # Create browser manager manager = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create crawler config crawler_config = CrawlerRunConfig(url="https://example.com") # Get a page page, context = await manager.get_page(crawler_config) logger.info("Got page successfully", tag="TEST") # Navigate to a website await page.goto("https://example.com") logger.info("Navigated to example.com", tag="TEST") # Get page title title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False finally: # Clean up the temp directory if os.path.exists(temp_dir): shutil.rmtree(temp_dir) async def test_docker_launch_mode(): """Test Docker browser in launch mode. This tests launching a Chrome browser within a Docker container on demand with custom settings. """ logger.info("Testing Docker browser in launch mode", tag="TEST") # Create temp directory for user data temp_dir = os.path.join(os.path.dirname(__file__), "tmp_user_data_launch") os.makedirs(temp_dir, exist_ok=True) try: # Create Docker configuration docker_config = DockerConfig( mode="launch", persistent=False, remove_on_exit=True, user_data_dir=temp_dir ) # Create browser configuration browser_config = BrowserConfig( browser_mode="docker", headless=True, text_mode=True, # Enable text mode for faster operation docker_config=docker_config ) # Create browser manager manager = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create crawler config crawler_config = CrawlerRunConfig(url="https://example.com") # Get a page page, context = await manager.get_page(crawler_config) logger.info("Got page successfully", tag="TEST") # Navigate to a website await page.goto("https://example.com") logger.info("Navigated to example.com", tag="TEST") # Get page title title = await page.title() logger.info(f"Page title: {title}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False finally: # Clean up the temp directory if os.path.exists(temp_dir): shutil.rmtree(temp_dir) async def test_docker_persistent_storage(): """Test Docker browser with persistent storage. This tests creating localStorage data in one session and verifying it persists to another session when using persistent storage. """ logger.info("Testing Docker browser with persistent storage", tag="TEST") # Create a unique temp directory test_id = uuid.uuid4().hex[:8] temp_dir = os.path.join(os.path.dirname(__file__), f"tmp_user_data_persist_{test_id}") os.makedirs(temp_dir, exist_ok=True) manager1 = None manager2 = None try: # Create Docker configuration with persistence docker_config = DockerConfig( mode="connect", persistent=True, # Keep container running between sessions user_data_dir=temp_dir, container_user_data_dir="/data" ) # Create browser configuration browser_config = BrowserConfig( browser_mode="docker", headless=True, docker_config=docker_config ) # Create first browser manager manager1 = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser await manager1.start() logger.info("First browser started successfully", tag="TEST") # Create crawler config crawler_config = CrawlerRunConfig() # Get a page page1, context1 = await manager1.get_page(crawler_config) # Navigate to example.com await page1.goto("https://example.com") # Set localStorage item test_value = f"test_value_{test_id}" await page1.evaluate(f"localStorage.setItem('test_key', '{test_value}')") logger.info(f"Set localStorage test_key = {test_value}", tag="TEST") # Close the first browser manager await manager1.close() logger.info("First browser closed", tag="TEST") # Create second browser manager with same config manager2 = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser await manager2.start() logger.info("Second browser started successfully", tag="TEST") # Get a page page2, context2 = await manager2.get_page(crawler_config) # Navigate to same site await page2.goto("https://example.com") # Get localStorage item value = await page2.evaluate("localStorage.getItem('test_key')") logger.info(f"Retrieved localStorage test_key = {value}", tag="TEST") # Check if persistence worked if value == test_value: logger.success("Storage persistence verified!", tag="TEST") else: logger.error(f"Storage persistence failed! Expected {test_value}, got {value}", tag="TEST") # Clean up await manager2.close() logger.info("Second browser closed successfully", tag="TEST") return value == test_value except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: if manager1: await manager1.close() if manager2: await manager2.close() except: pass return False finally: # Clean up the temp directory if os.path.exists(temp_dir): shutil.rmtree(temp_dir) async def test_docker_parallel_pages(): """Test Docker browser with parallel page creation. This tests the ability to create and use multiple pages in parallel from a single Docker browser instance. """ logger.info("Testing Docker browser with parallel pages", tag="TEST") try: # Create Docker configuration docker_config = DockerConfig( mode="connect", persistent=False, remove_on_exit=True ) # Create browser configuration browser_config = BrowserConfig( browser_mode="docker", headless=True, docker_config=docker_config ) # Create browser manager manager = BrowserManager(browser_config=browser_config, logger=logger) # Start the browser await manager.start() logger.info("Browser started successfully", tag="TEST") # Create crawler config crawler_config = CrawlerRunConfig() # Get multiple pages page_count = 3 pages = await manager.get_pages(crawler_config, count=page_count) logger.info(f"Got {len(pages)} pages successfully", tag="TEST") if len(pages) != page_count: logger.error(f"Expected {page_count} pages, got {len(pages)}", tag="TEST") await manager.close() return False # Navigate to different sites with each page tasks = [] for i, (page, _) in enumerate(pages): tasks.append(page.goto(f"https://example.com?page={i}")) # Wait for all navigations to complete await asyncio.gather(*tasks) logger.info("All pages navigated successfully", tag="TEST") # Get titles from all pages titles = [] for i, (page, _) in enumerate(pages): title = await page.title() titles.append(title) logger.info(f"Page {i+1} title: {title}", tag="TEST") # Clean up await manager.close() logger.info("Browser closed successfully", tag="TEST") return True except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: await manager.close() except: pass return False async def test_docker_registry_reuse(): """Test Docker container reuse via registry. This tests that containers with matching configurations are reused rather than creating new ones. """ logger.info("Testing Docker container reuse via registry", tag="TEST") # Create registry for this test registry_dir = os.path.join(os.path.dirname(__file__), "registry_reuse_test") registry_file = os.path.join(registry_dir, "registry.json") os.makedirs(registry_dir, exist_ok=True) manager1 = None manager2 = None container_id1 = None try: # Create identical Docker configurations with custom registry docker_config1 = DockerConfig( mode="connect", persistent=True, # Keep container running after closing registry_file=registry_file ) # Create first browser configuration browser_config1 = BrowserConfig( browser_mode="docker", headless=True, docker_config=docker_config1 ) # Create first browser manager manager1 = BrowserManager(browser_config=browser_config1, logger=logger) # Start the first browser await manager1.start() logger.info("First browser started successfully", tag="TEST") # Get container ID from the strategy docker_strategy1 = manager1.strategy container_id1 = docker_strategy1.container_id logger.info(f"First browser container ID: {container_id1[:12]}", tag="TEST") # Close the first manager but keep container running await manager1.close() logger.info("First browser closed", tag="TEST") # Create second Docker configuration identical to first docker_config2 = DockerConfig( mode="connect", persistent=True, registry_file=registry_file ) # Create second browser configuration browser_config2 = BrowserConfig( browser_mode="docker", headless=True, docker_config=docker_config2 ) # Create second browser manager manager2 = BrowserManager(browser_config=browser_config2, logger=logger) # Start the second browser - should reuse existing container await manager2.start() logger.info("Second browser started successfully", tag="TEST") # Get container ID from the second strategy docker_strategy2 = manager2.strategy container_id2 = docker_strategy2.container_id logger.info(f"Second browser container ID: {container_id2[:12]}", tag="TEST") # Verify container reuse if container_id1 == container_id2: logger.success("Container reuse successful - using same container!", tag="TEST") else: logger.error("Container reuse failed - new container created!", tag="TEST") # Clean up docker_strategy2.docker_config.persistent = False docker_strategy2.docker_config.remove_on_exit = True await manager2.close() logger.info("Second browser closed and container removed", tag="TEST") return container_id1 == container_id2 except Exception as e: logger.error(f"Test failed: {str(e)}", tag="TEST") # Ensure cleanup try: if manager1: await manager1.close() if manager2: await manager2.close() # Make sure container is removed if container_id1: await docker_utils.remove_container(container_id1, force=True) except: pass return False finally: # Clean up registry directory if os.path.exists(registry_dir): shutil.rmtree(registry_dir) async def run_tests(): """Run all tests sequentially.""" results = [] logger.info("Starting Docker Browser Strategy tests", tag="TEST") # Check if Docker is available if not await check_docker_available(): logger.error("Docker is not available - skipping tests", tag="TEST") return # First test Docker components # setup_result = await test_docker_components() # if not setup_result: # logger.error("Docker component tests failed - skipping browser tests", tag="TEST") # return # Run browser tests results.append(await test_docker_connect_mode()) results.append(await test_docker_launch_mode()) results.append(await test_docker_persistent_storage()) results.append(await test_docker_parallel_pages()) results.append(await test_docker_registry_reuse()) # Print summary total = len(results) passed = sum(1 for r in results if r) logger.info(f"Tests complete: {passed}/{total} passed", tag="SUMMARY") if passed == total: logger.success("All tests passed!", tag="SUMMARY") else: logger.error(f"{total - passed} tests failed", tag="SUMMARY") async def check_docker_available() -> bool: """Check if Docker is available on the system. Returns: bool: True if Docker is available, False otherwise """ try: proc = await asyncio.create_subprocess_exec( "docker", "--version", stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE ) stdout, _ = await proc.communicate() return proc.returncode == 0 and stdout except: return False if __name__ == "__main__": asyncio.run(run_tests())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/browser/manager/demo_browser_manager.py
tests/browser/manager/demo_browser_manager.py
"""Demo script for testing the enhanced BrowserManager. This script demonstrates the browser pooling capabilities of the enhanced BrowserManager with various configurations and usage patterns. """ import asyncio import time import random from crawl4ai.browser.manager import BrowserManager, UnavailableBehavior from crawl4ai.async_configs import BrowserConfig, CrawlerRunConfig from crawl4ai.async_logger import AsyncLogger import playwright SAFE_URLS = [ "https://example.com", "https://example.com/page1", "https://httpbin.org/get", "https://httpbin.org/html", "https://httpbin.org/ip", "https://httpbin.org/user-agent", "https://httpbin.org/headers", "https://httpbin.org/cookies", "https://httpstat.us/200", "https://httpstat.us/301", "https://httpstat.us/404", "https://httpstat.us/500", "https://jsonplaceholder.typicode.com/posts/1", "https://jsonplaceholder.typicode.com/posts/2", "https://jsonplaceholder.typicode.com/posts/3", "https://jsonplaceholder.typicode.com/posts/4", "https://jsonplaceholder.typicode.com/posts/5", "https://jsonplaceholder.typicode.com/comments/1", "https://jsonplaceholder.typicode.com/comments/2", "https://jsonplaceholder.typicode.com/users/1", "https://jsonplaceholder.typicode.com/users/2", "https://jsonplaceholder.typicode.com/albums/1", "https://jsonplaceholder.typicode.com/albums/2", "https://jsonplaceholder.typicode.com/photos/1", "https://jsonplaceholder.typicode.com/photos/2", "https://jsonplaceholder.typicode.com/todos/1", "https://jsonplaceholder.typicode.com/todos/2", "https://www.iana.org", "https://www.iana.org/domains", "https://www.iana.org/numbers", "https://www.iana.org/protocols", "https://www.iana.org/about", "https://www.iana.org/time-zones", "https://www.data.gov", "https://catalog.data.gov/dataset", "https://www.archives.gov", "https://www.usa.gov", "https://www.loc.gov", "https://www.irs.gov", "https://www.census.gov", "https://www.bls.gov", "https://www.gpo.gov", "https://www.w3.org", "https://www.w3.org/standards", "https://www.w3.org/WAI", "https://www.rfc-editor.org", "https://www.ietf.org", "https://www.icann.org", "https://www.internetsociety.org", "https://www.python.org" ] async def basic_pooling_demo(): """Demonstrate basic browser pooling functionality.""" print("\n=== Basic Browser Pooling Demo ===") # Create logger logger = AsyncLogger(verbose=True) # Create browser configurations config1 = BrowserConfig( browser_type="chromium", headless=True, browser_mode="playwright" ) config2 = BrowserConfig( browser_type="chromium", headless=True, browser_mode="cdp" ) # Create browser manager with on-demand behavior manager = BrowserManager( browser_config=config1, logger=logger, unavailable_behavior=UnavailableBehavior.ON_DEMAND, max_browsers_per_config=3 ) try: # Initialize pool with both configurations print("Initializing browser pool...") await manager.initialize_pool( browser_configs=[config1, config2], browsers_per_config=2 ) # Display initial pool status status = await manager.get_pool_status() print(f"Initial pool status: {status}") # Create crawler run configurations run_config1 = CrawlerRunConfig() run_config2 = CrawlerRunConfig() # Simulate concurrent page requests print("\nGetting pages for parallel crawling...") # Function to simulate crawling async def simulate_crawl(index: int, config: BrowserConfig, run_config: CrawlerRunConfig): print(f"Crawler {index}: Requesting page...") page, context, strategy = await manager.get_page(run_config, config) print(f"Crawler {index}: Got page, navigating to example.com...") try: await page.goto("https://example.com") title = await page.title() print(f"Crawler {index}: Page title: {title}") # Simulate work await asyncio.sleep(random.uniform(1, 3)) print(f"Crawler {index}: Work completed, releasing page...") # Check dynamic page content content = await page.content() content_length = len(content) print(f"Crawler {index}: Page content length: {content_length}") except Exception as e: print(f"Crawler {index}: Error: {str(e)}") finally: # Release the page await manager.release_page(page, strategy, config) print(f"Crawler {index}: Page released") # Create 5 parallel crawls crawl_tasks = [] for i in range(5): # Alternate between configurations config = config1 if i % 2 == 0 else config2 run_config = run_config1 if i % 2 == 0 else run_config2 task = asyncio.create_task(simulate_crawl(i+1, config, run_config)) crawl_tasks.append(task) # Wait for all crawls to complete await asyncio.gather(*crawl_tasks) # Display final pool status status = await manager.get_pool_status() print(f"\nFinal pool status: {status}") finally: # Clean up print("\nClosing browser manager...") await manager.close() print("Browser manager closed") async def prewarm_pages_demo(): """Demonstrate page pre-warming functionality.""" print("\n=== Page Pre-warming Demo ===") # Create logger logger = AsyncLogger(verbose=True) # Create browser configuration config = BrowserConfig( browser_type="chromium", headless=True, browser_mode="playwright" ) # Create crawler run configurations for pre-warming run_config1 = CrawlerRunConfig( user_agent="Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36" ) run_config2 = CrawlerRunConfig( user_agent="Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/14.0 Safari/605.1.15" ) # Create page pre-warm configurations page_configs = [ (config, run_config1, 2), # 2 pages with run_config1 (config, run_config2, 3) # 3 pages with run_config2 ] # Create browser manager manager = BrowserManager( browser_config=config, logger=logger, unavailable_behavior=UnavailableBehavior.EXCEPTION ) try: # Initialize pool with pre-warmed pages print("Initializing browser pool with pre-warmed pages...") await manager.initialize_pool( browser_configs=[config], browsers_per_config=2, page_configs=page_configs ) # Display pool status status = await manager.get_pool_status() print(f"Pool status after pre-warming: {status}") # Simulate using pre-warmed pages print("\nUsing pre-warmed pages...") async def use_prewarm_page(index: int, run_config: CrawlerRunConfig): print(f"Task {index}: Requesting pre-warmed page...") page, context, strategy = await manager.get_page(run_config, config) try: print(f"Task {index}: Got page, navigating to example.com...") await page.goto("https://example.com") # Verify user agent was applied correctly user_agent = await page.evaluate("() => navigator.userAgent") print(f"Task {index}: User agent: {user_agent}") # Get page title title = await page.title() print(f"Task {index}: Page title: {title}") # Simulate work await asyncio.sleep(1) finally: # Release the page print(f"Task {index}: Releasing page...") await manager.release_page(page, strategy, config) # Create tasks to use pre-warmed pages tasks = [] # Use run_config1 pages for i in range(2): tasks.append(asyncio.create_task(use_prewarm_page(i+1, run_config1))) # Use run_config2 pages for i in range(3): tasks.append(asyncio.create_task(use_prewarm_page(i+3, run_config2))) # Wait for all tasks to complete await asyncio.gather(*tasks) # Try to use more pages than we pre-warmed (should raise exception) print("\nTrying to use more pages than pre-warmed...") try: page, context, strategy = await manager.get_page(run_config1, config) try: print("Got extra page (unexpected)") await page.goto("https://example.com") finally: await manager.release_page(page, strategy, config) except Exception as e: print(f"Expected exception when requesting more pages: {str(e)}") finally: # Clean up print("\nClosing browser manager...") await manager.close() print("Browser manager closed") async def prewarm_on_demand_demo(): """Demonstrate pre-warming with on-demand browser creation.""" print("\n=== Pre-warming with On-Demand Browser Creation Demo ===") # Create logger logger = AsyncLogger(verbose=True) # Create browser configuration config = BrowserConfig( browser_type="chromium", headless=True, browser_mode="playwright" ) # Create crawler run configurations run_config = CrawlerRunConfig( user_agent="Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36" ) # Create page pre-warm configurations - just pre-warm 2 pages page_configs = [ (config, run_config, 2) ] # Create browser manager with ON_DEMAND behavior manager = BrowserManager( browser_config=config, logger=logger, unavailable_behavior=UnavailableBehavior.ON_DEMAND, max_browsers_per_config=5 # Allow up to 5 browsers ) try: # Initialize pool with pre-warmed pages print("Initializing browser pool with pre-warmed pages...") await manager.initialize_pool( browser_configs=[config], browsers_per_config=1, # Start with just 1 browser page_configs=page_configs ) # Display initial pool status status = await manager.get_pool_status() print(f"Initial pool status: {status}") # Simulate using more pages than pre-warmed - should create browsers on demand print("\nUsing more pages than pre-warmed (should create on demand)...") async def use_page(index: int): print(f"Task {index}: Requesting page...") page, context, strategy = await manager.get_page(run_config, config) try: print(f"Task {index}: Got page, navigating to example.com...") await page.goto("https://example.com") # Get page title title = await page.title() print(f"Task {index}: Page title: {title}") # Simulate work for a varying amount of time work_time = 1 + (index * 0.5) # Stagger completion times print(f"Task {index}: Working for {work_time} seconds...") await asyncio.sleep(work_time) print(f"Task {index}: Work completed") finally: # Release the page print(f"Task {index}: Releasing page...") await manager.release_page(page, strategy, config) # Create more tasks than pre-warmed pages tasks = [] for i in range(5): # Try to use 5 pages when only 2 are pre-warmed tasks.append(asyncio.create_task(use_page(i+1))) # Wait for all tasks to complete await asyncio.gather(*tasks) # Display final pool status - should show on-demand created browsers status = await manager.get_pool_status() print(f"\nFinal pool status: {status}") finally: # Clean up print("\nClosing browser manager...") await manager.close() print("Browser manager closed") async def high_volume_demo(): """Demonstrate high-volume access to pre-warmed pages.""" print("\n=== High Volume Pre-warmed Pages Demo ===") # Create logger logger = AsyncLogger(verbose=True) # Create browser configuration config = BrowserConfig( browser_type="chromium", headless=True, browser_mode="playwright" ) # Create crawler run configuration run_config = CrawlerRunConfig( user_agent="Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.124 Safari/537.36" ) # Set up dimensions browser_count = 10 pages_per_browser = 5 total_pages = browser_count * pages_per_browser # Create page pre-warm configuration page_configs = [ (config, run_config, total_pages) ] print(f"Preparing {browser_count} browsers with {pages_per_browser} pages each ({total_pages} total pages)") # Create browser manager with ON_DEMAND behavior as fallback # No need to specify max_browsers_per_config as it will be calculated automatically manager = BrowserManager( browser_config=config, logger=logger, unavailable_behavior=UnavailableBehavior.ON_DEMAND ) try: # Initialize pool with browsers and pre-warmed pages print(f"Pre-warming {total_pages} pages...") start_time = time.time() await manager.initialize_pool( browser_configs=[config], browsers_per_config=browser_count, page_configs=page_configs ) warmup_time = time.time() - start_time print(f"Pre-warming completed in {warmup_time:.2f} seconds") # Display pool status status = await manager.get_pool_status() print(f"Pool status after pre-warming: {status}") # Simulate using all pre-warmed pages simultaneously print(f"\nSending {total_pages} crawl requests simultaneously...") async def crawl_page(index: int): # url = f"https://example.com/page{index}" url = SAFE_URLS[index % len(SAFE_URLS)] print(f"Page {index}: Requesting page...") # Measure time to acquire page page_start = time.time() page, context, strategy = await manager.get_page(run_config, config) page_acquisition_time = time.time() - page_start try: # Navigate to the URL nav_start = time.time() await page.goto(url, timeout=5000) navigation_time = time.time() - nav_start # Get the page title title = await page.title() return { "index": index, "url": url, "title": title, "page_acquisition_time": page_acquisition_time, "navigation_time": navigation_time } except playwright._impl._errors.TimeoutError as e: # print(f"Page {index}: Navigation timed out - {e}") return { "index": index, "url": url, "title": "Navigation timed out", "page_acquisition_time": page_acquisition_time, "navigation_time": 0 } finally: # Release the page await manager.release_page(page, strategy, config) # Create and execute all tasks simultaneously start_time = time.time() # Non-parallel way # for i in range(total_pages): # await crawl_page(i+1) tasks = [crawl_page(i+1) for i in range(total_pages)] results = await asyncio.gather(*tasks) total_time = time.time() - start_time # # Print all titles # for result in results: # print(f"Page {result['index']} ({result['url']}): Title: {result['title']}") # print(f" Page acquisition time: {result['page_acquisition_time']:.4f}s") # print(f" Navigation time: {result['navigation_time']:.4f}s") # print(f" Total time: {result['page_acquisition_time'] + result['navigation_time']:.4f}s") # print("-" * 40) # Report results print(f"\nAll {total_pages} crawls completed in {total_time:.2f} seconds") # Calculate statistics acquisition_times = [r["page_acquisition_time"] for r in results] navigation_times = [r["navigation_time"] for r in results] avg_acquisition = sum(acquisition_times) / len(acquisition_times) max_acquisition = max(acquisition_times) min_acquisition = min(acquisition_times) avg_navigation = sum(navigation_times) / len(navigation_times) max_navigation = max(navigation_times) min_navigation = min(navigation_times) print("\nPage acquisition times:") print(f" Average: {avg_acquisition:.4f}s") print(f" Min: {min_acquisition:.4f}s") print(f" Max: {max_acquisition:.4f}s") print("\nPage navigation times:") print(f" Average: {avg_navigation:.4f}s") print(f" Min: {min_navigation:.4f}s") print(f" Max: {max_navigation:.4f}s") # Display final pool status status = await manager.get_pool_status() print(f"\nFinal pool status: {status}") finally: # Clean up print("\nClosing browser manager...") await manager.close() print("Browser manager closed") async def main(): """Run all demos.""" # await basic_pooling_demo() # await prewarm_pages_demo() # await prewarm_on_demand_demo() await high_volume_demo() # Additional demo functions can be added here if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/mcp/test_mcp_socket.py
tests/mcp/test_mcp_socket.py
# pip install "mcp-sdk[ws]" anyio import anyio, json from mcp.client.websocket import websocket_client from mcp.client.session import ClientSession async def test_list(): async with websocket_client("ws://localhost:8020/mcp/ws") as (r, w): async with ClientSession(r, w) as s: await s.initialize() print("tools :", [t.name for t in (await s.list_tools()).tools]) print("resources :", [r.name for r in (await s.list_resources()).resources]) print("templates :", [t.name for t in (await s.list_resource_templates()).resource_templates]) async def test_crawl(s: ClientSession) -> None: """Hit the @mcp_tool('crawl') endpoint.""" res = await s.call_tool( "crawl", { "urls": ["https://example.com"], "browser_config": {}, "crawler_config": {}, }, ) print("crawl →", json.loads(res.content[0].text)) async def test_md(s: ClientSession) -> None: """Hit the @mcp_tool('md') endpoint.""" res = await s.call_tool( "md", { "url": "https://example.com", "f": "fit", # or RAW, BM25, LLM "q": None, "c": "0", }, ) result = json.loads(res.content[0].text) print("md →", result['markdown'][:100], "...") async def test_screenshot(s: ClientSession): res = await s.call_tool( "screenshot", { "url": "https://example.com", "screenshot_wait_for": 1.0, }, ) png_b64 = json.loads(res.content[0].text)["screenshot"] print("screenshot →", png_b64[:60], "… (base64)") async def test_pdf(s: ClientSession): res = await s.call_tool( "pdf", { "url": "https://example.com", }, ) pdf_b64 = json.loads(res.content[0].text)["pdf"] print("pdf →", pdf_b64[:60], "… (base64)") async def test_execute_js(s: ClientSession): # click the “More” link on Hacker News front page and wait 1 s res = await s.call_tool( "execute_js", { "url": "https://news.ycombinator.com/news", "js_code": [ "await page.click('a.morelink')", "await page.waitForTimeout(1000)", ], }, ) crawl_result = json.loads(res.content[0].text) print("execute_js → status", crawl_result["success"], "| html len:", len(crawl_result["html"])) async def test_html(s: ClientSession): # click the “More” link on Hacker News front page and wait 1 s res = await s.call_tool( "html", { "url": "https://news.ycombinator.com/news", }, ) crawl_result = json.loads(res.content[0].text) print("execute_js → status", crawl_result["success"], "| html len:", len(crawl_result["html"])) async def test_context(s: ClientSession): # click the “More” link on Hacker News front page and wait 1 s res = await s.call_tool( "ask", { "query": "I hv a question about Crawl4ai library, how to extract internal links when crawling a page?" }, ) crawl_result = json.loads(res.content[0].text) print("execute_js → status", crawl_result["success"], "| html len:", len(crawl_result["html"])) async def main() -> None: async with websocket_client("ws://localhost:11235/mcp/ws") as (r, w): async with ClientSession(r, w) as s: await s.initialize() # handshake tools = (await s.list_tools()).tools print("tools:", [t.name for t in tools]) # await test_list() await test_crawl(s) await test_md(s) await test_screenshot(s) await test_pdf(s) await test_execute_js(s) await test_html(s) await test_context(s) anyio.run(main)
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/mcp/test_mcp_sse.py
tests/mcp/test_mcp_sse.py
from mcp.client.sse import sse_client from mcp.client.session import ClientSession async def main(): async with sse_client("http://127.0.0.1:8020/mcp") as (r, w): async with ClientSession(r, w) as sess: print(await sess.list_tools()) # now works if __name__ == "__main__": import asyncio asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/tests/loggers/test_logger.py
tests/loggers/test_logger.py
import asyncio from crawl4ai import AsyncWebCrawler, BrowserConfig, CrawlerRunConfig, CacheMode, AsyncLoggerBase import os from datetime import datetime class AsyncFileLogger(AsyncLoggerBase): """ File-only asynchronous logger that writes logs to a specified file. """ def __init__(self, log_file: str): """ Initialize the file logger. Args: log_file: File path for logging """ self.log_file = log_file os.makedirs(os.path.dirname(os.path.abspath(log_file)), exist_ok=True) def _write_to_file(self, level: str, message: str, tag: str): """Write a message to the log file.""" timestamp = datetime.now().strftime("%Y-%m-%d %H:%M:%S.%f")[:-3] with open(self.log_file, "a", encoding="utf-8") as f: f.write(f"[{timestamp}] [{level}] [{tag}] {message}\n") def debug(self, message: str, tag: str = "DEBUG", **kwargs): """Log a debug message to file.""" self._write_to_file("DEBUG", message, tag) def info(self, message: str, tag: str = "INFO", **kwargs): """Log an info message to file.""" self._write_to_file("INFO", message, tag) def success(self, message: str, tag: str = "SUCCESS", **kwargs): """Log a success message to file.""" self._write_to_file("SUCCESS", message, tag) def warning(self, message: str, tag: str = "WARNING", **kwargs): """Log a warning message to file.""" self._write_to_file("WARNING", message, tag) def error(self, message: str, tag: str = "ERROR", **kwargs): """Log an error message to file.""" self._write_to_file("ERROR", message, tag) def url_status(self, url: str, success: bool, timing: float, tag: str = "FETCH", url_length: int = 50): """Log URL fetch status to file.""" status = "SUCCESS" if success else "FAILED" message = f"{url[:url_length]}... | Status: {status} | Time: {timing:.2f}s" self._write_to_file("URL_STATUS", message, tag) def error_status(self, url: str, error: str, tag: str = "ERROR", url_length: int = 50): """Log error status to file.""" message = f"{url[:url_length]}... | Error: {error}" self._write_to_file("ERROR", message, tag) async def main(): browser_config = BrowserConfig(headless=True, verbose=True) crawler = AsyncWebCrawler(config=browser_config, logger=AsyncFileLogger("/Users/unclecode/devs/crawl4ai/.private/tmp/crawl.log")) await crawler.start() try: crawl_config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS, ) # Use the crawler multiple times result = await crawler.arun( url='https://kidocode.com/', config=crawl_config ) if result.success: print("First crawl - Raw Markdown Length:", len(result.markdown.raw_markdown)) finally: # Always ensure we close the crawler await crawler.close() if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/docker_client.py
crawl4ai/docker_client.py
from typing import List, Optional, Union, AsyncGenerator, Dict, Any, Callable import httpx import json from urllib.parse import urljoin import asyncio from .async_configs import BrowserConfig, CrawlerRunConfig from .models import CrawlResult from .async_logger import AsyncLogger, LogLevel from .utils import hooks_to_string class Crawl4aiClientError(Exception): """Base exception for Crawl4ai Docker client errors.""" pass class ConnectionError(Crawl4aiClientError): """Raised when connection to the Docker server fails.""" pass class RequestError(Crawl4aiClientError): """Raised when the server returns an error response.""" pass class Crawl4aiDockerClient: """Client for interacting with Crawl4AI Docker server with token authentication.""" def __init__( self, base_url: str = "http://localhost:8000", timeout: float = 30.0, verify_ssl: bool = True, verbose: bool = True, log_file: Optional[str] = None ): self.base_url = base_url.rstrip('/') self.timeout = timeout self.logger = AsyncLogger(log_file=log_file, log_level=LogLevel.DEBUG, verbose=verbose) self._http_client = httpx.AsyncClient( timeout=timeout, verify=verify_ssl, headers={"Content-Type": "application/json"} ) self._token: Optional[str] = None async def authenticate(self, email: str) -> None: """Authenticate with the server and store the token.""" url = urljoin(self.base_url, "/token") try: self.logger.info(f"Authenticating with email: {email}", tag="AUTH") response = await self._http_client.post(url, json={"email": email}) response.raise_for_status() data = response.json() self._token = data["access_token"] self._http_client.headers["Authorization"] = f"Bearer {self._token}" self.logger.success("Authentication successful", tag="AUTH") except (httpx.RequestError, httpx.HTTPStatusError) as e: error_msg = f"Authentication failed: {str(e)}" self.logger.error(error_msg, tag="ERROR") raise ConnectionError(error_msg) async def _check_server(self) -> None: """Check if server is reachable, raising an error if not.""" try: await self._http_client.get(urljoin(self.base_url, "/health")) self.logger.success(f"Connected to {self.base_url}", tag="READY") except httpx.RequestError as e: self.logger.error(f"Server unreachable: {str(e)}", tag="ERROR") raise ConnectionError(f"Cannot connect to server: {str(e)}") def _prepare_request( self, urls: List[str], browser_config: Optional[BrowserConfig] = None, crawler_config: Optional[CrawlerRunConfig] = None, hooks: Optional[Union[Dict[str, Callable], Dict[str, str]]] = None, hooks_timeout: int = 30 ) -> Dict[str, Any]: """Prepare request data from configs.""" if self._token: self._http_client.headers["Authorization"] = f"Bearer {self._token}" request_data = { "urls": urls, "browser_config": browser_config.dump() if browser_config else {}, "crawler_config": crawler_config.dump() if crawler_config else {} } # Handle hooks if provided if hooks: # Check if hooks are already strings or need conversion if any(callable(v) for v in hooks.values()): # Convert function objects to strings hooks_code = hooks_to_string(hooks) else: # Already in string format hooks_code = hooks request_data["hooks"] = { "code": hooks_code, "timeout": hooks_timeout } return request_data async def _request(self, method: str, endpoint: str, **kwargs) -> httpx.Response: """Make an HTTP request with error handling.""" url = urljoin(self.base_url, endpoint) try: response = await self._http_client.request(method, url, **kwargs) response.raise_for_status() return response except httpx.TimeoutException as e: raise ConnectionError(f"Request timed out: {str(e)}") except httpx.RequestError as e: raise ConnectionError(f"Failed to connect: {str(e)}") except httpx.HTTPStatusError as e: error_msg = (e.response.json().get("detail", str(e)) if "application/json" in e.response.headers.get("content-type", "") else str(e)) raise RequestError(f"Server error {e.response.status_code}: {error_msg}") async def crawl( self, urls: List[str], browser_config: Optional[BrowserConfig] = None, crawler_config: Optional[CrawlerRunConfig] = None, hooks: Optional[Union[Dict[str, Callable], Dict[str, str]]] = None, hooks_timeout: int = 30 ) -> Union[CrawlResult, List[CrawlResult], AsyncGenerator[CrawlResult, None]]: """ Execute a crawl operation. Args: urls: List of URLs to crawl browser_config: Browser configuration crawler_config: Crawler configuration hooks: Optional hooks - can be either: - Dict[str, Callable]: Function objects that will be converted to strings - Dict[str, str]: Already stringified hook code hooks_timeout: Timeout in seconds for each hook execution (1-120) Returns: Single CrawlResult, list of results, or async generator for streaming Example with function hooks: >>> async def my_hook(page, context, **kwargs): ... await page.set_viewport_size({"width": 1920, "height": 1080}) ... return page >>> >>> result = await client.crawl( ... ["https://example.com"], ... hooks={"on_page_context_created": my_hook} ... ) """ await self._check_server() data = self._prepare_request(urls, browser_config, crawler_config, hooks, hooks_timeout) is_streaming = crawler_config and crawler_config.stream self.logger.info(f"Crawling {len(urls)} URLs {'(streaming)' if is_streaming else ''}", tag="CRAWL") if is_streaming: async def stream_results() -> AsyncGenerator[CrawlResult, None]: async with self._http_client.stream("POST", f"{self.base_url}/crawl/stream", json=data) as response: response.raise_for_status() async for line in response.aiter_lines(): if line.strip(): result = json.loads(line) if "error" in result: self.logger.error_status(url=result.get("url", "unknown"), error=result["error"]) continue self.logger.url_status(url=result.get("url", "unknown"), success=True, timing=result.get("timing", 0.0)) if result.get("status") == "completed": continue else: yield CrawlResult(**result) return stream_results() response = await self._request("POST", "/crawl", json=data, timeout=hooks_timeout) result_data = response.json() if not result_data.get("success", False): raise RequestError(f"Crawl failed: {result_data.get('msg', 'Unknown error')}") results = [CrawlResult(**r) for r in result_data.get("results", [])] self.logger.success(f"Crawl completed with {len(results)} results", tag="CRAWL") return results[0] if len(results) == 1 else results async def get_schema(self) -> Dict[str, Any]: """Retrieve configuration schemas.""" response = await self._request("GET", "/schema") return response.json() async def close(self) -> None: """Close the HTTP client session.""" self.logger.info("Closing client", tag="CLOSE") await self._http_client.aclose() async def __aenter__(self) -> "Crawl4aiDockerClient": return self async def __aexit__(self, exc_type: Optional[type], exc_val: Optional[Exception], exc_tb: Optional[Any]) -> None: await self.close() # Example usage async def main(): async with Crawl4aiDockerClient(verbose=True) as client: await client.authenticate("user@example.com") result = await client.crawl(["https://example.com"]) print(result) schema = await client.get_schema() print(schema) if __name__ == "__main__": asyncio.run(main())
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/prompts.py
crawl4ai/prompts.py
PROMPT_EXTRACT_BLOCKS = """Here is the URL of the webpage: <url>{URL}</url> And here is the cleaned HTML content of that webpage: <html> {HTML} </html> Your task is to break down this HTML content into semantically relevant blocks, and for each block, generate a JSON object with the following keys: - index: an integer representing the index of the block in the content - tags: a list of semantic tags that are relevant to the content of the block - content: a list of strings containing the text content of the block - questions: a list of 3 questions that a user may ask about the content in this block To generate the JSON objects: 1. Carefully read through the HTML content and identify logical breaks or shifts in the content that would warrant splitting it into separate blocks. 2. For each block: a. Assign it an index based on its order in the content. b. Analyze the content and generate a list of relevant semantic tags that describe what the block is about. c. Extract the text content, clean it up if needed, and store it as a list of strings in the "content" field. d. Come up with 3 questions that a user might ask about this specific block of content, based on the tags and content. The questions should be relevant and answerable by the content in the block. 3. Ensure that the order of the JSON objects matches the order of the blocks as they appear in the original HTML content. 4. Double-check that each JSON object includes all required keys (index, tags, content, questions) and that the values are in the expected format (integer, list of strings, etc.). 5. Make sure the generated JSON is complete and parsable, with no errors or omissions. 6. Make sure to escape any special characters in the HTML content, and also single or double quote to avoid JSON parsing issues. Please provide your output within <blocks> tags, like this: <blocks> [{ "index": 0, "tags": ["introduction", "overview"], "content": ["This is the first paragraph of the article, which provides an introduction and overview of the main topic."], "questions": [ "What is the main topic of this article?", "What can I expect to learn from reading this article?", "Is this article suitable for beginners or experts in the field?" ] }, { "index": 1, "tags": ["history", "background"], "content": ["This is the second paragraph, which delves into the history and background of the topic.", "It provides context and sets the stage for the rest of the article."], "questions": [ "What historical events led to the development of this topic?", "How has the understanding of this topic evolved over time?", "What are some key milestones in the history of this topic?" ] }] </blocks> Remember, the output should be a complete, parsable JSON wrapped in <blocks> tags, with no omissions or errors. The JSON objects should semantically break down the content into relevant blocks, maintaining the original order.""" PROMPT_EXTRACT_BLOCKS = """Here is the URL of the webpage: <url>{URL}</url> And here is the cleaned HTML content of that webpage: <html> {HTML} </html> Your task is to break down this HTML content into semantically relevant blocks, and for each block, generate a JSON object with the following keys: - index: an integer representing the index of the block in the content - content: a list of strings containing the text content of the block To generate the JSON objects: 1. Carefully read through the HTML content and identify logical breaks or shifts in the content that would warrant splitting it into separate blocks. 2. For each block: a. Assign it an index based on its order in the content. b. Analyze the content and generate ONE semantic tag that describe what the block is about. c. Extract the text content, EXACTLY SAME AS THE GIVE DATA, clean it up if needed, and store it as a list of strings in the "content" field. 3. Ensure that the order of the JSON objects matches the order of the blocks as they appear in the original HTML content. 4. Double-check that each JSON object includes all required keys (index, tag, content) and that the values are in the expected format (integer, list of strings, etc.). 5. Make sure the generated JSON is complete and parsable, with no errors or omissions. 6. Make sure to escape any special characters in the HTML content, and also single or double quote to avoid JSON parsing issues. 7. Never alter the extracted content, just copy and paste it as it is. Please provide your output within <blocks> tags, like this: <blocks> [{ "index": 0, "tags": ["introduction"], "content": ["This is the first paragraph of the article, which provides an introduction and overview of the main topic."] }, { "index": 1, "tags": ["background"], "content": ["This is the second paragraph, which delves into the history and background of the topic.", "It provides context and sets the stage for the rest of the article."] }] </blocks> Remember, the output should be a complete, parsable JSON wrapped in <blocks> tags, with no omissions or errors. The JSON objects should semantically break down the content into relevant blocks, maintaining the original order.""" PROMPT_EXTRACT_BLOCKS_WITH_INSTRUCTION = """Here is the URL of the webpage: <url>{URL}</url> And here is the cleaned HTML content of that webpage: <html> {HTML} </html> Your task is to break down this HTML content into semantically relevant blocks, following the provided user's REQUEST, and for each block, generate a JSON object with the following keys: - index: an integer representing the index of the block in the content - content: a list of strings containing the text content of the block This is the user's REQUEST, pay attention to it: <request> {REQUEST} </request> To generate the JSON objects: 1. Carefully read through the HTML content and identify logical breaks or shifts in the content that would warrant splitting it into separate blocks. 2. For each block: a. Assign it an index based on its order in the content. b. Analyze the content and generate ONE semantic tag that describe what the block is about. c. Extract the text content, EXACTLY SAME AS GIVE DATA, clean it up if needed, and store it as a list of strings in the "content" field. 3. Ensure that the order of the JSON objects matches the order of the blocks as they appear in the original HTML content. 4. Double-check that each JSON object includes all required keys (index, tag, content) and that the values are in the expected format (integer, list of strings, etc.). 5. Make sure the generated JSON is complete and parsable, with no errors or omissions. 6. Make sure to escape any special characters in the HTML content, and also single or double quote to avoid JSON parsing issues. 7. Never alter the extracted content, just copy and paste it as it is. Please provide your output within <blocks> tags, like this: <blocks> [{ "index": 0, "tags": ["introduction"], "content": ["This is the first paragraph of the article, which provides an introduction and overview of the main topic."] }, { "index": 1, "tags": ["background"], "content": ["This is the second paragraph, which delves into the history and background of the topic.", "It provides context and sets the stage for the rest of the article."] }] </blocks> **Make sure to follow the user instruction to extract blocks aligin with the instruction.** Remember, the output should be a complete, parsable JSON wrapped in <blocks> tags, with no omissions or errors. The JSON objects should semantically break down the content into relevant blocks, maintaining the original order.""" PROMPT_EXTRACT_SCHEMA_WITH_INSTRUCTION = """Here is the content from the URL: <url>{URL}</url> <url_content> {HTML} </url_content> The user has made the following request for what information to extract from the above content: <user_request> {REQUEST} </user_request> <schema_block> {SCHEMA} </schema_block> Please carefully read the URL content and the user's request. If the user provided a desired JSON schema in the <schema_block> above, extract the requested information from the URL content according to that schema. If no schema was provided, infer an appropriate JSON schema based on the user's request that will best capture the key information they are looking for. Extraction instructions: Return the extracted information as a list of JSON objects, with each object in the list corresponding to a block of content from the URL, in the same order as it appears on the page. Wrap the entire JSON list in <blocks>...</blocks> XML tags. Quality Reflection: Before outputting your final answer, double check that the JSON you are returning is complete, containing all the information requested by the user, and is valid JSON that could be parsed by json.loads() with no errors or omissions. The outputted JSON objects should fully match the schema, either provided or inferred. Quality Score: After reflecting, score the quality and completeness of the JSON data you are about to return on a scale of 1 to 5. Write the score inside <score> tags. Avoid Common Mistakes: - Do NOT add any comments using "//" or "#" in the JSON output. It causes parsing errors. - Make sure the JSON is properly formatted with curly braces, square brackets, and commas in the right places. - Do not miss closing </blocks> tag at the end of the JSON output. - Do not generate the Python code show me how to do the task, this is your task to extract the information and return it in JSON format. Result Output the final list of JSON objects, wrapped in <blocks>...</blocks> XML tags. Make sure to close the tag properly.""" PROMPT_EXTRACT_INFERRED_SCHEMA = """Here is the content from the URL: <url>{URL}</url> <url_content> {HTML} </url_content> Please carefully read the URL content and the user's request. Analyze the page structure and infer the most appropriate JSON schema based on the content and request. Extraction Strategy: 1. First, determine if the page contains repetitive items (like multiple products, articles, etc.) or a single content item (like a single article or page). 2. For repetitive items: Identify the common pattern and extract each instance as a separate JSON object in an array. 3. For single content: Extract the key information into a comprehensive JSON object that captures the essential details. Extraction instructions: Return the extracted information as a list of JSON objects. For repetitive content, each object in the list should correspond to a distinct item. For single content, you may return just one detailed JSON object. Wrap the entire JSON list in <blocks>...</blocks> XML tags. Schema Design Guidelines: - Create meaningful property names that clearly describe the data they contain - Use nested objects for hierarchical information - Use arrays for lists of related items - Include all information requested by the user - Maintain consistency in property names and data structures - Only include properties that are actually present in the content - For dates, prefer ISO format (YYYY-MM-DD) - For prices or numeric values, extract them without currency symbols when possible Quality Reflection: Before outputting your final answer, double check that: 1. The inferred schema makes logical sense for the type of content 2. All requested information is included 3. The JSON is valid and could be parsed without errors 4. Property names are consistent and descriptive 5. The structure is optimal for the type of data being represented Avoid Common Mistakes: - Do NOT add any comments using "//" or "#" in the JSON output. It causes parsing errors. - Make sure the JSON is properly formatted with curly braces, square brackets, and commas in the right places. - Do not miss closing </blocks> tag at the end of the JSON output. - Do not generate Python code showing how to do the task; this is your task to extract the information and return it in JSON format. - Ensure consistency in property names across all objects - Don't include empty properties or null values unless they're meaningful - For repetitive content, ensure all objects follow the same schema Important: If user specific instruction is provided, then stress significantly on what user is requesting and describing about the schema of end result (if any). If user is requesting to extract specific information, then focus on that and ignore the rest of the content. <user_request> {REQUEST} </user_request> Result: Output the final list of JSON objects, wrapped in <blocks>...</blocks> XML tags. Make sure to close the tag properly. DO NOT ADD ANY PRE OR POST COMMENTS. JUST RETURN THE JSON OBJECTS INSIDE <blocks>...</blocks> TAGS. CRITICAL: The content inside the <blocks> tags MUST be a direct array of JSON objects (starting with '[' and ending with ']'), not a dictionary/object containing an array. For example, use <blocks>[{...}, {...}]</blocks> instead of <blocks>{"items": [{...}, {...}]}</blocks>. This is essential for proper parsing. """ PROMPT_FILTER_CONTENT = """Your task is to filter and convert HTML content into clean, focused markdown that's optimized for use with LLMs and information retrieval systems. TASK DETAILS: 1. Content Selection - DO: Keep essential information, main content, key details - DO: Preserve hierarchical structure using markdown headers - DO: Keep code blocks, tables, key lists - DON'T: Include navigation menus, ads, footers, cookie notices - DON'T: Keep social media widgets, sidebars, related content 2. Content Transformation - DO: Use proper markdown syntax (#, ##, **, `, etc) - DO: Convert tables to markdown tables - DO: Preserve code formatting with ```language blocks - DO: Maintain link texts but remove tracking parameters - DON'T: Include HTML tags in output - DON'T: Keep class names, ids, or other HTML attributes 3. Content Organization - DO: Maintain logical flow of information - DO: Group related content under appropriate headers - DO: Use consistent header levels - DON'T: Fragment related content - DON'T: Duplicate information IMPORTANT: If user specific instruction is provided, ignore above guideline and prioritize those requirements over these general guidelines. OUTPUT FORMAT: Wrap your response in <content> tags. Use proper markdown throughout. <content> [Your markdown content here] </content> Begin filtering now. -------------------------------------------- <|HTML_CONTENT_START|> {HTML} <|HTML_CONTENT_END|> <|USER_INSTRUCTION_START|> {REQUEST} <|USER_INSTRUCTION_END|> """ JSON_SCHEMA_BUILDER= """ # HTML Schema Generation Instructions You are a specialized model designed to analyze HTML patterns and generate extraction schemas. Your primary job is to create structured JSON schemas that can be used to extract data from HTML in a consistent and reliable way. When presented with HTML content, you must analyze its structure and generate a schema that captures all relevant data points. ## Your Core Responsibilities: 1. Analyze HTML structure to identify repeating patterns and important data points 2. Generate valid JSON schemas following the specified format 3. Create appropriate selectors that will work reliably for data extraction 4. Name fields meaningfully based on their content and purpose 5. Handle both specific user requests and autonomous pattern detection ## Available Schema Types You Can Generate: <schema_types> 1. Basic Single-Level Schema - Use for simple, flat data structures - Example: Product cards, user profiles - Direct field extractions 2. Nested Object Schema - Use for hierarchical data - Example: Articles with author details - Contains objects within objects 3. List Schema - Use for repeating elements - Example: Comment sections, product lists - Handles arrays of similar items 4. Complex Nested Lists - Use for multi-level data - Example: Categories with subcategories - Multiple levels of nesting 5. Transformation Schema - Use for data requiring processing - Supports regex and text transformations - Special attribute handling </schema_types> <schema_structure> Your output must always be a JSON object with this structure: { "name": "Descriptive name of the pattern", "baseSelector": "CSS selector for the repeating element", "fields": [ { "name": "field_name", "selector": "CSS selector", "type": "text|attribute|nested|list|regex", "attribute": "attribute_name", // Optional "transform": "transformation_type", // Optional "pattern": "regex_pattern", // Optional "fields": [] // For nested/list types } ] } </schema_structure> <type_definitions> Available field types: - text: Direct text extraction - attribute: HTML attribute extraction - nested: Object containing other fields - list: Array of similar items - regex: Pattern-based extraction </type_definitions> <behavior_rules> 1. When given a specific query: - Focus on extracting requested data points - Use most specific selectors possible - Include all fields mentioned in the query 2. When no query is provided: - Identify main content areas - Extract all meaningful data points - Use semantic structure to determine importance - Include prices, dates, titles, and other common data types 3. Always: - Use reliable CSS selectors - Handle dynamic class names appropriately - Create descriptive field names - Follow consistent naming conventions </behavior_rules> <examples> 1. Basic Product Card Example: <html> <div class="product-card" data-cat-id="electronics" data-subcat-id="laptops"> <h2 class="product-title">Gaming Laptop</h2> <span class="price">$999.99</span> <img src="laptop.jpg" alt="Gaming Laptop"> </div> </html> Generated Schema: { "name": "Product Cards", "baseSelector": ".product-card", "baseFields": [ {"name": "data_cat_id", "type": "attribute", "attribute": "data-cat-id"}, {"name": "data_subcat_id", "type": "attribute", "attribute": "data-subcat-id"} ], "fields": [ { "name": "title", "selector": ".product-title", "type": "text" }, { "name": "price", "selector": ".price", "type": "text" }, { "name": "image_url", "selector": "img", "type": "attribute", "attribute": "src" } ] } 2. Article with Author Details Example: <html> <article> <h1>The Future of AI</h1> <div class="author-info"> <span class="author-name">Dr. Smith</span> <img src="author.jpg" alt="Dr. Smith"> </div> </article> </html> Generated Schema: { "name": "Article Details", "baseSelector": "article", "fields": [ { "name": "title", "selector": "h1", "type": "text" }, { "name": "author", "type": "nested", "selector": ".author-info", "fields": [ { "name": "name", "selector": ".author-name", "type": "text" }, { "name": "avatar", "selector": "img", "type": "attribute", "attribute": "src" } ] } ] } 3. Comments Section Example: <html> <div class="comments-container"> <div class="comment" data-user-id="123"> <div class="user-name">John123</div> <p class="comment-text">Great article!</p> </div> <div class="comment" data-user-id="456"> <div class="user-name">Alice456</div> <p class="comment-text">Thanks for sharing.</p> </div> </div> </html> Generated Schema: { "name": "Comment Section", "baseSelector": ".comments-container", "baseFields": [ {"name": "data_user_id", "type": "attribute", "attribute": "data-user-id"} ], "fields": [ { "name": "comments", "type": "list", "selector": ".comment", "fields": [ { "name": "user", "selector": ".user-name", "type": "text" }, { "name": "content", "selector": ".comment-text", "type": "text" } ] } ] } 4. E-commerce Categories Example: <html> <div class="category-section" data-category="electronics"> <h2>Electronics</h2> <div class="subcategory"> <h3>Laptops</h3> <div class="product"> <span class="product-name">MacBook Pro</span> <span class="price">$1299</span> </div> <div class="product"> <span class="product-name">Dell XPS</span> <span class="price">$999</span> </div> </div> </div> </html> Generated Schema: { "name": "E-commerce Categories", "baseSelector": ".category-section", "baseFields": [ {"name": "data_category", "type": "attribute", "attribute": "data-category"} ], "fields": [ { "name": "category_name", "selector": "h2", "type": "text" }, { "name": "subcategories", "type": "nested_list", "selector": ".subcategory", "fields": [ { "name": "name", "selector": "h3", "type": "text" }, { "name": "products", "type": "list", "selector": ".product", "fields": [ { "name": "name", "selector": ".product-name", "type": "text" }, { "name": "price", "selector": ".price", "type": "text" } ] } ] } ] } 5. Job Listings with Transformations Example: <html> <div class="job-post"> <h3 class="job-title">Senior Developer</h3> <span class="salary-text">Salary: $120,000/year</span> <span class="location"> New York, NY </span> </div> </html> Generated Schema: { "name": "Job Listings", "baseSelector": ".job-post", "fields": [ { "name": "title", "selector": ".job-title", "type": "text", "transform": "uppercase" }, { "name": "salary", "selector": ".salary-text", "type": "regex", "pattern": "\\$([\\d,]+)" }, { "name": "location", "selector": ".location", "type": "text", "transform": "strip" } ] } 6. Skyscanner Place Card Example: <html> <div class="PlaceCard_descriptionContainer__M2NjN" data-testid="description-container"> <div class="PlaceCard_nameContainer__ZjZmY" tabindex="0" role="link"> <div class="PlaceCard_nameContent__ODUwZ"> <span class="BpkText_bpk-text__MjhhY BpkText_bpk-text--heading-4__Y2FlY">Doha</span> </div> <span class="BpkText_bpk-text__MjhhY BpkText_bpk-text--heading-4__Y2FlY PlaceCard_subName__NTVkY">Qatar</span> </div> <span class="PlaceCard_advertLabel__YTM0N">Sunny days and the warmest welcome awaits</span> <a class="BpkLink_bpk-link__MmQwY PlaceCard_descriptionLink__NzYwN" href="/flights/del/doha/" data-testid="flights-link"> <div class="PriceDescription_container__NjEzM"> <span class="BpkText_bpk-text--heading-5__MTRjZ">₹17,559</span> </div> </a> </div> </html> Generated Schema: { "name": "Skyscanner Place Cards", "baseSelector": "div[class^='PlaceCard_descriptionContainer__']", "baseFields": [ {"name": "data_testid", "type": "attribute", "attribute": "data-testid"} ], "fields": [ { "name": "city_name", "selector": "div[class^='PlaceCard_nameContent__'] .BpkText_bpk-text--heading-4__", "type": "text" }, { "name": "country_name", "selector": "span[class*='PlaceCard_subName__']", "type": "text" }, { "name": "description", "selector": "span[class*='PlaceCard_advertLabel__']", "type": "text" }, { "name": "flight_price", "selector": "a[data-testid='flights-link'] .BpkText_bpk-text--heading-5__", "type": "text" }, { "name": "flight_url", "selector": "a[data-testid='flights-link']", "type": "attribute", "attribute": "href" } ] } </examples> <output_requirements> Your output must: 1. Be valid JSON only 2. Include no explanatory text 3. Follow the exact schema structure provided 4. Use appropriate field types 5. Include all required fields 6. Use valid CSS selectors </output_requirements> """ JSON_SCHEMA_BUILDER_XPATH = """ # HTML Schema Generation Instructions You are a specialized model designed to analyze HTML patterns and generate extraction schemas. Your primary job is to create structured JSON schemas that can be used to extract data from HTML in a consistent and reliable way. When presented with HTML content, you must analyze its structure and generate a schema that captures all relevant data points. ## Your Core Responsibilities: 1. Analyze HTML structure to identify repeating patterns and important data points 2. Generate valid JSON schemas following the specified format 3. Create appropriate XPath selectors that will work reliably for data extraction 4. Name fields meaningfully based on their content and purpose 5. Handle both specific user requests and autonomous pattern detection ## Available Schema Types You Can Generate: <schema_types> 1. Basic Single-Level Schema - Use for simple, flat data structures - Example: Product cards, user profiles - Direct field extractions 2. Nested Object Schema - Use for hierarchical data - Example: Articles with author details - Contains objects within objects 3. List Schema - Use for repeating elements - Example: Comment sections, product lists - Handles arrays of similar items 4. Complex Nested Lists - Use for multi-level data - Example: Categories with subcategories - Multiple levels of nesting 5. Transformation Schema - Use for data requiring processing - Supports regex and text transformations - Special attribute handling </schema_types> <schema_structure> Your output must always be a JSON object with this structure: { "name": "Descriptive name of the pattern", "baseSelector": "XPath selector for the repeating element", "fields": [ { "name": "field_name", "selector": "XPath selector", "type": "text|attribute|nested|list|regex", "attribute": "attribute_name", // Optional "transform": "transformation_type", // Optional "pattern": "regex_pattern", // Optional "fields": [] // For nested/list types } ] } </schema_structure> <type_definitions> Available field types: - text: Direct text extraction - attribute: HTML attribute extraction - nested: Object containing other fields - list: Array of similar items - regex: Pattern-based extraction </type_definitions> <behavior_rules> 1. When given a specific query: - Focus on extracting requested data points - Use most specific selectors possible - Include all fields mentioned in the query 2. When no query is provided: - Identify main content areas - Extract all meaningful data points - Use semantic structure to determine importance - Include prices, dates, titles, and other common data types 3. Always: - Use reliable XPath selectors - Handle dynamic element IDs appropriately - Create descriptive field names - Follow consistent naming conventions </behavior_rules> <examples> 1. Basic Product Card Example: <html> <div class="product-card" data-cat-id="electronics" data-subcat-id="laptops"> <h2 class="product-title">Gaming Laptop</h2> <span class="price">$999.99</span> <img src="laptop.jpg" alt="Gaming Laptop"> </div> </html> Generated Schema: { "name": "Product Cards", "baseSelector": "//div[@class='product-card']", "baseFields": [ {"name": "data_cat_id", "type": "attribute", "attribute": "data-cat-id"}, {"name": "data_subcat_id", "type": "attribute", "attribute": "data-subcat-id"} ], "fields": [ { "name": "title", "selector": ".//h2[@class='product-title']", "type": "text" }, { "name": "price", "selector": ".//span[@class='price']", "type": "text" }, { "name": "image_url", "selector": ".//img", "type": "attribute", "attribute": "src" } ] } 2. Article with Author Details Example: <html> <article> <h1>The Future of AI</h1> <div class="author-info"> <span class="author-name">Dr. Smith</span> <img src="author.jpg" alt="Dr. Smith"> </div> </article> </html> Generated Schema: { "name": "Article Details", "baseSelector": "//article", "fields": [ { "name": "title", "selector": ".//h1", "type": "text" }, { "name": "author", "type": "nested", "selector": ".//div[@class='author-info']", "fields": [ { "name": "name", "selector": ".//span[@class='author-name']", "type": "text" }, { "name": "avatar", "selector": ".//img", "type": "attribute", "attribute": "src" } ] } ] } 3. Comments Section Example: <html> <div class="comments-container"> <div class="comment" data-user-id="123"> <div class="user-name">John123</div> <p class="comment-text">Great article!</p> </div> <div class="comment" data-user-id="456"> <div class="user-name">Alice456</div> <p class="comment-text">Thanks for sharing.</p> </div> </div> </html> Generated Schema: { "name": "Comment Section", "baseSelector": "//div[@class='comments-container']", "fields": [ { "name": "comments", "type": "list", "selector": ".//div[@class='comment']", "baseFields": [ {"name": "data_user_id", "type": "attribute", "attribute": "data-user-id"} ], "fields": [ { "name": "user", "selector": ".//div[@class='user-name']", "type": "text" }, { "name": "content", "selector": ".//p[@class='comment-text']", "type": "text" } ] } ] } 4. E-commerce Categories Example: <html> <div class="category-section" data-category="electronics"> <h2>Electronics</h2> <div class="subcategory"> <h3>Laptops</h3> <div class="product"> <span class="product-name">MacBook Pro</span> <span class="price">$1299</span> </div> <div class="product"> <span class="product-name">Dell XPS</span> <span class="price">$999</span> </div> </div> </div> </html> Generated Schema: { "name": "E-commerce Categories", "baseSelector": "//div[@class='category-section']", "baseFields": [ {"name": "data_category", "type": "attribute", "attribute": "data-category"} ], "fields": [ { "name": "category_name", "selector": ".//h2", "type": "text" }, { "name": "subcategories", "type": "nested_list", "selector": ".//div[@class='subcategory']", "fields": [ { "name": "name", "selector": ".//h3", "type": "text" }, { "name": "products", "type": "list", "selector": ".//div[@class='product']", "fields": [ { "name": "name", "selector": ".//span[@class='product-name']", "type": "text" }, { "name": "price", "selector": ".//span[@class='price']", "type": "text" } ] } ] } ] } 5. Job Listings with Transformations Example: <html> <div class="job-post"> <h3 class="job-title">Senior Developer</h3> <span class="salary-text">Salary: $120,000/year</span> <span class="location"> New York, NY </span> </div> </html> Generated Schema: { "name": "Job Listings", "baseSelector": "//div[@class='job-post']", "fields": [ { "name": "title", "selector": ".//h3[@class='job-title']", "type": "text", "transform": "uppercase" }, { "name": "salary", "selector": ".//span[@class='salary-text']", "type": "regex", "pattern": "\\$([\\d,]+)" }, { "name": "location", "selector": ".//span[@class='location']", "type": "text", "transform": "strip" } ] } 6. Skyscanner Place Card Example: <html> <div class="PlaceCard_descriptionContainer__M2NjN" data-testid="description-container"> <div class="PlaceCard_nameContainer__ZjZmY" tabindex="0" role="link"> <div class="PlaceCard_nameContent__ODUwZ"> <span class="BpkText_bpk-text__MjhhY BpkText_bpk-text--heading-4__Y2FlY">Doha</span> </div> <span class="BpkText_bpk-text__MjhhY BpkText_bpk-text--heading-4__Y2FlY PlaceCard_subName__NTVkY">Qatar</span> </div> <span class="PlaceCard_advertLabel__YTM0N">Sunny days and the warmest welcome awaits</span> <a class="BpkLink_bpk-link__MmQwY PlaceCard_descriptionLink__NzYwN" href="/flights/del/doha/" data-testid="flights-link"> <div class="PriceDescription_container__NjEzM"> <span class="BpkText_bpk-text--heading-5__MTRjZ">₹17,559</span> </div>
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_crawler_strategy.back.py
crawl4ai/async_crawler_strategy.back.py
from __future__ import annotations import asyncio import base64 import time from abc import ABC, abstractmethod from typing import Callable, Dict, Any, List, Union from typing import Optional, AsyncGenerator, Final import os from playwright.async_api import Page, Error from playwright.async_api import TimeoutError as PlaywrightTimeoutError from io import BytesIO from PIL import Image, ImageDraw, ImageFont import hashlib import uuid from .js_snippet import load_js_script from .models import AsyncCrawlResponse from .config import SCREENSHOT_HEIGHT_TRESHOLD from .async_configs import BrowserConfig, CrawlerRunConfig, HTTPCrawlerConfig from .async_logger import AsyncLogger from .ssl_certificate import SSLCertificate from .user_agent_generator import ValidUAGenerator from .browser_manager import BrowserManager import aiofiles import aiohttp import chardet from aiohttp.client import ClientTimeout from urllib.parse import urlparse from types import MappingProxyType import contextlib from functools import partial class AsyncCrawlerStrategy(ABC): """ Abstract base class for crawler strategies. Subclasses must implement the crawl method. """ @abstractmethod async def crawl(self, url: str, **kwargs) -> AsyncCrawlResponse: pass # 4 + 3 class AsyncPlaywrightCrawlerStrategy(AsyncCrawlerStrategy): """ Crawler strategy using Playwright. Attributes: browser_config (BrowserConfig): Configuration object containing browser settings. logger (AsyncLogger): Logger instance for recording events and errors. _downloaded_files (List[str]): List of downloaded file paths. hooks (Dict[str, Callable]): Dictionary of hooks for custom behavior. browser_manager (BrowserManager): Manager for browser creation and management. Methods: __init__(self, browser_config=None, logger=None, **kwargs): Initialize the AsyncPlaywrightCrawlerStrategy with a browser configuration. __aenter__(self): Start the browser and initialize the browser manager. __aexit__(self, exc_type, exc_val, exc_tb): Close the browser and clean up resources. start(self): Start the browser and initialize the browser manager. close(self): Close the browser and clean up resources. kill_session(self, session_id): Kill a browser session and clean up resources. crawl(self, url, **kwargs): Run the crawler for a single URL. """ def __init__( self, browser_config: BrowserConfig = None, logger: AsyncLogger = None, **kwargs ): """ Initialize the AsyncPlaywrightCrawlerStrategy with a browser configuration. Args: browser_config (BrowserConfig): Configuration object containing browser settings. If None, will be created from kwargs for backwards compatibility. logger: Logger instance for recording events and errors. **kwargs: Additional arguments for backwards compatibility and extending functionality. """ # Initialize browser config, either from provided object or kwargs self.browser_config = browser_config or BrowserConfig.from_kwargs(kwargs) self.logger = logger # Initialize session management self._downloaded_files = [] # Initialize hooks system self.hooks = { "on_browser_created": None, "on_page_context_created": None, "on_user_agent_updated": None, "on_execution_started": None, "on_execution_ended": None, "before_goto": None, "after_goto": None, "before_return_html": None, "before_retrieve_html": None, } # Initialize browser manager with config self.browser_manager = BrowserManager( browser_config=self.browser_config, logger=self.logger ) async def __aenter__(self): await self.start() return self async def __aexit__(self, exc_type, exc_val, exc_tb): await self.close() async def start(self): """ Start the browser and initialize the browser manager. """ await self.browser_manager.start() await self.execute_hook( "on_browser_created", self.browser_manager.browser, context=self.browser_manager.default_context, ) async def close(self): """ Close the browser and clean up resources. """ await self.browser_manager.close() # Explicitly reset the static Playwright instance BrowserManager._playwright_instance = None async def kill_session(self, session_id: str): """ Kill a browser session and clean up resources. Args: session_id (str): The ID of the session to kill. Returns: None """ # Log a warning message and no need kill session, in new version auto kill session self.logger.warning( message="Session auto-kill is enabled in the new version. No need to manually kill sessions.", tag="WARNING", ) await self.browser_manager.kill_session(session_id) def set_hook(self, hook_type: str, hook: Callable): """ Set a hook function for a specific hook type. Following are list of hook types: - on_browser_created: Called when a new browser instance is created. - on_page_context_created: Called when a new page context is created. - on_user_agent_updated: Called when the user agent is updated. - on_execution_started: Called when the execution starts. - before_goto: Called before a goto operation. - after_goto: Called after a goto operation. - before_return_html: Called before returning HTML content. - before_retrieve_html: Called before retrieving HTML content. All hooks except on_browser_created accepts a context and a page as arguments and **kwargs. However, on_browser_created accepts a browser and a context as arguments and **kwargs. Args: hook_type (str): The type of the hook. hook (Callable): The hook function to set. Returns: None """ if hook_type in self.hooks: self.hooks[hook_type] = hook else: raise ValueError(f"Invalid hook type: {hook_type}") async def execute_hook(self, hook_type: str, *args, **kwargs): """ Execute a hook function for a specific hook type. Args: hook_type (str): The type of the hook. *args: Variable length positional arguments. **kwargs: Keyword arguments. Returns: The return value of the hook function, if any. """ hook = self.hooks.get(hook_type) if hook: if asyncio.iscoroutinefunction(hook): return await hook(*args, **kwargs) else: return hook(*args, **kwargs) return args[0] if args else None def update_user_agent(self, user_agent: str): """ Update the user agent for the browser. Args: user_agent (str): The new user agent string. Returns: None """ self.user_agent = user_agent def set_custom_headers(self, headers: Dict[str, str]): """ Set custom headers for the browser. Args: headers (Dict[str, str]): A dictionary of headers to set. Returns: None """ self.headers = headers async def smart_wait(self, page: Page, wait_for: str, timeout: float = 30000): """ Wait for a condition in a smart way. This functions works as below: 1. If wait_for starts with 'js:', it assumes it's a JavaScript function and waits for it to return true. 2. If wait_for starts with 'css:', it assumes it's a CSS selector and waits for it to be present. 3. Otherwise, it tries to evaluate wait_for as a JavaScript function and waits for it to return true. 4. If it's not a JavaScript function, it assumes it's a CSS selector and waits for it to be present. This is a more advanced version of the wait_for parameter in CrawlerStrategy.crawl(). Args: page: Playwright page object wait_for (str): The condition to wait for. Can be a CSS selector, a JavaScript function, or explicitly prefixed with 'js:' or 'css:'. timeout (float): Maximum time to wait in milliseconds Returns: None """ wait_for = wait_for.strip() if wait_for.startswith("js:"): # Explicitly specified JavaScript js_code = wait_for[3:].strip() return await self.csp_compliant_wait(page, js_code, timeout) elif wait_for.startswith("css:"): # Explicitly specified CSS selector css_selector = wait_for[4:].strip() try: await page.wait_for_selector(css_selector, timeout=timeout) except Error as e: if "Timeout" in str(e): raise TimeoutError( f"Timeout after {timeout}ms waiting for selector '{css_selector}'" ) else: raise ValueError(f"Invalid CSS selector: '{css_selector}'") else: # Auto-detect based on content if wait_for.startswith("()") or wait_for.startswith("function"): # It's likely a JavaScript function return await self.csp_compliant_wait(page, wait_for, timeout) else: # Assume it's a CSS selector first try: await page.wait_for_selector(wait_for, timeout=timeout) except Error as e: if "Timeout" in str(e): raise TimeoutError( f"Timeout after {timeout}ms waiting for selector '{wait_for}'" ) else: # If it's not a timeout error, it might be an invalid selector # Let's try to evaluate it as a JavaScript function as a fallback try: return await self.csp_compliant_wait( page, f"() => {{{wait_for}}}", timeout ) except Error: raise ValueError( f"Invalid wait_for parameter: '{wait_for}'. " "It should be either a valid CSS selector, a JavaScript function, " "or explicitly prefixed with 'js:' or 'css:'." ) async def csp_compliant_wait( self, page: Page, user_wait_function: str, timeout: float = 30000 ): """ Wait for a condition in a CSP-compliant way. Args: page: Playwright page object user_wait_function: JavaScript function as string that returns boolean timeout: Maximum time to wait in milliseconds Returns: bool: True if condition was met, False if timed out Raises: RuntimeError: If there's an error evaluating the condition """ wrapper_js = f""" async () => {{ const userFunction = {user_wait_function}; const startTime = Date.now(); try {{ while (true) {{ if (await userFunction()) {{ return true; }} if (Date.now() - startTime > {timeout}) {{ return false; // Return false instead of throwing }} await new Promise(resolve => setTimeout(resolve, 100)); }} }} catch (error) {{ throw new Error(`Error evaluating condition: ${{error.message}}`); }} }} """ try: result = await page.evaluate(wrapper_js) return result except Exception as e: if "Error evaluating condition" in str(e): raise RuntimeError(f"Failed to evaluate wait condition: {str(e)}") # For timeout or other cases, just return False return False async def process_iframes(self, page): """ Process iframes on a page. This function will extract the content of each iframe and replace it with a div containing the extracted content. Args: page: Playwright page object Returns: Playwright page object """ # Find all iframes iframes = await page.query_selector_all("iframe") for i, iframe in enumerate(iframes): try: # Add a unique identifier to the iframe await iframe.evaluate(f'(element) => element.id = "iframe-{i}"') # Get the frame associated with this iframe frame = await iframe.content_frame() if frame: # Wait for the frame to load await frame.wait_for_load_state( "load", timeout=30000 ) # 30 seconds timeout # Extract the content of the iframe's body iframe_content = await frame.evaluate( "() => document.body.innerHTML" ) # Generate a unique class name for this iframe class_name = f"extracted-iframe-content-{i}" # Replace the iframe with a div containing the extracted content _iframe = iframe_content.replace("`", "\\`") await page.evaluate( f""" () => {{ const iframe = document.getElementById('iframe-{i}'); const div = document.createElement('div'); div.innerHTML = `{_iframe}`; div.className = '{class_name}'; iframe.replaceWith(div); }} """ ) else: self.logger.warning( message="Could not access content frame for iframe {index}", tag="SCRAPE", params={"index": i}, ) except Exception as e: self.logger.error( message="Error processing iframe {index}: {error}", tag="ERROR", params={"index": i, "error": str(e)}, ) # Return the page object return page async def create_session(self, **kwargs) -> str: """ Creates a new browser session and returns its ID. A browse session is a unique openned page can be reused for multiple crawls. This function is asynchronous and returns a string representing the session ID. Args: **kwargs: Optional keyword arguments to configure the session. Returns: str: The session ID. """ await self.start() session_id = kwargs.get("session_id") or str(uuid.uuid4()) user_agent = kwargs.get("user_agent", self.user_agent) # Use browser_manager to get a fresh page & context assigned to this session_id page, context = await self.browser_manager.get_page(CrawlerRunConfig( session_id=session_id, user_agent=user_agent, **kwargs, )) return session_id async def crawl( self, url: str, config: CrawlerRunConfig, **kwargs ) -> AsyncCrawlResponse: """ Crawls a given URL or processes raw HTML/local file content based on the URL prefix. Args: url (str): The URL to crawl. Supported prefixes: - 'http://' or 'https://': Web URL to crawl. - 'file://': Local file path to process. - 'raw://': Raw HTML content to process. **kwargs: Additional parameters: - 'screenshot' (bool): Whether to take a screenshot. - ... [other existing parameters] Returns: AsyncCrawlResponse: The response containing HTML, headers, status code, and optional screenshot. """ config = config or CrawlerRunConfig.from_kwargs(kwargs) response_headers = {} status_code = 200 # Default for local/raw HTML screenshot_data = None if url.startswith(("http://", "https://", "view-source:")): return await self._crawl_web(url, config) elif url.startswith("file://"): # initialize empty lists for console messages captured_console = [] # Process local file local_file_path = url[7:] # Remove 'file://' prefix if not os.path.exists(local_file_path): raise FileNotFoundError(f"Local file not found: {local_file_path}") with open(local_file_path, "r", encoding="utf-8") as f: html = f.read() if config.screenshot: screenshot_data = await self._generate_screenshot_from_html(html) if config.capture_console_messages: page, context = await self.browser_manager.get_page(crawlerRunConfig=config) captured_console = await self._capture_console_messages(page, url) return AsyncCrawlResponse( html=html, response_headers=response_headers, status_code=status_code, screenshot=screenshot_data, get_delayed_content=None, console_messages=captured_console, ) ##### # Since both "raw:" and "raw://" start with "raw:", the first condition is always true for both, so "raw://" will be sliced as "//...", which is incorrect. # Fix: Check for "raw://" first, then "raw:" # Also, the prefix "raw://" is actually 6 characters long, not 7, so it should be sliced accordingly: url[6:] ##### elif url.startswith("raw://") or url.startswith("raw:"): # Process raw HTML content # raw_html = url[4:] if url[:4] == "raw:" else url[7:] raw_html = url[6:] if url.startswith("raw://") else url[4:] html = raw_html if config.screenshot: screenshot_data = await self._generate_screenshot_from_html(html) return AsyncCrawlResponse( html=html, response_headers=response_headers, status_code=status_code, screenshot=screenshot_data, get_delayed_content=None, ) else: raise ValueError( "URL must start with 'http://', 'https://', 'file://', or 'raw:'" ) async def _crawl_web( self, url: str, config: CrawlerRunConfig ) -> AsyncCrawlResponse: """ Internal method to crawl web URLs with the specified configuration. Includes optional network and console capturing. Args: url (str): The web URL to crawl config (CrawlerRunConfig): Configuration object controlling the crawl behavior Returns: AsyncCrawlResponse: The response containing HTML, headers, status code, and optional data """ config.url = url response_headers = {} execution_result = None status_code = None redirected_url = url # Reset downloaded files list for new crawl self._downloaded_files = [] # Initialize capture lists captured_requests = [] captured_console = [] # Handle user agent with magic mode user_agent_to_override = config.user_agent if user_agent_to_override: self.browser_config.user_agent = user_agent_to_override elif config.magic or config.user_agent_mode == "random": self.browser_config.user_agent = ValidUAGenerator().generate( **(config.user_agent_generator_config or {}) ) # Get page for session page, context = await self.browser_manager.get_page(crawlerRunConfig=config) # await page.goto(URL) # Add default cookie # await context.add_cookies( # [{"name": "cookiesEnabled", "value": "true", "url": url}] # ) # Handle navigator overrides if config.override_navigator or config.simulate_user or config.magic: await context.add_init_script(load_js_script("navigator_overrider")) # Call hook after page creation await self.execute_hook("on_page_context_created", page, context=context, config=config) # Network Request Capturing if config.capture_network_requests: async def handle_request_capture(request): try: post_data_str = None try: # Be cautious with large post data post_data = request.post_data_buffer if post_data: # Attempt to decode, fallback to base64 or size indication try: post_data_str = post_data.decode('utf-8', errors='replace') except UnicodeDecodeError: post_data_str = f"[Binary data: {len(post_data)} bytes]" except Exception: post_data_str = "[Error retrieving post data]" captured_requests.append({ "event_type": "request", "url": request.url, "method": request.method, "headers": dict(request.headers), # Convert Header dict "post_data": post_data_str, "resource_type": request.resource_type, "is_navigation_request": request.is_navigation_request(), "timestamp": time.time() }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing request details for {request.url}: {e}", tag="CAPTURE") captured_requests.append({"event_type": "request_capture_error", "url": request.url, "error": str(e), "timestamp": time.time()}) async def handle_response_capture(response): try: try: # body = await response.body() # json_body = await response.json() text_body = await response.text() except Exception as e: body = None # json_body = None # text_body = None captured_requests.append({ "event_type": "response", "url": response.url, "status": response.status, "status_text": response.status_text, "headers": dict(response.headers), # Convert Header dict "from_service_worker": response.from_service_worker, "request_timing": response.request.timing, # Detailed timing info "timestamp": time.time(), "body" : { # "raw": body, # "json": json_body, "text": text_body } }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing response details for {response.url}: {e}", tag="CAPTURE") captured_requests.append({"event_type": "response_capture_error", "url": response.url, "error": str(e), "timestamp": time.time()}) async def handle_request_failed_capture(request): try: captured_requests.append({ "event_type": "request_failed", "url": request.url, "method": request.method, "resource_type": request.resource_type, "failure_text": str(request.failure) if request.failure else "Unknown failure", "timestamp": time.time() }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing request failed details for {request.url}: {e}", tag="CAPTURE") captured_requests.append({"event_type": "request_failed_capture_error", "url": request.url, "error": str(e), "timestamp": time.time()}) page.on("request", handle_request_capture) page.on("response", handle_response_capture) page.on("requestfailed", handle_request_failed_capture) # Console Message Capturing if config.capture_console_messages: def handle_console_capture(msg): try: message_type = "unknown" try: message_type = msg.type except: pass message_text = "unknown" try: message_text = msg.text except: pass # Basic console message with minimal content entry = { "type": message_type, "text": message_text, "timestamp": time.time() } captured_console.append(entry) except Exception as e: if self.logger: self.logger.warning(f"Error capturing console message: {e}", tag="CAPTURE") # Still add something to the list even on error captured_console.append({ "type": "console_capture_error", "error": str(e), "timestamp": time.time() }) def handle_pageerror_capture(err): try: error_message = "Unknown error" try: error_message = err.message except: pass error_stack = "" try: error_stack = err.stack except: pass captured_console.append({ "type": "error", "text": error_message, "stack": error_stack, "timestamp": time.time() }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing page error: {e}", tag="CAPTURE") captured_console.append({ "type": "pageerror_capture_error", "error": str(e), "timestamp": time.time() }) # Add event listeners directly page.on("console", handle_console_capture) page.on("pageerror", handle_pageerror_capture) # Set up console logging if requested if config.log_console: def log_consol( msg, console_log_type="debug" ): # Corrected the parameter syntax if console_log_type == "error": self.logger.error( message=f"Console error: {msg}", # Use f-string for variable interpolation tag="CONSOLE" ) elif console_log_type == "debug": self.logger.debug( message=f"Console: {msg}", # Use f-string for variable interpolation tag="CONSOLE" ) page.on("console", log_consol) page.on("pageerror", lambda e: log_consol(e, "error")) try: # Get SSL certificate information if requested and URL is HTTPS ssl_cert = None if config.fetch_ssl_certificate: ssl_cert = SSLCertificate.from_url(url) # Set up download handling if self.browser_config.accept_downloads: page.on( "download", lambda download: asyncio.create_task( self._handle_download(download) ), ) # Handle page navigation and content loading if not config.js_only: await self.execute_hook("before_goto", page, context=context, url=url, config=config) try: # Generate a unique nonce for this request if config.experimental.get("use_csp_nonce", False): nonce = hashlib.sha256(os.urandom(32)).hexdigest() # Add CSP headers to the request await page.set_extra_http_headers( { "Content-Security-Policy": f"default-src 'self'; script-src 'self' 'nonce-{nonce}' 'strict-dynamic'" } ) response = await page.goto( url, wait_until=config.wait_until, timeout=config.page_timeout ) redirected_url = page.url except Error as e: # Allow navigation to be aborted when downloading files # This is expected behavior for downloads in some browser engines if 'net::ERR_ABORTED' in str(e) and self.browser_config.accept_downloads: self.logger.info( message=f"Navigation aborted, likely due to file download: {url}", tag="GOTO", params={"url": url}, ) response = None else: raise RuntimeError(f"Failed on navigating ACS-GOTO:\n{str(e)}") await self.execute_hook( "after_goto", page, context=context, url=url, response=response, config=config ) # ────────────────────────────────────────────────────────────── # Walk the redirect chain. Playwright returns only the last # hop, so we trace the `request.redirected_from` links until the # first response that differs from the final one and surface its # status-code. # ────────────────────────────────────────────────────────────── if response is None: status_code = 200 response_headers = {} else: first_resp = response req = response.request while req and req.redirected_from: prev_req = req.redirected_from prev_resp = await prev_req.response() if prev_resp: # keep earliest first_resp = prev_resp req = prev_req status_code = first_resp.status
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_dispatcher.py
crawl4ai/async_dispatcher.py
from typing import Dict, Optional, List, Tuple, Union from .async_configs import CrawlerRunConfig from .models import ( CrawlResult, CrawlerTaskResult, CrawlStatus, DomainState, ) from .components.crawler_monitor import CrawlerMonitor from .types import AsyncWebCrawler from collections.abc import AsyncGenerator import time import psutil import asyncio import uuid from urllib.parse import urlparse import random from abc import ABC, abstractmethod from .utils import get_true_memory_usage_percent class RateLimiter: def __init__( self, base_delay: Tuple[float, float] = (1.0, 3.0), max_delay: float = 60.0, max_retries: int = 3, rate_limit_codes: List[int] = None, ): self.base_delay = base_delay self.max_delay = max_delay self.max_retries = max_retries self.rate_limit_codes = rate_limit_codes or [429, 503] self.domains: Dict[str, DomainState] = {} def get_domain(self, url: str) -> str: return urlparse(url).netloc async def wait_if_needed(self, url: str) -> None: domain = self.get_domain(url) state = self.domains.get(domain) if not state: self.domains[domain] = DomainState() state = self.domains[domain] now = time.time() if state.last_request_time: wait_time = max(0, state.current_delay - (now - state.last_request_time)) if wait_time > 0: await asyncio.sleep(wait_time) # Random delay within base range if no current delay if state.current_delay == 0: state.current_delay = random.uniform(*self.base_delay) state.last_request_time = time.time() def update_delay(self, url: str, status_code: int) -> bool: domain = self.get_domain(url) state = self.domains[domain] if status_code in self.rate_limit_codes: state.fail_count += 1 if state.fail_count > self.max_retries: return False # Exponential backoff with random jitter state.current_delay = min( state.current_delay * 2 * random.uniform(0.75, 1.25), self.max_delay ) else: # Gradually reduce delay on success state.current_delay = max( random.uniform(*self.base_delay), state.current_delay * 0.75 ) state.fail_count = 0 return True class BaseDispatcher(ABC): def __init__( self, rate_limiter: Optional[RateLimiter] = None, monitor: Optional[CrawlerMonitor] = None, ): self.crawler = None self._domain_last_hit: Dict[str, float] = {} self.concurrent_sessions = 0 self.rate_limiter = rate_limiter self.monitor = monitor def select_config(self, url: str, configs: Union[CrawlerRunConfig, List[CrawlerRunConfig]]) -> Optional[CrawlerRunConfig]: """Select the appropriate config for a given URL. Args: url: The URL to match against configs: Single config or list of configs to choose from Returns: The matching config, or None if no match found """ # Single config - return as is if isinstance(configs, CrawlerRunConfig): return configs # Empty list - return None if not configs: return None # Find first matching config for config in configs: if config.is_match(url): return config # No match found - return None to indicate URL should be skipped return None @abstractmethod async def crawl_url( self, url: str, config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], task_id: str, monitor: Optional[CrawlerMonitor] = None, ) -> CrawlerTaskResult: pass @abstractmethod async def run_urls( self, urls: List[str], crawler: AsyncWebCrawler, # noqa: F821 config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], monitor: Optional[CrawlerMonitor] = None, ) -> List[CrawlerTaskResult]: pass class MemoryAdaptiveDispatcher(BaseDispatcher): def __init__( self, memory_threshold_percent: float = 90.0, critical_threshold_percent: float = 95.0, # New critical threshold recovery_threshold_percent: float = 85.0, # New recovery threshold check_interval: float = 1.0, max_session_permit: int = 20, fairness_timeout: float = 600.0, # 10 minutes before prioritizing long-waiting URLs memory_wait_timeout: Optional[float] = 600.0, rate_limiter: Optional[RateLimiter] = None, monitor: Optional[CrawlerMonitor] = None, ): super().__init__(rate_limiter, monitor) self.memory_threshold_percent = memory_threshold_percent self.critical_threshold_percent = critical_threshold_percent self.recovery_threshold_percent = recovery_threshold_percent self.check_interval = check_interval self.max_session_permit = max_session_permit self.fairness_timeout = fairness_timeout self.memory_wait_timeout = memory_wait_timeout self.result_queue = asyncio.Queue() self.task_queue = asyncio.PriorityQueue() # Priority queue for better management self.memory_pressure_mode = False # Flag to indicate when we're in memory pressure mode self.current_memory_percent = 0.0 # Track current memory usage self._high_memory_start_time: Optional[float] = None async def _memory_monitor_task(self): """Background task to continuously monitor memory usage and update state""" while True: self.current_memory_percent = get_true_memory_usage_percent() # Enter memory pressure mode if we cross the threshold if self.current_memory_percent >= self.memory_threshold_percent: if not self.memory_pressure_mode: self.memory_pressure_mode = True self._high_memory_start_time = time.time() if self.monitor: self.monitor.update_memory_status("PRESSURE") else: if self._high_memory_start_time is None: self._high_memory_start_time = time.time() if ( self.memory_wait_timeout is not None and self._high_memory_start_time is not None and time.time() - self._high_memory_start_time >= self.memory_wait_timeout ): raise MemoryError( "Memory usage exceeded threshold for" f" {self.memory_wait_timeout} seconds" ) # Exit memory pressure mode if we go below recovery threshold elif self.memory_pressure_mode and self.current_memory_percent <= self.recovery_threshold_percent: self.memory_pressure_mode = False self._high_memory_start_time = None if self.monitor: self.monitor.update_memory_status("NORMAL") elif self.current_memory_percent < self.memory_threshold_percent: self._high_memory_start_time = None # In critical mode, we might need to take more drastic action if self.current_memory_percent >= self.critical_threshold_percent: if self.monitor: self.monitor.update_memory_status("CRITICAL") # We could implement additional memory-saving measures here await asyncio.sleep(self.check_interval) def _get_priority_score(self, wait_time: float, retry_count: int) -> float: """Calculate priority score (lower is higher priority) - URLs waiting longer than fairness_timeout get higher priority - More retry attempts decreases priority """ if wait_time > self.fairness_timeout: # High priority for long-waiting URLs return -wait_time # Standard priority based on retries return retry_count async def crawl_url( self, url: str, config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], task_id: str, retry_count: int = 0, ) -> CrawlerTaskResult: start_time = time.time() error_message = "" memory_usage = peak_memory = 0.0 # Select appropriate config for this URL selected_config = self.select_config(url, config) # If no config matches, return failed result if selected_config is None: error_message = f"No matching configuration found for URL: {url}" if self.monitor: self.monitor.update_task( task_id, status=CrawlStatus.FAILED, error_message=error_message ) return CrawlerTaskResult( task_id=task_id, url=url, result=CrawlResult( url=url, html="", metadata={"status": "no_config_match"}, success=False, error_message=error_message ), memory_usage=0, peak_memory=0, start_time=start_time, end_time=time.time(), error_message=error_message, retry_count=retry_count ) # Get starting memory for accurate measurement process = psutil.Process() start_memory = process.memory_info().rss / (1024 * 1024) try: if self.monitor: self.monitor.update_task( task_id, status=CrawlStatus.IN_PROGRESS, start_time=start_time, retry_count=retry_count ) self.concurrent_sessions += 1 if self.rate_limiter: await self.rate_limiter.wait_if_needed(url) # Check if we're in critical memory state if self.current_memory_percent >= self.critical_threshold_percent: # Requeue this task with increased priority and retry count enqueue_time = time.time() priority = self._get_priority_score(enqueue_time - start_time, retry_count + 1) await self.task_queue.put((priority, (url, task_id, retry_count + 1, enqueue_time))) # Update monitoring if self.monitor: self.monitor.update_task( task_id, status=CrawlStatus.QUEUED, error_message="Requeued due to critical memory pressure" ) # Return placeholder result with requeued status return CrawlerTaskResult( task_id=task_id, url=url, result=CrawlResult( url=url, html="", metadata={"status": "requeued"}, success=False, error_message="Requeued due to critical memory pressure" ), memory_usage=0, peak_memory=0, start_time=start_time, end_time=time.time(), error_message="Requeued due to critical memory pressure", retry_count=retry_count + 1 ) # Execute the crawl with selected config result = await self.crawler.arun(url, config=selected_config, session_id=task_id) # Measure memory usage end_memory = process.memory_info().rss / (1024 * 1024) memory_usage = peak_memory = end_memory - start_memory # Handle rate limiting if self.rate_limiter and result.status_code: if not self.rate_limiter.update_delay(url, result.status_code): error_message = f"Rate limit retry count exceeded for domain {urlparse(url).netloc}" if self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.FAILED) # Update status based on result if not result.success: error_message = result.error_message if self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.FAILED) elif self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.COMPLETED) except Exception as e: error_message = str(e) if self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.FAILED) result = CrawlResult( url=url, html="", metadata={}, success=False, error_message=str(e) ) finally: end_time = time.time() if self.monitor: self.monitor.update_task( task_id, end_time=end_time, memory_usage=memory_usage, peak_memory=peak_memory, error_message=error_message, retry_count=retry_count ) self.concurrent_sessions -= 1 return CrawlerTaskResult( task_id=task_id, url=url, result=result, memory_usage=memory_usage, peak_memory=peak_memory, start_time=start_time, end_time=end_time, error_message=error_message, retry_count=retry_count ) async def run_urls( self, urls: List[str], crawler: AsyncWebCrawler, config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], ) -> List[CrawlerTaskResult]: self.crawler = crawler # Start the memory monitor task memory_monitor = asyncio.create_task(self._memory_monitor_task()) if self.monitor: self.monitor.start() results = [] try: # Initialize task queue for url in urls: task_id = str(uuid.uuid4()) if self.monitor: self.monitor.add_task(task_id, url) # Add to queue with initial priority 0, retry count 0, and current time await self.task_queue.put((0, (url, task_id, 0, time.time()))) active_tasks = [] # Process until both queues are empty while not self.task_queue.empty() or active_tasks: if memory_monitor.done(): exc = memory_monitor.exception() if exc: for t in active_tasks: t.cancel() raise exc # If memory pressure is low, greedily fill all available slots if not self.memory_pressure_mode: slots = self.max_session_permit - len(active_tasks) while slots > 0: try: # Use get_nowait() to immediately get tasks without blocking priority, (url, task_id, retry_count, enqueue_time) = self.task_queue.get_nowait() # Create and start the task task = asyncio.create_task( self.crawl_url(url, config, task_id, retry_count) ) active_tasks.append(task) # Update waiting time in monitor if self.monitor: wait_time = time.time() - enqueue_time self.monitor.update_task( task_id, wait_time=wait_time, status=CrawlStatus.IN_PROGRESS ) slots -= 1 except asyncio.QueueEmpty: # No more tasks in queue, exit the loop break # Wait for completion even if queue is starved if active_tasks: done, pending = await asyncio.wait( active_tasks, timeout=0.1, return_when=asyncio.FIRST_COMPLETED ) # Process completed tasks for completed_task in done: result = await completed_task results.append(result) # Update active tasks list active_tasks = list(pending) else: # If no active tasks but still waiting, sleep briefly await asyncio.sleep(self.check_interval / 2) # Update priorities for waiting tasks if needed await self._update_queue_priorities() except Exception as e: if self.monitor: self.monitor.update_memory_status(f"QUEUE_ERROR: {str(e)}") finally: # Clean up memory_monitor.cancel() if self.monitor: self.monitor.stop() return results async def _update_queue_priorities(self): """Periodically update priorities of items in the queue to prevent starvation""" # Skip if queue is empty if self.task_queue.empty(): return # Use a drain-and-refill approach to update all priorities temp_items = [] # Drain the queue (with a safety timeout to prevent blocking) try: drain_start = time.time() while not self.task_queue.empty() and time.time() - drain_start < 5.0: # 5 second safety timeout try: # Get item from queue with timeout priority, (url, task_id, retry_count, enqueue_time) = await asyncio.wait_for( self.task_queue.get(), timeout=0.1 ) # Calculate new priority based on current wait time current_time = time.time() wait_time = current_time - enqueue_time new_priority = self._get_priority_score(wait_time, retry_count) # Store with updated priority temp_items.append((new_priority, (url, task_id, retry_count, enqueue_time))) # Update monitoring stats for this task if self.monitor and task_id in self.monitor.stats: self.monitor.update_task(task_id, wait_time=wait_time) except asyncio.TimeoutError: # Queue might be empty or very slow break except Exception as e: # If anything goes wrong, make sure we refill the queue with what we've got self.monitor.update_memory_status(f"QUEUE_ERROR: {str(e)}") # Calculate queue statistics if temp_items and self.monitor: total_queued = len(temp_items) wait_times = [item[1][3] for item in temp_items] highest_wait_time = time.time() - min(wait_times) if wait_times else 0 avg_wait_time = sum(time.time() - t for t in wait_times) / len(wait_times) if wait_times else 0 # Update queue statistics in monitor self.monitor.update_queue_statistics( total_queued=total_queued, highest_wait_time=highest_wait_time, avg_wait_time=avg_wait_time ) # Sort by priority (lowest number = highest priority) temp_items.sort(key=lambda x: x[0]) # Refill the queue with updated priorities for item in temp_items: await self.task_queue.put(item) async def run_urls_stream( self, urls: List[str], crawler: AsyncWebCrawler, config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], ) -> AsyncGenerator[CrawlerTaskResult, None]: self.crawler = crawler # Start the memory monitor task memory_monitor = asyncio.create_task(self._memory_monitor_task()) if self.monitor: self.monitor.start() try: # Initialize task queue for url in urls: task_id = str(uuid.uuid4()) if self.monitor: self.monitor.add_task(task_id, url) # Add to queue with initial priority 0, retry count 0, and current time await self.task_queue.put((0, (url, task_id, 0, time.time()))) active_tasks = [] completed_count = 0 total_urls = len(urls) while completed_count < total_urls: if memory_monitor.done(): exc = memory_monitor.exception() if exc: for t in active_tasks: t.cancel() raise exc # If memory pressure is low, greedily fill all available slots if not self.memory_pressure_mode: slots = self.max_session_permit - len(active_tasks) while slots > 0: try: # Use get_nowait() to immediately get tasks without blocking priority, (url, task_id, retry_count, enqueue_time) = self.task_queue.get_nowait() # Create and start the task task = asyncio.create_task( self.crawl_url(url, config, task_id, retry_count) ) active_tasks.append(task) # Update waiting time in monitor if self.monitor: wait_time = time.time() - enqueue_time self.monitor.update_task( task_id, wait_time=wait_time, status=CrawlStatus.IN_PROGRESS ) slots -= 1 except asyncio.QueueEmpty: # No more tasks in queue, exit the loop break # Process completed tasks and yield results if active_tasks: done, pending = await asyncio.wait( active_tasks, timeout=0.1, return_when=asyncio.FIRST_COMPLETED ) for completed_task in done: result = await completed_task # Only count as completed if it wasn't requeued if "requeued" not in result.error_message: completed_count += 1 yield result # Update active tasks list active_tasks = list(pending) else: # If no active tasks but still waiting, sleep briefly await asyncio.sleep(self.check_interval / 2) # Update priorities for waiting tasks if needed await self._update_queue_priorities() finally: # Clean up memory_monitor.cancel() if self.monitor: self.monitor.stop() class SemaphoreDispatcher(BaseDispatcher): def __init__( self, semaphore_count: int = 5, max_session_permit: int = 20, rate_limiter: Optional[RateLimiter] = None, monitor: Optional[CrawlerMonitor] = None, ): super().__init__(rate_limiter, monitor) self.semaphore_count = semaphore_count self.max_session_permit = max_session_permit async def crawl_url( self, url: str, config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], task_id: str, semaphore: asyncio.Semaphore = None, ) -> CrawlerTaskResult: start_time = time.time() error_message = "" memory_usage = peak_memory = 0.0 # Select appropriate config for this URL selected_config = self.select_config(url, config) # If no config matches, return failed result if selected_config is None: error_message = f"No matching configuration found for URL: {url}" if self.monitor: self.monitor.update_task( task_id, status=CrawlStatus.FAILED, error_message=error_message ) return CrawlerTaskResult( task_id=task_id, url=url, result=CrawlResult( url=url, html="", metadata={"status": "no_config_match"}, success=False, error_message=error_message ), memory_usage=0, peak_memory=0, start_time=start_time, end_time=time.time(), error_message=error_message ) try: if self.monitor: self.monitor.update_task( task_id, status=CrawlStatus.IN_PROGRESS, start_time=start_time ) if self.rate_limiter: await self.rate_limiter.wait_if_needed(url) async with semaphore: process = psutil.Process() start_memory = process.memory_info().rss / (1024 * 1024) result = await self.crawler.arun(url, config=selected_config, session_id=task_id) end_memory = process.memory_info().rss / (1024 * 1024) memory_usage = peak_memory = end_memory - start_memory if self.rate_limiter and result.status_code: if not self.rate_limiter.update_delay(url, result.status_code): error_message = f"Rate limit retry count exceeded for domain {urlparse(url).netloc}" if self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.FAILED) return CrawlerTaskResult( task_id=task_id, url=url, result=result, memory_usage=memory_usage, peak_memory=peak_memory, start_time=start_time, end_time=time.time(), error_message=error_message, ) if not result.success: error_message = result.error_message if self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.FAILED) elif self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.COMPLETED) except Exception as e: error_message = str(e) if self.monitor: self.monitor.update_task(task_id, status=CrawlStatus.FAILED) result = CrawlResult( url=url, html="", metadata={}, success=False, error_message=str(e) ) finally: end_time = time.time() if self.monitor: self.monitor.update_task( task_id, end_time=end_time, memory_usage=memory_usage, peak_memory=peak_memory, error_message=error_message, ) return CrawlerTaskResult( task_id=task_id, url=url, result=result, memory_usage=memory_usage, peak_memory=peak_memory, start_time=start_time, end_time=end_time, error_message=error_message, ) async def run_urls( self, crawler: AsyncWebCrawler, # noqa: F821 urls: List[str], config: Union[CrawlerRunConfig, List[CrawlerRunConfig]], ) -> List[CrawlerTaskResult]: self.crawler = crawler if self.monitor: self.monitor.start() try: semaphore = asyncio.Semaphore(self.semaphore_count) tasks = [] for url in urls: task_id = str(uuid.uuid4()) if self.monitor: self.monitor.add_task(task_id, url) task = asyncio.create_task( self.crawl_url(url, config, task_id, semaphore) ) tasks.append(task) return await asyncio.gather(*tasks, return_exceptions=True) finally: if self.monitor: self.monitor.stop()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/cli.py
crawl4ai/cli.py
import click import os import sys import time import humanize from typing import Dict, Any, Optional, List import json import yaml import anyio from rich.console import Console from rich.table import Table from rich.panel import Panel from rich.prompt import Prompt, Confirm from crawl4ai import ( CacheMode, AsyncWebCrawler, CrawlResult, BrowserConfig, CrawlerRunConfig, LLMExtractionStrategy, LXMLWebScrapingStrategy, JsonCssExtractionStrategy, JsonXPathExtractionStrategy, BM25ContentFilter, PruningContentFilter, BrowserProfiler, DefaultMarkdownGenerator, LLMConfig, BFSDeepCrawlStrategy, DFSDeepCrawlStrategy, BestFirstCrawlingStrategy, ) from crawl4ai.config import USER_SETTINGS from litellm import completion from pathlib import Path # Initialize rich console console = Console() def get_global_config() -> dict: config_dir = Path.home() / ".crawl4ai" config_file = config_dir / "global.yml" if not config_file.exists(): config_dir.mkdir(parents=True, exist_ok=True) return {} with open(config_file) as f: return yaml.safe_load(f) or {} def save_global_config(config: dict): config_file = Path.home() / ".crawl4ai" / "global.yml" with open(config_file, "w") as f: yaml.dump(config, f) def setup_llm_config() -> tuple[str, str]: config = get_global_config() provider = config.get("DEFAULT_LLM_PROVIDER") token = config.get("DEFAULT_LLM_PROVIDER_TOKEN") if not provider: click.echo("\nNo default LLM provider configured.") click.echo("Provider format: 'company/model' (e.g., 'openai/gpt-4o', 'anthropic/claude-3-sonnet')") click.echo("See available providers at: https://docs.litellm.ai/docs/providers") provider = click.prompt("Enter provider") if not provider.startswith("ollama/"): if not token: token = click.prompt("Enter API token for " + provider, hide_input=True) else: token = "no-token" if not config.get("DEFAULT_LLM_PROVIDER") or not config.get("DEFAULT_LLM_PROVIDER_TOKEN"): config["DEFAULT_LLM_PROVIDER"] = provider config["DEFAULT_LLM_PROVIDER_TOKEN"] = token save_global_config(config) click.echo("\nConfiguration saved to ~/.crawl4ai/global.yml") return provider, token async def stream_llm_response(url: str, markdown: str, query: str, provider: str, token: str): response = completion( model=provider, api_key=token, messages=[ { "content": f"You are Crawl4ai assistant, answering user question based on the provided context which is crawled from {url}.", "role": "system" }, { "content": f"<|start of context|>\n{markdown}\n<|end of context|>\n\n{query}", "role": "user" }, ], stream=True, ) for chunk in response: if content := chunk["choices"][0]["delta"].get("content"): print(content, end="", flush=True) print() # New line at end def parse_key_values(ctx, param, value) -> Dict[str, Any]: if not value: return {} result = {} pairs = value.split(',') for pair in pairs: try: k, v = pair.split('=', 1) # Handle common value types if v.lower() == 'true': v = True elif v.lower() == 'false': v = False elif v.isdigit(): v = int(v) elif v.replace('.','',1).isdigit(): v = float(v) elif v.startswith('[') and v.endswith(']'): v = [x.strip() for x in v[1:-1].split(',') if x.strip()] elif v.startswith('{') and v.endswith('}'): try: v = json.loads(v) except json.JSONDecodeError: raise click.BadParameter(f'Invalid JSON object: {v}') result[k.strip()] = v except ValueError: raise click.BadParameter(f'Invalid key=value pair: {pair}') return result def load_config_file(path: Optional[str]) -> dict: if not path: return {} try: with open(path) as f: if path.endswith((".yaml", ".yml")): return yaml.safe_load(f) return json.load(f) except Exception as e: raise click.BadParameter(f'Error loading config file {path}: {str(e)}') def load_schema_file(path: Optional[str]) -> dict: if not path: return None return load_config_file(path) async def run_crawler(url: str, browser_cfg: BrowserConfig, crawler_cfg: CrawlerRunConfig, verbose: bool): if verbose: click.echo("Starting crawler with configurations:") click.echo(f"Browser config: {browser_cfg.dump()}") click.echo(f"Crawler config: {crawler_cfg.dump()}") async with AsyncWebCrawler(config=browser_cfg) as crawler: try: result = await crawler.arun(url=url, config=crawler_cfg) return result except Exception as e: raise click.ClickException(f"Crawling failed: {str(e)}") def show_examples(): examples = """ 🚀 Crawl4AI CLI Examples 1️⃣ Basic Usage: # Simple crawl with default settings crwl https://example.com # Get markdown output crwl https://example.com -o markdown # Verbose JSON output with cache bypass crwl https://example.com -o json -v --bypass-cache 2️⃣ Using Config Files: # Using browser and crawler configs crwl https://example.com -B browser.yml -C crawler.yml # CSS-based extraction crwl https://example.com -e extract_css.yml -s css_schema.json -o json # LLM-based extraction with config file crwl https://example.com -e extract_llm.yml -s llm_schema.json -o json # Quick LLM-based JSON extraction (prompts for LLM provider first time) crwl https://example.com -j # Auto-extracts structured data crwl https://example.com -j "Extract product details including name, price, and features" # With specific instructions 3️⃣ Direct Parameters: # Browser settings crwl https://example.com -b "headless=true,viewport_width=1280,user_agent_mode=random" # Crawler settings crwl https://example.com -c "css_selector=#main,delay_before_return_html=2,scan_full_page=true" 4️⃣ Profile Management for Identity-Based Crawling: # Launch interactive profile manager crwl profiles # Create, list, and delete browser profiles for identity-based crawling # Use a profile for crawling (keeps you logged in) crwl https://example.com -p my-profile-name # Example: Crawl a site that requires login # 1. First create a profile and log in: crwl profiles # 2. Then use that profile to crawl the authenticated site: crwl https://site-requiring-login.com/dashboard -p my-profile-name 5️⃣ CDP Mode for Browser Automation: # Launch browser with CDP debugging on default port 9222 crwl cdp # Use a specific profile and custom port crwl cdp -p my-profile -P 9223 # Launch headless browser with CDP enabled crwl cdp --headless # Launch in incognito mode (ignores profile) crwl cdp --incognito # Use the CDP URL with other tools (Puppeteer, Playwright, etc.) # The URL will be displayed in the terminal when the browser starts 6️⃣ Sample Config Files: browser.yml: headless: true viewport_width: 1280 user_agent_mode: "random" verbose: true ignore_https_errors: true extract_css.yml: type: "json-css" params: verbose: true css_schema.json: { "name": "ArticleExtractor", "baseSelector": ".article", "fields": [ { "name": "title", "selector": "h1.title", "type": "text" }, { "name": "link", "selector": "a.read-more", "type": "attribute", "attribute": "href" } ] } extract_llm.yml: type: "llm" provider: "openai/gpt-4" instruction: "Extract all articles with their titles and links" api_token: "your-token" params: temperature: 0.3 max_tokens: 1000 llm_schema.json: { "title": "Article", "type": "object", "properties": { "title": { "type": "string", "description": "The title of the article" }, "link": { "type": "string", "description": "URL to the full article" } } } 7️⃣ Advanced Usage: # Combine configs with direct parameters crwl https://example.com -B browser.yml -b "headless=false,viewport_width=1920" # Full extraction pipeline with config files crwl https://example.com \\ -B browser.yml \\ -C crawler.yml \\ -e extract_llm.yml \\ -s llm_schema.json \\ -o json \\ -v # Quick LLM-based extraction with specific instructions crwl https://amazon.com/dp/B01DFKC2SO \\ -j "Extract product title, current price, original price, rating, and all product specifications" \\ -b "headless=true,viewport_width=1280" \\ -v # Content filtering with BM25 crwl https://example.com \\ -f filter_bm25.yml \\ -o markdown-fit # Authenticated crawling with profile crwl https://login-required-site.com \\ -p my-authenticated-profile \\ -c "css_selector=.dashboard-content" \\ -o markdown For more documentation visit: https://github.com/unclecode/crawl4ai 8️⃣ Q&A with LLM: # Ask a question about the content crwl https://example.com -q "What is the main topic discussed?" # First view content, then ask questions crwl https://example.com -o markdown # See the crawled content first crwl https://example.com -q "Summarize the key points" crwl https://example.com -q "What are the conclusions?" # Advanced crawling with Q&A crwl https://example.com \\ -B browser.yml \\ -c "css_selector=article,scan_full_page=true" \\ -q "What are the pros and cons mentioned?" Note: First time using -q will prompt for LLM provider and API token. These will be saved in ~/.crawl4ai/global.yml for future use. Supported provider format: 'company/model' Examples: - ollama/llama3.3 - openai/gpt-4 - anthropic/claude-3-sonnet - cohere/command - google/gemini-pro See full list of providers: https://docs.litellm.ai/docs/providers # Set default LLM provider and token in advance crwl config set DEFAULT_LLM_PROVIDER "anthropic/claude-3-sonnet" crwl config set DEFAULT_LLM_PROVIDER_TOKEN "your-api-token-here" # Set default browser behavior crwl config set BROWSER_HEADLESS false # Always show browser window crwl config set USER_AGENT_MODE random # Use random user agent 9️⃣ Profile Management: # Launch interactive profile manager crwl profiles # Create a profile and use it for crawling crwl profiles # Create and set up your profile interactively crwl https://example.com -p my-profile-name # Use profile for crawling # Example workflow for authenticated site # 1. First create a profile and log in to the site: crwl profiles # Select "Create new profile" option # 2. Then use that profile to crawl authenticated content: crwl https://site-requiring-login.com/dashboard -p my-profile-name 🔄 Builtin Browser Management: # Start a builtin browser (runs in the background) crwl browser start # Check builtin browser status crwl browser status # Open a visible window to see the browser crwl browser view --url https://example.com # Stop the builtin browser crwl browser stop # Restart with different options crwl browser restart --browser-type chromium --port 9223 --no-headless # Use the builtin browser in your code # (Just set browser_mode="builtin" in your BrowserConfig) browser_config = BrowserConfig( browser_mode="builtin", headless=True ) # Usage via CLI: crwl https://example.com -b "browser_mode=builtin" """ click.echo(examples) def get_directory_size(path: str) -> int: """Calculate the total size of a directory in bytes""" total_size = 0 for dirpath, _, filenames in os.walk(path): for f in filenames: fp = os.path.join(dirpath, f) if not os.path.islink(fp): total_size += os.path.getsize(fp) return total_size def display_profiles_table(profiles: List[Dict[str, Any]]): """Display a rich table of browser profiles""" if not profiles: console.print(Panel("[yellow]No profiles found. Create one with the 'create' command.[/yellow]", title="Browser Profiles", border_style="blue")) return table = Table(title="Browser Profiles", show_header=True, header_style="bold cyan", border_style="blue") table.add_column("#", style="dim", width=4) table.add_column("Name", style="cyan", no_wrap=True) table.add_column("Path", style="green") table.add_column("Created", style="yellow") table.add_column("Browser", style="magenta") table.add_column("Size", style="blue", justify="right") for i, profile in enumerate(profiles): # Calculate folder size size = get_directory_size(profile["path"]) human_size = humanize.naturalsize(size) # Format creation date created = profile["created"].strftime("%Y-%m-%d %H:%M") # Add row to table table.add_row( str(i+1), profile["name"], profile["path"], created, profile["type"].capitalize(), human_size ) console.print(table) async def create_profile_interactive(profiler: BrowserProfiler): """Interactive profile creation wizard""" console.print(Panel("[bold cyan]Create Browser Profile[/bold cyan]\n" "This will open a browser window for you to set up your identity.\n" "Log in to sites, adjust settings, then press 'q' to save.", border_style="cyan")) profile_name = Prompt.ask("[cyan]Enter profile name[/cyan]", default=f"profile_{int(time.time())}") console.print("[cyan]Creating profile...[/cyan]") console.print("[yellow]A browser window will open. After logging in to sites, press 'q' in this terminal to save.[/yellow]") # Create the profile try: profile_path = await profiler.create_profile(profile_name) if profile_path: console.print(f"[green]Profile successfully created at:[/green] {profile_path}") else: console.print("[red]Failed to create profile.[/red]") except Exception as e: console.print(f"[red]Error creating profile: {str(e)}[/red]") def delete_profile_interactive(profiler: BrowserProfiler): """Interactive profile deletion""" profiles = profiler.list_profiles() if not profiles: console.print("[yellow]No profiles found to delete.[/yellow]") return # Display profiles display_profiles_table(profiles) # Get profile selection idx = Prompt.ask( "[red]Enter number of profile to delete[/red]", console=console, choices=[str(i+1) for i in range(len(profiles))], show_choices=False ) try: idx = int(idx) - 1 profile = profiles[idx] # Confirm deletion if Confirm.ask(f"[red]Are you sure you want to delete profile '{profile['name']}'?[/red]"): success = profiler.delete_profile(profile["path"]) if success: console.print(f"[green]Profile '{profile['name']}' deleted successfully.[/green]") else: console.print(f"[red]Failed to delete profile '{profile['name']}'.[/red]") except (ValueError, IndexError): console.print("[red]Invalid selection.[/red]") async def crawl_with_profile_cli(profile_path, url): """Use a profile to crawl a website via CLI""" console.print(f"[cyan]Crawling [bold]{url}[/bold] using profile at [bold]{profile_path}[/bold][/cyan]") # Create browser config with the profile browser_cfg = BrowserConfig( headless=False, # Set to False to see the browser in action use_managed_browser=True, user_data_dir=profile_path ) # Default crawler config crawler_cfg = CrawlerRunConfig() # Ask for output format output_format = Prompt.ask( "[cyan]Output format[/cyan]", choices=["all", "json", "markdown", "md", "title"], default="markdown" ) try: # Run the crawler result = await run_crawler(url, browser_cfg, crawler_cfg, True) # Handle output if output_format == "all": console.print(json.dumps(result.model_dump(), indent=2)) elif output_format == "json": console.print(json.dumps(json.loads(result.extracted_content), indent=2)) elif output_format in ["markdown", "md"]: console.print(result.markdown.raw_markdown) elif output_format == "title": console.print(result.metadata.get("title", "No title found")) console.print(f"[green]Successfully crawled[/green] {url}") return result except Exception as e: console.print(f"[red]Error crawling:[/red] {str(e)}") return None async def use_profile_to_crawl(): """Interactive profile selection for crawling""" profiler = BrowserProfiler() profiles = profiler.list_profiles() if not profiles: console.print("[yellow]No profiles found. Create one first.[/yellow]") return # Display profiles display_profiles_table(profiles) # Get profile selection idx = Prompt.ask( "[cyan]Enter number of profile to use[/cyan]", console=console, choices=[str(i+1) for i in range(len(profiles))], show_choices=False ) try: idx = int(idx) - 1 profile = profiles[idx] # Get URL url = Prompt.ask("[cyan]Enter URL to crawl[/cyan]") if url: # Crawl with the selected profile await crawl_with_profile_cli(profile["path"], url) else: console.print("[red]No URL provided[/red]") except (ValueError, IndexError): console.print("[red]Invalid selection[/red]") async def manage_profiles(): """Interactive profile management menu""" profiler = BrowserProfiler() options = { "1": "List profiles", "2": "Create new profile", "3": "Delete profile", "4": "Use a profile to crawl a website", "5": "Exit", } while True: console.print(Panel("[bold cyan]Browser Profile Manager[/bold cyan]", border_style="cyan")) for key, value in options.items(): color = "green" if key == "1" else "yellow" if key == "2" else "red" if key == "3" else "blue" if key == "4" else "cyan" console.print(f"[{color}]{key}[/{color}]. {value}") choice = Prompt.ask("Enter choice", choices=list(options.keys()), default="1") if choice == "1": # List profiles profiles = profiler.list_profiles() display_profiles_table(profiles) elif choice == "2": # Create profile await create_profile_interactive(profiler) elif choice == "3": # Delete profile delete_profile_interactive(profiler) elif choice == "4": # Use profile to crawl await use_profile_to_crawl() elif choice == "5": # Exit console.print("[cyan]Exiting profile manager.[/cyan]") break # Add a separator between operations console.print("\n") @click.group(context_settings={"help_option_names": ["-h", "--help"]}) def cli(): """Crawl4AI CLI - Web content extraction and browser profile management tool""" pass @cli.group("browser") def browser_cmd(): """Manage browser instances for Crawl4AI Commands to manage browser instances for Crawl4AI, including: - status - Check status of the builtin browser - start - Start a new builtin browser - stop - Stop the running builtin browser - restart - Restart the builtin browser """ pass @browser_cmd.command("status") def browser_status_cmd(): """Show status of the builtin browser""" profiler = BrowserProfiler() try: status = anyio.run(profiler.get_builtin_browser_status) if status["running"]: info = status["info"] console.print(Panel( f"[green]Builtin browser is running[/green]\n\n" f"CDP URL: [cyan]{info['cdp_url']}[/cyan]\n" f"Process ID: [yellow]{info['pid']}[/yellow]\n" f"Browser type: [blue]{info['browser_type']}[/blue]\n" f"User data directory: [magenta]{info['user_data_dir']}[/magenta]\n" f"Started: [cyan]{time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(info['start_time']))}[/cyan]", title="Builtin Browser Status", border_style="green" )) else: console.print(Panel( "[yellow]Builtin browser is not running[/yellow]\n\n" "Use 'crwl browser start' to start a builtin browser", title="Builtin Browser Status", border_style="yellow" )) except Exception as e: console.print(f"[red]Error checking browser status: {str(e)}[/red]") sys.exit(1) @browser_cmd.command("start") @click.option("--browser-type", "-b", type=click.Choice(["chromium", "firefox"]), default="chromium", help="Browser type (default: chromium)") @click.option("--port", "-p", type=int, default=9222, help="Debugging port (default: 9222)") @click.option("--headless/--no-headless", default=True, help="Run browser in headless mode") def browser_start_cmd(browser_type: str, port: int, headless: bool): """Start a builtin browser instance This will start a persistent browser instance that can be used by Crawl4AI by setting browser_mode="builtin" in BrowserConfig. """ profiler = BrowserProfiler() # First check if browser is already running status = anyio.run(profiler.get_builtin_browser_status) if status["running"]: console.print(Panel( "[yellow]Builtin browser is already running[/yellow]\n\n" f"CDP URL: [cyan]{status['cdp_url']}[/cyan]\n\n" "Use 'crwl browser restart' to restart the browser", title="Builtin Browser Start", border_style="yellow" )) return try: console.print(Panel( f"[cyan]Starting builtin browser[/cyan]\n\n" f"Browser type: [green]{browser_type}[/green]\n" f"Debugging port: [yellow]{port}[/yellow]\n" f"Headless: [cyan]{'Yes' if headless else 'No'}[/cyan]", title="Builtin Browser Start", border_style="cyan" )) cdp_url = anyio.run( profiler.launch_builtin_browser, browser_type, port, headless ) if cdp_url: console.print(Panel( f"[green]Builtin browser started successfully[/green]\n\n" f"CDP URL: [cyan]{cdp_url}[/cyan]\n\n" "This browser will be used automatically when setting browser_mode='builtin'", title="Builtin Browser Start", border_style="green" )) else: console.print(Panel( "[red]Failed to start builtin browser[/red]", title="Builtin Browser Start", border_style="red" )) sys.exit(1) except Exception as e: console.print(f"[red]Error starting builtin browser: {str(e)}[/red]") sys.exit(1) @browser_cmd.command("stop") def browser_stop_cmd(): """Stop the running builtin browser""" profiler = BrowserProfiler() try: # First check if browser is running status = anyio.run(profiler.get_builtin_browser_status) if not status["running"]: console.print(Panel( "[yellow]No builtin browser is currently running[/yellow]", title="Builtin Browser Stop", border_style="yellow" )) return console.print(Panel( "[cyan]Stopping builtin browser...[/cyan]", title="Builtin Browser Stop", border_style="cyan" )) success = anyio.run(profiler.kill_builtin_browser) if success: console.print(Panel( "[green]Builtin browser stopped successfully[/green]", title="Builtin Browser Stop", border_style="green" )) else: console.print(Panel( "[red]Failed to stop builtin browser[/red]", title="Builtin Browser Stop", border_style="red" )) sys.exit(1) except Exception as e: console.print(f"[red]Error stopping builtin browser: {str(e)}[/red]") sys.exit(1) @browser_cmd.command("view") @click.option("--url", "-u", help="URL to navigate to (defaults to about:blank)") def browser_view_cmd(url: Optional[str]): """ Open a visible window of the builtin browser This command connects to the running builtin browser and opens a visible window, allowing you to see what the browser is currently viewing or navigate to a URL. """ profiler = BrowserProfiler() try: # First check if browser is running status = anyio.run(profiler.get_builtin_browser_status) if not status["running"]: console.print(Panel( "[yellow]No builtin browser is currently running[/yellow]\n\n" "Use 'crwl browser start' to start a builtin browser first", title="Builtin Browser View", border_style="yellow" )) return info = status["info"] cdp_url = info["cdp_url"] console.print(Panel( f"[cyan]Opening visible window connected to builtin browser[/cyan]\n\n" f"CDP URL: [green]{cdp_url}[/green]\n" f"URL to load: [yellow]{url or 'about:blank'}[/yellow]", title="Builtin Browser View", border_style="cyan" )) # Use the CDP URL to launch a new visible window import subprocess import os # Determine the browser command based on platform if sys.platform == "darwin": # macOS browser_cmd = ["/Applications/Google Chrome.app/Contents/MacOS/Google Chrome"] elif sys.platform == "win32": # Windows browser_cmd = ["C:\\Program Files\\Google\\Chrome\\Application\\chrome.exe"] else: # Linux browser_cmd = ["google-chrome"] # Add arguments browser_args = [ f"--remote-debugging-port={info['debugging_port']}", "--remote-debugging-address=localhost", "--no-first-run", "--no-default-browser-check" ] # Add URL if provided if url: browser_args.append(url) # Launch browser try: subprocess.Popen(browser_cmd + browser_args) console.print("[green]Browser window opened. Close it when finished viewing.[/green]") except Exception as e: console.print(f"[red]Error launching browser: {str(e)}[/red]") console.print(f"[yellow]Try connecting manually to {cdp_url} in Chrome or using the '--remote-debugging-port' flag.[/yellow]") except Exception as e: console.print(f"[red]Error viewing builtin browser: {str(e)}[/red]") sys.exit(1) @browser_cmd.command("restart") @click.option("--browser-type", "-b", type=click.Choice(["chromium", "firefox"]), default=None, help="Browser type (defaults to same as current)") @click.option("--port", "-p", type=int, default=None, help="Debugging port (defaults to same as current)") @click.option("--headless/--no-headless", default=None, help="Run browser in headless mode") def browser_restart_cmd(browser_type: Optional[str], port: Optional[int], headless: Optional[bool]): """Restart the builtin browser Stops the current builtin browser if running and starts a new one. By default, uses the same configuration as the current browser. """ profiler = BrowserProfiler() try: # First check if browser is running and get its config status = anyio.run(profiler.get_builtin_browser_status) current_config = {} if status["running"]: info = status["info"] current_config = { "browser_type": info["browser_type"], "port": info["debugging_port"], "headless": True # Default assumption } # Stop the browser console.print(Panel( "[cyan]Stopping current builtin browser...[/cyan]", title="Builtin Browser Restart", border_style="cyan" )) success = anyio.run(profiler.kill_builtin_browser) if not success: console.print(Panel( "[red]Failed to stop current browser[/red]", title="Builtin Browser Restart", border_style="red" )) sys.exit(1) # Use provided options or defaults from current config browser_type = browser_type or current_config.get("browser_type", "chromium") port = port or current_config.get("port", 9222) headless = headless if headless is not None else current_config.get("headless", True) # Start a new browser console.print(Panel( f"[cyan]Starting new builtin browser[/cyan]\n\n" f"Browser type: [green]{browser_type}[/green]\n" f"Debugging port: [yellow]{port}[/yellow]\n" f"Headless: [cyan]{'Yes' if headless else 'No'}[/cyan]", title="Builtin Browser Restart", border_style="cyan" )) cdp_url = anyio.run( profiler.launch_builtin_browser, browser_type, port, headless ) if cdp_url: console.print(Panel( f"[green]Builtin browser restarted successfully[/green]\n\n" f"CDP URL: [cyan]{cdp_url}[/cyan]", title="Builtin Browser Restart", border_style="green" )) else: console.print(Panel( "[red]Failed to restart builtin browser[/red]", title="Builtin Browser Restart", border_style="red" )) sys.exit(1) except Exception as e: console.print(f"[red]Error restarting builtin browser: {str(e)}[/red]") sys.exit(1) @cli.command("cdp") @click.option("--user-data-dir", "-d", help="Directory to use for browser data (will be created if it doesn't exist)") @click.option("--port", "-P", type=int, default=9222, help="Debugging port (default: 9222)") @click.option("--browser-type", "-b", type=click.Choice(["chromium", "firefox"]), default="chromium", help="Browser type (default: chromium)") @click.option("--headless", is_flag=True, help="Run browser in headless mode") @click.option("--incognito", is_flag=True, help="Run in incognito/private mode (ignores user-data-dir)") def cdp_cmd(user_data_dir: Optional[str], port: int, browser_type: str, headless: bool, incognito: bool): """Launch a standalone browser with CDP debugging enabled
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_logger.py
crawl4ai/async_logger.py
from abc import ABC, abstractmethod from enum import Enum from typing import Optional, Dict, Any, List import os from datetime import datetime from urllib.parse import unquote from rich.console import Console from rich.text import Text from .utils import create_box_message class LogLevel(Enum): DEFAULT = 0 DEBUG = 1 INFO = 2 SUCCESS = 3 WARNING = 4 ERROR = 5 CRITICAL = 6 ALERT = 7 NOTICE = 8 EXCEPTION = 9 FATAL = 10 def __str__(self): return self.name.lower() class LogColor(str, Enum): """Enum for log colors.""" DEBUG = "bright_black" INFO = "cyan" SUCCESS = "green" WARNING = "yellow" ERROR = "red" CYAN = "cyan" GREEN = "green" YELLOW = "yellow" MAGENTA = "magenta" DIM_MAGENTA = "dim magenta" RED = "red" def __str__(self): """Automatically convert rich color to string.""" return self.value class AsyncLoggerBase(ABC): @abstractmethod def debug(self, message: str, tag: str = "DEBUG", **kwargs): pass @abstractmethod def info(self, message: str, tag: str = "INFO", **kwargs): pass @abstractmethod def success(self, message: str, tag: str = "SUCCESS", **kwargs): pass @abstractmethod def warning(self, message: str, tag: str = "WARNING", **kwargs): pass @abstractmethod def error(self, message: str, tag: str = "ERROR", **kwargs): pass @abstractmethod def url_status(self, url: str, success: bool, timing: float, tag: str = "FETCH", url_length: int = 100): pass @abstractmethod def error_status(self, url: str, error: str, tag: str = "ERROR", url_length: int = 100): pass class AsyncLogger(AsyncLoggerBase): """ Asynchronous logger with support for colored console output and file logging. Supports templated messages with colored components. """ DEFAULT_ICONS = { "INIT": "→", "READY": "✓", "FETCH": "↓", "SCRAPE": "◆", "EXTRACT": "■", "COMPLETE": "●", "ERROR": "×", "DEBUG": "⋯", "INFO": "ℹ", "WARNING": "⚠", "SUCCESS": "✔", "CRITICAL": "‼", "ALERT": "⚡", "NOTICE": "ℹ", "EXCEPTION": "❗", "FATAL": "☠", "DEFAULT": "•", } DEFAULT_COLORS = { LogLevel.DEBUG: LogColor.DEBUG, LogLevel.INFO: LogColor.INFO, LogLevel.SUCCESS: LogColor.SUCCESS, LogLevel.WARNING: LogColor.WARNING, LogLevel.ERROR: LogColor.ERROR, } def __init__( self, log_file: Optional[str] = None, log_level: LogLevel = LogLevel.DEBUG, tag_width: int = 10, icons: Optional[Dict[str, str]] = None, colors: Optional[Dict[LogLevel, LogColor]] = None, verbose: bool = True, ): """ Initialize the logger. Args: log_file: Optional file path for logging log_level: Minimum log level to display tag_width: Width for tag formatting icons: Custom icons for different tags colors: Custom colors for different log levels verbose: Whether to output to console """ self.log_file = log_file self.log_level = log_level self.tag_width = tag_width self.icons = icons or self.DEFAULT_ICONS self.colors = colors or self.DEFAULT_COLORS self.verbose = verbose self.console = Console() # Create log file directory if needed if log_file: os.makedirs(os.path.dirname(os.path.abspath(log_file)), exist_ok=True) def _format_tag(self, tag: str) -> str: """Format a tag with consistent width.""" return f"[{tag}]".ljust(self.tag_width, ".") def _get_icon(self, tag: str) -> str: """Get the icon for a tag, defaulting to info icon if not found.""" return self.icons.get(tag, self.icons["INFO"]) def _shorten(self, text, length, placeholder="..."): """Truncate text in the middle if longer than length, or pad if shorter.""" if len(text) <= length: return text.ljust(length) # Pad with spaces to reach desired length half = (length - len(placeholder)) // 2 shortened = text[:half] + placeholder + text[-half:] return shortened.ljust(length) # Also pad shortened text to consistent length def _write_to_file(self, message: str): """Write a message to the log file if configured.""" if self.log_file: text = Text.from_markup(message) plain_text = text.plain timestamp = datetime.now().strftime("%Y-%m-%d %H:%M:%S.%f")[:-3] with open(self.log_file, "a", encoding="utf-8") as f: f.write(f"[{timestamp}] {plain_text}\n") def _log( self, level: LogLevel, message: str, tag: str, params: Optional[Dict[str, Any]] = None, colors: Optional[Dict[str, LogColor]] = None, boxes: Optional[List[str]] = None, base_color: Optional[LogColor] = None, **kwargs, ): """ Core logging method that handles message formatting and output. Args: level: Log level for this message message: Message template string tag: Tag for the message params: Parameters to format into the message colors: Color overrides for specific parameters boxes: Box overrides for specific parameters base_color: Base color for the entire message """ if level.value < self.log_level.value: return # avoid conflict with rich formatting parsed_message = message.replace("[", "[[").replace("]", "]]") if params: # FIXME: If there are formatting strings in floating point format, # this may result in colors and boxes not being applied properly. # such as {value:.2f}, the value is 0.23333 format it to 0.23, # but we replace("0.23333", "[color]0.23333[/color]") formatted_message = parsed_message.format(**params) for key, value in params.items(): # value_str may discard `[` and `]`, so we need to replace it. value_str = str(value).replace("[", "[[").replace("]", "]]") # check is need apply color if colors and key in colors: color_str = f"[{colors[key]}]{value_str}[/{colors[key]}]" formatted_message = formatted_message.replace(value_str, color_str) value_str = color_str # check is need apply box if boxes and key in boxes: formatted_message = formatted_message.replace(value_str, create_box_message(value_str, type=str(level))) else: formatted_message = parsed_message # Construct the full log line color: LogColor = base_color or self.colors[level] log_line = f"[{color}]{self._format_tag(tag)} {self._get_icon(tag)} {formatted_message} [/{color}]" # Output to console if verbose if self.verbose or kwargs.get("force_verbose", False): self.console.print(log_line) # Write to file if configured self._write_to_file(log_line) def debug(self, message: str, tag: str = "DEBUG", **kwargs): """Log a debug message.""" self._log(LogLevel.DEBUG, message, tag, **kwargs) def info(self, message: str, tag: str = "INFO", **kwargs): """Log an info message.""" self._log(LogLevel.INFO, message, tag, **kwargs) def success(self, message: str, tag: str = "SUCCESS", **kwargs): """Log a success message.""" self._log(LogLevel.SUCCESS, message, tag, **kwargs) def warning(self, message: str, tag: str = "WARNING", **kwargs): """Log a warning message.""" self._log(LogLevel.WARNING, message, tag, **kwargs) def critical(self, message: str, tag: str = "CRITICAL", **kwargs): """Log a critical message.""" self._log(LogLevel.ERROR, message, tag, **kwargs) def exception(self, message: str, tag: str = "EXCEPTION", **kwargs): """Log an exception message.""" self._log(LogLevel.ERROR, message, tag, **kwargs) def fatal(self, message: str, tag: str = "FATAL", **kwargs): """Log a fatal message.""" self._log(LogLevel.ERROR, message, tag, **kwargs) def alert(self, message: str, tag: str = "ALERT", **kwargs): """Log an alert message.""" self._log(LogLevel.ERROR, message, tag, **kwargs) def notice(self, message: str, tag: str = "NOTICE", **kwargs): """Log a notice message.""" self._log(LogLevel.INFO, message, tag, **kwargs) def error(self, message: str, tag: str = "ERROR", **kwargs): """Log an error message.""" self._log(LogLevel.ERROR, message, tag, **kwargs) def url_status( self, url: str, success: bool, timing: float, tag: str = "FETCH", url_length: int = 100, ): """ Convenience method for logging URL fetch status. Args: url: The URL being processed success: Whether the operation was successful timing: Time taken for the operation tag: Tag for the message url_length: Maximum length for URL in log """ decoded_url = unquote(url) readable_url = self._shorten(decoded_url, url_length) self._log( level=LogLevel.SUCCESS if success else LogLevel.ERROR, message="{url} | {status} | ⏱: {timing:.2f}s", tag=tag, params={ "url": readable_url, "status": "✓" if success else "✗", "timing": timing, }, colors={ "status": LogColor.SUCCESS if success else LogColor.ERROR, "timing": LogColor.WARNING, }, ) def error_status( self, url: str, error: str, tag: str = "ERROR", url_length: int = 50 ): """ Convenience method for logging error status. Args: url: The URL being processed error: Error message tag: Tag for the message url_length: Maximum length for URL in log """ decoded_url = unquote(url) readable_url = self._shorten(decoded_url, url_length) self._log( level=LogLevel.ERROR, message="{url} | Error: {error}", tag=tag, params={"url": readable_url, "error": error}, ) class AsyncFileLogger(AsyncLoggerBase): """ File-only asynchronous logger that writes logs to a specified file. """ def __init__(self, log_file: str): """ Initialize the file logger. Args: log_file: File path for logging """ self.log_file = log_file os.makedirs(os.path.dirname(os.path.abspath(log_file)), exist_ok=True) def _write_to_file(self, level: str, message: str, tag: str): """Write a message to the log file.""" timestamp = datetime.now().strftime("%Y-%m-%d %H:%M:%S.%f")[:-3] with open(self.log_file, "a", encoding="utf-8") as f: f.write(f"[{timestamp}] [{level}] [{tag}] {message}\n") def debug(self, message: str, tag: str = "DEBUG", **kwargs): """Log a debug message to file.""" self._write_to_file("DEBUG", message, tag) def info(self, message: str, tag: str = "INFO", **kwargs): """Log an info message to file.""" self._write_to_file("INFO", message, tag) def success(self, message: str, tag: str = "SUCCESS", **kwargs): """Log a success message to file.""" self._write_to_file("SUCCESS", message, tag) def warning(self, message: str, tag: str = "WARNING", **kwargs): """Log a warning message to file.""" self._write_to_file("WARNING", message, tag) def error(self, message: str, tag: str = "ERROR", **kwargs): """Log an error message to file.""" self._write_to_file("ERROR", message, tag) def url_status(self, url: str, success: bool, timing: float, tag: str = "FETCH", url_length: int = 100): """Log URL fetch status to file.""" status = "SUCCESS" if success else "FAILED" message = f"{url[:url_length]}... | Status: {status} | Time: {timing:.2f}s" self._write_to_file("URL_STATUS", message, tag) def error_status(self, url: str, error: str, tag: str = "ERROR", url_length: int = 100): """Log error status to file.""" message = f"{url[:url_length]}... | Error: {error}" self._write_to_file("ERROR", message, tag)
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/models.py
crawl4ai/models.py
from pydantic import BaseModel, HttpUrl, PrivateAttr, Field, ConfigDict from typing import List, Dict, Optional, Callable, Awaitable, Union, Any from typing import AsyncGenerator from typing import Generic, TypeVar from enum import Enum from dataclasses import dataclass from .ssl_certificate import SSLCertificate from datetime import datetime from datetime import timedelta ############################### # Dispatcher Models ############################### @dataclass class DomainState: last_request_time: float = 0 current_delay: float = 0 fail_count: int = 0 @dataclass class CrawlerTaskResult: task_id: str url: str result: "CrawlResult" memory_usage: float peak_memory: float start_time: Union[datetime, float] end_time: Union[datetime, float] error_message: str = "" retry_count: int = 0 wait_time: float = 0.0 @property def success(self) -> bool: return self.result.success class CrawlStatus(Enum): QUEUED = "QUEUED" IN_PROGRESS = "IN_PROGRESS" COMPLETED = "COMPLETED" FAILED = "FAILED" @dataclass class CrawlStats: task_id: str url: str status: CrawlStatus start_time: Optional[Union[datetime, float]] = None end_time: Optional[Union[datetime, float]] = None memory_usage: float = 0.0 peak_memory: float = 0.0 error_message: str = "" wait_time: float = 0.0 retry_count: int = 0 counted_requeue: bool = False @property def duration(self) -> str: if not self.start_time: return "0:00" # Convert start_time to datetime if it's a float start = self.start_time if isinstance(start, float): start = datetime.fromtimestamp(start) # Get end time or use current time end = self.end_time or datetime.now() # Convert end_time to datetime if it's a float if isinstance(end, float): end = datetime.fromtimestamp(end) duration = end - start return str(timedelta(seconds=int(duration.total_seconds()))) class DisplayMode(Enum): DETAILED = "DETAILED" AGGREGATED = "AGGREGATED" ############################### # Crawler Models ############################### @dataclass class TokenUsage: completion_tokens: int = 0 prompt_tokens: int = 0 total_tokens: int = 0 completion_tokens_details: Optional[dict] = None prompt_tokens_details: Optional[dict] = None class UrlModel(BaseModel): url: HttpUrl forced: bool = False @dataclass class TraversalStats: """Statistics for the traversal process""" start_time: datetime = datetime.now() urls_processed: int = 0 urls_failed: int = 0 urls_skipped: int = 0 total_depth_reached: int = 0 current_depth: int = 0 class DispatchResult(BaseModel): task_id: str memory_usage: float peak_memory: float start_time: Union[datetime, float] end_time: Union[datetime, float] error_message: str = "" class MarkdownGenerationResult(BaseModel): raw_markdown: str markdown_with_citations: str references_markdown: str fit_markdown: Optional[str] = None fit_html: Optional[str] = None def __str__(self): return self.raw_markdown class CrawlResult(BaseModel): url: str html: str fit_html: Optional[str] = None success: bool cleaned_html: Optional[str] = None media: Dict[str, List[Dict]] = {} links: Dict[str, List[Dict]] = {} downloaded_files: Optional[List[str]] = None js_execution_result: Optional[Dict[str, Any]] = None screenshot: Optional[str] = None pdf: Optional[bytes] = None mhtml: Optional[str] = None _markdown: Optional[MarkdownGenerationResult] = PrivateAttr(default=None) extracted_content: Optional[str] = None metadata: Optional[dict] = None error_message: Optional[str] = None session_id: Optional[str] = None response_headers: Optional[dict] = None status_code: Optional[int] = None ssl_certificate: Optional[SSLCertificate] = None dispatch_result: Optional[DispatchResult] = None redirected_url: Optional[str] = None network_requests: Optional[List[Dict[str, Any]]] = None console_messages: Optional[List[Dict[str, Any]]] = None tables: List[Dict] = Field(default_factory=list) # NEW – [{headers,rows,caption,summary}] model_config = ConfigDict(arbitrary_types_allowed=True) # NOTE: The StringCompatibleMarkdown class, custom __init__ method, property getters/setters, # and model_dump override all exist to support a smooth transition from markdown as a string # to markdown as a MarkdownGenerationResult object, while maintaining backward compatibility. # # This allows code that expects markdown to be a string to continue working, while also # providing access to the full MarkdownGenerationResult object's properties. # # The markdown_v2 property is deprecated and raises an error directing users to use markdown. # # When backward compatibility is no longer needed in future versions, this entire mechanism # can be simplified to a standard field with no custom accessors or serialization logic. def __init__(self, **data): markdown_result = data.pop('markdown', None) super().__init__(**data) if markdown_result is not None: self._markdown = ( MarkdownGenerationResult(**markdown_result) if isinstance(markdown_result, dict) else markdown_result ) @property def markdown(self): """ Property that returns a StringCompatibleMarkdown object that behaves like a string but also provides access to MarkdownGenerationResult attributes. This approach allows backward compatibility with code that expects 'markdown' to be a string, while providing access to the full MarkdownGenerationResult. """ if self._markdown is None: return None return StringCompatibleMarkdown(self._markdown) @markdown.setter def markdown(self, value): """ Setter for the markdown property. """ self._markdown = value @property def markdown_v2(self): """ Deprecated property that raises an AttributeError when accessed. This property exists to inform users that 'markdown_v2' has been deprecated and they should use 'markdown' instead. """ raise AttributeError( "The 'markdown_v2' attribute is deprecated and has been removed. " """Please use 'markdown' instead, which now returns a MarkdownGenerationResult, with following properties: - raw_markdown: The raw markdown string - markdown_with_citations: The markdown string with citations - references_markdown: The markdown string with references - fit_markdown: The markdown string with fit text """ ) @property def fit_markdown(self): """ Deprecated property that raises an AttributeError when accessed. """ raise AttributeError( "The 'fit_markdown' attribute is deprecated and has been removed. " "Please use 'markdown.fit_markdown' instead." ) @property def fit_html(self): """ Deprecated property that raises an AttributeError when accessed. """ raise AttributeError( "The 'fit_html' attribute is deprecated and has been removed. " "Please use 'markdown.fit_html' instead." ) def model_dump(self, *args, **kwargs): """ Override model_dump to include the _markdown private attribute in serialization. This override is necessary because: 1. PrivateAttr fields are excluded from serialization by default 2. We need to maintain backward compatibility by including the 'markdown' field in the serialized output 3. We're transitioning from 'markdown_v2' to enhancing 'markdown' to hold the same type of data Future developers: This method ensures that the markdown content is properly serialized despite being stored in a private attribute. If the serialization requirements change, this is where you would update the logic. """ result = super().model_dump(*args, **kwargs) # Remove any property descriptors that might have been included # These deprecated properties should not be in the serialized output for key in ['fit_html', 'fit_markdown', 'markdown_v2']: if key in result and isinstance(result[key], property): # del result[key] # Nasrin: I decided to convert it to string instead of removing it. result[key] = str(result[key]) # Add the markdown field properly if self._markdown is not None: result["markdown"] = self._markdown.model_dump() return result class StringCompatibleMarkdown(str): """A string subclass that also provides access to MarkdownGenerationResult attributes""" def __new__(cls, markdown_result): return super().__new__(cls, markdown_result.raw_markdown) def __init__(self, markdown_result): self._markdown_result = markdown_result def __getattr__(self, name): return getattr(self._markdown_result, name) CrawlResultT = TypeVar('CrawlResultT', bound=CrawlResult) class CrawlResultContainer(Generic[CrawlResultT]): def __init__(self, results: Union[CrawlResultT, List[CrawlResultT]]): # Normalize to a list if isinstance(results, list): self._results = results else: self._results = [results] def __iter__(self): return iter(self._results) def __getitem__(self, index): return self._results[index] def __len__(self): return len(self._results) def __getattr__(self, attr): # Delegate attribute access to the first element. if self._results: return getattr(self._results[0], attr) raise AttributeError(f"{self.__class__.__name__} object has no attribute '{attr}'") def __repr__(self): return f"{self.__class__.__name__}({self._results!r})" RunManyReturn = Union[ CrawlResultContainer[CrawlResultT], AsyncGenerator[CrawlResultT, None] ] # END of backward compatibility code for markdown/markdown_v2. # When removing this code in the future, make sure to: # 1. Replace the private attribute and property with a standard field # 2. Update any serialization logic that might depend on the current behavior class AsyncCrawlResponse(BaseModel): html: str response_headers: Dict[str, str] js_execution_result: Optional[Dict[str, Any]] = None status_code: int screenshot: Optional[str] = None pdf_data: Optional[bytes] = None mhtml_data: Optional[str] = None get_delayed_content: Optional[Callable[[Optional[float]], Awaitable[str]]] = None downloaded_files: Optional[List[str]] = None ssl_certificate: Optional[SSLCertificate] = None redirected_url: Optional[str] = None network_requests: Optional[List[Dict[str, Any]]] = None console_messages: Optional[List[Dict[str, Any]]] = None model_config = ConfigDict(arbitrary_types_allowed=True) ############################### # Scraping Models ############################### class MediaItem(BaseModel): src: Optional[str] = "" data: Optional[str] = "" alt: Optional[str] = "" desc: Optional[str] = "" score: Optional[int] = 0 type: str = "image" group_id: Optional[int] = 0 format: Optional[str] = None width: Optional[int] = None class Link(BaseModel): href: Optional[str] = "" text: Optional[str] = "" title: Optional[str] = "" base_domain: Optional[str] = "" head_data: Optional[Dict[str, Any]] = None # Head metadata extracted from link target head_extraction_status: Optional[str] = None # "success", "failed", "skipped" head_extraction_error: Optional[str] = None # Error message if extraction failed intrinsic_score: Optional[float] = None # Quality score based on URL structure, text, and context contextual_score: Optional[float] = None # BM25 relevance score based on query and head content total_score: Optional[float] = None # Combined score from intrinsic and contextual scores class Media(BaseModel): images: List[MediaItem] = [] videos: List[ MediaItem ] = [] # Using MediaItem model for now, can be extended with Video model if needed audios: List[ MediaItem ] = [] # Using MediaItem model for now, can be extended with Audio model if needed tables: List[Dict] = [] # Table data extracted from HTML tables class Links(BaseModel): internal: List[Link] = [] external: List[Link] = [] class ScrapingResult(BaseModel): cleaned_html: str success: bool media: Media = Media() links: Links = Links() metadata: Dict[str, Any] = {}
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_database.py
crawl4ai/async_database.py
import os from pathlib import Path import aiosqlite import asyncio from typing import Optional, Dict from contextlib import asynccontextmanager import json from .models import CrawlResult, MarkdownGenerationResult, StringCompatibleMarkdown import aiofiles from .async_logger import AsyncLogger from .utils import ensure_content_dirs, generate_content_hash from .utils import VersionManager from .utils import get_error_context, create_box_message base_directory = DB_PATH = os.path.join( os.getenv("CRAWL4_AI_BASE_DIRECTORY", Path.home()), ".crawl4ai" ) os.makedirs(DB_PATH, exist_ok=True) DB_PATH = os.path.join(base_directory, "crawl4ai.db") class AsyncDatabaseManager: def __init__(self, pool_size: int = 10, max_retries: int = 3): self.db_path = DB_PATH self.content_paths = ensure_content_dirs(os.path.dirname(DB_PATH)) self.pool_size = pool_size self.max_retries = max_retries self.connection_pool: Dict[int, aiosqlite.Connection] = {} self.pool_lock = asyncio.Lock() self.init_lock = asyncio.Lock() self.connection_semaphore = asyncio.Semaphore(pool_size) self._initialized = False self.version_manager = VersionManager() self.logger = AsyncLogger( log_file=os.path.join(base_directory, ".crawl4ai", "crawler_db.log"), verbose=False, tag_width=10, ) async def initialize(self): """Initialize the database and connection pool""" try: self.logger.info("Initializing database", tag="INIT") # Ensure the database file exists os.makedirs(os.path.dirname(self.db_path), exist_ok=True) # Check if version update is needed needs_update = self.version_manager.needs_update() # Always ensure base table exists await self.ainit_db() # Verify the table exists async with aiosqlite.connect(self.db_path, timeout=30.0) as db: async with db.execute( "SELECT name FROM sqlite_master WHERE type='table' AND name='crawled_data'" ) as cursor: result = await cursor.fetchone() if not result: raise Exception("crawled_data table was not created") # If version changed or fresh install, run updates if needs_update: self.logger.info("New version detected, running updates", tag="INIT") await self.update_db_schema() from .migrations import ( run_migration, ) # Import here to avoid circular imports await run_migration() self.version_manager.update_version() # Update stored version after successful migration self.logger.success( "Version update completed successfully", tag="COMPLETE" ) else: self.logger.success( "Database initialization completed successfully", tag="COMPLETE" ) except Exception as e: self.logger.error( message="Database initialization error: {error}", tag="ERROR", params={"error": str(e)}, ) self.logger.info( message="Database will be initialized on first use", tag="INIT" ) raise async def cleanup(self): """Cleanup connections when shutting down""" async with self.pool_lock: for conn in self.connection_pool.values(): await conn.close() self.connection_pool.clear() @asynccontextmanager async def get_connection(self): """Connection pool manager with enhanced error handling""" if not self._initialized: async with self.init_lock: if not self._initialized: try: await self.initialize() self._initialized = True except Exception as e: import sys error_context = get_error_context(sys.exc_info()) self.logger.error( message="Database initialization failed:\n{error}\n\nContext:\n{context}\n\nTraceback:\n{traceback}", tag="ERROR", force_verbose=True, params={ "error": str(e), "context": error_context["code_context"], "traceback": error_context["full_traceback"], }, ) raise await self.connection_semaphore.acquire() task_id = id(asyncio.current_task()) try: async with self.pool_lock: if task_id not in self.connection_pool: try: conn = await aiosqlite.connect(self.db_path, timeout=30.0) await conn.execute("PRAGMA journal_mode = WAL") await conn.execute("PRAGMA busy_timeout = 5000") # Verify database structure async with conn.execute( "PRAGMA table_info(crawled_data)" ) as cursor: columns = await cursor.fetchall() column_names = [col[1] for col in columns] expected_columns = { "url", "html", "cleaned_html", "markdown", "extracted_content", "success", "media", "links", "metadata", "screenshot", "response_headers", "downloaded_files", } missing_columns = expected_columns - set(column_names) if missing_columns: raise ValueError( f"Database missing columns: {missing_columns}" ) self.connection_pool[task_id] = conn except Exception as e: import sys error_context = get_error_context(sys.exc_info()) error_message = ( f"Unexpected error in db get_connection at line {error_context['line_no']} " f"in {error_context['function']} ({error_context['filename']}):\n" f"Error: {str(e)}\n\n" f"Code context:\n{error_context['code_context']}" ) self.logger.error( message="{error}", tag="ERROR", params={"error": str(error_message)}, boxes=["error"], ) raise yield self.connection_pool[task_id] except Exception as e: import sys error_context = get_error_context(sys.exc_info()) error_message = ( f"Unexpected error in db get_connection at line {error_context['line_no']} " f"in {error_context['function']} ({error_context['filename']}):\n" f"Error: {str(e)}\n\n" f"Code context:\n{error_context['code_context']}" ) self.logger.error( message="{error}", tag="ERROR", params={"error": str(error_message)}, boxes=["error"], ) raise finally: async with self.pool_lock: if task_id in self.connection_pool: await self.connection_pool[task_id].close() del self.connection_pool[task_id] self.connection_semaphore.release() async def execute_with_retry(self, operation, *args): """Execute database operations with retry logic""" for attempt in range(self.max_retries): try: async with self.get_connection() as db: result = await operation(db, *args) await db.commit() return result except Exception as e: if attempt == self.max_retries - 1: self.logger.error( message="Operation failed after {retries} attempts: {error}", tag="ERROR", force_verbose=True, params={"retries": self.max_retries, "error": str(e)}, ) raise await asyncio.sleep(1 * (attempt + 1)) # Exponential backoff async def ainit_db(self): """Initialize database schema""" async with aiosqlite.connect(self.db_path, timeout=30.0) as db: await db.execute( """ CREATE TABLE IF NOT EXISTS crawled_data ( url TEXT PRIMARY KEY, html TEXT, cleaned_html TEXT, markdown TEXT, extracted_content TEXT, success BOOLEAN, media TEXT DEFAULT "{}", links TEXT DEFAULT "{}", metadata TEXT DEFAULT "{}", screenshot TEXT DEFAULT "", response_headers TEXT DEFAULT "{}", downloaded_files TEXT DEFAULT "{}" -- New column added ) """ ) await db.commit() async def update_db_schema(self): """Update database schema if needed""" async with aiosqlite.connect(self.db_path, timeout=30.0) as db: cursor = await db.execute("PRAGMA table_info(crawled_data)") columns = await cursor.fetchall() column_names = [column[1] for column in columns] # List of new columns to add new_columns = [ "media", "links", "metadata", "screenshot", "response_headers", "downloaded_files", ] for column in new_columns: if column not in column_names: await self.aalter_db_add_column(column, db) await db.commit() async def aalter_db_add_column(self, new_column: str, db): """Add new column to the database""" if new_column == "response_headers": await db.execute( f'ALTER TABLE crawled_data ADD COLUMN {new_column} TEXT DEFAULT "{{}}"' ) else: await db.execute( f'ALTER TABLE crawled_data ADD COLUMN {new_column} TEXT DEFAULT ""' ) self.logger.info( message="Added column '{column}' to the database", tag="INIT", params={"column": new_column}, ) async def aget_cached_url(self, url: str) -> Optional[CrawlResult]: """Retrieve cached URL data as CrawlResult""" async def _get(db): async with db.execute( "SELECT * FROM crawled_data WHERE url = ?", (url,) ) as cursor: row = await cursor.fetchone() if not row: return None # Get column names columns = [description[0] for description in cursor.description] # Create dict from row data row_dict = dict(zip(columns, row)) # Load content from files using stored hashes content_fields = { "html": row_dict["html"], "cleaned_html": row_dict["cleaned_html"], "markdown": row_dict["markdown"], "extracted_content": row_dict["extracted_content"], "screenshot": row_dict["screenshot"], "screenshots": row_dict["screenshot"], } for field, hash_value in content_fields.items(): if hash_value: content = await self._load_content( hash_value, field.split("_")[0], # Get content type from field name ) row_dict[field] = content or "" else: row_dict[field] = "" # Parse JSON fields json_fields = [ "media", "links", "metadata", "response_headers", "markdown", ] for field in json_fields: try: row_dict[field] = ( json.loads(row_dict[field]) if row_dict[field] else {} ) except json.JSONDecodeError: # Very UGLY, never mention it to me please if field == "markdown" and isinstance(row_dict[field], str): row_dict[field] = MarkdownGenerationResult( raw_markdown=row_dict[field] or "", markdown_with_citations="", references_markdown="", fit_markdown="", fit_html="", ) else: row_dict[field] = {} if isinstance(row_dict["markdown"], Dict): if row_dict["markdown"].get("raw_markdown"): row_dict["markdown"] = row_dict["markdown"]["raw_markdown"] # Parse downloaded_files try: row_dict["downloaded_files"] = ( json.loads(row_dict["downloaded_files"]) if row_dict["downloaded_files"] else [] ) except json.JSONDecodeError: row_dict["downloaded_files"] = [] # Remove any fields not in CrawlResult model valid_fields = CrawlResult.__annotations__.keys() filtered_dict = {k: v for k, v in row_dict.items() if k in valid_fields} filtered_dict["markdown"] = row_dict["markdown"] return CrawlResult(**filtered_dict) try: return await self.execute_with_retry(_get) except Exception as e: self.logger.error( message="Error retrieving cached URL: {error}", tag="ERROR", force_verbose=True, params={"error": str(e)}, ) return None async def acache_url(self, result: CrawlResult): """Cache CrawlResult data""" # Store content files and get hashes content_map = { "html": (result.html, "html"), "cleaned_html": (result.cleaned_html or "", "cleaned"), "markdown": None, "extracted_content": (result.extracted_content or "", "extracted"), "screenshot": (result.screenshot or "", "screenshots"), } try: if isinstance(result.markdown, StringCompatibleMarkdown): content_map["markdown"] = ( result.markdown, "markdown", ) elif isinstance(result.markdown, MarkdownGenerationResult): content_map["markdown"] = ( result.markdown.model_dump_json(), "markdown", ) elif isinstance(result.markdown, str): markdown_result = MarkdownGenerationResult(raw_markdown=result.markdown) content_map["markdown"] = ( markdown_result.model_dump_json(), "markdown", ) else: content_map["markdown"] = ( MarkdownGenerationResult().model_dump_json(), "markdown", ) except Exception as e: self.logger.warning( message=f"Error processing markdown content: {str(e)}", tag="WARNING" ) # Fallback to empty markdown result content_map["markdown"] = ( MarkdownGenerationResult().model_dump_json(), "markdown", ) content_hashes = {} for field, (content, content_type) in content_map.items(): content_hashes[field] = await self._store_content(content, content_type) async def _cache(db): await db.execute( """ INSERT INTO crawled_data ( url, html, cleaned_html, markdown, extracted_content, success, media, links, metadata, screenshot, response_headers, downloaded_files ) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?) ON CONFLICT(url) DO UPDATE SET html = excluded.html, cleaned_html = excluded.cleaned_html, markdown = excluded.markdown, extracted_content = excluded.extracted_content, success = excluded.success, media = excluded.media, links = excluded.links, metadata = excluded.metadata, screenshot = excluded.screenshot, response_headers = excluded.response_headers, downloaded_files = excluded.downloaded_files """, ( result.url, content_hashes["html"], content_hashes["cleaned_html"], content_hashes["markdown"], content_hashes["extracted_content"], result.success, json.dumps(result.media), json.dumps(result.links), json.dumps(result.metadata or {}), content_hashes["screenshot"], json.dumps(result.response_headers or {}), json.dumps(result.downloaded_files or []), ), ) try: await self.execute_with_retry(_cache) except Exception as e: self.logger.error( message="Error caching URL: {error}", tag="ERROR", force_verbose=True, params={"error": str(e)}, ) async def aget_total_count(self) -> int: """Get total number of cached URLs""" async def _count(db): async with db.execute("SELECT COUNT(*) FROM crawled_data") as cursor: result = await cursor.fetchone() return result[0] if result else 0 try: return await self.execute_with_retry(_count) except Exception as e: self.logger.error( message="Error getting total count: {error}", tag="ERROR", force_verbose=True, params={"error": str(e)}, ) return 0 async def aclear_db(self): """Clear all data from the database""" async def _clear(db): await db.execute("DELETE FROM crawled_data") try: await self.execute_with_retry(_clear) except Exception as e: self.logger.error( message="Error clearing database: {error}", tag="ERROR", force_verbose=True, params={"error": str(e)}, ) async def aflush_db(self): """Drop the entire table""" async def _flush(db): await db.execute("DROP TABLE IF EXISTS crawled_data") try: await self.execute_with_retry(_flush) except Exception as e: self.logger.error( message="Error flushing database: {error}", tag="ERROR", force_verbose=True, params={"error": str(e)}, ) async def _store_content(self, content: str, content_type: str) -> str: """Store content in filesystem and return hash""" if not content: return "" content_hash = generate_content_hash(content) file_path = os.path.join(self.content_paths[content_type], content_hash) # Only write if file doesn't exist if not os.path.exists(file_path): async with aiofiles.open(file_path, "w", encoding="utf-8") as f: await f.write(content) return content_hash async def _load_content( self, content_hash: str, content_type: str ) -> Optional[str]: """Load content from filesystem by hash""" if not content_hash: return None file_path = os.path.join(self.content_paths[content_type], content_hash) try: async with aiofiles.open(file_path, "r", encoding="utf-8") as f: return await f.read() except: self.logger.error( message="Failed to load content: {file_path}", tag="ERROR", force_verbose=True, params={"file_path": file_path}, ) return None # Create a singleton instance async_db_manager = AsyncDatabaseManager()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/ssl_certificate.py
crawl4ai/ssl_certificate.py
"""SSL Certificate class for handling certificate operations.""" import ssl import socket import base64 import json from typing import Dict, Any, Optional from urllib.parse import urlparse import OpenSSL.crypto from pathlib import Path # === Inherit from dict === class SSLCertificate(dict): """ A class representing an SSL certificate, behaving like a dictionary for direct JSON serialization. It stores the certificate information internally and provides methods for export and property access. Inherits from dict, so instances are directly JSON serializable. """ # Use __slots__ for potential memory optimization if desired, though less common when inheriting dict # __slots__ = ("_cert_info",) # If using slots, be careful with dict inheritance interaction def __init__(self, cert_info: Dict[str, Any]): """ Initializes the SSLCertificate object. Args: cert_info (Dict[str, Any]): The raw certificate dictionary. """ # 1. Decode the data (handle bytes -> str) decoded_info = self._decode_cert_data(cert_info) # 2. Store the decoded info internally (optional but good practice) # self._cert_info = decoded_info # You can keep this if methods rely on it # 3. Initialize the dictionary part of the object with the decoded data super().__init__(decoded_info) @staticmethod def _decode_cert_data(data: Any) -> Any: """Helper method to decode bytes in certificate data.""" if isinstance(data, bytes): try: # Try UTF-8 first, fallback to latin-1 for arbitrary bytes return data.decode("utf-8") except UnicodeDecodeError: return data.decode("latin-1") # Or handle as needed, maybe hex representation elif isinstance(data, dict): return { ( k.decode("utf-8") if isinstance(k, bytes) else k ): SSLCertificate._decode_cert_data(v) for k, v in data.items() } elif isinstance(data, list): return [SSLCertificate._decode_cert_data(item) for item in data] return data @staticmethod def from_url(url: str, timeout: int = 10) -> Optional["SSLCertificate"]: """ Create SSLCertificate instance from a URL. Fetches cert info and initializes. (Fetching logic remains the same) """ cert_info_raw = None # Variable to hold the fetched dict try: hostname = urlparse(url).netloc if ":" in hostname: hostname = hostname.split(":")[0] context = ssl.create_default_context() # Set check_hostname to False and verify_mode to CERT_NONE temporarily # for potentially problematic certificates during fetch, but parse the result regardless. # context.check_hostname = False # context.verify_mode = ssl.CERT_NONE with socket.create_connection((hostname, 443), timeout=timeout) as sock: with context.wrap_socket(sock, server_hostname=hostname) as ssock: cert_binary = ssock.getpeercert(binary_form=True) if not cert_binary: print(f"Warning: No certificate returned for {hostname}") return None x509 = OpenSSL.crypto.load_certificate( OpenSSL.crypto.FILETYPE_ASN1, cert_binary ) # Create the dictionary directly cert_info_raw = { "subject": dict(x509.get_subject().get_components()), "issuer": dict(x509.get_issuer().get_components()), "version": x509.get_version(), "serial_number": hex(x509.get_serial_number()), "not_before": x509.get_notBefore(), # Keep as bytes initially, _decode handles it "not_after": x509.get_notAfter(), # Keep as bytes initially "fingerprint": x509.digest("sha256").hex(), # hex() is already string "signature_algorithm": x509.get_signature_algorithm(), # Keep as bytes "raw_cert": base64.b64encode(cert_binary), # Base64 is bytes, _decode handles it } # Add extensions extensions = [] for i in range(x509.get_extension_count()): ext = x509.get_extension(i) # get_short_name() returns bytes, str(ext) handles value conversion extensions.append( {"name": ext.get_short_name(), "value": str(ext)} ) cert_info_raw["extensions"] = extensions except ssl.SSLCertVerificationError as e: print(f"SSL Verification Error for {url}: {e}") # Decide if you want to proceed or return None based on your needs # You might try fetching without verification here if needed, but be cautious. return None except socket.gaierror: print(f"Could not resolve hostname: {hostname}") return None except socket.timeout: print(f"Connection timed out for {url}") return None except Exception as e: print(f"Error fetching/processing certificate for {url}: {e}") # Log the full error details if needed: logging.exception("Cert fetch error") return None # If successful, create the SSLCertificate instance from the dictionary if cert_info_raw: return SSLCertificate(cert_info_raw) else: return None # --- Properties now access the dictionary items directly via self[] --- @property def issuer(self) -> Dict[str, str]: return self.get("issuer", {}) # Use self.get for safety @property def subject(self) -> Dict[str, str]: return self.get("subject", {}) @property def valid_from(self) -> str: return self.get("not_before", "") @property def valid_until(self) -> str: return self.get("not_after", "") @property def fingerprint(self) -> str: return self.get("fingerprint", "") # --- Export methods can use `self` directly as it is the dict --- def to_json(self, filepath: Optional[str] = None) -> Optional[str]: """Export certificate as JSON.""" # `self` is already the dictionary we want to serialize json_str = json.dumps(self, indent=2, ensure_ascii=False) if filepath: Path(filepath).write_text(json_str, encoding="utf-8") return None return json_str def to_pem(self, filepath: Optional[str] = None) -> Optional[str]: """Export certificate as PEM.""" try: # Decode the raw_cert (which should be string due to _decode) raw_cert_bytes = base64.b64decode(self.get("raw_cert", "")) x509 = OpenSSL.crypto.load_certificate( OpenSSL.crypto.FILETYPE_ASN1, raw_cert_bytes ) pem_data = OpenSSL.crypto.dump_certificate( OpenSSL.crypto.FILETYPE_PEM, x509 ).decode("utf-8") if filepath: Path(filepath).write_text(pem_data, encoding="utf-8") return None return pem_data except Exception as e: print(f"Error converting to PEM: {e}") return None def to_der(self, filepath: Optional[str] = None) -> Optional[bytes]: """Export certificate as DER.""" try: # Decode the raw_cert (which should be string due to _decode) der_data = base64.b64decode(self.get("raw_cert", "")) if filepath: Path(filepath).write_bytes(der_data) return None return der_data except Exception as e: print(f"Error converting to DER: {e}") return None # Optional: Add __repr__ for better debugging def __repr__(self) -> str: subject_cn = self.subject.get('CN', 'N/A') issuer_cn = self.issuer.get('CN', 'N/A') return f"<SSLCertificate Subject='{subject_cn}' Issuer='{issuer_cn}'>"
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/chunking_strategy.py
crawl4ai/chunking_strategy.py
from abc import ABC, abstractmethod import re from collections import Counter import string from .model_loader import load_nltk_punkt # Define the abstract base class for chunking strategies class ChunkingStrategy(ABC): """ Abstract base class for chunking strategies. """ @abstractmethod def chunk(self, text: str) -> list: """ Abstract method to chunk the given text. Args: text (str): The text to chunk. Returns: list: A list of chunks. """ pass # Create an identity chunking strategy f(x) = [x] class IdentityChunking(ChunkingStrategy): """ Chunking strategy that returns the input text as a single chunk. """ def chunk(self, text: str) -> list: return [text] # Regex-based chunking class RegexChunking(ChunkingStrategy): """ Chunking strategy that splits text based on regular expression patterns. """ def __init__(self, patterns=None, **kwargs): """ Initialize the RegexChunking object. Args: patterns (list): A list of regular expression patterns to split text. """ if patterns is None: patterns = [r"\n\n"] # Default split pattern self.patterns = patterns def chunk(self, text: str) -> list: paragraphs = [text] for pattern in self.patterns: new_paragraphs = [] for paragraph in paragraphs: new_paragraphs.extend(re.split(pattern, paragraph)) paragraphs = new_paragraphs return paragraphs # NLP-based sentence chunking class NlpSentenceChunking(ChunkingStrategy): """ Chunking strategy that splits text into sentences using NLTK's sentence tokenizer. """ def __init__(self, **kwargs): """ Initialize the NlpSentenceChunking object. """ from crawl4ai.le.legacy.model_loader import load_nltk_punkt load_nltk_punkt() def chunk(self, text: str) -> list: # Improved regex for sentence splitting # sentence_endings = re.compile( # r'(?<!\w\.\w.)(?<![A-Z][a-z]\.)(?<![A-Z][A-Z]\.)(?<![A-Za-z]\.)(?<=\.|\?|\!|\n)\s' # ) # sentences = sentence_endings.split(text) # sens = [sent.strip() for sent in sentences if sent] from nltk.tokenize import sent_tokenize sentences = sent_tokenize(text) sens = [sent.strip() for sent in sentences] return list(set(sens)) # Topic-based segmentation using TextTiling class TopicSegmentationChunking(ChunkingStrategy): """ Chunking strategy that segments text into topics using NLTK's TextTilingTokenizer. How it works: 1. Segment the text into topics using TextTilingTokenizer 2. Extract keywords for each topic segment """ def __init__(self, num_keywords=3, **kwargs): """ Initialize the TopicSegmentationChunking object. Args: num_keywords (int): The number of keywords to extract for each topic segment. """ import nltk as nl self.tokenizer = nl.tokenize.TextTilingTokenizer() self.num_keywords = num_keywords def chunk(self, text: str) -> list: # Use the TextTilingTokenizer to segment the text segmented_topics = self.tokenizer.tokenize(text) return segmented_topics def extract_keywords(self, text: str) -> list: # Tokenize and remove stopwords and punctuation import nltk as nl tokens = nl.toknize.word_tokenize(text) tokens = [ token.lower() for token in tokens if token not in nl.corpus.stopwords.words("english") and token not in string.punctuation ] # Calculate frequency distribution freq_dist = Counter(tokens) keywords = [word for word, freq in freq_dist.most_common(self.num_keywords)] return keywords def chunk_with_topics(self, text: str) -> list: # Segment the text into topics segments = self.chunk(text) # Extract keywords for each topic segment segments_with_topics = [ (segment, self.extract_keywords(segment)) for segment in segments ] return segments_with_topics # Fixed-length word chunks class FixedLengthWordChunking(ChunkingStrategy): """ Chunking strategy that splits text into fixed-length word chunks. How it works: 1. Split the text into words 2. Create chunks of fixed length 3. Return the list of chunks """ def __init__(self, chunk_size=100, **kwargs): """ Initialize the fixed-length word chunking strategy with the given chunk size. Args: chunk_size (int): The size of each chunk in words. """ self.chunk_size = chunk_size def chunk(self, text: str) -> list: words = text.split() return [ " ".join(words[i : i + self.chunk_size]) for i in range(0, len(words), self.chunk_size) ] # Sliding window chunking class SlidingWindowChunking(ChunkingStrategy): """ Chunking strategy that splits text into overlapping word chunks. How it works: 1. Split the text into words 2. Create chunks of fixed length 3. Return the list of chunks """ def __init__(self, window_size=100, step=50, **kwargs): """ Initialize the sliding window chunking strategy with the given window size and step size. Args: window_size (int): The size of the sliding window in words. step (int): The step size for sliding the window in words. """ self.window_size = window_size self.step = step def chunk(self, text: str) -> list: words = text.split() chunks = [] if len(words) <= self.window_size: return [text] for i in range(0, len(words) - self.window_size + 1, self.step): chunk = " ".join(words[i : i + self.window_size]) chunks.append(chunk) # Handle the last chunk if it doesn't align perfectly if i + self.window_size < len(words): chunks.append(" ".join(words[-self.window_size :])) return chunks class OverlappingWindowChunking(ChunkingStrategy): """ Chunking strategy that splits text into overlapping word chunks. How it works: 1. Split the text into words using whitespace 2. Create chunks of fixed length equal to the window size 3. Slide the window by the overlap size 4. Return the list of chunks """ def __init__(self, window_size=1000, overlap=100, **kwargs): """ Initialize the overlapping window chunking strategy with the given window size and overlap size. Args: window_size (int): The size of the window in words. overlap (int): The size of the overlap between consecutive chunks in words. """ self.window_size = window_size self.overlap = overlap def chunk(self, text: str) -> list: words = text.split() chunks = [] if len(words) <= self.window_size: return [text] start = 0 while start < len(words): end = start + self.window_size chunk = " ".join(words[start:end]) chunks.append(chunk) if end >= len(words): break start = end - self.overlap return chunks
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/browser_manager.py
crawl4ai/browser_manager.py
import asyncio import time from typing import List, Optional import os import sys import shutil import tempfile import psutil import signal import subprocess import shlex from playwright.async_api import BrowserContext import hashlib from .js_snippet import load_js_script from .config import DOWNLOAD_PAGE_TIMEOUT from .async_configs import BrowserConfig, CrawlerRunConfig from .utils import get_chromium_path import warnings BROWSER_DISABLE_OPTIONS = [ "--disable-background-networking", "--disable-background-timer-throttling", "--disable-backgrounding-occluded-windows", "--disable-breakpad", "--disable-client-side-phishing-detection", "--disable-component-extensions-with-background-pages", "--disable-default-apps", "--disable-extensions", "--disable-features=TranslateUI", "--disable-hang-monitor", "--disable-ipc-flooding-protection", "--disable-popup-blocking", "--disable-prompt-on-repost", "--disable-sync", "--force-color-profile=srgb", "--metrics-recording-only", "--no-first-run", "--password-store=basic", "--use-mock-keychain", ] class ManagedBrowser: """ Manages the browser process and context. This class allows to connect to the browser using CDP protocol. Attributes: browser_type (str): The type of browser to launch. Supported values: "chromium", "firefox", "webkit". Default: "chromium". user_data_dir (str or None): Path to a user data directory for persistent sessions. If None, a temporary directory may be used. Default: None. headless (bool): Whether to run the browser in headless mode (no visible GUI). Default: True. browser_process (subprocess.Popen): The process object for the browser. temp_dir (str): Temporary directory for user data if not provided. debugging_port (int): Port for debugging the browser. host (str): Host for debugging the browser. Methods: start(): Starts the browser process and returns the CDP endpoint URL. _get_browser_path(): Returns the browser executable path based on OS and browser type. _get_browser_args(): Returns browser-specific command line arguments. _get_user_data_dir(): Returns the user data directory path. _cleanup(): Terminates the browser process and removes the temporary directory. create_profile(): Static method to create a user profile by launching a browser for user interaction. """ @staticmethod def build_browser_flags(config: BrowserConfig) -> List[str]: """Common CLI flags for launching Chromium""" flags = [ "--disable-gpu", "--disable-gpu-compositing", "--disable-software-rasterizer", "--no-sandbox", "--disable-dev-shm-usage", "--no-first-run", "--no-default-browser-check", "--disable-infobars", "--window-position=0,0", "--ignore-certificate-errors", "--ignore-certificate-errors-spki-list", "--disable-blink-features=AutomationControlled", "--window-position=400,0", "--disable-renderer-backgrounding", "--disable-ipc-flooding-protection", "--force-color-profile=srgb", "--mute-audio", "--disable-background-timer-throttling", ] if config.light_mode: flags.extend(BROWSER_DISABLE_OPTIONS) if config.text_mode: flags.extend([ "--blink-settings=imagesEnabled=false", "--disable-remote-fonts", "--disable-images", "--disable-javascript", "--disable-software-rasterizer", "--disable-dev-shm-usage", ]) # proxy support if config.proxy: flags.append(f"--proxy-server={config.proxy}") elif config.proxy_config: creds = "" if config.proxy_config.username and config.proxy_config.password: creds = f"{config.proxy_config.username}:{config.proxy_config.password}@" flags.append(f"--proxy-server={creds}{config.proxy_config.server}") # dedupe return list(dict.fromkeys(flags)) browser_type: str user_data_dir: str headless: bool browser_process: subprocess.Popen temp_dir: str debugging_port: int host: str def __init__( self, browser_type: str = "chromium", user_data_dir: Optional[str] = None, headless: bool = False, logger=None, host: str = "localhost", debugging_port: int = 9222, cdp_url: Optional[str] = None, browser_config: Optional[BrowserConfig] = None, ): """ Initialize the ManagedBrowser instance. Args: browser_type (str): The type of browser to launch. Supported values: "chromium", "firefox", "webkit". Default: "chromium". user_data_dir (str or None): Path to a user data directory for persistent sessions. If None, a temporary directory may be used. Default: None. headless (bool): Whether to run the browser in headless mode (no visible GUI). Default: True. logger (logging.Logger): Logger instance for logging messages. Default: None. host (str): Host for debugging the browser. Default: "localhost". debugging_port (int): Port for debugging the browser. Default: 9222. cdp_url (str or None): CDP URL to connect to the browser. Default: None. browser_config (BrowserConfig): Configuration object containing all browser settings. Default: None. """ self.browser_type = browser_config.browser_type self.user_data_dir = browser_config.user_data_dir self.headless = browser_config.headless self.browser_process = None self.temp_dir = None self.debugging_port = browser_config.debugging_port self.host = browser_config.host self.logger = logger self.shutting_down = False self.cdp_url = browser_config.cdp_url self.browser_config = browser_config async def start(self) -> str: """ Starts the browser process or returns CDP endpoint URL. If cdp_url is provided, returns it directly. If user_data_dir is not provided for local browser, creates a temporary directory. Returns: str: CDP endpoint URL """ # If CDP URL provided, just return it if self.cdp_url: return self.cdp_url # Create temp dir if needed if not self.user_data_dir: self.temp_dir = tempfile.mkdtemp(prefix="browser-profile-") self.user_data_dir = self.temp_dir # Get browser path and args based on OS and browser type # browser_path = self._get_browser_path() args = await self._get_browser_args() if self.browser_config.extra_args: args.extend(self.browser_config.extra_args) # ── make sure no old Chromium instance is owning the same port/profile ── try: if sys.platform == "win32": if psutil is None: raise RuntimeError("psutil not available, cannot clean old browser") for p in psutil.process_iter(["pid", "name", "cmdline"]): cl = " ".join(p.info.get("cmdline") or []) if ( f"--remote-debugging-port={self.debugging_port}" in cl and f"--user-data-dir={self.user_data_dir}" in cl ): p.kill() p.wait(timeout=5) else: # macOS / Linux # kill any process listening on the same debugging port pids = ( subprocess.check_output(shlex.split(f"lsof -t -i:{self.debugging_port}")) .decode() .strip() .splitlines() ) for pid in pids: try: os.kill(int(pid), signal.SIGTERM) except ProcessLookupError: pass # remove Chromium singleton locks, or new launch exits with # “Opening in existing browser session.” for f in ("SingletonLock", "SingletonSocket", "SingletonCookie"): fp = os.path.join(self.user_data_dir, f) if os.path.exists(fp): os.remove(fp) except Exception as _e: # non-fatal — we'll try to start anyway, but log what happened self.logger.warning(f"pre-launch cleanup failed: {_e}", tag="BROWSER") # Start browser process try: # Use DETACHED_PROCESS flag on Windows to fully detach the process # On Unix, we'll use preexec_fn=os.setpgrp to start the process in a new process group if sys.platform == "win32": self.browser_process = subprocess.Popen( args, stdout=subprocess.PIPE, stderr=subprocess.PIPE, creationflags=subprocess.DETACHED_PROCESS | subprocess.CREATE_NEW_PROCESS_GROUP ) else: self.browser_process = subprocess.Popen( args, stdout=subprocess.PIPE, stderr=subprocess.PIPE, preexec_fn=os.setpgrp # Start in a new process group ) # If verbose is True print args used to run the process if self.logger and self.browser_config.verbose: self.logger.debug( f"Starting browser with args: {' '.join(args)}", tag="BROWSER" ) # We'll monitor for a short time to make sure it starts properly, but won't keep monitoring await asyncio.sleep(0.5) # Give browser time to start await self._initial_startup_check() await asyncio.sleep(2) # Give browser time to start return f"http://{self.host}:{self.debugging_port}" except Exception as e: await self.cleanup() raise Exception(f"Failed to start browser: {e}") async def _initial_startup_check(self): """ Perform a quick check to make sure the browser started successfully. This only runs once at startup rather than continuously monitoring. """ if not self.browser_process: return # Check that process started without immediate termination await asyncio.sleep(0.5) if self.browser_process.poll() is not None: # Process already terminated stdout, stderr = b"", b"" try: stdout, stderr = self.browser_process.communicate(timeout=0.5) except subprocess.TimeoutExpired: pass self.logger.error( message="Browser process terminated during startup | Code: {code} | STDOUT: {stdout} | STDERR: {stderr}", tag="ERROR", params={ "code": self.browser_process.returncode, "stdout": stdout.decode() if stdout else "", "stderr": stderr.decode() if stderr else "", }, ) async def _monitor_browser_process(self): """ Monitor the browser process for unexpected termination. How it works: 1. Read stdout and stderr from the browser process. 2. If the process has terminated, log the error message and terminate the browser. 3. If the shutting_down flag is set, log the normal termination message. 4. If any other error occurs, log the error message. Note: This method should be called in a separate task to avoid blocking the main event loop. This is DEPRECATED and should not be used for builtin browsers that need to outlive the Python process. """ if self.browser_process: try: stdout, stderr = await asyncio.gather( asyncio.to_thread(self.browser_process.stdout.read), asyncio.to_thread(self.browser_process.stderr.read), ) # Check shutting_down flag BEFORE logging anything if self.browser_process.poll() is not None: if not self.shutting_down: self.logger.error( message="Browser process terminated unexpectedly | Code: {code} | STDOUT: {stdout} | STDERR: {stderr}", tag="ERROR", params={ "code": self.browser_process.returncode, "stdout": stdout.decode(), "stderr": stderr.decode(), }, ) await self.cleanup() else: self.logger.info( message="Browser process terminated normally | Code: {code}", tag="INFO", params={"code": self.browser_process.returncode}, ) except Exception as e: if not self.shutting_down: self.logger.error( message="Error monitoring browser process: {error}", tag="ERROR", params={"error": str(e)}, ) def _get_browser_path_WIP(self) -> str: """Returns the browser executable path based on OS and browser type""" if sys.platform == "darwin": # macOS paths = { "chromium": "/Applications/Google Chrome.app/Contents/MacOS/Google Chrome", "firefox": "/Applications/Firefox.app/Contents/MacOS/firefox", "webkit": "/Applications/Safari.app/Contents/MacOS/Safari", } elif sys.platform == "win32": # Windows paths = { "chromium": "C:\\Program Files\\Google\\Chrome\\Application\\chrome.exe", "firefox": "C:\\Program Files\\Mozilla Firefox\\firefox.exe", "webkit": None, # WebKit not supported on Windows } else: # Linux paths = { "chromium": "google-chrome", "firefox": "firefox", "webkit": None, # WebKit not supported on Linux } return paths.get(self.browser_type) async def _get_browser_path(self) -> str: browser_path = await get_chromium_path(self.browser_type) return browser_path async def _get_browser_args(self) -> List[str]: """Returns full CLI args for launching the browser""" base = [await self._get_browser_path()] if self.browser_type == "chromium": flags = [ f"--remote-debugging-port={self.debugging_port}", f"--user-data-dir={self.user_data_dir}", ] if self.headless: flags.append("--headless=new") # Add viewport flag if specified in config if self.browser_config.viewport_height and self.browser_config.viewport_width: flags.append(f"--window-size={self.browser_config.viewport_width},{self.browser_config.viewport_height}") # merge common launch flags flags.extend(self.build_browser_flags(self.browser_config)) elif self.browser_type == "firefox": flags = [ "--remote-debugging-port", str(self.debugging_port), "--profile", self.user_data_dir, ] if self.headless: flags.append("--headless") else: raise NotImplementedError(f"Browser type {self.browser_type} not supported") return base + flags async def cleanup(self): """Cleanup browser process and temporary directory""" # Set shutting_down flag BEFORE any termination actions self.shutting_down = True if self.browser_process: try: # For builtin browsers that should persist, we should check if it's a detached process # Only terminate if we have proper control over the process if not self.browser_process.poll(): # Process is still running self.browser_process.terminate() # Wait for process to end gracefully for _ in range(10): # 10 attempts, 100ms each if self.browser_process.poll() is not None: break await asyncio.sleep(0.1) # Force kill if still running if self.browser_process.poll() is None: if sys.platform == "win32": # On Windows we might need taskkill for detached processes try: subprocess.run(["taskkill", "/F", "/PID", str(self.browser_process.pid)]) except Exception: self.browser_process.kill() else: self.browser_process.kill() await asyncio.sleep(0.1) # Brief wait for kill to take effect except Exception as e: self.logger.error( message="Error terminating browser: {error}", tag="ERROR", params={"error": str(e)}, ) if self.temp_dir and os.path.exists(self.temp_dir): try: shutil.rmtree(self.temp_dir) except Exception as e: self.logger.error( message="Error removing temporary directory: {error}", tag="ERROR", params={"error": str(e)}, ) # These methods have been moved to BrowserProfiler class @staticmethod async def create_profile(browser_config=None, profile_name=None, logger=None): """ This method has been moved to the BrowserProfiler class. Creates a browser profile by launching a browser for interactive user setup and waits until the user closes it. The profile is stored in a directory that can be used later with BrowserConfig.user_data_dir. Please use BrowserProfiler.create_profile() instead. Example: ```python from crawl4ai.browser_profiler import BrowserProfiler profiler = BrowserProfiler() profile_path = await profiler.create_profile(profile_name="my-login-profile") ``` """ from .browser_profiler import BrowserProfiler # Create a BrowserProfiler instance and delegate to it profiler = BrowserProfiler(logger=logger) return await profiler.create_profile(profile_name=profile_name, browser_config=browser_config) @staticmethod def list_profiles(): """ This method has been moved to the BrowserProfiler class. Lists all available browser profiles in the Crawl4AI profiles directory. Please use BrowserProfiler.list_profiles() instead. Example: ```python from crawl4ai.browser_profiler import BrowserProfiler profiler = BrowserProfiler() profiles = profiler.list_profiles() ``` """ from .browser_profiler import BrowserProfiler # Create a BrowserProfiler instance and delegate to it profiler = BrowserProfiler() return profiler.list_profiles() @staticmethod def delete_profile(profile_name_or_path): """ This method has been moved to the BrowserProfiler class. Delete a browser profile by name or path. Please use BrowserProfiler.delete_profile() instead. Example: ```python from crawl4ai.browser_profiler import BrowserProfiler profiler = BrowserProfiler() success = profiler.delete_profile("my-profile") ``` """ from .browser_profiler import BrowserProfiler # Create a BrowserProfiler instance and delegate to it profiler = BrowserProfiler() return profiler.delete_profile(profile_name_or_path) async def clone_runtime_state( src: BrowserContext, dst: BrowserContext, crawlerRunConfig: CrawlerRunConfig | None = None, browserConfig: BrowserConfig | None = None, ) -> None: """ Bring everything that *can* be changed at runtime from `src` → `dst`. 1. Cookies 2. localStorage (and sessionStorage, same API) 3. Extra headers, permissions, geolocation if supplied in configs """ # ── 1. cookies ──────────────────────────────────────────────────────────── cookies = await src.cookies() if cookies: await dst.add_cookies(cookies) # ── 2. localStorage / sessionStorage ────────────────────────────────────── state = await src.storage_state() for origin in state.get("origins", []): url = origin["origin"] kvs = origin.get("localStorage", []) if not kvs: continue page = dst.pages[0] if dst.pages else await dst.new_page() await page.goto(url, wait_until="domcontentloaded") for k, v in kvs: await page.evaluate("(k,v)=>localStorage.setItem(k,v)", k, v) # ── 3. runtime-mutable extras from configs ──────────────────────────────── # headers if browserConfig and browserConfig.headers: await dst.set_extra_http_headers(browserConfig.headers) # geolocation if crawlerRunConfig and crawlerRunConfig.geolocation: await dst.grant_permissions(["geolocation"]) await dst.set_geolocation( { "latitude": crawlerRunConfig.geolocation.latitude, "longitude": crawlerRunConfig.geolocation.longitude, "accuracy": crawlerRunConfig.geolocation.accuracy, } ) return dst class BrowserManager: """ Manages the browser instance and context. Attributes: config (BrowserConfig): Configuration object containing all browser settings logger: Logger instance for recording events and errors browser (Browser): The browser instance default_context (BrowserContext): The default browser context managed_browser (ManagedBrowser): The managed browser instance playwright (Playwright): The Playwright instance sessions (dict): Dictionary to store session information session_ttl (int): Session timeout in seconds """ _playwright_instance = None @classmethod async def get_playwright(cls, use_undetected: bool = False): if use_undetected: from patchright.async_api import async_playwright else: from playwright.async_api import async_playwright cls._playwright_instance = await async_playwright().start() return cls._playwright_instance def __init__(self, browser_config: BrowserConfig, logger=None, use_undetected: bool = False): """ Initialize the BrowserManager with a browser configuration. Args: browser_config (BrowserConfig): Configuration object containing all browser settings logger: Logger instance for recording events and errors use_undetected (bool): Whether to use undetected browser (Patchright) """ self.config: BrowserConfig = browser_config self.logger = logger self.use_undetected = use_undetected # Browser state self.browser = None self.default_context = None self.managed_browser = None self.playwright = None # Session management self.sessions = {} self.session_ttl = 1800 # 30 minutes # Keep track of contexts by a "config signature," so each unique config reuses a single context self.contexts_by_config = {} self._contexts_lock = asyncio.Lock() # Serialize context.new_page() across concurrent tasks to avoid races # when using a shared persistent context (context.pages may be empty # for all racers). Prevents 'Target page/context closed' errors. self._page_lock = asyncio.Lock() # Stealth adapter for stealth mode self._stealth_adapter = None if self.config.enable_stealth and not self.use_undetected: from .browser_adapter import StealthAdapter self._stealth_adapter = StealthAdapter() # Initialize ManagedBrowser if needed if self.config.use_managed_browser: self.managed_browser = ManagedBrowser( browser_type=self.config.browser_type, user_data_dir=self.config.user_data_dir, headless=self.config.headless, logger=self.logger, debugging_port=self.config.debugging_port, cdp_url=self.config.cdp_url, browser_config=self.config, ) async def start(self): """ Start the browser instance and set up the default context. How it works: 1. Check if Playwright is already initialized. 2. If not, initialize Playwright. 3. If managed browser is used, start it and connect to the CDP endpoint. 4. If managed browser is not used, launch the browser and set up the default context. Note: This method should be called in a separate task to avoid blocking the main event loop. """ if self.playwright is not None: await self.close() if self.use_undetected: from patchright.async_api import async_playwright else: from playwright.async_api import async_playwright # Initialize playwright self.playwright = await async_playwright().start() if self.config.cdp_url or self.config.use_managed_browser: self.config.use_managed_browser = True cdp_url = await self.managed_browser.start() if not self.config.cdp_url else self.config.cdp_url # Add CDP endpoint verification before connecting if not await self._verify_cdp_ready(cdp_url): raise Exception(f"CDP endpoint at {cdp_url} is not ready after startup") self.browser = await self.playwright.chromium.connect_over_cdp(cdp_url) contexts = self.browser.contexts if contexts: self.default_context = contexts[0] else: self.default_context = await self.create_browser_context() await self.setup_context(self.default_context) else: browser_args = self._build_browser_args() # Launch appropriate browser type if self.config.browser_type == "firefox": self.browser = await self.playwright.firefox.launch(**browser_args) elif self.config.browser_type == "webkit": self.browser = await self.playwright.webkit.launch(**browser_args) else: self.browser = await self.playwright.chromium.launch(**browser_args) self.default_context = self.browser async def _verify_cdp_ready(self, cdp_url: str) -> bool: """Verify CDP endpoint is ready with exponential backoff""" import aiohttp self.logger.debug(f"Starting CDP verification for {cdp_url}", tag="BROWSER") for attempt in range(5): try: async with aiohttp.ClientSession() as session: async with session.get(f"{cdp_url}/json/version", timeout=aiohttp.ClientTimeout(total=2)) as response: if response.status == 200: self.logger.debug(f"CDP endpoint ready after {attempt + 1} attempts", tag="BROWSER") return True except Exception as e: self.logger.debug(f"CDP check attempt {attempt + 1} failed: {e}", tag="BROWSER") delay = 0.5 * (1.4 ** attempt) self.logger.debug(f"Waiting {delay:.2f}s before next CDP check...", tag="BROWSER") await asyncio.sleep(delay) self.logger.debug(f"CDP verification failed after 5 attempts", tag="BROWSER") return False def _build_browser_args(self) -> dict: """Build browser launch arguments from config.""" args = [ "--disable-gpu", "--disable-gpu-compositing", "--disable-software-rasterizer", "--no-sandbox", "--disable-dev-shm-usage", "--no-first-run", "--no-default-browser-check", "--disable-infobars", "--window-position=0,0", "--ignore-certificate-errors", "--ignore-certificate-errors-spki-list", "--disable-blink-features=AutomationControlled", "--window-position=400,0", "--disable-renderer-backgrounding", "--disable-ipc-flooding-protection", "--force-color-profile=srgb", "--mute-audio", "--disable-background-timer-throttling", # "--single-process", f"--window-size={self.config.viewport_width},{self.config.viewport_height}", ] if self.config.light_mode: args.extend(BROWSER_DISABLE_OPTIONS) if self.config.text_mode: args.extend( [ "--blink-settings=imagesEnabled=false", "--disable-remote-fonts", "--disable-images", "--disable-javascript", "--disable-software-rasterizer", "--disable-dev-shm-usage", ] ) if self.config.extra_args: args.extend(self.config.extra_args) # Deduplicate args args = list(dict.fromkeys(args)) browser_args = {"headless": self.config.headless, "args": args} if self.config.chrome_channel: browser_args["channel"] = self.config.chrome_channel if self.config.accept_downloads: browser_args["downloads_path"] = self.config.downloads_path or os.path.join( os.getcwd(), "downloads" ) os.makedirs(browser_args["downloads_path"], exist_ok=True) if self.config.proxy: warnings.warn( "BrowserConfig.proxy is deprecated and ignored. Use proxy_config instead.", DeprecationWarning, ) if self.config.proxy_config: from playwright.async_api import ProxySettings proxy_settings = ProxySettings( server=self.config.proxy_config.server, username=self.config.proxy_config.username, password=self.config.proxy_config.password, ) browser_args["proxy"] = proxy_settings return browser_args async def setup_context( self, context: BrowserContext, crawlerRunConfig: CrawlerRunConfig = None, is_default=False, ): """ Set up a browser context with the configured options. How it works: 1. Set extra HTTP headers if provided. 2. Add cookies if provided.
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_webcrawler.py
crawl4ai/async_webcrawler.py
from .__version__ import __version__ as crawl4ai_version import os import sys import time from pathlib import Path from typing import Optional, List import json import asyncio # from contextlib import nullcontext, asynccontextmanager from contextlib import asynccontextmanager from .models import ( CrawlResult, MarkdownGenerationResult, DispatchResult, ScrapingResult, CrawlResultContainer, RunManyReturn ) from .async_database import async_db_manager from .chunking_strategy import * # noqa: F403 from .chunking_strategy import IdentityChunking from .content_filter_strategy import * # noqa: F403 from .extraction_strategy import * # noqa: F403 from .extraction_strategy import NoExtractionStrategy from .async_crawler_strategy import ( AsyncCrawlerStrategy, AsyncPlaywrightCrawlerStrategy, AsyncCrawlResponse, ) from .cache_context import CacheMode, CacheContext from .markdown_generation_strategy import ( DefaultMarkdownGenerator, MarkdownGenerationStrategy, ) from .deep_crawling import DeepCrawlDecorator from .async_logger import AsyncLogger, AsyncLoggerBase from .async_configs import BrowserConfig, CrawlerRunConfig, ProxyConfig, SeedingConfig from .async_dispatcher import * # noqa: F403 from .async_dispatcher import BaseDispatcher, MemoryAdaptiveDispatcher, RateLimiter from .async_url_seeder import AsyncUrlSeeder from .utils import ( sanitize_input_encode, InvalidCSSSelectorError, fast_format_html, get_error_context, RobotsParser, preprocess_html_for_schema, ) class AsyncWebCrawler: """ Asynchronous web crawler with flexible caching capabilities. There are two ways to use the crawler: 1. Using context manager (recommended for simple cases): ```python async with AsyncWebCrawler() as crawler: result = await crawler.arun(url="https://example.com") ``` 2. Using explicit lifecycle management (recommended for long-running applications): ```python crawler = AsyncWebCrawler() await crawler.start() # Use the crawler multiple times result1 = await crawler.arun(url="https://example.com") result2 = await crawler.arun(url="https://another.com") await crawler.close() ``` Attributes: browser_config (BrowserConfig): Configuration object for browser settings. crawler_strategy (AsyncCrawlerStrategy): Strategy for crawling web pages. logger (AsyncLogger): Logger instance for recording events and errors. crawl4ai_folder (str): Directory for storing cache. base_directory (str): Base directory for storing cache. ready (bool): Whether the crawler is ready for use. Methods: start(): Start the crawler explicitly without using context manager. close(): Close the crawler explicitly without using context manager. arun(): Run the crawler for a single source: URL (web, local file, or raw HTML). awarmup(): Perform warmup sequence. arun_many(): Run the crawler for multiple sources. aprocess_html(): Process HTML content. Typical Usage: async with AsyncWebCrawler() as crawler: result = await crawler.arun(url="https://example.com") print(result.markdown) Using configuration: browser_config = BrowserConfig(browser_type="chromium", headless=True) async with AsyncWebCrawler(config=browser_config) as crawler: crawler_config = CrawlerRunConfig( cache_mode=CacheMode.BYPASS ) result = await crawler.arun(url="https://example.com", config=crawler_config) print(result.markdown) """ _domain_last_hit = {} def __init__( self, crawler_strategy: AsyncCrawlerStrategy = None, config: BrowserConfig = None, base_directory: str = str( os.getenv("CRAWL4_AI_BASE_DIRECTORY", Path.home())), thread_safe: bool = False, logger: AsyncLoggerBase = None, **kwargs, ): """ Initialize the AsyncWebCrawler. Args: crawler_strategy: Strategy for crawling web pages. Default AsyncPlaywrightCrawlerStrategy config: Configuration object for browser settings. Default BrowserConfig() base_directory: Base directory for storing cache thread_safe: Whether to use thread-safe operations **kwargs: Additional arguments for backwards compatibility """ # Handle browser configuration browser_config = config or BrowserConfig() self.browser_config = browser_config # Initialize logger first since other components may need it self.logger = logger or AsyncLogger( log_file=os.path.join(base_directory, ".crawl4ai", "crawler.log"), verbose=self.browser_config.verbose, tag_width=10, ) # Initialize crawler strategy params = {k: v for k, v in kwargs.items() if k in [ "browser_config", "logger"]} self.crawler_strategy = crawler_strategy or AsyncPlaywrightCrawlerStrategy( browser_config=browser_config, logger=self.logger, **params, # Pass remaining kwargs for backwards compatibility ) # Thread safety setup self._lock = asyncio.Lock() if thread_safe else None # Initialize directories self.crawl4ai_folder = os.path.join(base_directory, ".crawl4ai") os.makedirs(self.crawl4ai_folder, exist_ok=True) os.makedirs(f"{self.crawl4ai_folder}/cache", exist_ok=True) # Initialize robots parser self.robots_parser = RobotsParser() self.ready = False # Decorate arun method with deep crawling capabilities self._deep_handler = DeepCrawlDecorator(self) self.arun = self._deep_handler(self.arun) self.url_seeder: Optional[AsyncUrlSeeder] = None async def start(self): """ Start the crawler explicitly without using context manager. This is equivalent to using 'async with' but gives more control over the lifecycle. Returns: AsyncWebCrawler: The initialized crawler instance """ await self.crawler_strategy.__aenter__() self.logger.info(f"Crawl4AI {crawl4ai_version}", tag="INIT") self.ready = True return self async def close(self): """ Close the crawler explicitly without using context manager. This should be called when you're done with the crawler if you used start(). This method will: 1. Clean up browser resources 2. Close any open pages and contexts """ await self.crawler_strategy.__aexit__(None, None, None) async def __aenter__(self): return await self.start() async def __aexit__(self, exc_type, exc_val, exc_tb): await self.close() @asynccontextmanager async def nullcontext(self): """异步空上下文管理器""" yield async def arun( self, url: str, config: CrawlerRunConfig = None, **kwargs, ) -> RunManyReturn: """ Runs the crawler for a single source: URL (web, local file, or raw HTML). Migration Guide: Old way (deprecated): result = await crawler.arun( url="https://example.com", word_count_threshold=200, screenshot=True, ... ) New way (recommended): config = CrawlerRunConfig( word_count_threshold=200, screenshot=True, ... ) result = await crawler.arun(url="https://example.com", crawler_config=config) Args: url: The URL to crawl (http://, https://, file://, or raw:) crawler_config: Configuration object controlling crawl behavior [other parameters maintained for backwards compatibility] Returns: CrawlResult: The result of crawling and processing """ # Auto-start if not ready if not self.ready: await self.start() config = config or CrawlerRunConfig() if not isinstance(url, str) or not url: raise ValueError( "Invalid URL, make sure the URL is a non-empty string") async with self._lock or self.nullcontext(): try: self.logger.verbose = config.verbose # Default to ENABLED if no cache mode specified if config.cache_mode is None: config.cache_mode = CacheMode.ENABLED # Create cache context cache_context = CacheContext(url, config.cache_mode, False) # Initialize processing variables async_response: AsyncCrawlResponse = None cached_result: CrawlResult = None screenshot_data = None pdf_data = None extracted_content = None start_time = time.perf_counter() # Try to get cached result if appropriate if cache_context.should_read(): cached_result = await async_db_manager.aget_cached_url(url) if cached_result: html = sanitize_input_encode(cached_result.html) extracted_content = sanitize_input_encode( cached_result.extracted_content or "" ) extracted_content = ( None if not extracted_content or extracted_content == "[]" else extracted_content ) # If screenshot is requested but its not in cache, then set cache_result to None screenshot_data = cached_result.screenshot pdf_data = cached_result.pdf # if config.screenshot and not screenshot or config.pdf and not pdf: if config.screenshot and not screenshot_data: cached_result = None if config.pdf and not pdf_data: cached_result = None self.logger.url_status( url=cache_context.display_url, success=bool(html), timing=time.perf_counter() - start_time, tag="FETCH", ) # Update proxy configuration from rotation strategy if available if config and config.proxy_rotation_strategy: next_proxy: ProxyConfig = await config.proxy_rotation_strategy.get_next_proxy() if next_proxy: self.logger.info( message="Switch proxy: {proxy}", tag="PROXY", params={"proxy": next_proxy.server} ) config.proxy_config = next_proxy # config = config.clone(proxy_config=next_proxy) # Fetch fresh content if needed if not cached_result or not html: t1 = time.perf_counter() if config.user_agent: self.crawler_strategy.update_user_agent( config.user_agent) # Check robots.txt if enabled if config and config.check_robots_txt: if not await self.robots_parser.can_fetch( url, self.browser_config.user_agent ): return CrawlResult( url=url, html="", success=False, status_code=403, error_message="Access denied by robots.txt", response_headers={ "X-Robots-Status": "Blocked by robots.txt" }, ) ############################## # Call CrawlerStrategy.crawl # ############################## async_response = await self.crawler_strategy.crawl( url, config=config, # Pass the entire config object ) html = sanitize_input_encode(async_response.html) screenshot_data = async_response.screenshot pdf_data = async_response.pdf_data js_execution_result = async_response.js_execution_result t2 = time.perf_counter() self.logger.url_status( url=cache_context.display_url, success=bool(html), timing=t2 - t1, tag="FETCH", ) ############################################################### # Process the HTML content, Call CrawlerStrategy.process_html # ############################################################### from urllib.parse import urlparse crawl_result: CrawlResult = await self.aprocess_html( url=url, html=html, extracted_content=extracted_content, config=config, # Pass the config object instead of individual parameters screenshot_data=screenshot_data, pdf_data=pdf_data, verbose=config.verbose, is_raw_html=True if url.startswith("raw:") else False, redirected_url=async_response.redirected_url, original_scheme=urlparse(url).scheme, **kwargs, ) crawl_result.status_code = async_response.status_code crawl_result.redirected_url = async_response.redirected_url or url crawl_result.response_headers = async_response.response_headers crawl_result.downloaded_files = async_response.downloaded_files crawl_result.js_execution_result = js_execution_result crawl_result.mhtml = async_response.mhtml_data crawl_result.ssl_certificate = async_response.ssl_certificate # Add captured network and console data if available crawl_result.network_requests = async_response.network_requests crawl_result.console_messages = async_response.console_messages crawl_result.success = bool(html) crawl_result.session_id = getattr( config, "session_id", None) self.logger.url_status( url=cache_context.display_url, success=crawl_result.success, timing=time.perf_counter() - start_time, tag="COMPLETE", ) # Update cache if appropriate if cache_context.should_write() and not bool(cached_result): await async_db_manager.acache_url(crawl_result) return CrawlResultContainer(crawl_result) else: self.logger.url_status( url=cache_context.display_url, success=True, timing=time.perf_counter() - start_time, tag="COMPLETE" ) cached_result.success = bool(html) cached_result.session_id = getattr( config, "session_id", None) cached_result.redirected_url = cached_result.redirected_url or url return CrawlResultContainer(cached_result) except Exception as e: error_context = get_error_context(sys.exc_info()) error_message = ( f"Unexpected error in _crawl_web at line {error_context['line_no']} " f"in {error_context['function']} ({error_context['filename']}):\n" f"Error: {str(e)}\n\n" f"Code context:\n{error_context['code_context']}" ) self.logger.error_status( url=url, error=error_message, tag="ERROR", ) return CrawlResultContainer( CrawlResult( url=url, html="", success=False, error_message=error_message ) ) async def aprocess_html( self, url: str, html: str, extracted_content: str, config: CrawlerRunConfig, screenshot_data: str, pdf_data: str, verbose: bool, **kwargs, ) -> CrawlResult: """ Process HTML content using the provided configuration. Args: url: The URL being processed html: Raw HTML content extracted_content: Previously extracted content (if any) config: Configuration object controlling processing behavior screenshot_data: Screenshot data (if any) pdf_data: PDF data (if any) verbose: Whether to enable verbose logging **kwargs: Additional parameters for backwards compatibility Returns: CrawlResult: Processed result containing extracted and formatted content """ cleaned_html = "" try: _url = url if not kwargs.get("is_raw_html", False) else "Raw HTML" t1 = time.perf_counter() # Get scraping strategy and ensure it has a logger scraping_strategy = config.scraping_strategy if not scraping_strategy.logger: scraping_strategy.logger = self.logger # Process HTML content params = config.__dict__.copy() params.pop("url", None) # add keys from kwargs to params that doesn't exist in params params.update({k: v for k, v in kwargs.items() if k not in params.keys()}) ################################ # Scraping Strategy Execution # ################################ result: ScrapingResult = scraping_strategy.scrap( url, html, **params) if result is None: raise ValueError( f"Process HTML, Failed to extract content from the website: {url}" ) except InvalidCSSSelectorError as e: raise ValueError(str(e)) except Exception as e: raise ValueError( f"Process HTML, Failed to extract content from the website: {url}, error: {str(e)}" ) # Extract results - handle both dict and ScrapingResult if isinstance(result, dict): cleaned_html = sanitize_input_encode( result.get("cleaned_html", "")) media = result.get("media", {}) tables = media.pop("tables", []) if isinstance(media, dict) else [] links = result.get("links", {}) metadata = result.get("metadata", {}) else: cleaned_html = sanitize_input_encode(result.cleaned_html) # media = result.media.model_dump() # tables = media.pop("tables", []) # links = result.links.model_dump() media = result.media.model_dump() if hasattr(result.media, 'model_dump') else result.media tables = media.pop("tables", []) if isinstance(media, dict) else [] links = result.links.model_dump() if hasattr(result.links, 'model_dump') else result.links metadata = result.metadata fit_html = preprocess_html_for_schema(html_content=html, text_threshold= 500, max_size= 300_000) ################################ # Generate Markdown # ################################ markdown_generator: Optional[MarkdownGenerationStrategy] = ( config.markdown_generator or DefaultMarkdownGenerator() ) # --- SELECT HTML SOURCE BASED ON CONTENT_SOURCE --- # Get the desired source from the generator config, default to 'cleaned_html' selected_html_source = getattr(markdown_generator, 'content_source', 'cleaned_html') # Define the source selection logic using dict dispatch html_source_selector = { "raw_html": lambda: html, # The original raw HTML "cleaned_html": lambda: cleaned_html, # The HTML after scraping strategy "fit_html": lambda: fit_html, # The HTML after preprocessing for schema } markdown_input_html = cleaned_html # Default to cleaned_html try: # Get the appropriate lambda function, default to returning cleaned_html if key not found source_lambda = html_source_selector.get(selected_html_source, lambda: cleaned_html) # Execute the lambda to get the selected HTML markdown_input_html = source_lambda() # Log which source is being used (optional, but helpful for debugging) # if self.logger and verbose: # actual_source_used = selected_html_source if selected_html_source in html_source_selector else 'cleaned_html (default)' # self.logger.debug(f"Using '{actual_source_used}' as source for Markdown generation for {url}", tag="MARKDOWN_SRC") except Exception as e: # Handle potential errors, especially from preprocess_html_for_schema if self.logger: self.logger.warning( f"Error getting/processing '{selected_html_source}' for markdown source: {e}. Falling back to cleaned_html.", tag="MARKDOWN_SRC" ) # Ensure markdown_input_html is still the default cleaned_html in case of error markdown_input_html = cleaned_html # --- END: HTML SOURCE SELECTION --- # Uncomment if by default we want to use PruningContentFilter # if not config.content_filter and not markdown_generator.content_filter: # markdown_generator.content_filter = PruningContentFilter() markdown_result: MarkdownGenerationResult = ( markdown_generator.generate_markdown( input_html=markdown_input_html, base_url=params.get("redirected_url", url) # html2text_options=kwargs.get('html2text', {}) ) ) # Log processing completion self.logger.url_status( url=_url, success=True, timing=int((time.perf_counter() - t1) * 1000) / 1000, tag="SCRAPE" ) # self.logger.info( # message="{url:.50}... | Time: {timing}s", # tag="SCRAPE", # params={"url": _url, "timing": int((time.perf_counter() - t1) * 1000) / 1000}, # ) ################################ # Structured Content Extraction # ################################ if ( not bool(extracted_content) and config.extraction_strategy and not isinstance(config.extraction_strategy, NoExtractionStrategy) ): t1 = time.perf_counter() # Choose content based on input_format content_format = config.extraction_strategy.input_format if content_format == "fit_markdown" and not markdown_result.fit_markdown: self.logger.url_status( url=_url, success=bool(html), timing=time.perf_counter() - t1, tag="EXTRACT", ) content_format = "markdown" content = { "markdown": markdown_result.raw_markdown, "html": html, "fit_html": fit_html, "cleaned_html": cleaned_html, "fit_markdown": markdown_result.fit_markdown, }.get(content_format, markdown_result.raw_markdown) # Use IdentityChunking for HTML input, otherwise use provided chunking strategy chunking = ( IdentityChunking() if content_format in ["html", "cleaned_html", "fit_html"] else config.chunking_strategy ) sections = chunking.chunk(content) # extracted_content = config.extraction_strategy.run(_url, sections) # Use async version if available for better parallelism if hasattr(config.extraction_strategy, 'arun'): extracted_content = await config.extraction_strategy.arun(_url, sections) else: # Fallback to sync version run in thread pool to avoid blocking extracted_content = await asyncio.to_thread( config.extraction_strategy.run, url, sections ) extracted_content = json.dumps( extracted_content, indent=4, default=str, ensure_ascii=False ) # Log extraction completion self.logger.url_status( url=_url, success=bool(html), timing=time.perf_counter() - t1, tag="EXTRACT", ) # Apply HTML formatting if requested if config.prettiify: cleaned_html = fast_format_html(cleaned_html) # Return complete crawl result return CrawlResult( url=url, html=html, fit_html=fit_html, cleaned_html=cleaned_html, markdown=markdown_result, media=media, tables=tables, # NEW links=links, metadata=metadata, screenshot=screenshot_data, pdf=pdf_data, extracted_content=extracted_content, success=True, error_message="", ) async def arun_many( self, urls: List[str], config: Optional[Union[CrawlerRunConfig, List[CrawlerRunConfig]]] = None, dispatcher: Optional[BaseDispatcher] = None, # Legacy parameters maintained for backwards compatibility # word_count_threshold=MIN_WORD_THRESHOLD, # extraction_strategy: ExtractionStrategy = None, # chunking_strategy: ChunkingStrategy = RegexChunking(), # content_filter: RelevantContentFilter = None, # cache_mode: Optional[CacheMode] = None, # bypass_cache: bool = False, # css_selector: str = None, # screenshot: bool = False, # pdf: bool = False, # user_agent: str = None, # verbose=True, **kwargs, ) -> RunManyReturn: """ Runs the crawler for multiple URLs concurrently using a configurable dispatcher strategy. Args: urls: List of URLs to crawl config: Configuration object(s) controlling crawl behavior. Can be: - Single CrawlerRunConfig: Used for all URLs - List[CrawlerRunConfig]: Configs with url_matcher for URL-specific settings dispatcher: The dispatcher strategy instance to use. Defaults to MemoryAdaptiveDispatcher [other parameters maintained for backwards compatibility] Returns: Union[List[CrawlResult], AsyncGenerator[CrawlResult, None]]: Either a list of all results or an async generator yielding results Examples: # Batch processing (default) results = await crawler.arun_many( urls=["https://example1.com", "https://example2.com"], config=CrawlerRunConfig(cache_mode=CacheMode.BYPASS) ) for result in results: print(f"Processed {result.url}: {len(result.markdown)} chars") # Streaming results async for result in await crawler.arun_many( urls=["https://example1.com", "https://example2.com"], config=CrawlerRunConfig(cache_mode=CacheMode.BYPASS, stream=True), ): print(f"Processed {result.url}: {len(result.markdown)} chars") """ config = config or CrawlerRunConfig() # if config is None: # config = CrawlerRunConfig( # word_count_threshold=word_count_threshold, # extraction_strategy=extraction_strategy, # chunking_strategy=chunking_strategy, # content_filter=content_filter, # cache_mode=cache_mode, # bypass_cache=bypass_cache, # css_selector=css_selector, # screenshot=screenshot, # pdf=pdf, # verbose=verbose, # **kwargs, # ) if dispatcher is None: dispatcher = MemoryAdaptiveDispatcher( rate_limiter=RateLimiter( base_delay=(1.0, 3.0), max_delay=60.0, max_retries=3 ), ) def transform_result(task_result): return ( setattr( task_result.result, "dispatch_result", DispatchResult( task_id=task_result.task_id, memory_usage=task_result.memory_usage, peak_memory=task_result.peak_memory, start_time=task_result.start_time, end_time=task_result.end_time, error_message=task_result.error_message, ), ) or task_result.result ) # Handle stream setting - use first config's stream setting if config is a list if isinstance(config, list): stream = config[0].stream if config else False else: stream = config.stream if stream: async def result_transformer(): async for task_result in dispatcher.run_urls_stream( crawler=self, urls=urls, config=config ): yield transform_result(task_result) return result_transformer() else: _results = await dispatcher.run_urls(crawler=self, urls=urls, config=config) return [transform_result(res) for res in _results] async def aseed_urls( self, domain_or_domains: Union[str, List[str]], config: Optional[SeedingConfig] = None, **kwargs ) -> Union[List[str], Dict[str, List[Union[str, Dict[str, Any]]]]]: """ Discovers, filters, and optionally validates URLs for a given domain(s) using sitemaps and Common Crawl archives. Args: domain_or_domains: A single domain string (e.g., "iana.org") or a list of domains. config: A SeedingConfig object to control the seeding process. Parameters passed directly via kwargs will override those in 'config'. **kwargs: Additional parameters (e.g., `source`, `live_check`, `extract_head`, `pattern`, `concurrency`, `hits_per_sec`, `force_refresh`, `verbose`) that will be used to construct or update the SeedingConfig. Returns: If `extract_head` is False: - For a single domain: `List[str]` of discovered URLs. - For multiple domains: `Dict[str, List[str]]` mapping each domain to its URLs. If `extract_head` is True: - For a single domain: `List[Dict[str, Any]]` where each dict contains 'url' and 'head_data' (parsed <head> metadata). - For multiple domains: `Dict[str, List[Dict[str, Any]]]` mapping each domain to a list of URL data dictionaries. Raises:
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/migrations.py
crawl4ai/migrations.py
import os import asyncio from pathlib import Path import aiosqlite from typing import Optional import xxhash import aiofiles import shutil from datetime import datetime from .async_logger import AsyncLogger, LogLevel # Initialize logger logger = AsyncLogger(log_level=LogLevel.DEBUG, verbose=True) # logging.basicConfig(level=logging.INFO) # logger = logging.getLogger(__name__) class DatabaseMigration: def __init__(self, db_path: str): self.db_path = db_path self.content_paths = self._ensure_content_dirs(os.path.dirname(db_path)) def _ensure_content_dirs(self, base_path: str) -> dict: dirs = { "html": "html_content", "cleaned": "cleaned_html", "markdown": "markdown_content", "extracted": "extracted_content", "screenshots": "screenshots", } content_paths = {} for key, dirname in dirs.items(): path = os.path.join(base_path, dirname) os.makedirs(path, exist_ok=True) content_paths[key] = path return content_paths def _generate_content_hash(self, content: str) -> str: x = xxhash.xxh64() x.update(content.encode()) content_hash = x.hexdigest() return content_hash # return hashlib.sha256(content.encode()).hexdigest() async def _store_content(self, content: str, content_type: str) -> str: if not content: return "" content_hash = self._generate_content_hash(content) file_path = os.path.join(self.content_paths[content_type], content_hash) if not os.path.exists(file_path): async with aiofiles.open(file_path, "w", encoding="utf-8") as f: await f.write(content) return content_hash async def migrate_database(self): """Migrate existing database to file-based storage""" # logger.info("Starting database migration...") logger.info("Starting database migration...", tag="INIT") try: async with aiosqlite.connect(self.db_path) as db: # Get all rows async with db.execute( """SELECT url, html, cleaned_html, markdown, extracted_content, screenshot FROM crawled_data""" ) as cursor: rows = await cursor.fetchall() migrated_count = 0 for row in rows: ( url, html, cleaned_html, markdown, extracted_content, screenshot, ) = row # Store content in files and get hashes html_hash = await self._store_content(html, "html") cleaned_hash = await self._store_content(cleaned_html, "cleaned") markdown_hash = await self._store_content(markdown, "markdown") extracted_hash = await self._store_content( extracted_content, "extracted" ) screenshot_hash = await self._store_content( screenshot, "screenshots" ) # Update database with hashes await db.execute( """ UPDATE crawled_data SET html = ?, cleaned_html = ?, markdown = ?, extracted_content = ?, screenshot = ? WHERE url = ? """, ( html_hash, cleaned_hash, markdown_hash, extracted_hash, screenshot_hash, url, ), ) migrated_count += 1 if migrated_count % 100 == 0: logger.info(f"Migrated {migrated_count} records...", tag="INIT") await db.commit() logger.success( f"Migration completed. {migrated_count} records processed.", tag="COMPLETE", ) except Exception as e: # logger.error(f"Migration failed: {e}") logger.error( message="Migration failed: {error}", tag="ERROR", params={"error": str(e)}, ) raise e async def backup_database(db_path: str) -> str: """Create backup of existing database""" if not os.path.exists(db_path): logger.info("No existing database found. Skipping backup.", tag="INIT") return None # Create backup with timestamp timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") backup_path = f"{db_path}.backup_{timestamp}" try: # Wait for any potential write operations to finish await asyncio.sleep(1) # Create backup shutil.copy2(db_path, backup_path) logger.info(f"Database backup created at: {backup_path}", tag="COMPLETE") return backup_path except Exception as e: # logger.error(f"Backup failed: {e}") logger.error( message="Migration failed: {error}", tag="ERROR", params={"error": str(e)} ) raise e async def run_migration(db_path: Optional[str] = None): """Run database migration""" if db_path is None: db_path = os.path.join(Path.home(), ".crawl4ai", "crawl4ai.db") if not os.path.exists(db_path): logger.info("No existing database found. Skipping migration.", tag="INIT") return # Create backup first backup_path = await backup_database(db_path) if not backup_path: return migration = DatabaseMigration(db_path) await migration.migrate_database() def main(): """CLI entry point for migration""" import argparse parser = argparse.ArgumentParser( description="Migrate Crawl4AI database to file-based storage" ) parser.add_argument("--db-path", help="Custom database path") args = parser.parse_args() asyncio.run(run_migration(args.db_path)) if __name__ == "__main__": main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/model_loader.py
crawl4ai/model_loader.py
from functools import lru_cache from pathlib import Path import subprocess, os import shutil from .model_loader import * import argparse from crawl4ai.config import MODEL_REPO_BRANCH __location__ = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) @lru_cache() def get_available_memory(device): import torch if device.type == "cuda": return torch.cuda.get_device_properties(device).total_memory elif device.type == "mps": return 48 * 1024**3 # Assuming 8GB for MPS, as a conservative estimate else: return 0 @lru_cache() def calculate_batch_size(device): available_memory = get_available_memory(device) if device.type == "cpu": return 16 elif device.type in ["cuda", "mps"]: # Adjust these thresholds based on your model size and available memory if available_memory >= 31 * 1024**3: # > 32GB return 256 elif available_memory >= 15 * 1024**3: # > 16GB to 32GB return 128 elif available_memory >= 8 * 1024**3: # 8GB to 16GB return 64 else: return 32 else: return 16 # Default batch size @lru_cache() def get_device(): import torch if torch.cuda.is_available(): device = torch.device("cuda") elif torch.backends.mps.is_available(): device = torch.device("mps") else: device = torch.device("cpu") return device def set_model_device(model): device = get_device() model.to(device) return model, device @lru_cache() def get_home_folder(): home_folder = os.path.join( os.getenv("CRAWL4_AI_BASE_DIRECTORY", Path.home()), ".crawl4ai" ) os.makedirs(home_folder, exist_ok=True) os.makedirs(f"{home_folder}/cache", exist_ok=True) os.makedirs(f"{home_folder}/models", exist_ok=True) return home_folder @lru_cache() def load_bert_base_uncased(): from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained("bert-base-uncased", resume_download=None) model = BertModel.from_pretrained("bert-base-uncased", resume_download=None) model.eval() model, device = set_model_device(model) return tokenizer, model @lru_cache() def load_HF_embedding_model(model_name="BAAI/bge-small-en-v1.5") -> tuple: """Load the Hugging Face model for embedding. Args: model_name (str, optional): The model name to load. Defaults to "BAAI/bge-small-en-v1.5". Returns: tuple: The tokenizer and model. """ from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained(model_name, resume_download=None) model = AutoModel.from_pretrained(model_name, resume_download=None) model.eval() model, device = set_model_device(model) return tokenizer, model @lru_cache() def load_text_classifier(): from transformers import AutoTokenizer, AutoModelForSequenceClassification from transformers import pipeline tokenizer = AutoTokenizer.from_pretrained( "dstefa/roberta-base_topic_classification_nyt_news" ) model = AutoModelForSequenceClassification.from_pretrained( "dstefa/roberta-base_topic_classification_nyt_news" ) model.eval() model, device = set_model_device(model) pipe = pipeline("text-classification", model=model, tokenizer=tokenizer) return pipe @lru_cache() def load_text_multilabel_classifier(): from transformers import AutoModelForSequenceClassification, AutoTokenizer from scipy.special import expit import torch # # Check for available device: CUDA, MPS (for Apple Silicon), or CPU # if torch.cuda.is_available(): # device = torch.device("cuda") # elif torch.backends.mps.is_available(): # device = torch.device("mps") # else: # device = torch.device("cpu") # # return load_spacy_model(), torch.device("cpu") MODEL = "cardiffnlp/tweet-topic-21-multi" tokenizer = AutoTokenizer.from_pretrained(MODEL, resume_download=None) model = AutoModelForSequenceClassification.from_pretrained( MODEL, resume_download=None ) model.eval() model, device = set_model_device(model) class_mapping = model.config.id2label def _classifier(texts, threshold=0.5, max_length=64): tokens = tokenizer( texts, return_tensors="pt", padding=True, truncation=True, max_length=max_length, ) tokens = { key: val.to(device) for key, val in tokens.items() } # Move tokens to the selected device with torch.no_grad(): output = model(**tokens) scores = output.logits.detach().cpu().numpy() scores = expit(scores) predictions = (scores >= threshold) * 1 batch_labels = [] for prediction in predictions: labels = [ class_mapping[i] for i, value in enumerate(prediction) if value == 1 ] batch_labels.append(labels) return batch_labels return _classifier, device @lru_cache() def load_nltk_punkt(): import nltk try: nltk.data.find("tokenizers/punkt") except LookupError: nltk.download("punkt") return nltk.data.find("tokenizers/punkt") @lru_cache() def load_spacy_model(): import spacy name = "models/reuters" home_folder = get_home_folder() model_folder = Path(home_folder) / name # Check if the model directory already exists if not (model_folder.exists() and any(model_folder.iterdir())): repo_url = "https://github.com/unclecode/crawl4ai.git" branch = MODEL_REPO_BRANCH repo_folder = Path(home_folder) / "crawl4ai" print("[LOG] ⏬ Downloading Spacy model for the first time...") # Remove existing repo folder if it exists if repo_folder.exists(): try: shutil.rmtree(repo_folder) if model_folder.exists(): shutil.rmtree(model_folder) except PermissionError: print( "[WARNING] Unable to remove existing folders. Please manually delete the following folders and try again:" ) print(f"- {repo_folder}") print(f"- {model_folder}") return None try: # Clone the repository subprocess.run( ["git", "clone", "-b", branch, repo_url, str(repo_folder)], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, check=True, ) # Create the models directory if it doesn't exist models_folder = Path(home_folder) / "models" models_folder.mkdir(parents=True, exist_ok=True) # Copy the reuters model folder to the models directory source_folder = repo_folder / "models" / "reuters" shutil.copytree(source_folder, model_folder) # Remove the cloned repository shutil.rmtree(repo_folder) print("[LOG] ✅ Spacy Model downloaded successfully") except subprocess.CalledProcessError as e: print(f"An error occurred while cloning the repository: {e}") return None except Exception as e: print(f"An error occurred: {e}") return None try: return spacy.load(str(model_folder)) except Exception as e: print(f"Error loading spacy model: {e}") return None def download_all_models(remove_existing=False): """Download all models required for Crawl4AI.""" if remove_existing: print("[LOG] Removing existing models...") home_folder = get_home_folder() model_folders = [ os.path.join(home_folder, "models/reuters"), os.path.join(home_folder, "models"), ] for folder in model_folders: if Path(folder).exists(): shutil.rmtree(folder) print("[LOG] Existing models removed.") # Load each model to trigger download # print("[LOG] Downloading BERT Base Uncased...") # load_bert_base_uncased() # print("[LOG] Downloading BGE Small EN v1.5...") # load_bge_small_en_v1_5() # print("[LOG] Downloading ONNX model...") # load_onnx_all_MiniLM_l6_v2() print("[LOG] Downloading text classifier...") _, device = load_text_multilabel_classifier() print(f"[LOG] Text classifier loaded on {device}") print("[LOG] Downloading custom NLTK Punkt model...") load_nltk_punkt() print("[LOG] ✅ All models downloaded successfully.") def main(): print("[LOG] Welcome to the Crawl4AI Model Downloader!") print("[LOG] This script will download all the models required for Crawl4AI.") parser = argparse.ArgumentParser(description="Crawl4AI Model Downloader") parser.add_argument( "--remove-existing", action="store_true", help="Remove existing models before downloading", ) args = parser.parse_args() download_all_models(remove_existing=args.remove_existing) if __name__ == "__main__": main()
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/table_extraction.py
crawl4ai/table_extraction.py
""" Table extraction strategies for Crawl4AI. This module provides various strategies for detecting and extracting tables from HTML content. The strategy pattern allows for flexible table extraction methods while maintaining a consistent interface. """ from abc import ABC, abstractmethod from typing import Dict, List, Optional, Any, Union, Tuple from lxml import etree import re import json from .types import LLMConfig, create_llm_config from .utils import perform_completion_with_backoff, sanitize_html import os from concurrent.futures import ThreadPoolExecutor, as_completed import time import tiktoken class TableExtractionStrategy(ABC): """ Abstract base class for all table extraction strategies. This class defines the interface that all table extraction strategies must implement. It provides a consistent way to detect and extract tables from HTML content. """ def __init__(self, **kwargs): """ Initialize the table extraction strategy. Args: **kwargs: Additional keyword arguments for specific strategies """ self.verbose = kwargs.get("verbose", False) self.logger = kwargs.get("logger", None) @abstractmethod def extract_tables(self, element: etree.Element, **kwargs) -> List[Dict[str, Any]]: """ Extract tables from the given HTML element. Args: element: The HTML element (typically the body or a container element) **kwargs: Additional parameters for extraction Returns: List of dictionaries containing table data, each with: - headers: List of column headers - rows: List of row data (each row is a list) - caption: Table caption if present - summary: Table summary attribute if present - metadata: Additional metadata about the table """ pass def _log(self, level: str, message: str, tag: str = "TABLE", **kwargs): """Helper method to safely use logger.""" if self.logger: log_method = getattr(self.logger, level, None) if log_method: log_method(message=message, tag=tag, **kwargs) class DefaultTableExtraction(TableExtractionStrategy): """ Default table extraction strategy that implements the current Crawl4AI table extraction logic. This strategy uses a scoring system to identify data tables (vs layout tables) and extracts structured data including headers, rows, captions, and summaries. It handles colspan and rowspan attributes to preserve table structure. """ def __init__(self, **kwargs): """ Initialize the default table extraction strategy. Args: table_score_threshold (int): Minimum score for a table to be considered a data table (default: 7) min_rows (int): Minimum number of rows for a valid table (default: 0) min_cols (int): Minimum number of columns for a valid table (default: 0) **kwargs: Additional parameters passed to parent class """ super().__init__(**kwargs) self.table_score_threshold = kwargs.get("table_score_threshold", 7) self.min_rows = kwargs.get("min_rows", 0) self.min_cols = kwargs.get("min_cols", 0) def extract_tables(self, element: etree.Element, **kwargs) -> List[Dict[str, Any]]: """ Extract all data tables from the HTML element. Args: element: The HTML element to search for tables **kwargs: Additional parameters (can override instance settings) Returns: List of dictionaries containing extracted table data """ tables_data = [] # Allow kwargs to override instance settings score_threshold = kwargs.get("table_score_threshold", self.table_score_threshold) # Find all table elements tables = element.xpath(".//table") for table in tables: # Check if this is a data table (not a layout table) if self.is_data_table(table, table_score_threshold=score_threshold): try: table_data = self.extract_table_data(table) # Apply minimum size filters if specified if self.min_rows > 0 and len(table_data.get("rows", [])) < self.min_rows: continue if self.min_cols > 0: col_count = len(table_data.get("headers", [])) or ( max(len(row) for row in table_data.get("rows", [])) if table_data.get("rows") else 0 ) if col_count < self.min_cols: continue tables_data.append(table_data) except Exception as e: self._log("error", f"Error extracting table data: {str(e)}", "TABLE_EXTRACT") continue return tables_data def is_data_table(self, table: etree.Element, **kwargs) -> bool: """ Determine if a table is a data table (vs. layout table) using a scoring system. Args: table: The table element to evaluate **kwargs: Additional parameters (e.g., table_score_threshold) Returns: True if the table scores above the threshold, False otherwise """ score = 0 # Check for thead and tbody has_thead = len(table.xpath(".//thead")) > 0 has_tbody = len(table.xpath(".//tbody")) > 0 if has_thead: score += 2 if has_tbody: score += 1 # Check for th elements th_count = len(table.xpath(".//th")) if th_count > 0: score += 2 if has_thead or table.xpath(".//tr[1]/th"): score += 1 # Check for nested tables (negative indicator) if len(table.xpath(".//table")) > 0: score -= 3 # Role attribute check role = table.get("role", "").lower() if role in {"presentation", "none"}: score -= 3 # Column consistency rows = table.xpath(".//tr") if not rows: return False col_counts = [len(row.xpath(".//td|.//th")) for row in rows] if col_counts: avg_cols = sum(col_counts) / len(col_counts) variance = sum((c - avg_cols)**2 for c in col_counts) / len(col_counts) if variance < 1: score += 2 # Caption and summary if table.xpath(".//caption"): score += 2 if table.get("summary"): score += 1 # Text density total_text = sum( len(''.join(cell.itertext()).strip()) for row in rows for cell in row.xpath(".//td|.//th") ) total_tags = sum(1 for _ in table.iterdescendants()) text_ratio = total_text / (total_tags + 1e-5) if text_ratio > 20: score += 3 elif text_ratio > 10: score += 2 # Data attributes data_attrs = sum(1 for attr in table.attrib if attr.startswith('data-')) score += data_attrs * 0.5 # Size check if col_counts and len(rows) >= 2: avg_cols = sum(col_counts) / len(col_counts) if avg_cols >= 2: score += 2 threshold = kwargs.get("table_score_threshold", self.table_score_threshold) return score >= threshold def extract_table_data(self, table: etree.Element) -> Dict[str, Any]: """ Extract structured data from a table element. Args: table: The table element to extract data from Returns: Dictionary containing: - headers: List of column headers - rows: List of row data (each row is a list) - caption: Table caption if present - summary: Table summary attribute if present - metadata: Additional metadata about the table """ # Extract caption and summary caption = table.xpath(".//caption/text()") caption = caption[0].strip() if caption else "" summary = table.get("summary", "").strip() # Extract headers with colspan handling headers = [] thead_rows = table.xpath(".//thead/tr") if thead_rows: header_cells = thead_rows[0].xpath(".//th") for cell in header_cells: text = cell.text_content().strip() colspan = int(cell.get("colspan", 1)) headers.extend([text] * colspan) else: # Check first row for headers first_row = table.xpath(".//tr[1]") if first_row: for cell in first_row[0].xpath(".//th|.//td"): text = cell.text_content().strip() colspan = int(cell.get("colspan", 1)) headers.extend([text] * colspan) # Extract rows with colspan handling rows = [] for row in table.xpath(".//tr[not(ancestor::thead)]"): row_data = [] for cell in row.xpath(".//td"): text = cell.text_content().strip() colspan = int(cell.get("colspan", 1)) row_data.extend([text] * colspan) if row_data: rows.append(row_data) # Align rows with headers max_columns = len(headers) if headers else ( max(len(row) for row in rows) if rows else 0 ) aligned_rows = [] for row in rows: aligned = row[:max_columns] + [''] * (max_columns - len(row)) aligned_rows.append(aligned) # Generate default headers if none found if not headers and max_columns > 0: headers = [f"Column {i+1}" for i in range(max_columns)] # Build metadata metadata = { "row_count": len(aligned_rows), "column_count": max_columns, "has_headers": bool(thead_rows) or bool(table.xpath(".//tr[1]/th")), "has_caption": bool(caption), "has_summary": bool(summary) } # Add table attributes that might be useful if table.get("id"): metadata["id"] = table.get("id") if table.get("class"): metadata["class"] = table.get("class") return { "headers": headers, "rows": aligned_rows, "caption": caption, "summary": summary, "metadata": metadata } class NoTableExtraction(TableExtractionStrategy): """ A strategy that does not extract any tables. This can be used to explicitly disable table extraction when needed. """ def extract_tables(self, element: etree.Element, **kwargs) -> List[Dict[str, Any]]: """ Return an empty list (no tables extracted). Args: element: The HTML element (ignored) **kwargs: Additional parameters (ignored) Returns: Empty list """ return [] class LLMTableExtraction(TableExtractionStrategy): """ LLM-based table extraction strategy that uses language models to intelligently extract and structure table data, handling complex cases like rowspan, colspan, and nested tables. This strategy uses an LLM to understand table structure semantically and convert it to structured data that can be easily consumed by pandas DataFrames. """ # System prompt for table extraction TABLE_EXTRACTION_PROMPT = """You are a specialized table extraction system that converts complex HTML tables into structured JSON data. Your primary goal is to handle difficult, irregular HTML tables that cannot be easily parsed by standard tools, transforming them into clean, tabulated data. ## Critical Requirements **IMPORTANT**: You must extract **EVERY SINGLE ROW** from the table, regardless of size. Tables often contain hundreds of rows, and omitting data is unacceptable. The reason we use an LLM for this task is because these tables have complex structures that standard parsers cannot handle properly. ## Output Format **Your response must be valid JSON**. The output must be properly formatted, parseable JSON with: - Proper escaping of quotes in strings - Valid JSON syntax (commas, brackets, etc.) - No trailing commas - Proper handling of special characters ## Table Structure Every table should be extracted as a JSON object with this structure: ```json { "headers": ["Column 1", "Column 2", ...], "rows": [ ["Row 1 Col 1", "Row 1 Col 2", ...], ["Row 2 Col 1", "Row 2 Col 2", ...], // ... continue for ALL rows ... ], "caption": "Table caption if present", "summary": "Table summary attribute if present", "metadata": { "row_count": <actual_number_of_rows>, "column_count": <number>, "has_headers": <boolean>, "has_merged_cells": <boolean>, "nested_tables": <boolean>, "table_type": "data|pivot|matrix|nested" } } ``` ## Handling Complex Structures ### Why This Matters Standard HTML parsers fail on tables with: - Complex colspan/rowspan arrangements - Nested tables - Irregular structures - Mixed header patterns Your job is to intelligently interpret these structures and produce clean, regular data. ### Colspan (Merged Columns) When a cell spans multiple columns, duplicate the value across all spanned columns to maintain rectangular data structure. Example HTML: ```html <tr> <td colspan="3">Quarterly Report</td> <td>Total</td> </tr> ``` Becomes: ["Quarterly Report", "Quarterly Report", "Quarterly Report", "Total"] ### Rowspan (Merged Rows) When a cell spans multiple rows, duplicate the value down all affected rows. Example with many rows: ```html <tr> <td rowspan="50">Category A</td> <td>Item 1</td> <td>$100</td> </tr> <tr> <td>Item 2</td> <td>$200</td> </tr> <!-- ... 48 more rows ... --> ``` Result structure (response must be valid JSON): ```json { "headers": ["Category", "Item", "Price"], "rows": [ ["Category A", "Item 1", "$100"], ["Category A", "Item 2", "$200"], ["Category A", "Item 3", "$300"], ["Category A", "Item 4", "$400"], ["Category A", "Item 5", "$500"], // ... ALL 50 rows must be included ... ["Category A", "Item 50", "$5000"] ], "metadata": { "row_count": 50, "column_count": 3, "has_headers": true, "has_merged_cells": true, "nested_tables": false, "table_type": "data" } } ``` ### Nested Tables For tables containing other tables: 1. Extract the outer table structure 2. Represent nested tables as a JSON string or structured representation 3. Ensure the data remains usable ## Complete Examples ### Example 1: Large Table with Complex Structure Input HTML (abbreviated for documentation): ```html <table> <thead> <tr> <th rowspan="2">Department</th> <th colspan="4">2024 Performance</th> </tr> <tr> <th>Q1</th> <th>Q2</th> <th>Q3</th> <th>Q4</th> </tr> </thead> <tbody> <tr> <td rowspan="15">Sales</td> <td>Region North</td> <td>$1.2M</td> <td>$1.5M</td> <td>$1.8M</td> </tr> <tr> <td>Region South</td> <td>$0.9M</td> <td>$1.1M</td> <td>$1.3M</td> </tr> <!-- ... 13 more regions ... --> <tr> <td rowspan="20">Engineering</td> <td>Team Alpha</td> <td>85%</td> <td>88%</td> <td>92%</td> </tr> <!-- ... 19 more teams ... --> <!-- ... continue for 200+ total rows ... --> </tbody> </table> ``` Output (showing structure with all rows) - must be valid JSON: ```json { "headers": ["Department", "Team/Region", "Q1", "Q2", "Q3", "Q4"], "rows": [ ["Sales", "Region North", "$1.2M", "$1.5M", "$1.8M"], ["Sales", "Region South", "$0.9M", "$1.1M", "$1.3M"], ["Sales", "Region East", "$1.1M", "$1.4M", "$1.6M"], ["Sales", "Region West", "$1.0M", "$1.2M", "$1.5M"], ["Sales", "Region Central", "$0.8M", "$1.0M", "$1.2M"], // ... ALL 15 Sales rows must be included ... ["Engineering", "Team Alpha", "85%", "88%", "92%"], ["Engineering", "Team Beta", "82%", "85%", "89%"], ["Engineering", "Team Gamma", "88%", "90%", "93%"], // ... ALL 20 Engineering rows must be included ... // ... Continue for EVERY row in the table ... ], "caption": "", "summary": "", "metadata": { "row_count": 235, "column_count": 6, "has_headers": true, "has_merged_cells": true, "nested_tables": false, "table_type": "data" } } ``` ### Example 2: Pivot Table with Hundreds of Rows Input structure: ```html <table> <tr> <th>Product ID</th> <th>Jan</th> <th>Feb</th> <!-- ... all 12 months ... --> </tr> <tr> <td>PROD-001</td> <td>1,234</td> <td>1,456</td> <!-- ... --> </tr> <!-- ... 500+ product rows ... --> </table> ``` Output must include ALL rows and be valid JSON: ```json { "headers": ["Product ID", "Jan", "Feb", "Mar", "Apr", "May", "Jun", "Jul", "Aug", "Sep", "Oct", "Nov", "Dec"], "rows": [ ["PROD-001", "1,234", "1,456", "1,789", "2,012", "2,234", "2,456", "2,678", "2,890", "3,123", "3,345", "3,567", "3,789"], ["PROD-002", "2,345", "2,567", "2,789", "3,012", "3,234", "3,456", "3,678", "3,890", "4,123", "4,345", "4,567", "4,789"], ["PROD-003", "3,456", "3,678", "3,890", "4,123", "4,345", "4,567", "4,789", "5,012", "5,234", "5,456", "5,678", "5,890"], // ... ALL 500+ rows MUST be included ... ["PROD-547", "9,876", "10,098", "10,321", "10,543", "10,765", "10,987", "11,210", "11,432", "11,654", "11,876", "12,098", "12,321"] ], "metadata": { "row_count": 547, "column_count": 13, "has_headers": true, "has_merged_cells": false, "nested_tables": false, "table_type": "pivot" } } ``` ## Critical Data Integrity Rules 1. **COMPLETENESS**: Extract EVERY row, no matter how many (10, 100, 1000+) 2. **ACCURACY**: Preserve exact values, including formatting 3. **STRUCTURE**: Maintain consistent column count across all rows 4. **VALIDATION**: Ensure output is valid JSON that can be parsed 5. **ESCAPING**: Properly escape quotes and special characters in cell values ## Special Handling Instructions ### Large Tables - Never abbreviate or summarize - Never use "..." to indicate omitted rows - Process every row even if it takes significant time - The metadata row_count must match actual extracted rows ### Complex Merged Cells - Track rowspan/colspan values carefully - Ensure proper cell duplication - Maintain data alignment across all rows ### Data Types - Keep numbers as strings to preserve formatting - Preserve currency symbols, percentages, etc. - Handle empty cells as empty strings "" ### Error Prevention - If a cell contains quotes, escape them properly - Handle newlines within cells appropriately - Ensure no JSON syntax errors ## Output Validation Before returning results: 1. Verify JSON is valid and parseable 2. Confirm row count matches actual data 3. Check that all rows have same column count 4. Ensure all data is preserved without truncation ## JSON Schema Definition Your output must conform to the following JSON schema (OpenAPI 3.0 format): { "components": { "schemas": { "ExtractedTable": { "type": "object", "required": [ "headers", "rows", "metadata" ], "properties": { "headers": { "type": "array", "description": "Column headers for the table", "items": { "type": "string" }, "minItems": 1 }, "rows": { "type": "array", "description": "All table rows - must include every single row", "items": { "type": "array", "items": { "type": "string" }, "minItems": 1 } }, "caption": { "type": "string", "description": "Table caption if present", "default": "" }, "summary": { "type": "string", "description": "Table summary attribute if present", "default": "" }, "metadata": { "type": "object", "required": [ "row_count", "column_count", "has_headers", "has_merged_cells", "nested_tables", "table_type" ], "properties": { "row_count": { "type": "integer", "description": "Actual count of rows extracted", "minimum": 0 }, "column_count": { "type": "integer", "description": "Number of columns in the table", "minimum": 1 }, "has_headers": { "type": "boolean", "description": "Whether table has identified headers" }, "has_merged_cells": { "type": "boolean", "description": "Whether table contains colspan or rowspan" }, "nested_tables": { "type": "boolean", "description": "Whether table contains nested tables" }, "table_type": { "type": "string", "enum": ["data", "pivot", "matrix", "nested"], "description": "Classification of table structure" } } } } } } } } **CRITICAL**: Your response must be a valid JSON object that conforms to this schema. The entire purpose of using an LLM for this task is to handle complex HTML tables that standard parsers cannot process correctly. Your value lies in intelligently interpreting complex structures and returning complete, clean, tabulated data in valid JSON format.""" def __init__(self, llm_config: Optional[LLMConfig] = None, css_selector: Optional[str] = None, max_tries: int = 3, enable_chunking: bool = True, chunk_token_threshold: int = 3000, min_rows_per_chunk: int = 10, max_parallel_chunks: int = 5, verbose: bool = False, **kwargs): """ Initialize the LLM-based table extraction strategy. Args: llm_config: LLM configuration for the extraction css_selector: Optional CSS selector to focus on specific page areas max_tries: Maximum number of retries if LLM fails to extract tables (default: 3) enable_chunking: Enable smart chunking for large tables (default: True) chunk_token_threshold: Token threshold for triggering chunking (default: 3000) min_rows_per_chunk: Minimum rows per chunk (default: 10) max_parallel_chunks: Maximum parallel chunk processing (default: 5) verbose: Enable verbose logging **kwargs: Additional parameters passed to parent class """ super().__init__(verbose=verbose, **kwargs) # Set up LLM configuration self.llm_config = llm_config if not self.llm_config: # Use default configuration if not provided self.llm_config = create_llm_config( provider=os.getenv("DEFAULT_PROVIDER", "openai/gpt-4o-mini"), api_token=os.getenv("OPENAI_API_KEY"), ) self.css_selector = css_selector self.max_tries = max(1, max_tries) # Ensure at least 1 try self.enable_chunking = enable_chunking self.chunk_token_threshold = chunk_token_threshold self.min_rows_per_chunk = max(5, min_rows_per_chunk) # At least 5 rows per chunk self.max_parallel_chunks = max(1, max_parallel_chunks) self.extra_args = kwargs.get("extra_args", {}) def extract_tables(self, element: etree.Element, **kwargs) -> List[Dict[str, Any]]: """ Extract tables from HTML using LLM. Args: element: The HTML element to search for tables **kwargs: Additional parameters Returns: List of dictionaries containing extracted table data """ # Allow CSS selector override via kwargs css_selector = kwargs.get("css_selector", self.css_selector) # Get the HTML content to process if css_selector: # Use XPath to convert CSS selector (basic conversion) # For more complex CSS selectors, we might need a proper CSS to XPath converter selected_elements = self._css_to_xpath_select(element, css_selector) if not selected_elements: self._log("warning", f"No elements found for CSS selector: {css_selector}") return [] html_content = ''.join(etree.tostring(elem, encoding='unicode') for elem in selected_elements) else: # Process entire element html_content = etree.tostring(element, encoding='unicode') # Check if there are any tables in the content if '<table' not in html_content.lower(): if self.verbose: self._log("info", f"No <table> tags found in HTML content") return [] if self.verbose: self._log("info", f"Found table tags in HTML, content length: {len(html_content)}") # Check if chunking is needed if self.enable_chunking and self._needs_chunking(html_content): if self.verbose: self._log("info", "Content exceeds token threshold, using chunked extraction") return self._extract_with_chunking(html_content) # Single extraction for small content # Prepare the prompt user_prompt = f"""GENERATE THE TABULATED DATA from the following HTML content: ```html {sanitize_html(html_content)} ``` Return only a JSON array of extracted tables following the specified format.""" # Try extraction with retries for attempt in range(1, self.max_tries + 1): try: if self.verbose and attempt > 1: self._log("info", f"Retry attempt {attempt}/{self.max_tries} for table extraction") # Call LLM with the extraction prompt response = perform_completion_with_backoff( provider=self.llm_config.provider, prompt_with_variables=self.TABLE_EXTRACTION_PROMPT + "\n\n" + user_prompt + "\n\n MAKE SURE TO EXTRACT ALL DATA, DO NOT LEAVE ANYTHING FOR BRAVITY, YOUR GOAL IS TO RETURN ALL, NO MATTER HOW LONG IS DATA", api_token=self.llm_config.api_token, base_url=self.llm_config.base_url, json_response=True, base_delay=self.llm_config.backoff_base_delay, max_attempts=self.llm_config.backoff_max_attempts, exponential_factor=self.llm_config.backoff_exponential_factor, extra_args=self.extra_args ) # Parse the response if response and response.choices: content = response.choices[0].message.content if self.verbose: self._log("debug", f"LLM response type: {type(content)}") if isinstance(content, str): self._log("debug", f"LLM response preview: {content[:200]}...") # Parse JSON response if isinstance(content, str): tables_data = json.loads(content) else: tables_data = content # Handle various response formats from LLM # Sometimes LLM wraps response in "result" or other keys if isinstance(tables_data, dict): # Check for common wrapper keys if 'result' in tables_data: tables_data = tables_data['result'] elif 'tables' in tables_data: tables_data = tables_data['tables'] elif 'data' in tables_data: tables_data = tables_data['data'] else: # If it's a single table dict, wrap in list tables_data = [tables_data] # Flatten nested lists if needed while isinstance(tables_data, list) and len(tables_data) == 1 and isinstance(tables_data[0], list): tables_data = tables_data[0] # Ensure we have a list if not isinstance(tables_data, list): tables_data = [tables_data] if self.verbose: self._log("debug", f"Parsed {len(tables_data)} table(s) from LLM response") # Validate and clean the extracted tables validated_tables = [] for table in tables_data: if self._validate_table_structure(table): validated_tables.append(self._ensure_table_format(table)) elif self.verbose: self._log("warning", f"Table failed validation: {table}") # Check if we got valid tables if validated_tables: if self.verbose: self._log("info", f"Successfully extracted {len(validated_tables)} tables using LLM on attempt {attempt}") return validated_tables # If no valid tables but we still have attempts left, retry if attempt < self.max_tries: if self.verbose: self._log("warning", f"No valid tables extracted on attempt {attempt}, retrying...") continue else: if self.verbose: self._log("warning", f"No valid tables extracted after {self.max_tries} attempts") return [] except json.JSONDecodeError as e: if self.verbose: self._log("error", f"JSON parsing error on attempt {attempt}: {str(e)}") if attempt < self.max_tries: continue else: return [] except Exception as e: if self.verbose: self._log("error", f"Error in LLM table extraction on attempt {attempt}: {str(e)}") if attempt == self.max_tries: import traceback self._log("debug", f"Traceback: {traceback.format_exc()}") # For unexpected errors, retry if we have attempts left if attempt < self.max_tries: # Add a small delay before retry for rate limiting import time time.sleep(1) continue else: return []
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/utils.py
crawl4ai/utils.py
import time from concurrent.futures import ThreadPoolExecutor, as_completed from bs4 import BeautifulSoup, Comment, element, Tag, NavigableString import json import html import lxml import re import os import subprocess import platform from .prompts import PROMPT_EXTRACT_BLOCKS from array import array from .html2text import html2text, CustomHTML2Text # from .config import * from .config import MIN_WORD_THRESHOLD, IMAGE_DESCRIPTION_MIN_WORD_THRESHOLD, IMAGE_SCORE_THRESHOLD, DEFAULT_PROVIDER, PROVIDER_MODELS import httpx from socket import gaierror from pathlib import Path from typing import Dict, Any, List, Optional, Callable, Generator, Tuple, Iterable from urllib.parse import urljoin import requests from requests.exceptions import InvalidSchema import xxhash import textwrap import cProfile import pstats from functools import wraps import asyncio from lxml import etree, html as lhtml import sqlite3 import hashlib from urllib.robotparser import RobotFileParser import aiohttp from functools import lru_cache from packaging import version from . import __version__ from typing import Sequence from itertools import chain from collections import deque import psutil import numpy as np from urllib.parse import ( urljoin, urlparse, urlunparse, parse_qsl, urlencode, quote, unquote ) import inspect # Monkey patch to fix wildcard handling in urllib.robotparser from urllib.robotparser import RuleLine import re original_applies_to = RuleLine.applies_to def patched_applies_to(self, filename): # Handle wildcards in paths if '*' in self.path or '%2A' in self.path or self.path in ("*", "%2A"): pattern = self.path.replace('%2A', '*') pattern = re.escape(pattern).replace('\\*', '.*') pattern = '^' + pattern if pattern.endswith('\\$'): pattern = pattern[:-2] + '$' try: return bool(re.match(pattern, filename)) except re.error: return original_applies_to(self, filename) return original_applies_to(self, filename) RuleLine.applies_to = patched_applies_to # Monkey patch ends def chunk_documents( documents: Iterable[str], chunk_token_threshold: int, overlap: int, word_token_rate: float = 0.75, tokenizer: Optional[Callable[[str], List[str]]] = None, ) -> Generator[str, None, None]: """ Efficiently chunks documents into token-limited sections with overlap between chunks. Args: documents: Iterable of document strings chunk_token_threshold: Maximum tokens per chunk overlap: Number of tokens to overlap between chunks word_token_rate: Token estimate per word when not using a tokenizer tokenizer: Function that splits text into tokens (if available) Yields: Text chunks as strings """ token_queue = deque() contribution_queue = deque() current_token_count = 0.0 for doc in documents: # Tokenize document if tokenizer: tokens = tokenizer(doc) contributions = [1.0] * len(tokens) else: tokens = doc.split() contributions = [word_token_rate] * len(tokens) # Add to processing queues token_queue.extend(tokens) contribution_queue.extend(contributions) current_token_count += sum(contributions) # Process full chunks while current_token_count >= chunk_token_threshold: # Find chunk split point chunk_tokens = [] chunk_contrib = [] chunk_total = 0.0 # Build chunk up to threshold while contribution_queue: next_contrib = contribution_queue[0] if chunk_total + next_contrib > chunk_token_threshold: break chunk_total += next_contrib chunk_contrib.append(contribution_queue.popleft()) chunk_tokens.append(token_queue.popleft()) # Handle edge case where first token exceeds threshold if not chunk_contrib: # Single token exceeds threshold chunk_contrib.append(contribution_queue.popleft()) chunk_tokens.append(token_queue.popleft()) # Calculate overlap overlap_total = 0.0 overlap_idx = 0 for contrib in reversed(chunk_contrib): if overlap_total + contrib > overlap: break overlap_total += contrib overlap_idx += 1 # Prepend overlap to queues if overlap_idx > 0: overlap_tokens = chunk_tokens[-overlap_idx:] overlap_contrib = chunk_contrib[-overlap_idx:] token_queue.extendleft(reversed(overlap_tokens)) contribution_queue.extendleft(reversed(overlap_contrib)) current_token_count += overlap_total # Update current token count and yield chunk current_token_count -= sum(chunk_contrib) yield " ".join(chunk_tokens[:len(chunk_tokens)-overlap_idx] if overlap_idx else chunk_tokens) # Yield remaining tokens if token_queue: yield " ".join(token_queue) def merge_chunks( docs: Sequence[str], target_size: int, overlap: int = 0, word_token_ratio: float = 1.0, splitter: Callable = None ) -> List[str]: """ Merges a sequence of documents into chunks based on a target token count, with optional overlap. Each document is split into tokens using the provided splitter function (defaults to str.split). Tokens are distributed into chunks aiming for the specified target size, with optional overlapping tokens between consecutive chunks. Returns a list of non-empty merged chunks as strings. Args: docs: Sequence of input document strings to be merged. target_size: Target number of tokens per chunk. overlap: Number of tokens to overlap between consecutive chunks. word_token_ratio: Multiplier to estimate token count from word count. splitter: Callable used to split each document into tokens. Returns: List of merged document chunks as strings, each not exceeding the target token size. """ # Pre-tokenize all docs and store token counts splitter = splitter or str.split token_counts = array('I') all_tokens: List[List[str]] = [] total_tokens = 0 for doc in docs: tokens = splitter(doc) count = int(len(tokens) * word_token_ratio) if count: # Skip empty docs token_counts.append(count) all_tokens.append(tokens) total_tokens += count if not total_tokens: return [] # Pre-allocate chunks num_chunks = max(1, (total_tokens + target_size - 1) // target_size) chunks: List[List[str]] = [[] for _ in range(num_chunks)] curr_chunk = 0 curr_size = 0 # Distribute tokens for tokens in chain.from_iterable(all_tokens): if curr_size >= target_size and curr_chunk < num_chunks - 1: if overlap > 0: overlap_tokens = chunks[curr_chunk][-overlap:] curr_chunk += 1 chunks[curr_chunk].extend(overlap_tokens) curr_size = len(overlap_tokens) else: curr_chunk += 1 curr_size = 0 chunks[curr_chunk].append(tokens) curr_size += 1 # Return only non-empty chunks return [' '.join(chunk) for chunk in chunks if chunk] class VersionManager: def __init__(self): self.home_dir = Path.home() / ".crawl4ai" self.version_file = self.home_dir / "version.txt" def get_installed_version(self): """Get the version recorded in home directory""" if not self.version_file.exists(): return None try: return version.parse(self.version_file.read_text().strip()) except Exception as _ex: return None def update_version(self): """Update the version file to current library version""" self.version_file.write_text(__version__.__version__) def needs_update(self): """Check if database needs update based on version""" installed = self.get_installed_version() current = version.parse(__version__.__version__) return installed is None or installed < current class RobotsParser: # Default 7 days cache TTL CACHE_TTL = 7 * 24 * 60 * 60 def __init__(self, cache_dir=None, cache_ttl=None): self.cache_dir = cache_dir or os.path.join(get_home_folder(), ".crawl4ai", "robots") self.cache_ttl = cache_ttl or self.CACHE_TTL os.makedirs(self.cache_dir, exist_ok=True) self.db_path = os.path.join(self.cache_dir, "robots_cache.db") self._init_db() def _init_db(self): # Use WAL mode for better concurrency and performance with sqlite3.connect(self.db_path) as conn: conn.execute("PRAGMA journal_mode=WAL") conn.execute(""" CREATE TABLE IF NOT EXISTS robots_cache ( domain TEXT PRIMARY KEY, rules TEXT NOT NULL, fetch_time INTEGER NOT NULL, hash TEXT NOT NULL ) """) conn.execute("CREATE INDEX IF NOT EXISTS idx_domain ON robots_cache(domain)") def _get_cached_rules(self, domain: str) -> tuple[str, bool]: """Get cached rules. Returns (rules, is_fresh)""" with sqlite3.connect(self.db_path) as conn: cursor = conn.execute( "SELECT rules, fetch_time, hash FROM robots_cache WHERE domain = ?", (domain,) ) result = cursor.fetchone() if not result: return None, False rules, fetch_time, _ = result # Check if cache is still fresh based on TTL return rules, (time.time() - fetch_time) < self.cache_ttl def _cache_rules(self, domain: str, content: str): """Cache robots.txt content with hash for change detection""" hash_val = hashlib.md5(content.encode()).hexdigest() with sqlite3.connect(self.db_path) as conn: # Check if content actually changed cursor = conn.execute( "SELECT hash FROM robots_cache WHERE domain = ?", (domain,) ) result = cursor.fetchone() # Only update if hash changed or no previous entry if not result or result[0] != hash_val: conn.execute( """INSERT OR REPLACE INTO robots_cache (domain, rules, fetch_time, hash) VALUES (?, ?, ?, ?)""", (domain, content, int(time.time()), hash_val) ) async def can_fetch(self, url: str, user_agent: str = "*") -> bool: """ Check if URL can be fetched according to robots.txt rules. Args: url: The URL to check user_agent: User agent string to check against (default: "*") Returns: bool: True if allowed, False if disallowed by robots.txt """ # Handle empty/invalid URLs try: parsed = urlparse(url) domain = parsed.netloc if not domain: return True except Exception as _ex: return True # Fast path - check cache first rules, is_fresh = self._get_cached_rules(domain) # If rules not found or stale, fetch new ones if not is_fresh: try: # Ensure we use the same scheme as the input URL scheme = parsed.scheme or 'http' robots_url = f"{scheme}://{domain}/robots.txt" async with aiohttp.ClientSession() as session: async with session.get(robots_url, timeout=2, ssl=False) as response: if response.status == 200: rules = await response.text() self._cache_rules(domain, rules) else: return True except Exception as _ex: # On any error (timeout, connection failed, etc), allow access return True if not rules: return True # Create parser for this check parser = RobotFileParser() parser.parse(rules.splitlines()) # If parser can't read rules, allow access if not parser.mtime(): return True return parser.can_fetch(user_agent, url) def clear_cache(self): """Clear all cached robots.txt entries""" with sqlite3.connect(self.db_path) as conn: conn.execute("DELETE FROM robots_cache") def clear_expired(self): """Remove only expired entries from cache""" with sqlite3.connect(self.db_path) as conn: expire_time = int(time.time()) - self.cache_ttl conn.execute("DELETE FROM robots_cache WHERE fetch_time < ?", (expire_time,)) class InvalidCSSSelectorError(Exception): pass SPLITS = bytearray([ # Control chars (0-31) + space (32) 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, # Special chars (33-47): ! " # $ % & ' ( ) * + , - . / 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, # Numbers (48-57): Treat as non-splits 0,0,0,0,0,0,0,0,0,0, # More special chars (58-64): : ; < = > ? @ 1,1,1,1,1,1,1, # Uppercase (65-90): Keep 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, # More special chars (91-96): [ \ ] ^ _ ` 1,1,1,1,1,1, # Lowercase (97-122): Keep 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, # Special chars (123-126): { | } ~ 1,1,1,1, # Extended ASCII *([1] * 128) ]) # Additional split chars for HTML/code HTML_CODE_CHARS = { # HTML specific '•', '►', '▼', '©', '®', '™', '→', '⇒', '≈', '≤', '≥', # Programming symbols '+=', '-=', '*=', '/=', '=>', '<=>', '!=', '==', '===', '++', '--', '<<', '>>', '&&', '||', '??', '?:', '?.', # Common Unicode '…', '"', '"', ''', ''', '«', '»', '—', '–', # Additional splits '+', '=', '~', '@', '#', '$', '%', '^', '&', '*', '(', ')', '{', '}', '[', ']', '|', '\\', '/', '`', '<', '>', ',', '.', '?', '!', ':', ';', '-', '_' } def advanced_split(text: str) -> list[str]: result = [] word = array('u') i = 0 text_len = len(text) while i < text_len: char = text[i] o = ord(char) # Fast path for ASCII if o < 256 and SPLITS[o]: if word: result.append(word.tounicode()) word = array('u') # Check for multi-char symbols elif i < text_len - 1: two_chars = char + text[i + 1] if two_chars in HTML_CODE_CHARS: if word: result.append(word.tounicode()) word = array('u') i += 1 # Skip next char since we used it else: word.append(char) else: word.append(char) i += 1 if word: result.append(word.tounicode()) return result def create_box_message( message: str, type: str = "info", width: int = 120, add_newlines: bool = True, double_line: bool = False, ) -> str: """ Create a styled message box with colored borders and formatted text. How it works: 1. Determines box style and colors based on the message type (e.g., info, warning). 2. Wraps text to fit within the specified width. 3. Constructs a box using characters (single or double lines) with appropriate formatting. 4. Adds optional newlines before and after the box. Args: message (str): The message to display inside the box. type (str): Type of the message (e.g., "info", "warning", "error", "success"). Defaults to "info". width (int): Width of the box. Defaults to 120. add_newlines (bool): Whether to add newlines before and after the box. Defaults to True. double_line (bool): Whether to use double lines for the box border. Defaults to False. Returns: str: A formatted string containing the styled message box. """ # Define border and text colors for different types styles = { "warning": ("yellow", "bright_yellow", "⚠"), "info": ("blue", "bright_blue", "ℹ"), "debug": ("lightblack", "bright_black", "⋯"), "success": ("green", "bright_green", "✓"), "error": ("red", "bright_red", "×"), } border_color, text_color, prefix = styles.get(type.lower(), styles["info"]) # Define box characters based on line style box_chars = { "single": ("─", "│", "┌", "┐", "└", "┘"), "double": ("═", "║", "╔", "╗", "╚", "╝"), } line_style = "double" if double_line else "single" h_line, v_line, tl, tr, bl, br = box_chars[line_style] # Process lines with lighter text color formatted_lines = [] raw_lines = message.split("\n") if raw_lines: first_line = f"{prefix} {raw_lines[0].strip()}" wrapped_first = textwrap.fill(first_line, width=width - 4) formatted_lines.extend(wrapped_first.split("\n")) for line in raw_lines[1:]: if line.strip(): wrapped = textwrap.fill(f" {line.strip()}", width=width - 4) formatted_lines.extend(wrapped.split("\n")) else: formatted_lines.append("") # Create the box with colored borders and lighter text horizontal_line = h_line * (width - 1) box = [ f"[{border_color}]{tl}{horizontal_line}{tr}[/{border_color}]", *[ f"[{border_color}]{v_line}[{text_color}] {line:<{width-2}}[/{text_color}][{border_color}]{v_line}[/{border_color}]" for line in formatted_lines ], f"[{border_color}]{bl}{horizontal_line}{br}[/{border_color}]", ] result = "\n".join(box) if add_newlines: result = f"\n{result}\n" return result def calculate_semaphore_count(): """ Calculate the optimal semaphore count based on system resources. How it works: 1. Determines the number of CPU cores and total system memory. 2. Sets a base count as half of the available CPU cores. 3. Limits the count based on memory, assuming 2GB per semaphore instance. 4. Returns the minimum value between CPU and memory-based limits. Returns: int: The calculated semaphore count. """ cpu_count = os.cpu_count() memory_gb = get_system_memory() / (1024**3) # Convert to GB base_count = max(1, cpu_count // 2) memory_based_cap = int(memory_gb / 2) # Assume 2GB per instance return min(base_count, memory_based_cap) def get_system_memory(): """ Get the total system memory in bytes. How it works: 1. Detects the operating system. 2. Reads memory information from system-specific commands or files. 3. Converts the memory to bytes for uniformity. Returns: int: The total system memory in bytes. Raises: OSError: If the operating system is unsupported. """ system = platform.system() if system == "Linux": with open("/proc/meminfo", "r") as mem: for line in mem: if line.startswith("MemTotal:"): return int(line.split()[1]) * 1024 # Convert KB to bytes elif system == "Darwin": # macOS import subprocess output = subprocess.check_output(["sysctl", "-n", "hw.memsize"]).decode("utf-8") return int(output.strip()) elif system == "Windows": import ctypes kernel32 = ctypes.windll.kernel32 c_ulonglong = ctypes.c_ulonglong class MEMORYSTATUSEX(ctypes.Structure): _fields_ = [ ("dwLength", ctypes.c_ulong), ("dwMemoryLoad", ctypes.c_ulong), ("ullTotalPhys", c_ulonglong), ("ullAvailPhys", c_ulonglong), ("ullTotalPageFile", c_ulonglong), ("ullAvailPageFile", c_ulonglong), ("ullTotalVirtual", c_ulonglong), ("ullAvailVirtual", c_ulonglong), ("ullAvailExtendedVirtual", c_ulonglong), ] memoryStatus = MEMORYSTATUSEX() memoryStatus.dwLength = ctypes.sizeof(MEMORYSTATUSEX) kernel32.GlobalMemoryStatusEx(ctypes.byref(memoryStatus)) return memoryStatus.ullTotalPhys else: raise OSError("Unsupported operating system") def get_home_folder(): """ Get or create the home folder for Crawl4AI configuration and cache. How it works: 1. Uses environment variables or defaults to the user's home directory. 2. Creates `.crawl4ai` and its subdirectories (`cache`, `models`) if they don't exist. 3. Returns the path to the home folder. Returns: str: The path to the Crawl4AI home folder. """ home_folder = os.path.join( os.getenv( "CRAWL4_AI_BASE_DIRECTORY", os.getenv("CRAWL4_AI_BASE_DIRECTORY", Path.home()), ), ".crawl4ai", ) os.makedirs(home_folder, exist_ok=True) os.makedirs(f"{home_folder}/cache", exist_ok=True) os.makedirs(f"{home_folder}/models", exist_ok=True) return home_folder async def get_chromium_path(browser_type) -> str: """Returns the browser executable path using playwright's browser management. Uses playwright's built-in browser management to get the correct browser executable path regardless of platform. This ensures we're using the same browser version that playwright is tested with. Returns: str: Path to browser executable Raises: RuntimeError: If browser executable cannot be found """ browser_types = { "chromium": "chromium", "firefox": "firefox", "webkit": "webkit" } browser_type = browser_types.get(browser_type) if not browser_type: raise RuntimeError(f"Unsupported browser type: {browser_type}") # Check if a path has already been saved for this browser type home_folder = get_home_folder() path_file = os.path.join(home_folder, f"{browser_type.lower()}.path") if os.path.exists(path_file): with open(path_file, "r") as f: return f.read() from playwright.async_api import async_playwright async with async_playwright() as p: browsers = { 'chromium': p.chromium, 'firefox': p.firefox, 'webkit': p.webkit } if browser_type.lower() not in browsers: raise ValueError( f"Invalid browser type. Must be one of: {', '.join(browsers.keys())}" ) # Save the path int the crawl4ai home folder home_folder = get_home_folder() browser_path = browsers[browser_type.lower()].executable_path if not browser_path: raise RuntimeError(f"Browser executable not found for type: {browser_type}") # Save the path in a text file with browser type name with open(os.path.join(home_folder, f"{browser_type.lower()}.path"), "w") as f: f.write(browser_path) return browser_path def beautify_html(escaped_html): """ Beautifies an escaped HTML string. Parameters: escaped_html (str): A string containing escaped HTML. Returns: str: A beautifully formatted HTML string. """ # Unescape the HTML string unescaped_html = html.unescape(escaped_html) # Use BeautifulSoup to parse and prettify the HTML soup = BeautifulSoup(unescaped_html, "html.parser") pretty_html = soup.prettify() return pretty_html def split_and_parse_json_objects(json_string): """ Splits a JSON string which is a list of objects and tries to parse each object. Parameters: json_string (str): A string representation of a list of JSON objects, e.g., '[{...}, {...}, ...]'. Returns: tuple: A tuple containing two lists: - First list contains all successfully parsed JSON objects. - Second list contains the string representations of all segments that couldn't be parsed. """ # Trim the leading '[' and trailing ']' if json_string.startswith("[") and json_string.endswith("]"): json_string = json_string[1:-1].strip() # Split the string into segments that look like individual JSON objects segments = [] depth = 0 start_index = 0 for i, char in enumerate(json_string): if char == "{": if depth == 0: start_index = i depth += 1 elif char == "}": depth -= 1 if depth == 0: segments.append(json_string[start_index : i + 1]) # Try parsing each segment parsed_objects = [] unparsed_segments = [] for segment in segments: try: obj = json.loads(segment) parsed_objects.append(obj) except json.JSONDecodeError: unparsed_segments.append(segment) return parsed_objects, unparsed_segments def sanitize_html(html): """ Sanitize an HTML string by escaping quotes. How it works: 1. Replaces all unwanted and special characters with an empty string. 2. Escapes double and single quotes for safe usage. Args: html (str): The HTML string to sanitize. Returns: str: The sanitized HTML string. """ # Replace all unwanted and special characters with an empty string sanitized_html = html # sanitized_html = re.sub(r'[^\w\s.,;:!?=\[\]{}()<>\/\\\-"]', '', html) # Escape all double and single quotes sanitized_html = sanitized_html.replace('"', '\\"').replace("'", "\\'") return sanitized_html def sanitize_input_encode(text: str) -> str: """Sanitize input to handle potential encoding issues.""" try: try: if not text: return "" # Attempt to encode and decode as UTF-8 to handle potential encoding issues return text.encode("utf-8", errors="ignore").decode("utf-8") except UnicodeEncodeError as e: print( f"Warning: Encoding issue detected. Some characters may be lost. Error: {e}" ) # Fall back to ASCII if UTF-8 fails return text.encode("ascii", errors="ignore").decode("ascii") except Exception as e: raise ValueError(f"Error sanitizing input: {str(e)}") from e def escape_json_string(s): """ Escapes characters in a string to be JSON safe. Parameters: s (str): The input string to be escaped. Returns: str: The escaped string, safe for JSON encoding. """ # Replace problematic backslash first s = s.replace("\\", "\\\\") # Replace the double quote s = s.replace('"', '\\"') # Escape control characters s = s.replace("\b", "\\b") s = s.replace("\f", "\\f") s = s.replace("\n", "\\n") s = s.replace("\r", "\\r") s = s.replace("\t", "\\t") # Additional problematic characters # Unicode control characters s = re.sub(r"[\x00-\x1f\x7f-\x9f]", lambda x: "\\u{:04x}".format(ord(x.group())), s) return s def replace_inline_tags(soup, tags, only_text=False): """ Replace inline HTML tags with Markdown-style equivalents. How it works: 1. Maps specific tags (e.g., <b>, <i>) to Markdown syntax. 2. Finds and replaces all occurrences of these tags in the provided BeautifulSoup object. 3. Optionally replaces tags with their text content only. Args: soup (BeautifulSoup): Parsed HTML content. tags (List[str]): List of tags to replace. only_text (bool): Whether to replace tags with plain text. Defaults to False. Returns: BeautifulSoup: Updated BeautifulSoup object with replaced tags. """ tag_replacements = { "b": lambda tag: f"**{tag.text}**", "i": lambda tag: f"*{tag.text}*", "u": lambda tag: f"__{tag.text}__", "span": lambda tag: f"{tag.text}", "del": lambda tag: f"~~{tag.text}~~", "ins": lambda tag: f"++{tag.text}++", "sub": lambda tag: f"~{tag.text}~", "sup": lambda tag: f"^^{tag.text}^^", "strong": lambda tag: f"**{tag.text}**", "em": lambda tag: f"*{tag.text}*", "code": lambda tag: f"`{tag.text}`", "kbd": lambda tag: f"`{tag.text}`", "var": lambda tag: f"_{tag.text}_", "s": lambda tag: f"~~{tag.text}~~", "q": lambda tag: f'"{tag.text}"', "abbr": lambda tag: f"{tag.text} ({tag.get('title', '')})", "cite": lambda tag: f"_{tag.text}_", "dfn": lambda tag: f"_{tag.text}_", "time": lambda tag: f"{tag.text}", "small": lambda tag: f"<small>{tag.text}</small>", "mark": lambda tag: f"=={tag.text}==", } replacement_data = [ (tag, tag_replacements.get(tag, lambda t: t.text)) for tag in tags ] for tag_name, replacement_func in replacement_data: for tag in soup.find_all(tag_name): replacement_text = tag.text if only_text else replacement_func(tag) tag.replace_with(replacement_text) return soup # for tag_name in tags: # for tag in soup.find_all(tag_name): # if not only_text: # replacement_text = tag_replacements.get(tag_name, lambda t: t.text)(tag) # tag.replace_with(replacement_text) # else: # tag.replace_with(tag.text) # return soup def get_content_of_website( url, html, word_count_threshold=MIN_WORD_THRESHOLD, css_selector=None, **kwargs ): """ Extract structured content, media, and links from website HTML. How it works: 1. Parses the HTML content using BeautifulSoup. 2. Extracts internal/external links and media (images, videos, audios). 3. Cleans the content by removing unwanted tags and attributes. 4. Converts cleaned HTML to Markdown. 5. Collects metadata and returns the extracted information. Args: url (str): The website URL. html (str): The HTML content of the website. word_count_threshold (int): Minimum word count for content inclusion. Defaults to MIN_WORD_THRESHOLD. css_selector (Optional[str]): CSS selector to extract specific content. Defaults to None. Returns: Dict[str, Any]: Extracted content including Markdown, cleaned HTML, media, links, and metadata. """ try: if not html: return None # Parse HTML content with BeautifulSoup soup = BeautifulSoup(html, "html.parser") # Get the content within the <body> tag body = soup.body # If css_selector is provided, extract content based on the selector if css_selector: selected_elements = body.select(css_selector) if not selected_elements: raise InvalidCSSSelectorError( f"Invalid CSS selector , No elements found for CSS selector: {css_selector}" ) div_tag = soup.new_tag("div") for el in selected_elements: div_tag.append(el) body = div_tag links = {"internal": [], "external": []} # Extract all internal and external links for a in body.find_all("a", href=True): href = a["href"] url_base = url.split("/")[2] if href.startswith("http") and url_base not in href: links["external"].append({"href": href, "text": a.get_text()}) else: links["internal"].append({"href": href, "text": a.get_text()}) # Remove script, style, and other tags that don't carry useful content from body for tag in body.find_all(["script", "style", "link", "meta", "noscript"]): tag.decompose() # Remove all attributes from remaining tags in body, except for img tags for tag in body.find_all(): if tag.name != "img": tag.attrs = {} # Extract all img tgas int0 [{src: '', alt: ''}]
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/content_scraping_strategy.py
crawl4ai/content_scraping_strategy.py
import re from itertools import chain from abc import ABC, abstractmethod from typing import Dict, Any, Optional from bs4 import BeautifulSoup import asyncio import requests from .config import ( MIN_WORD_THRESHOLD, IMAGE_DESCRIPTION_MIN_WORD_THRESHOLD, IMAGE_SCORE_THRESHOLD, ONLY_TEXT_ELIGIBLE_TAGS, IMPORTANT_ATTRS, SOCIAL_MEDIA_DOMAINS, ) from bs4 import NavigableString, Comment from bs4 import PageElement, Tag from urllib.parse import urljoin from requests.exceptions import InvalidSchema from .utils import ( extract_metadata, normalize_url, is_external_url, get_base_domain, extract_metadata_using_lxml, extract_page_context, calculate_link_intrinsic_score, ) from lxml import etree from lxml import html as lhtml from typing import List from .models import ScrapingResult, MediaItem, Link, Media, Links import copy # Pre-compile regular expressions for Open Graph and Twitter metadata OG_REGEX = re.compile(r"^og:") TWITTER_REGEX = re.compile(r"^twitter:") DIMENSION_REGEX = re.compile(r"(\d+)(\D*)") # Function to parse srcset def parse_srcset(s: str) -> List[Dict]: if not s: return [] variants = [] for part in s.split(","): part = part.strip() if not part: continue parts = part.split() if len(parts) >= 1: url = parts[0] width = ( parts[1].rstrip("w").split('.')[0] if len(parts) > 1 and parts[1].endswith("w") else None ) variants.append({"url": url, "width": width}) return variants # Function to parse image height/width value and units def parse_dimension(dimension): if dimension: # match = re.match(r"(\d+)(\D*)", dimension) match = DIMENSION_REGEX.match(dimension) if match: number = int(match.group(1)) unit = match.group(2) or "px" # Default unit is 'px' if not specified return number, unit return None, None # Fetch image file metadata to extract size and extension def fetch_image_file_size(img, base_url): # If src is relative path construct full URL, if not it may be CDN URL img_url = urljoin(base_url, img.get("src")) try: response = requests.head(img_url) if response.status_code == 200: return response.headers.get("Content-Length", None) else: print(f"Failed to retrieve file size for {img_url}") return None except InvalidSchema: return None finally: return class ContentScrapingStrategy(ABC): @abstractmethod def scrap(self, url: str, html: str, **kwargs) -> ScrapingResult: pass @abstractmethod async def ascrap(self, url: str, html: str, **kwargs) -> ScrapingResult: pass class LXMLWebScrapingStrategy(ContentScrapingStrategy): """ LXML-based implementation for fast web content scraping. This is the primary scraping strategy in Crawl4AI, providing high-performance HTML parsing and content extraction using the lxml library. Note: WebScrapingStrategy is now an alias for this class to maintain backward compatibility. """ def __init__(self, logger=None): self.logger = logger self.DIMENSION_REGEX = re.compile(r"(\d+)(\D*)") self.BASE64_PATTERN = re.compile(r'data:image/[^;]+;base64,([^"]+)') def _log(self, level, message, tag="SCRAPE", **kwargs): """Helper method to safely use logger.""" if self.logger: log_method = getattr(self.logger, level) log_method(message=message, tag=tag, **kwargs) def scrap(self, url: str, html: str, **kwargs) -> ScrapingResult: """ Main entry point for content scraping. Args: url (str): The URL of the page to scrape. html (str): The HTML content of the page. **kwargs: Additional keyword arguments. Returns: ScrapingResult: A structured result containing the scraped content. """ actual_url = kwargs.get("redirected_url", url) raw_result = self._scrap(actual_url, html, **kwargs) if raw_result is None: return ScrapingResult( cleaned_html="", success=False, media=Media(), links=Links(), metadata={}, ) # Convert media items media = Media( images=[ MediaItem(**img) for img in raw_result.get("media", {}).get("images", []) if img ], videos=[ MediaItem(**vid) for vid in raw_result.get("media", {}).get("videos", []) if vid ], audios=[ MediaItem(**aud) for aud in raw_result.get("media", {}).get("audios", []) if aud ], tables=raw_result.get("media", {}).get("tables", []) ) # Convert links links = Links( internal=[ Link(**link) for link in raw_result.get("links", {}).get("internal", []) if link ], external=[ Link(**link) for link in raw_result.get("links", {}).get("external", []) if link ], ) return ScrapingResult( cleaned_html=raw_result.get("cleaned_html", ""), success=raw_result.get("success", False), media=media, links=links, metadata=raw_result.get("metadata", {}), ) async def ascrap(self, url: str, html: str, **kwargs) -> ScrapingResult: """ Main entry point for asynchronous content scraping. Args: url (str): The URL of the page to scrape. html (str): The HTML content of the page. **kwargs: Additional keyword arguments. Returns: ScrapingResult: A structured result containing the scraped content. """ return await asyncio.to_thread(self.scrap, url, html, **kwargs) def process_element(self, url, element: lhtml.HtmlElement, **kwargs) -> Dict[str, Any]: """ Process an HTML element. How it works: 1. Check if the element is an image, video, or audio. 2. Extract the element's attributes and content. 3. Process the element based on its type. 4. Return the processed element information. Args: url (str): The URL of the page containing the element. element (lhtml.HtmlElement): The HTML element to process. **kwargs: Additional keyword arguments. Returns: dict: A dictionary containing the processed element information. """ media = {"images": [], "videos": [], "audios": [], "tables": []} internal_links_dict = {} external_links_dict = {} self._process_element( url, element, media, internal_links_dict, external_links_dict, **kwargs ) return { "media": media, "internal_links_dict": internal_links_dict, "external_links_dict": external_links_dict, } def _process_element( self, url: str, element: lhtml.HtmlElement, media: Dict[str, List], internal_links_dict: Dict[str, Any], external_links_dict: Dict[str, Any], page_context: dict = None, **kwargs, ) -> bool: base_domain = kwargs.get("base_domain", get_base_domain(url)) exclude_domains = set(kwargs.get("exclude_domains", [])) # Process links try: base_element = element.xpath("//head/base[@href]") if base_element: base_href = base_element[0].get("href", "").strip() if base_href: url = base_href except Exception as e: self._log("error", f"Error extracting base URL: {str(e)}", "SCRAPE") pass for link in element.xpath(".//a[@href]"): href = link.get("href", "").strip() if not href: continue try: normalized_href = normalize_url( href, url, preserve_https=kwargs.get('preserve_https_for_internal_links', False), original_scheme=kwargs.get('original_scheme') ) link_data = { "href": normalized_href, "text": link.text_content().strip(), "title": link.get("title", "").strip(), "base_domain": base_domain, } # Add intrinsic scoring if enabled if kwargs.get("score_links", False) and page_context is not None: try: intrinsic_score = calculate_link_intrinsic_score( link_text=link_data["text"], url=normalized_href, title_attr=link_data["title"], class_attr=link.get("class", ""), rel_attr=link.get("rel", ""), page_context=page_context ) link_data["intrinsic_score"] = intrinsic_score except Exception: # Fail gracefully - assign default score link_data["intrinsic_score"] = 0 else: # No scoring enabled - assign infinity (all links equal priority) link_data["intrinsic_score"] = 0 is_external = is_external_url(normalized_href, base_domain) if is_external: link_base_domain = get_base_domain(normalized_href) link_data["base_domain"] = link_base_domain if ( kwargs.get("exclude_external_links", False) or link_base_domain in exclude_domains ): link.getparent().remove(link) continue if normalized_href not in external_links_dict: external_links_dict[normalized_href] = link_data else: if normalized_href not in internal_links_dict: internal_links_dict[normalized_href] = link_data except Exception as e: self._log("error", f"Error processing link: {str(e)}", "SCRAPE") continue # Process images images = element.xpath(".//img") total_images = len(images) for idx, img in enumerate(images): src = img.get("src") or "" img_domain = get_base_domain(src) # Decide if we need to exclude this image # 1) If its domain is in exclude_domains, remove. # 2) Or if exclude_external_images=True and it's an external domain, remove. if (img_domain in exclude_domains) or ( kwargs.get("exclude_external_images", False) and is_external_url(src, base_domain) ): parent = img.getparent() if parent is not None: parent.remove(img) continue # Otherwise, process the image as usual. try: processed_images = self.process_image( img, url, idx, total_images, **kwargs ) if processed_images: media["images"].extend(processed_images) except Exception as e: self._log("error", f"Error processing image: {str(e)}", "SCRAPE") # Process videos and audios for media_type in ["video", "audio"]: for elem in element.xpath(f".//{media_type}"): media_info = { "src": elem.get("src"), "alt": elem.get("alt"), "type": media_type, "description": self.find_closest_parent_with_useful_text( elem, **kwargs ), } media[f"{media_type}s"].append(media_info) # Process source tags within media elements for source in elem.xpath(".//source"): if src := source.get("src"): media[f"{media_type}s"].append({**media_info, "src": src}) # Clean up unwanted elements if kwargs.get("remove_forms", False): for form in element.xpath(".//form"): form.getparent().remove(form) if excluded_tags := kwargs.get("excluded_tags", []): for tag in excluded_tags: for elem in element.xpath(f".//{tag}"): elem.getparent().remove(elem) if excluded_selector := kwargs.get("excluded_selector", ""): try: for elem in element.cssselect(excluded_selector): elem.getparent().remove(elem) except Exception: pass # Invalid selector return True def find_closest_parent_with_useful_text( self, element: lhtml.HtmlElement, **kwargs ) -> Optional[str]: image_description_min_word_threshold = kwargs.get( "image_description_min_word_threshold", IMAGE_DESCRIPTION_MIN_WORD_THRESHOLD ) current = element while current is not None: if ( current.text and len(current.text_content().split()) >= image_description_min_word_threshold ): return current.text_content().strip() current = current.getparent() return None def flatten_nested_elements(self, element: lhtml.HtmlElement) -> lhtml.HtmlElement: """Flatten nested elements of the same type in LXML tree""" if len(element) == 1 and element.tag == element[0].tag: return self.flatten_nested_elements(element[0]) for child in element: child_idx = element.index(child) flattened_child = self.flatten_nested_elements(child) if flattened_child is not child: # Only replace if actually flattened element[child_idx] = flattened_child return element def process_image( self, img: lhtml.HtmlElement, url: str, index: int, total_images: int, **kwargs ) -> Optional[List[Dict]]: # Quick validation checks style = img.get("style", "") alt = img.get("alt", "") src = img.get("src", "") data_src = img.get("data-src", "") srcset = img.get("srcset", "") data_srcset = img.get("data-srcset", "") if "display:none" in style: return None parent = img.getparent() if parent.tag in ["button", "input"]: return None parent_classes = parent.get("class", "").split() if any( "button" in cls or "icon" in cls or "logo" in cls for cls in parent_classes ): return None # If src is in class or alt, likely an icon if (src and any(c in src for c in ["button", "icon", "logo"])) or ( alt and any(c in alt for c in ["button", "icon", "logo"]) ): return None # Score calculation score = 0 if (width := img.get("width")) and width.isdigit(): score += 1 if int(width) > 150 else 0 if (height := img.get("height")) and height.isdigit(): score += 1 if int(height) > 150 else 0 if alt: score += 1 score += index / total_images < 0.5 # Check formats in all possible sources image_formats = {"jpg", "jpeg", "png", "webp", "avif", "gif"} detected_format = None for url in [src, data_src, srcset, data_srcset]: if url: format_matches = [fmt for fmt in image_formats if fmt in url.lower()] if format_matches: detected_format = format_matches[0] score += 1 break if srcset or data_srcset: score += 1 if picture := img.xpath("./ancestor::picture[1]"): score += 1 if score <= kwargs.get("image_score_threshold", IMAGE_SCORE_THRESHOLD): return None # Process image variants unique_urls = set() image_variants = [] base_info = { "alt": alt, "desc": self.find_closest_parent_with_useful_text(img, **kwargs), "score": score, "type": "image", "group_id": index, "format": detected_format, } def add_variant(src: str, width: Optional[str] = None): if src and not src.startswith("data:") and src not in unique_urls: unique_urls.add(src) variant = {**base_info, "src": src} if width: variant["width"] = width image_variants.append(variant) # Add variants from different sources add_variant(src) add_variant(data_src) for srcset_attr in [srcset, data_srcset]: if srcset_attr: for source in parse_srcset(srcset_attr): add_variant(source["url"], source["width"]) # Handle picture element if picture: for source in picture[0].xpath(".//source[@srcset]"): if source_srcset := source.get("srcset"): for src_data in parse_srcset(source_srcset): add_variant(src_data["url"], src_data["width"]) # Check framework-specific attributes for attr, value in img.attrib.items(): if ( attr.startswith("data-") and ("src" in attr or "srcset" in attr) and "http" in value ): add_variant(value) return image_variants if image_variants else None def remove_empty_elements_fast(self, root, word_count_threshold=5): """ Remove elements that fall below the desired word threshold in a single pass from the bottom up. Skips non-element nodes like HtmlComment and bypasses certain tags that are allowed to have no content. """ bypass_tags = { "a", "img", "br", "hr", "input", "meta", "link", "source", "track", "wbr", "tr", "td", "th", } for el in reversed(list(root.iterdescendants())): if not isinstance(el, lhtml.HtmlElement): continue if el.tag in bypass_tags: continue # Skip elements inside <pre> or <code> tags where whitespace is significant # This preserves whitespace-only spans (e.g., <span class="w"> </span>) in code blocks is_in_code_block = False ancestor = el.getparent() while ancestor is not None: if ancestor.tag in ("pre", "code"): is_in_code_block = True break ancestor = ancestor.getparent() if is_in_code_block: continue text_content = (el.text_content() or "").strip() if ( len(text_content.split()) < word_count_threshold and not el.getchildren() ): parent = el.getparent() if parent is not None: parent.remove(el) return root def remove_unwanted_attributes_fast( self, root: lhtml.HtmlElement, important_attrs=None, keep_data_attributes=False ) -> lhtml.HtmlElement: """ Removes all attributes from each element (including root) except those in `important_attrs`. If `keep_data_attributes=True`, also retain any attribute starting with 'data-'. Returns the same root element, mutated in-place, for fluent usage. """ if important_attrs is None: important_attrs = set(IMPORTANT_ATTRS) # If you want to handle the root as well, use 'include_self=True' # so you don't miss attributes on the top-level element. # Manually include the root, then all its descendants for el in chain((root,), root.iterdescendants()): # We only remove attributes on HtmlElement nodes, skip comments or text nodes if not isinstance(el, lhtml.HtmlElement): continue old_attribs = dict(el.attrib) new_attribs = {} for attr_name, attr_val in old_attribs.items(): # If it's an important attribute, keep it if attr_name in important_attrs: new_attribs[attr_name] = attr_val # Or if keep_data_attributes is True and it's a 'data-*' attribute elif keep_data_attributes and attr_name.startswith("data-"): new_attribs[attr_name] = attr_val # Clear old attributes and set the filtered set el.attrib.clear() el.attrib.update(new_attribs) return root def _scrap( self, url: str, html: str, word_count_threshold: int = MIN_WORD_THRESHOLD, css_selector: str = None, target_elements: List[str] = None, **kwargs, ) -> Dict[str, Any]: if not html: return None success = True try: doc = lhtml.document_fromstring(html) # Match BeautifulSoup's behavior of using body or full doc # body = doc.xpath('//body')[0] if doc.xpath('//body') else doc body = doc base_domain = get_base_domain(url) # Extract page context for link scoring (if enabled) - do this BEFORE any removals page_context = None if kwargs.get("score_links", False): try: # Extract title title_elements = doc.xpath('//title') page_title = title_elements[0].text_content() if title_elements else "" # Extract headlines headlines = [] for tag in ['h1', 'h2', 'h3']: elements = doc.xpath(f'//{tag}') for el in elements: text = el.text_content().strip() if text: headlines.append(text) headlines_text = ' '.join(headlines) # Extract meta description meta_desc_elements = doc.xpath('//meta[@name="description"]/@content') meta_description = meta_desc_elements[0] if meta_desc_elements else "" # Create page context page_context = extract_page_context(page_title, headlines_text, meta_description, url) except Exception: page_context = {} # Fail gracefully # Early removal of all images if exclude_all_images is set # This is more efficient in lxml as we remove elements before any processing if kwargs.get("exclude_all_images", False): for img in body.xpath('//img'): if img.getparent() is not None: img.getparent().remove(img) # Add comment removal if kwargs.get("remove_comments", False): comments = body.xpath("//comment()") for comment in comments: comment.getparent().remove(comment) # Handle tag-based removal first excluded_tags = set(kwargs.get("excluded_tags", []) or []) if excluded_tags: for tag in excluded_tags: for element in body.xpath(f".//{tag}"): if element.getparent() is not None: element.getparent().remove(element) # Handle CSS selector-based exclusion excluded_selector = kwargs.get("excluded_selector", "") if excluded_selector: try: for element in body.cssselect(excluded_selector): if element.getparent() is not None: element.getparent().remove(element) except Exception as e: self._log( "error", f"Error with excluded CSS selector: {str(e)}", "SCRAPE" ) # Extract metadata before any content filtering try: meta = extract_metadata_using_lxml( "", doc ) # Using same function as BeautifulSoup version except Exception as e: self._log("error", f"Error extracting metadata: {str(e)}", "SCRAPE") meta = {} content_element = None if target_elements: try: for_content_targeted_element = [] for target_element in target_elements: for_content_targeted_element.extend(body.cssselect(target_element)) content_element = lhtml.Element("div") content_element.extend(copy.deepcopy(for_content_targeted_element)) except Exception as e: self._log("error", f"Error with target element detection: {str(e)}", "SCRAPE") return None else: content_element = body # Remove script and style tags for tag in ["script", "style", "link", "meta", "noscript"]: for element in body.xpath(f".//{tag}"): if element.getparent() is not None: element.getparent().remove(element) # Handle social media and domain exclusions kwargs["exclude_domains"] = set(kwargs.get("exclude_domains", [])) if kwargs.get("exclude_social_media_links", False): kwargs["exclude_social_media_domains"] = set( kwargs.get("exclude_social_media_domains", []) + SOCIAL_MEDIA_DOMAINS ) kwargs["exclude_domains"].update(kwargs["exclude_social_media_domains"]) # Process forms if needed if kwargs.get("remove_forms", False): for form in body.xpath(".//form"): if form.getparent() is not None: form.getparent().remove(form) # Process content media = {"images": [], "videos": [], "audios": [], "tables": []} internal_links_dict = {} external_links_dict = {} self._process_element( url, body, media, internal_links_dict, external_links_dict, page_context=page_context, base_domain=base_domain, **kwargs, ) # Extract tables using the table extraction strategy if provided if 'table' not in excluded_tags: table_extraction = kwargs.get('table_extraction') if table_extraction: # Pass logger to the strategy if it doesn't have one if not table_extraction.logger: table_extraction.logger = self.logger # Extract tables using the strategy extracted_tables = table_extraction.extract_tables(body, **kwargs) media["tables"].extend(extracted_tables) # Handle only_text option if kwargs.get("only_text", False): for tag in ONLY_TEXT_ELIGIBLE_TAGS: for element in body.xpath(f".//{tag}"): if element.text: new_text = lhtml.Element("span") new_text.text = element.text_content() if element.getparent() is not None: element.getparent().replace(element, new_text) # Clean base64 images for img in body.xpath(".//img[@src]"): src = img.get("src", "") if self.BASE64_PATTERN.match(src): img.set("src", self.BASE64_PATTERN.sub("", src)) # Remove empty elements self.remove_empty_elements_fast(body, 1) # Remove unneeded attributes self.remove_unwanted_attributes_fast( body, keep_data_attributes=kwargs.get("keep_data_attributes", False) ) # Generate output HTML cleaned_html = lhtml.tostring( # body, content_element, encoding="unicode", pretty_print=True, method="html", with_tail=False, ).strip() # Create links dictionary in the format expected by LinkPreview links = { "internal": list(internal_links_dict.values()), "external": list(external_links_dict.values()), } # Extract head content for links if configured link_preview_config = kwargs.get("link_preview_config") if link_preview_config is not None: try: import asyncio from .link_preview import LinkPreview from .models import Links, Link verbose = link_preview_config.verbose if verbose: self._log("info", "Starting link head extraction for {internal} internal and {external} external links", params={"internal": len(links["internal"]), "external": len(links["external"])}, tag="LINK_EXTRACT") # Convert dict links to Link objects internal_links = [Link(**link_data) for link_data in links["internal"]] external_links = [Link(**link_data) for link_data in links["external"]] links_obj = Links(internal=internal_links, external=external_links) # Create a config object for LinkPreview class TempCrawlerRunConfig: def __init__(self, link_config, score_links): self.link_preview_config = link_config self.score_links = score_links config = TempCrawlerRunConfig(link_preview_config, kwargs.get("score_links", False)) # Extract head content (run async operation in sync context) async def extract_links(): async with LinkPreview(self.logger) as extractor: return await extractor.extract_link_heads(links_obj, config) # Run the async operation try: # Check if we're already in an async context loop = asyncio.get_running_loop() # If we're in an async context, we need to run in a thread import concurrent.futures with concurrent.futures.ThreadPoolExecutor() as executor: future = executor.submit(asyncio.run, extract_links()) updated_links = future.result() except RuntimeError: # No running loop, we can use asyncio.run directly updated_links = asyncio.run(extract_links()) # Convert back to dict format links["internal"] = [link.dict() for link in updated_links.internal]
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_crawler_strategy.py
crawl4ai/async_crawler_strategy.py
from __future__ import annotations import asyncio import base64 import time from abc import ABC, abstractmethod from typing import Callable, Dict, Any, List, Union from typing import Optional, AsyncGenerator, Final import os from playwright.async_api import Page, Error from playwright.async_api import TimeoutError as PlaywrightTimeoutError from io import BytesIO from PIL import Image, ImageDraw, ImageFont import hashlib import uuid from .js_snippet import load_js_script from .models import AsyncCrawlResponse from .config import SCREENSHOT_HEIGHT_TRESHOLD from .async_configs import BrowserConfig, CrawlerRunConfig, HTTPCrawlerConfig from .async_logger import AsyncLogger from .ssl_certificate import SSLCertificate from .user_agent_generator import ValidUAGenerator from .browser_manager import BrowserManager from .browser_adapter import BrowserAdapter, PlaywrightAdapter, UndetectedAdapter import aiofiles import aiohttp import chardet from aiohttp.client import ClientTimeout from urllib.parse import urlparse from types import MappingProxyType import contextlib from functools import partial class AsyncCrawlerStrategy(ABC): """ Abstract base class for crawler strategies. Subclasses must implement the crawl method. """ @abstractmethod async def crawl(self, url: str, **kwargs) -> AsyncCrawlResponse: pass # 4 + 3 class AsyncPlaywrightCrawlerStrategy(AsyncCrawlerStrategy): """ Crawler strategy using Playwright. Attributes: browser_config (BrowserConfig): Configuration object containing browser settings. logger (AsyncLogger): Logger instance for recording events and errors. _downloaded_files (List[str]): List of downloaded file paths. hooks (Dict[str, Callable]): Dictionary of hooks for custom behavior. browser_manager (BrowserManager): Manager for browser creation and management. Methods: __init__(self, browser_config=None, logger=None, **kwargs): Initialize the AsyncPlaywrightCrawlerStrategy with a browser configuration. __aenter__(self): Start the browser and initialize the browser manager. __aexit__(self, exc_type, exc_val, exc_tb): Close the browser and clean up resources. start(self): Start the browser and initialize the browser manager. close(self): Close the browser and clean up resources. kill_session(self, session_id): Kill a browser session and clean up resources. crawl(self, url, **kwargs): Run the crawler for a single URL. """ def __init__( self, browser_config: BrowserConfig = None, logger: AsyncLogger = None, browser_adapter: BrowserAdapter = None, **kwargs ): """ Initialize the AsyncPlaywrightCrawlerStrategy with a browser configuration. Args: browser_config (BrowserConfig): Configuration object containing browser settings. If None, will be created from kwargs for backwards compatibility. logger: Logger instance for recording events and errors. browser_adapter (BrowserAdapter): Browser adapter for handling browser-specific operations. If None, defaults to PlaywrightAdapter. **kwargs: Additional arguments for backwards compatibility and extending functionality. """ # Initialize browser config, either from provided object or kwargs self.browser_config = browser_config or BrowserConfig.from_kwargs(kwargs) self.logger = logger # Initialize browser adapter self.adapter = browser_adapter or PlaywrightAdapter() # Initialize session management self._downloaded_files = [] # Initialize hooks system self.hooks = { "on_browser_created": None, "on_page_context_created": None, "on_user_agent_updated": None, "on_execution_started": None, "on_execution_ended": None, "before_goto": None, "after_goto": None, "before_return_html": None, "before_retrieve_html": None, } # Initialize browser manager with config self.browser_manager = BrowserManager( browser_config=self.browser_config, logger=self.logger, use_undetected=isinstance(self.adapter, UndetectedAdapter) ) async def __aenter__(self): await self.start() return self async def __aexit__(self, exc_type, exc_val, exc_tb): await self.close() async def start(self): """ Start the browser and initialize the browser manager. """ await self.browser_manager.start() await self.execute_hook( "on_browser_created", self.browser_manager.browser, context=self.browser_manager.default_context, ) async def close(self): """ Close the browser and clean up resources. """ await self.browser_manager.close() # Explicitly reset the static Playwright instance BrowserManager._playwright_instance = None async def kill_session(self, session_id: str): """ Kill a browser session and clean up resources. Args: session_id (str): The ID of the session to kill. Returns: None """ # Log a warning message and no need kill session, in new version auto kill session self.logger.warning( message="Session auto-kill is enabled in the new version. No need to manually kill sessions.", tag="WARNING", ) await self.browser_manager.kill_session(session_id) def set_hook(self, hook_type: str, hook: Callable): """ Set a hook function for a specific hook type. Following are list of hook types: - on_browser_created: Called when a new browser instance is created. - on_page_context_created: Called when a new page context is created. - on_user_agent_updated: Called when the user agent is updated. - on_execution_started: Called when the execution starts. - before_goto: Called before a goto operation. - after_goto: Called after a goto operation. - before_return_html: Called before returning HTML content. - before_retrieve_html: Called before retrieving HTML content. All hooks except on_browser_created accepts a context and a page as arguments and **kwargs. However, on_browser_created accepts a browser and a context as arguments and **kwargs. Args: hook_type (str): The type of the hook. hook (Callable): The hook function to set. Returns: None """ if hook_type in self.hooks: self.hooks[hook_type] = hook else: raise ValueError(f"Invalid hook type: {hook_type}") async def execute_hook(self, hook_type: str, *args, **kwargs): """ Execute a hook function for a specific hook type. Args: hook_type (str): The type of the hook. *args: Variable length positional arguments. **kwargs: Keyword arguments. Returns: The return value of the hook function, if any. """ hook = self.hooks.get(hook_type) if hook: if asyncio.iscoroutinefunction(hook): return await hook(*args, **kwargs) else: return hook(*args, **kwargs) return args[0] if args else None def update_user_agent(self, user_agent: str): """ Update the user agent for the browser. Args: user_agent (str): The new user agent string. Returns: None """ self.user_agent = user_agent def set_custom_headers(self, headers: Dict[str, str]): """ Set custom headers for the browser. Args: headers (Dict[str, str]): A dictionary of headers to set. Returns: None """ self.headers = headers async def smart_wait(self, page: Page, wait_for: str, timeout: float = 30000): """ Wait for a condition in a smart way. This functions works as below: 1. If wait_for starts with 'js:', it assumes it's a JavaScript function and waits for it to return true. 2. If wait_for starts with 'css:', it assumes it's a CSS selector and waits for it to be present. 3. Otherwise, it tries to evaluate wait_for as a JavaScript function and waits for it to return true. 4. If it's not a JavaScript function, it assumes it's a CSS selector and waits for it to be present. This is a more advanced version of the wait_for parameter in CrawlerStrategy.crawl(). Args: page: Playwright page object wait_for (str): The condition to wait for. Can be a CSS selector, a JavaScript function, or explicitly prefixed with 'js:' or 'css:'. timeout (float): Maximum time to wait in milliseconds Returns: None """ wait_for = wait_for.strip() if wait_for.startswith("js:"): # Explicitly specified JavaScript js_code = wait_for[3:].strip() return await self.csp_compliant_wait(page, js_code, timeout) elif wait_for.startswith("css:"): # Explicitly specified CSS selector css_selector = wait_for[4:].strip() try: await page.wait_for_selector(css_selector, timeout=timeout) except Error as e: if "Timeout" in str(e): raise TimeoutError( f"Timeout after {timeout}ms waiting for selector '{css_selector}'" ) else: raise ValueError(f"Invalid CSS selector: '{css_selector}'") else: # Auto-detect based on content if wait_for.startswith("()") or wait_for.startswith("function"): # It's likely a JavaScript function return await self.csp_compliant_wait(page, wait_for, timeout) else: # Assume it's a CSS selector first try: await page.wait_for_selector(wait_for, timeout=timeout) except Error as e: if "Timeout" in str(e): raise TimeoutError( f"Timeout after {timeout}ms waiting for selector '{wait_for}'" ) else: # If it's not a timeout error, it might be an invalid selector # Let's try to evaluate it as a JavaScript function as a fallback try: return await self.csp_compliant_wait( page, f"() => {{{wait_for}}}", timeout ) except Error: raise ValueError( f"Invalid wait_for parameter: '{wait_for}'. " "It should be either a valid CSS selector, a JavaScript function, " "or explicitly prefixed with 'js:' or 'css:'." ) async def csp_compliant_wait( self, page: Page, user_wait_function: str, timeout: float = 30000 ): """ Wait for a condition in a CSP-compliant way. Args: page: Playwright page object user_wait_function: JavaScript function as string that returns boolean timeout: Maximum time to wait in milliseconds Returns: bool: True if condition was met, False if timed out Raises: RuntimeError: If there's an error evaluating the condition """ wrapper_js = f""" async () => {{ const userFunction = {user_wait_function}; const startTime = Date.now(); try {{ while (true) {{ if (await userFunction()) {{ return true; }} if (Date.now() - startTime > {timeout}) {{ return false; // Return false instead of throwing }} await new Promise(resolve => setTimeout(resolve, 100)); }} }} catch (error) {{ throw new Error(`Error evaluating condition: ${{error.message}}`); }} }} """ try: result = await self.adapter.evaluate(page, wrapper_js) return result except Exception as e: if "Error evaluating condition" in str(e): raise RuntimeError(f"Failed to evaluate wait condition: {str(e)}") # For timeout or other cases, just return False return False async def process_iframes(self, page): """ Process iframes on a page. This function will extract the content of each iframe and replace it with a div containing the extracted content. Args: page: Playwright page object Returns: Playwright page object """ # Find all iframes iframes = await page.query_selector_all("iframe") for i, iframe in enumerate(iframes): try: # Add a unique identifier to the iframe await iframe.evaluate(f'(element) => element.id = "iframe-{i}"') # Get the frame associated with this iframe frame = await iframe.content_frame() if frame: # Wait for the frame to load await frame.wait_for_load_state( "load", timeout=30000 ) # 30 seconds timeout # Extract the content of the iframe's body iframe_content = await frame.evaluate( "() => document.body.innerHTML" ) # Generate a unique class name for this iframe class_name = f"extracted-iframe-content-{i}" # Replace the iframe with a div containing the extracted content _iframe = iframe_content.replace("`", "\\`") await self.adapter.evaluate(page, f""" () => {{ const iframe = document.getElementById('iframe-{i}'); const div = document.createElement('div'); div.innerHTML = `{_iframe}`; div.className = '{class_name}'; iframe.replaceWith(div); }} """ ) else: self.logger.warning( message="Could not access content frame for iframe {index}", tag="SCRAPE", params={"index": i}, ) except Exception as e: self.logger.error( message="Error processing iframe {index}: {error}", tag="ERROR", params={"index": i, "error": str(e)}, ) # Return the page object return page async def create_session(self, **kwargs) -> str: """ Creates a new browser session and returns its ID. A browse session is a unique openned page can be reused for multiple crawls. This function is asynchronous and returns a string representing the session ID. Args: **kwargs: Optional keyword arguments to configure the session. Returns: str: The session ID. """ await self.start() session_id = kwargs.get("session_id") or str(uuid.uuid4()) user_agent = kwargs.get("user_agent", self.user_agent) # Use browser_manager to get a fresh page & context assigned to this session_id page, context = await self.browser_manager.get_page(CrawlerRunConfig( session_id=session_id, user_agent=user_agent, **kwargs, )) return session_id async def crawl( self, url: str, config: CrawlerRunConfig, **kwargs ) -> AsyncCrawlResponse: """ Crawls a given URL or processes raw HTML/local file content based on the URL prefix. Args: url (str): The URL to crawl. Supported prefixes: - 'http://' or 'https://': Web URL to crawl. - 'file://': Local file path to process. - 'raw://': Raw HTML content to process. **kwargs: Additional parameters: - 'screenshot' (bool): Whether to take a screenshot. - ... [other existing parameters] Returns: AsyncCrawlResponse: The response containing HTML, headers, status code, and optional screenshot. """ config = config or CrawlerRunConfig.from_kwargs(kwargs) response_headers = {} status_code = 200 # Default for local/raw HTML screenshot_data = None if url.startswith(("http://", "https://", "view-source:")): return await self._crawl_web(url, config) elif url.startswith("file://"): # initialize empty lists for console messages captured_console = [] # Process local file local_file_path = url[7:] # Remove 'file://' prefix if not os.path.exists(local_file_path): raise FileNotFoundError(f"Local file not found: {local_file_path}") with open(local_file_path, "r", encoding="utf-8") as f: html = f.read() if config.screenshot: screenshot_data = await self._generate_screenshot_from_html(html) if config.capture_console_messages: page, context = await self.browser_manager.get_page(crawlerRunConfig=config) captured_console = await self._capture_console_messages(page, url) return AsyncCrawlResponse( html=html, response_headers=response_headers, status_code=status_code, screenshot=screenshot_data, get_delayed_content=None, console_messages=captured_console, ) ##### # Since both "raw:" and "raw://" start with "raw:", the first condition is always true for both, so "raw://" will be sliced as "//...", which is incorrect. # Fix: Check for "raw://" first, then "raw:" # Also, the prefix "raw://" is actually 6 characters long, not 7, so it should be sliced accordingly: url[6:] ##### elif url.startswith("raw://") or url.startswith("raw:"): # Process raw HTML content # raw_html = url[4:] if url[:4] == "raw:" else url[7:] raw_html = url[6:] if url.startswith("raw://") else url[4:] html = raw_html if config.screenshot: screenshot_data = await self._generate_screenshot_from_html(html) return AsyncCrawlResponse( html=html, response_headers=response_headers, status_code=status_code, screenshot=screenshot_data, get_delayed_content=None, ) else: raise ValueError( "URL must start with 'http://', 'https://', 'file://', or 'raw:'" ) async def _crawl_web( self, url: str, config: CrawlerRunConfig ) -> AsyncCrawlResponse: """ Internal method to crawl web URLs with the specified configuration. Includes optional network and console capturing. Args: url (str): The web URL to crawl config (CrawlerRunConfig): Configuration object controlling the crawl behavior Returns: AsyncCrawlResponse: The response containing HTML, headers, status code, and optional data """ config.url = url response_headers = {} execution_result = None status_code = None redirected_url = url # Reset downloaded files list for new crawl self._downloaded_files = [] # Initialize capture lists captured_requests = [] captured_console = [] # Handle user agent with magic mode user_agent_to_override = config.user_agent if user_agent_to_override: self.browser_config.user_agent = user_agent_to_override elif config.magic or config.user_agent_mode == "random": self.browser_config.user_agent = ValidUAGenerator().generate( **(config.user_agent_generator_config or {}) ) # Get page for session page, context = await self.browser_manager.get_page(crawlerRunConfig=config) # await page.goto(URL) # Add default cookie # await context.add_cookies( # [{"name": "cookiesEnabled", "value": "true", "url": url}] # ) # Handle navigator overrides if config.override_navigator or config.simulate_user or config.magic: await context.add_init_script(load_js_script("navigator_overrider")) # Call hook after page creation await self.execute_hook("on_page_context_created", page, context=context, config=config) # Network Request Capturing if config.capture_network_requests: async def handle_request_capture(request): try: post_data_str = None try: # Be cautious with large post data post_data = request.post_data_buffer if post_data: # Attempt to decode, fallback to base64 or size indication try: post_data_str = post_data.decode('utf-8', errors='replace') except UnicodeDecodeError: post_data_str = f"[Binary data: {len(post_data)} bytes]" except Exception: post_data_str = "[Error retrieving post data]" captured_requests.append({ "event_type": "request", "url": request.url, "method": request.method, "headers": dict(request.headers), # Convert Header dict "post_data": post_data_str, "resource_type": request.resource_type, "is_navigation_request": request.is_navigation_request(), "timestamp": time.time() }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing request details for {request.url}: {e}", tag="CAPTURE") captured_requests.append({"event_type": "request_capture_error", "url": request.url, "error": str(e), "timestamp": time.time()}) async def handle_response_capture(response): try: try: # body = await response.body() # json_body = await response.json() text_body = await response.text() except Exception as e: body = None # json_body = None # text_body = None captured_requests.append({ "event_type": "response", "url": response.url, "status": response.status, "status_text": response.status_text, "headers": dict(response.headers), # Convert Header dict "from_service_worker": response.from_service_worker, "request_timing": response.request.timing, # Detailed timing info "timestamp": time.time(), "body" : { # "raw": body, # "json": json_body, "text": text_body } }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing response details for {response.url}: {e}", tag="CAPTURE") captured_requests.append({"event_type": "response_capture_error", "url": response.url, "error": str(e), "timestamp": time.time()}) async def handle_request_failed_capture(request): try: captured_requests.append({ "event_type": "request_failed", "url": request.url, "method": request.method, "resource_type": request.resource_type, "failure_text": str(request.failure) if request.failure else "Unknown failure", "timestamp": time.time() }) except Exception as e: if self.logger: self.logger.warning(f"Error capturing request failed details for {request.url}: {e}", tag="CAPTURE") captured_requests.append({"event_type": "request_failed_capture_error", "url": request.url, "error": str(e), "timestamp": time.time()}) page.on("request", handle_request_capture) page.on("response", handle_response_capture) page.on("requestfailed", handle_request_failed_capture) # Console Message Capturing handle_console = None handle_error = None if config.capture_console_messages: # Set up console capture using adapter handle_console = await self.adapter.setup_console_capture(page, captured_console) handle_error = await self.adapter.setup_error_capture(page, captured_console) # Set up console logging if requested # Note: For undetected browsers, console logging won't work directly # but captured messages can still be logged after retrieval try: # Get SSL certificate information if requested and URL is HTTPS ssl_cert = None if config.fetch_ssl_certificate: ssl_cert = SSLCertificate.from_url(url) # Set up download handling if self.browser_config.accept_downloads: page.on( "download", lambda download: asyncio.create_task( self._handle_download(download) ), ) # Handle page navigation and content loading if not config.js_only: await self.execute_hook("before_goto", page, context=context, url=url, config=config) try: # Generate a unique nonce for this request if config.experimental.get("use_csp_nonce", False): nonce = hashlib.sha256(os.urandom(32)).hexdigest() # Add CSP headers to the request await page.set_extra_http_headers( { "Content-Security-Policy": f"default-src 'self'; script-src 'self' 'nonce-{nonce}' 'strict-dynamic'" } ) response = await page.goto( url, wait_until=config.wait_until, timeout=config.page_timeout ) redirected_url = page.url except Error as e: # Allow navigation to be aborted when downloading files # This is expected behavior for downloads in some browser engines if 'net::ERR_ABORTED' in str(e) and self.browser_config.accept_downloads: self.logger.info( message=f"Navigation aborted, likely due to file download: {url}", tag="GOTO", params={"url": url}, ) response = None else: raise RuntimeError(f"Failed on navigating ACS-GOTO:\n{str(e)}") await self.execute_hook( "after_goto", page, context=context, url=url, response=response, config=config ) # ────────────────────────────────────────────────────────────── # Walk the redirect chain. Playwright returns only the last # hop, so we trace the `request.redirected_from` links until the # first response that differs from the final one and surface its # status-code. # ────────────────────────────────────────────────────────────── if response is None: status_code = 200 response_headers = {} else: first_resp = response req = response.request while req and req.redirected_from: prev_req = req.redirected_from prev_resp = await prev_req.response() if prev_resp: # keep earliest first_resp = prev_resp req = prev_req status_code = first_resp.status response_headers = first_resp.headers # if response is None: # status_code = 200 # response_headers = {} # else: # status_code = response.status # response_headers = response.headers else: status_code = 200 response_headers = {} # Wait for body element and visibility try: await page.wait_for_selector("body", state="attached", timeout=30000) # Use the new check_visibility function with csp_compliant_wait is_visible = await self.csp_compliant_wait( page, """() => { const element = document.body; if (!element) return false; const style = window.getComputedStyle(element); const isVisible = style.display !== 'none' && style.visibility !== 'hidden' && style.opacity !== '0'; return isVisible; }""", timeout=30000, ) if not is_visible and not config.ignore_body_visibility: visibility_info = await self.check_visibility(page) raise Error(f"Body element is hidden: {visibility_info}") except Error: visibility_info = await self.check_visibility(page) if self.browser_config.verbose: self.logger.debug( message="Body visibility info: {info}", tag="DEBUG", params={"info": visibility_info}, ) if not config.ignore_body_visibility: raise Error(f"Body element is hidden: {visibility_info}") # try: # await page.wait_for_selector("body", state="attached", timeout=30000) # await page.wait_for_function( # """ # () => { # const body = document.body; # const style = window.getComputedStyle(body);
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/user_agent_generator.py
crawl4ai/user_agent_generator.py
import random from typing import Optional, Literal, List, Dict, Tuple import re from abc import ABC, abstractmethod from fake_useragent import UserAgent import requests from lxml import html import json from typing import Union class UAGen(ABC): @abstractmethod def generate(self, browsers: Optional[List[str]] = None, os: Optional[Union[str, List[str]]] = None, min_version: float = 0.0, platforms: Optional[Union[str, List[str]]] = None, pct_threshold: Optional[float] = None, fallback: str = "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 Chrome/116.0.0.0 Safari/537.36") -> Union[str, Dict]: pass @staticmethod def generate_client_hints( user_agent: str) -> str: """Generate Sec-CH-UA header value based on user agent string""" def _parse_user_agent(user_agent: str) -> Dict[str, str]: """Parse a user agent string to extract browser and version information""" browsers = { "chrome": r"Chrome/(\d+)", "edge": r"Edg/(\d+)", "safari": r"Version/(\d+)", "firefox": r"Firefox/(\d+)", } result = {} for browser, pattern in browsers.items(): match = re.search(pattern, user_agent) if match: result[browser] = match.group(1) return result browsers = _parse_user_agent(user_agent) # Client hints components hints = [] # Handle different browser combinations if "chrome" in browsers: hints.append(f'"Chromium";v="{browsers["chrome"]}"') hints.append('"Not_A Brand";v="8"') if "edge" in browsers: hints.append(f'"Microsoft Edge";v="{browsers["edge"]}"') else: hints.append(f'"Google Chrome";v="{browsers["chrome"]}"') elif "firefox" in browsers: # Firefox doesn't typically send Sec-CH-UA return '""' elif "safari" in browsers: # Safari's format for client hints hints.append(f'"Safari";v="{browsers["safari"]}"') hints.append('"Not_A Brand";v="8"') return ", ".join(hints) class ValidUAGenerator(UAGen): def __init__(self): self.ua = UserAgent() def generate(self, browsers: Optional[List[str]] = None, os: Optional[Union[str, List[str]]] = None, min_version: float = 0.0, platforms: Optional[Union[str, List[str]]] = None, pct_threshold: Optional[float] = None, fallback: str = "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 Chrome/116.0.0.0 Safari/537.36") -> str: self.ua = UserAgent( browsers=browsers or ['Chrome', 'Firefox', 'Edge'], os=os or ['Windows', 'Mac OS X'], min_version=min_version, platforms=platforms or ['desktop'], fallback=fallback ) return self.ua.random class OnlineUAGenerator(UAGen): def __init__(self): self.agents = [] self._fetch_agents() def _fetch_agents(self): try: response = requests.get( 'https://www.useragents.me/', timeout=5, headers={'Accept': 'text/html,application/xhtml+xml'} ) response.raise_for_status() tree = html.fromstring(response.content) json_text = tree.cssselect('#most-common-desktop-useragents-json-csv > div:nth-child(1) > textarea')[0].text self.agents = json.loads(json_text) except Exception as e: print(f"Error fetching agents: {e}") def generate(self, browsers: Optional[List[str]] = None, os: Optional[Union[str, List[str]]] = None, min_version: float = 0.0, platforms: Optional[Union[str, List[str]]] = None, pct_threshold: Optional[float] = None, fallback: str = "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 Chrome/116.0.0.0 Safari/537.36") -> Dict: if not self.agents: self._fetch_agents() filtered_agents = self.agents if pct_threshold: filtered_agents = [a for a in filtered_agents if a['pct'] >= pct_threshold] if browsers: filtered_agents = [a for a in filtered_agents if any(b.lower() in a['ua'].lower() for b in browsers)] if os: os_list = [os] if isinstance(os, str) else os filtered_agents = [a for a in filtered_agents if any(o.lower() in a['ua'].lower() for o in os_list)] if platforms: platform_list = [platforms] if isinstance(platforms, str) else platforms filtered_agents = [a for a in filtered_agents if any(p.lower() in a['ua'].lower() for p in platform_list)] return filtered_agents[0] if filtered_agents else {'ua': fallback, 'pct': 0} class UserAgentGenerator(): """ Generate random user agents with specified constraints. Attributes: desktop_platforms (dict): A dictionary of possible desktop platforms and their corresponding user agent strings. mobile_platforms (dict): A dictionary of possible mobile platforms and their corresponding user agent strings. browser_combinations (dict): A dictionary of possible browser combinations and their corresponding user agent strings. rendering_engines (dict): A dictionary of possible rendering engines and their corresponding user agent strings. chrome_versions (list): A list of possible Chrome browser versions. firefox_versions (list): A list of possible Firefox browser versions. edge_versions (list): A list of possible Edge browser versions. safari_versions (list): A list of possible Safari browser versions. ios_versions (list): A list of possible iOS browser versions. android_versions (list): A list of possible Android browser versions. Methods: generate_user_agent( platform: Literal["desktop", "mobile"] = "desktop", browser: str = "chrome", rendering_engine: str = "chrome_webkit", chrome_version: Optional[str] = None, firefox_version: Optional[str] = None, edge_version: Optional[str] = None, safari_version: Optional[str] = None, ios_version: Optional[str] = None, android_version: Optional[str] = None ): Generates a random user agent string based on the specified parameters. """ def __init__(self): # Previous platform definitions remain the same... self.desktop_platforms = { "windows": { "10_64": "(Windows NT 10.0; Win64; x64)", "10_32": "(Windows NT 10.0; WOW64)", }, "macos": { "intel": "(Macintosh; Intel Mac OS X 10_15_7)", "newer": "(Macintosh; Intel Mac OS X 10.15; rv:109.0)", }, "linux": { "generic": "(X11; Linux x86_64)", "ubuntu": "(X11; Ubuntu; Linux x86_64)", "chrome_os": "(X11; CrOS x86_64 14541.0.0)", }, } self.mobile_platforms = { "android": { "samsung": "(Linux; Android 13; SM-S901B)", "pixel": "(Linux; Android 12; Pixel 6)", "oneplus": "(Linux; Android 13; OnePlus 9 Pro)", "xiaomi": "(Linux; Android 12; M2102J20SG)", }, "ios": { "iphone": "(iPhone; CPU iPhone OS 16_5 like Mac OS X)", "ipad": "(iPad; CPU OS 16_5 like Mac OS X)", }, } # Browser Combinations self.browser_combinations = { 1: [["chrome"], ["firefox"], ["safari"], ["edge"]], 2: [["gecko", "firefox"], ["chrome", "safari"], ["webkit", "safari"]], 3: [["chrome", "safari", "edge"], ["webkit", "chrome", "safari"]], } # Rendering Engines with versions self.rendering_engines = { "chrome_webkit": "AppleWebKit/537.36", "safari_webkit": "AppleWebKit/605.1.15", "gecko": [ # Added Gecko versions "Gecko/20100101", "Gecko/20100101", # Firefox usually uses this constant version "Gecko/2010010", ], } # Browser Versions self.chrome_versions = [ "Chrome/119.0.6045.199", "Chrome/118.0.5993.117", "Chrome/117.0.5938.149", "Chrome/116.0.5845.187", "Chrome/115.0.5790.171", ] self.edge_versions = [ "Edg/119.0.2151.97", "Edg/118.0.2088.76", "Edg/117.0.2045.47", "Edg/116.0.1938.81", "Edg/115.0.1901.203", ] self.safari_versions = [ "Safari/537.36", # For Chrome-based "Safari/605.1.15", "Safari/604.1", "Safari/602.1", "Safari/601.5.17", ] # Added Firefox versions self.firefox_versions = [ "Firefox/119.0", "Firefox/118.0.2", "Firefox/117.0.1", "Firefox/116.0", "Firefox/115.0.3", "Firefox/114.0.2", "Firefox/113.0.1", "Firefox/112.0", "Firefox/111.0.1", "Firefox/110.0", ] def get_browser_stack(self, num_browsers: int = 1) -> List[str]: """ Get a valid combination of browser versions. How it works: 1. Check if the number of browsers is supported. 2. Randomly choose a combination of browsers. 3. Iterate through the combination and add browser versions. 4. Return the browser stack. Args: num_browsers: Number of browser specifications (1-3) Returns: List[str]: A list of browser versions. """ if num_browsers not in self.browser_combinations: raise ValueError(f"Unsupported number of browsers: {num_browsers}") combination = random.choice(self.browser_combinations[num_browsers]) browser_stack = [] for browser in combination: if browser == "chrome": browser_stack.append(random.choice(self.chrome_versions)) elif browser == "firefox": browser_stack.append(random.choice(self.firefox_versions)) elif browser == "safari": browser_stack.append(random.choice(self.safari_versions)) elif browser == "edge": browser_stack.append(random.choice(self.edge_versions)) elif browser == "gecko": browser_stack.append(random.choice(self.rendering_engines["gecko"])) elif browser == "webkit": browser_stack.append(self.rendering_engines["chrome_webkit"]) return browser_stack def generate( self, device_type: Optional[Literal["desktop", "mobile"]] = None, os_type: Optional[str] = None, device_brand: Optional[str] = None, browser_type: Optional[Literal["chrome", "edge", "safari", "firefox"]] = None, num_browsers: int = 3, ) -> str: """ Generate a random user agent with specified constraints. Args: device_type: 'desktop' or 'mobile' os_type: 'windows', 'macos', 'linux', 'android', 'ios' device_brand: Specific device brand browser_type: 'chrome', 'edge', 'safari', or 'firefox' num_browsers: Number of browser specifications (1-3) """ # Get platform string platform = self.get_random_platform(device_type, os_type, device_brand) # Start with Mozilla components = ["Mozilla/5.0", platform] # Add browser stack browser_stack = self.get_browser_stack(num_browsers) # Add appropriate legacy token based on browser stack if "Firefox" in str(browser_stack) or browser_type == "firefox": components.append(random.choice(self.rendering_engines["gecko"])) elif "Chrome" in str(browser_stack) or "Safari" in str(browser_stack) or browser_type == "chrome": components.append(self.rendering_engines["chrome_webkit"]) components.append("(KHTML, like Gecko)") elif "Edge" in str(browser_stack) or browser_type == "edge": components.append(self.rendering_engines["safari_webkit"]) components.append("(KHTML, like Gecko)") elif "Safari" in str(browser_stack) or browser_type == "safari": components.append(self.rendering_engines["chrome_webkit"]) components.append("(KHTML, like Gecko)") # Add browser versions components.extend(browser_stack) return " ".join(components) def generate_with_client_hints(self, **kwargs) -> Tuple[str, str]: """Generate both user agent and matching client hints""" user_agent = self.generate(**kwargs) client_hints = self.generate_client_hints(user_agent) return user_agent, client_hints def get_random_platform(self, device_type, os_type, device_brand): """Helper method to get random platform based on constraints""" platforms = ( self.desktop_platforms if device_type == "desktop" else self.mobile_platforms if device_type == "mobile" else {**self.desktop_platforms, **self.mobile_platforms} ) if os_type: for platform_group in [self.desktop_platforms, self.mobile_platforms]: if os_type in platform_group: platforms = {os_type: platform_group[os_type]} break os_key = random.choice(list(platforms.keys())) if device_brand and device_brand in platforms[os_key]: return platforms[os_key][device_brand] return random.choice(list(platforms[os_key].values())) def parse_user_agent(self, user_agent: str) -> Dict[str, str]: """Parse a user agent string to extract browser and version information""" browsers = { "chrome": r"Chrome/(\d+)", "edge": r"Edg/(\d+)", "safari": r"Version/(\d+)", "firefox": r"Firefox/(\d+)", } result = {} for browser, pattern in browsers.items(): match = re.search(pattern, user_agent) if match: result[browser] = match.group(1) return result def generate_client_hints(self, user_agent: str) -> str: """Generate Sec-CH-UA header value based on user agent string""" browsers = self.parse_user_agent(user_agent) # Client hints components hints = [] # Handle different browser combinations if "chrome" in browsers: hints.append(f'"Chromium";v="{browsers["chrome"]}"') hints.append('"Not_A Brand";v="8"') if "edge" in browsers: hints.append(f'"Microsoft Edge";v="{browsers["edge"]}"') else: hints.append(f'"Google Chrome";v="{browsers["chrome"]}"') elif "firefox" in browsers: # Firefox doesn't typically send Sec-CH-UA return '""' elif "safari" in browsers: # Safari's format for client hints hints.append(f'"Safari";v="{browsers["safari"]}"') hints.append('"Not_A Brand";v="8"') return ", ".join(hints) # Example usage: if __name__ == "__main__": # Usage example: generator = ValidUAGenerator() ua = generator.generate() print(ua) generator = OnlineUAGenerator() ua = generator.generate() print(ua)
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
false
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/adaptive_crawler copy.py
crawl4ai/adaptive_crawler copy.py
""" Adaptive Web Crawler for Crawl4AI This module implements adaptive information foraging for efficient web crawling. It determines when sufficient information has been gathered to answer a query, avoiding unnecessary crawls while ensuring comprehensive coverage. """ from abc import ABC, abstractmethod from typing import Dict, List, Optional, Set, Tuple, Any, Union from dataclasses import dataclass, field import asyncio import pickle import os import json import math from collections import defaultdict, Counter import re from pathlib import Path from crawl4ai.async_webcrawler import AsyncWebCrawler from crawl4ai.async_configs import CrawlerRunConfig, LinkPreviewConfig from crawl4ai.models import Link, CrawlResult @dataclass class CrawlState: """Tracks the current state of adaptive crawling""" crawled_urls: Set[str] = field(default_factory=set) knowledge_base: List[CrawlResult] = field(default_factory=list) pending_links: List[Link] = field(default_factory=list) query: str = "" metrics: Dict[str, float] = field(default_factory=dict) # Statistical tracking term_frequencies: Dict[str, int] = field(default_factory=lambda: defaultdict(int)) document_frequencies: Dict[str, int] = field(default_factory=lambda: defaultdict(int)) documents_with_terms: Dict[str, Set[int]] = field(default_factory=lambda: defaultdict(set)) total_documents: int = 0 # History tracking for saturation new_terms_history: List[int] = field(default_factory=list) crawl_order: List[str] = field(default_factory=list) # Embedding-specific tracking (only if strategy is embedding) kb_embeddings: Optional[Any] = None # Will be numpy array query_embeddings: Optional[Any] = None # Will be numpy array expanded_queries: List[str] = field(default_factory=list) coverage_shape: Optional[Any] = None # Alpha shape semantic_gaps: List[Tuple[List[float], float]] = field(default_factory=list) # Serializable embedding_model: str = "" def save(self, path: Union[str, Path]): """Save state to disk for persistence""" path = Path(path) path.parent.mkdir(parents=True, exist_ok=True) # Convert CrawlResult objects to dicts for serialization state_dict = { 'crawled_urls': list(self.crawled_urls), 'knowledge_base': [self._crawl_result_to_dict(cr) for cr in self.knowledge_base], 'pending_links': [link.model_dump() for link in self.pending_links], 'query': self.query, 'metrics': self.metrics, 'term_frequencies': dict(self.term_frequencies), 'document_frequencies': dict(self.document_frequencies), 'documents_with_terms': {k: list(v) for k, v in self.documents_with_terms.items()}, 'total_documents': self.total_documents, 'new_terms_history': self.new_terms_history, 'crawl_order': self.crawl_order, # Embedding-specific fields (convert numpy arrays to lists for JSON) 'kb_embeddings': self.kb_embeddings.tolist() if self.kb_embeddings is not None else None, 'query_embeddings': self.query_embeddings.tolist() if self.query_embeddings is not None else None, 'expanded_queries': self.expanded_queries, 'semantic_gaps': self.semantic_gaps, 'embedding_model': self.embedding_model } with open(path, 'w') as f: json.dump(state_dict, f, indent=2) @classmethod def load(cls, path: Union[str, Path]) -> 'CrawlState': """Load state from disk""" path = Path(path) with open(path, 'r') as f: state_dict = json.load(f) state = cls() state.crawled_urls = set(state_dict['crawled_urls']) state.knowledge_base = [cls._dict_to_crawl_result(d) for d in state_dict['knowledge_base']] state.pending_links = [Link(**link_dict) for link_dict in state_dict['pending_links']] state.query = state_dict['query'] state.metrics = state_dict['metrics'] state.term_frequencies = defaultdict(int, state_dict['term_frequencies']) state.document_frequencies = defaultdict(int, state_dict['document_frequencies']) state.documents_with_terms = defaultdict(set, {k: set(v) for k, v in state_dict['documents_with_terms'].items()}) state.total_documents = state_dict['total_documents'] state.new_terms_history = state_dict['new_terms_history'] state.crawl_order = state_dict['crawl_order'] # Load embedding-specific fields (convert lists back to numpy arrays) import numpy as np state.kb_embeddings = np.array(state_dict['kb_embeddings']) if state_dict.get('kb_embeddings') is not None else None state.query_embeddings = np.array(state_dict['query_embeddings']) if state_dict.get('query_embeddings') is not None else None state.expanded_queries = state_dict.get('expanded_queries', []) state.semantic_gaps = state_dict.get('semantic_gaps', []) state.embedding_model = state_dict.get('embedding_model', '') return state @staticmethod def _crawl_result_to_dict(cr: CrawlResult) -> Dict: """Convert CrawlResult to serializable dict""" # Extract markdown content safely markdown_content = "" if hasattr(cr, 'markdown') and cr.markdown: if hasattr(cr.markdown, 'raw_markdown'): markdown_content = cr.markdown.raw_markdown else: markdown_content = str(cr.markdown) return { 'url': cr.url, 'content': markdown_content, 'links': cr.links if hasattr(cr, 'links') else {}, 'metadata': cr.metadata if hasattr(cr, 'metadata') else {} } @staticmethod def _dict_to_crawl_result(d: Dict): """Convert dict back to CrawlResult""" # Create a mock object that has the minimal interface we need class MockMarkdown: def __init__(self, content): self.raw_markdown = content class MockCrawlResult: def __init__(self, url, content, links, metadata): self.url = url self.markdown = MockMarkdown(content) self.links = links self.metadata = metadata return MockCrawlResult( url=d['url'], content=d.get('content', ''), links=d.get('links', {}), metadata=d.get('metadata', {}) ) @dataclass class AdaptiveConfig: """Configuration for adaptive crawling""" confidence_threshold: float = 0.7 max_depth: int = 5 max_pages: int = 20 top_k_links: int = 3 min_gain_threshold: float = 0.1 strategy: str = "statistical" # statistical, embedding, llm # Advanced parameters saturation_threshold: float = 0.8 consistency_threshold: float = 0.7 coverage_weight: float = 0.4 consistency_weight: float = 0.3 saturation_weight: float = 0.3 # Link scoring parameters relevance_weight: float = 0.5 novelty_weight: float = 0.3 authority_weight: float = 0.2 # Persistence save_state: bool = False state_path: Optional[str] = None # Embedding strategy parameters embedding_model: str = "sentence-transformers/all-MiniLM-L6-v2" embedding_llm_config: Optional[Dict] = None # Separate config for embeddings n_query_variations: int = 10 coverage_threshold: float = 0.85 alpha_shape_alpha: float = 0.5 # Embedding confidence calculation parameters embedding_coverage_radius: float = 0.2 # Distance threshold for "covered" query points # Example: With radius=0.2, a query point is considered covered if ANY document # is within cosine distance 0.2 (very similar). Smaller = stricter coverage requirement embedding_k_exp: float = 3.0 # Exponential decay factor for distance-to-score mapping # Example: score = exp(-k_exp * distance). With k_exp=1, distance 0.2 → score 0.82, # distance 0.5 → score 0.61. Higher k_exp = steeper decay = more emphasis on very close matches embedding_nearest_weight: float = 0.7 # Weight for nearest neighbor in hybrid scoring embedding_top_k_weight: float = 0.3 # Weight for top-k average in hybrid scoring # Example: If nearest doc has score 0.9 and top-3 avg is 0.6, final = 0.7*0.9 + 0.3*0.6 = 0.81 # Higher nearest_weight = more focus on best match vs neighborhood density # Embedding link selection parameters embedding_overlap_threshold: float = 0.85 # Similarity threshold for penalizing redundant links # Example: Links with >0.85 similarity to existing KB get penalized to avoid redundancy # Lower = more aggressive deduplication, Higher = allow more similar content # Embedding stopping criteria parameters embedding_min_relative_improvement: float = 0.1 # Minimum relative improvement to continue # Example: If confidence is 0.6, need improvement > 0.06 per batch to continue crawling # Lower = more patient crawling, Higher = stop earlier when progress slows embedding_validation_min_score: float = 0.4 # Minimum validation score to trust convergence # Example: Even if learning converged, keep crawling if validation score < 0.4 # This prevents premature stopping when we haven't truly covered the query space # Quality confidence mapping parameters (for display to user) embedding_quality_min_confidence: float = 0.7 # Minimum confidence for validated systems embedding_quality_max_confidence: float = 0.95 # Maximum realistic confidence embedding_quality_scale_factor: float = 0.833 # Scaling factor for confidence mapping # Example: Validated system with learning_score=0.5 → confidence = 0.7 + (0.5-0.4)*0.833 = 0.78 # These control how internal scores map to user-friendly confidence percentages def validate(self): """Validate configuration parameters""" assert 0 <= self.confidence_threshold <= 1, "confidence_threshold must be between 0 and 1" assert self.max_depth > 0, "max_depth must be positive" assert self.max_pages > 0, "max_pages must be positive" assert self.top_k_links > 0, "top_k_links must be positive" assert 0 <= self.min_gain_threshold <= 1, "min_gain_threshold must be between 0 and 1" # Check weights sum to 1 weight_sum = self.coverage_weight + self.consistency_weight + self.saturation_weight assert abs(weight_sum - 1.0) < 0.001, f"Coverage weights must sum to 1, got {weight_sum}" weight_sum = self.relevance_weight + self.novelty_weight + self.authority_weight assert abs(weight_sum - 1.0) < 0.001, f"Link scoring weights must sum to 1, got {weight_sum}" # Validate embedding parameters assert 0 < self.embedding_coverage_radius < 1, "embedding_coverage_radius must be between 0 and 1" assert self.embedding_k_exp > 0, "embedding_k_exp must be positive" assert 0 <= self.embedding_nearest_weight <= 1, "embedding_nearest_weight must be between 0 and 1" assert 0 <= self.embedding_top_k_weight <= 1, "embedding_top_k_weight must be between 0 and 1" assert abs(self.embedding_nearest_weight + self.embedding_top_k_weight - 1.0) < 0.001, "Embedding weights must sum to 1" assert 0 <= self.embedding_overlap_threshold <= 1, "embedding_overlap_threshold must be between 0 and 1" assert 0 < self.embedding_min_relative_improvement < 1, "embedding_min_relative_improvement must be between 0 and 1" assert 0 <= self.embedding_validation_min_score <= 1, "embedding_validation_min_score must be between 0 and 1" assert 0 <= self.embedding_quality_min_confidence <= 1, "embedding_quality_min_confidence must be between 0 and 1" assert 0 <= self.embedding_quality_max_confidence <= 1, "embedding_quality_max_confidence must be between 0 and 1" assert self.embedding_quality_scale_factor > 0, "embedding_quality_scale_factor must be positive" class CrawlStrategy(ABC): """Abstract base class for crawling strategies""" @abstractmethod async def calculate_confidence(self, state: CrawlState) -> float: """Calculate overall confidence that we have sufficient information""" pass @abstractmethod async def rank_links(self, state: CrawlState, config: AdaptiveConfig) -> List[Tuple[Link, float]]: """Rank pending links by expected information gain""" pass @abstractmethod async def should_stop(self, state: CrawlState, config: AdaptiveConfig) -> bool: """Determine if crawling should stop""" pass @abstractmethod async def update_state(self, state: CrawlState, new_results: List[CrawlResult]) -> None: """Update state with new crawl results""" pass class StatisticalStrategy(CrawlStrategy): """Pure statistical approach - no LLM, no embeddings""" def __init__(self): self.idf_cache = {} self.bm25_k1 = 1.2 # BM25 parameter self.bm25_b = 0.75 # BM25 parameter async def calculate_confidence(self, state: CrawlState) -> float: """Calculate confidence using coverage, consistency, and saturation""" if not state.knowledge_base: return 0.0 coverage = self._calculate_coverage(state) consistency = self._calculate_consistency(state) saturation = self._calculate_saturation(state) # Store individual metrics state.metrics['coverage'] = coverage state.metrics['consistency'] = consistency state.metrics['saturation'] = saturation # Weighted combination (weights from config not accessible here, using defaults) confidence = 0.4 * coverage + 0.3 * consistency + 0.3 * saturation return confidence def _calculate_coverage(self, state: CrawlState) -> float: """Coverage scoring - measures query term presence across knowledge base Returns a score between 0 and 1, where: - 0 means no query terms found - 1 means excellent coverage of all query terms """ if not state.query or state.total_documents == 0: return 0.0 query_terms = self._tokenize(state.query.lower()) if not query_terms: return 0.0 term_scores = [] max_tf = max(state.term_frequencies.values()) if state.term_frequencies else 1 for term in query_terms: tf = state.term_frequencies.get(term, 0) df = state.document_frequencies.get(term, 0) if df > 0: # Document coverage: what fraction of docs contain this term doc_coverage = df / state.total_documents # Frequency signal: normalized log frequency freq_signal = math.log(1 + tf) / math.log(1 + max_tf) if max_tf > 0 else 0 # Combined score: document coverage with frequency boost term_score = doc_coverage * (1 + 0.5 * freq_signal) term_scores.append(term_score) else: term_scores.append(0.0) # Average across all query terms coverage = sum(term_scores) / len(term_scores) # Apply square root curve to make score more intuitive # This helps differentiate between partial and good coverage return min(1.0, math.sqrt(coverage)) def _calculate_consistency(self, state: CrawlState) -> float: """Information overlap between pages - high overlap suggests coherent topic coverage""" if len(state.knowledge_base) < 2: return 1.0 # Single or no documents are perfectly consistent # Calculate pairwise term overlap overlaps = [] for i in range(len(state.knowledge_base)): for j in range(i + 1, len(state.knowledge_base)): # Get terms from both documents terms_i = set(self._get_document_terms(state.knowledge_base[i])) terms_j = set(self._get_document_terms(state.knowledge_base[j])) if terms_i and terms_j: # Jaccard similarity overlap = len(terms_i & terms_j) / len(terms_i | terms_j) overlaps.append(overlap) if overlaps: # Average overlap as consistency measure consistency = sum(overlaps) / len(overlaps) else: consistency = 0.0 return consistency def _calculate_saturation(self, state: CrawlState) -> float: """Diminishing returns indicator - are we still discovering new information?""" if not state.new_terms_history: return 0.0 if len(state.new_terms_history) < 2: return 0.0 # Not enough history # Calculate rate of new term discovery recent_rate = state.new_terms_history[-1] if state.new_terms_history[-1] > 0 else 1 initial_rate = state.new_terms_history[0] if state.new_terms_history[0] > 0 else 1 # Saturation increases as rate decreases saturation = 1 - (recent_rate / initial_rate) return max(0.0, min(saturation, 1.0)) async def rank_links(self, state: CrawlState, config: AdaptiveConfig) -> List[Tuple[Link, float]]: """Rank links by expected information gain""" scored_links = [] for link in state.pending_links: # Skip already crawled URLs if link.href in state.crawled_urls: continue # Calculate component scores relevance = self._calculate_relevance(link, state) novelty = self._calculate_novelty(link, state) authority = 1.0 # authority = self._calculate_authority(link) # Combined score score = (config.relevance_weight * relevance + config.novelty_weight * novelty + config.authority_weight * authority) scored_links.append((link, score)) # Sort by score descending scored_links.sort(key=lambda x: x[1], reverse=True) return scored_links def _calculate_relevance(self, link: Link, state: CrawlState) -> float: """BM25 relevance score between link preview and query""" if not state.query or not link: return 0.0 # Combine available text from link link_text = ' '.join(filter(None, [ link.text or '', link.title or '', link.head_data.get('meta', {}).get('title', '') if link.head_data else '', link.head_data.get('meta', {}).get('description', '') if link.head_data else '', link.head_data.get('meta', {}).get('keywords', '') if link.head_data else '' ])).lower() if not link_text: return 0.0 # Use contextual score if available (from BM25 scoring during crawl) # if link.contextual_score is not None: if link.contextual_score and link.contextual_score > 0: return link.contextual_score # Otherwise, calculate simple term overlap query_terms = set(self._tokenize(state.query.lower())) link_terms = set(self._tokenize(link_text)) if not query_terms: return 0.0 overlap = len(query_terms & link_terms) / len(query_terms) return overlap def _calculate_novelty(self, link: Link, state: CrawlState) -> float: """Estimate how much new information this link might provide""" if not state.knowledge_base: return 1.0 # First links are maximally novel # Get terms from link preview link_text = ' '.join(filter(None, [ link.text or '', link.title or '', link.head_data.get('title', '') if link.head_data else '', link.head_data.get('description', '') if link.head_data else '', link.head_data.get('keywords', '') if link.head_data else '' ])).lower() link_terms = set(self._tokenize(link_text)) if not link_terms: return 0.5 # Unknown novelty # Calculate what percentage of link terms are new existing_terms = set(state.term_frequencies.keys()) new_terms = link_terms - existing_terms novelty = len(new_terms) / len(link_terms) if link_terms else 0.0 return novelty def _calculate_authority(self, link: Link) -> float: """Simple authority score based on URL structure and link attributes""" score = 0.5 # Base score if not link.href: return 0.0 url = link.href.lower() # Positive indicators if '/docs/' in url or '/documentation/' in url: score += 0.2 if '/api/' in url or '/reference/' in url: score += 0.2 if '/guide/' in url or '/tutorial/' in url: score += 0.1 # Check for file extensions if url.endswith('.pdf'): score += 0.1 elif url.endswith(('.jpg', '.png', '.gif')): score -= 0.3 # Reduce score for images # Use intrinsic score if available if link.intrinsic_score is not None: score = 0.7 * score + 0.3 * link.intrinsic_score return min(score, 1.0) async def should_stop(self, state: CrawlState, config: AdaptiveConfig) -> bool: """Determine if crawling should stop""" # Check confidence threshold confidence = state.metrics.get('confidence', 0.0) if confidence >= config.confidence_threshold: return True # Check resource limits if len(state.crawled_urls) >= config.max_pages: return True # Check if we have any links left if not state.pending_links: return True # Check saturation if state.metrics.get('saturation', 0.0) >= config.saturation_threshold: return True return False async def update_state(self, state: CrawlState, new_results: List[CrawlResult]) -> None: """Update state with new crawl results""" for result in new_results: # Track new terms old_term_count = len(state.term_frequencies) # Extract and process content - try multiple fields try: content = result.markdown.raw_markdown except AttributeError: print(f"Warning: CrawlResult {result.url} has no markdown content") content = "" # content = "" # if hasattr(result, 'extracted_content') and result.extracted_content: # content = result.extracted_content # elif hasattr(result, 'markdown') and result.markdown: # content = result.markdown.raw_markdown # elif hasattr(result, 'cleaned_html') and result.cleaned_html: # content = result.cleaned_html # elif hasattr(result, 'html') and result.html: # # Use raw HTML as last resort # content = result.html terms = self._tokenize(content.lower()) # Update term frequencies term_set = set() for term in terms: state.term_frequencies[term] += 1 term_set.add(term) # Update document frequencies doc_id = state.total_documents for term in term_set: if term not in state.documents_with_terms[term]: state.document_frequencies[term] += 1 state.documents_with_terms[term].add(doc_id) # Track new terms discovered new_term_count = len(state.term_frequencies) new_terms = new_term_count - old_term_count state.new_terms_history.append(new_terms) # Update document count state.total_documents += 1 # Add to crawl order state.crawl_order.append(result.url) def _tokenize(self, text: str) -> List[str]: """Simple tokenization - can be enhanced""" # Remove punctuation and split text = re.sub(r'[^\w\s]', ' ', text) tokens = text.split() # Filter short tokens and stop words (basic) tokens = [t for t in tokens if len(t) > 2] return tokens def _get_document_terms(self, crawl_result: CrawlResult) -> List[str]: """Extract terms from a crawl result""" content = crawl_result.markdown.raw_markdown or "" return self._tokenize(content.lower()) class EmbeddingStrategy(CrawlStrategy): """Embedding-based adaptive crawling using semantic space coverage""" def __init__(self, embedding_model: str = None, llm_config: Dict = None): self.embedding_model = embedding_model or "sentence-transformers/all-MiniLM-L6-v2" self.llm_config = llm_config self._embedding_cache = {} self._link_embedding_cache = {} # Cache for link embeddings self._validation_passed = False # Track if validation passed # Performance optimization caches self._distance_matrix_cache = None # Cache for query-KB distances self._kb_embeddings_hash = None # Track KB changes self._validation_embeddings_cache = None # Cache validation query embeddings self._kb_similarity_threshold = 0.95 # Threshold for deduplication async def _get_embeddings(self, texts: List[str]) -> Any: """Get embeddings using configured method""" from .utils import get_text_embeddings embedding_llm_config = { 'provider': 'openai/text-embedding-3-small', 'api_token': os.getenv('OPENAI_API_KEY') } return await get_text_embeddings( texts, embedding_llm_config, self.embedding_model ) def _compute_distance_matrix(self, query_embeddings: Any, kb_embeddings: Any) -> Any: """Compute distance matrix using vectorized operations""" import numpy as np if kb_embeddings is None or len(kb_embeddings) == 0: return None # Ensure proper shapes if len(query_embeddings.shape) == 1: query_embeddings = query_embeddings.reshape(1, -1) if len(kb_embeddings.shape) == 1: kb_embeddings = kb_embeddings.reshape(1, -1) # Vectorized cosine distance: 1 - cosine_similarity # Normalize vectors query_norm = query_embeddings / np.linalg.norm(query_embeddings, axis=1, keepdims=True) kb_norm = kb_embeddings / np.linalg.norm(kb_embeddings, axis=1, keepdims=True) # Compute cosine similarity matrix similarity_matrix = np.dot(query_norm, kb_norm.T) # Convert to distance distance_matrix = 1 - similarity_matrix return distance_matrix def _get_cached_distance_matrix(self, query_embeddings: Any, kb_embeddings: Any) -> Any: """Get distance matrix with caching""" import numpy as np if kb_embeddings is None or len(kb_embeddings) == 0: return None # Check if KB has changed kb_hash = hash(kb_embeddings.tobytes()) if kb_embeddings is not None else None if (self._distance_matrix_cache is None or kb_hash != self._kb_embeddings_hash): # Recompute matrix self._distance_matrix_cache = self._compute_distance_matrix(query_embeddings, kb_embeddings) self._kb_embeddings_hash = kb_hash return self._distance_matrix_cache async def map_query_semantic_space(self, query: str, n_synthetic: int = 10) -> Any: """Generate a point cloud representing the semantic neighborhood of the query""" from .utils import perform_completion_with_backoff # Generate more variations than needed for train/val split n_total = int(n_synthetic * 1.3) # Generate 30% more for validation # Generate variations using LLM prompt = f"""Generate {n_total} variations of this query that explore different aspects: '{query}' These should be queries a user might ask when looking for similar information. Include different phrasings, related concepts, and specific aspects. Return as a JSON array of strings.""" # Use the LLM for query generation provider = self.llm_config.get('provider', 'openai/gpt-4o-mini') if self.llm_config else 'openai/gpt-4o-mini' api_token = self.llm_config.get('api_token') if self.llm_config else None # response = perform_completion_with_backoff( # provider=provider, # prompt_with_variables=prompt, # api_token=api_token, # json_response=True # ) # variations = json.loads(response.choices[0].message.content) # # Mock data with more variations for split variations ={'queries': ['what are the best vegetables to use in fried rice?', 'how do I make vegetable fried rice from scratch?', 'can you provide a quick recipe for vegetable fried rice?', 'what cooking techniques are essential for perfect fried rice with vegetables?', 'how to add flavor to vegetable fried rice?', 'are there any tips for making healthy fried rice with vegetables?']} variations = {'queries': [ 'How do async and await work with coroutines in Python?', 'What is the role of event loops in asynchronous programming?', 'Can you explain the differences between async/await and traditional callback methods?', 'How do coroutines interact with event loops in JavaScript?', 'What are the benefits of using async await over promises in Node.js?', # 'How to manage multiple coroutines with an event loop?', # 'What are some common pitfalls when using async await with coroutines?', # 'How do different programming languages implement async await and event loops?', # 'What happens when an async function is called without await?', # 'How does the event loop handle blocking operations?', 'Can you nest async functions and how does that affect the event loop?', 'What is the performance impact of using async/await?' ]} # Split into train and validation # all_queries = [query] + variations['queries'] # Randomly shuffle for proper train/val split (keeping original query in training) import random # Keep original query always in training other_queries = variations['queries'].copy() random.shuffle(other_queries) # Split: 80% for training, 20% for validation n_validation = max(2, int(len(other_queries) * 0.2)) # At least 2 for validation val_queries = other_queries[-n_validation:] train_queries = [query] + other_queries[:-n_validation] # Embed only training queries for now (faster) train_embeddings = await self._get_embeddings(train_queries) # Store validation queries for later (don't embed yet to save time) self._validation_queries = val_queries return train_embeddings, train_queries def compute_coverage_shape(self, query_points: Any, alpha: float = 0.5): """Find the minimal shape that covers all query points using alpha shape""" try: import numpy as np if len(query_points) < 3: return None # For high-dimensional embeddings (e.g., 384-dim, 768-dim),
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/async_configs.py
crawl4ai/async_configs.py
import importlib import os import warnings import requests from .config import ( DEFAULT_PROVIDER, DEFAULT_PROVIDER_API_KEY, MIN_WORD_THRESHOLD, IMAGE_DESCRIPTION_MIN_WORD_THRESHOLD, PROVIDER_MODELS, PROVIDER_MODELS_PREFIXES, SCREENSHOT_HEIGHT_TRESHOLD, PAGE_TIMEOUT, IMAGE_SCORE_THRESHOLD, SOCIAL_MEDIA_DOMAINS, ) from .user_agent_generator import UAGen, ValidUAGenerator # , OnlineUAGenerator from .extraction_strategy import ExtractionStrategy, LLMExtractionStrategy from .chunking_strategy import ChunkingStrategy, RegexChunking from .markdown_generation_strategy import MarkdownGenerationStrategy, DefaultMarkdownGenerator from .content_scraping_strategy import ContentScrapingStrategy, LXMLWebScrapingStrategy from .deep_crawling import DeepCrawlStrategy from .table_extraction import TableExtractionStrategy, DefaultTableExtraction from .cache_context import CacheMode from .proxy_strategy import ProxyRotationStrategy import inspect from typing import Any, Callable, Dict, List, Optional, Union from enum import Enum # Type alias for URL matching UrlMatcher = Union[str, Callable[[str], bool], List[Union[str, Callable[[str], bool]]]] class MatchMode(Enum): OR = "or" AND = "and" # from .proxy_strategy import ProxyConfig def to_serializable_dict(obj: Any, ignore_default_value : bool = False): """ Recursively convert an object to a serializable dictionary using {type, params} structure for complex objects. """ if obj is None: return None # Handle basic types if isinstance(obj, (str, int, float, bool)): return obj # Handle Enum if isinstance(obj, Enum): return {"type": obj.__class__.__name__, "params": obj.value} # Handle datetime objects if hasattr(obj, "isoformat"): return obj.isoformat() # Handle lists, tuples, and sets, and basically any iterable if isinstance(obj, (list, tuple, set)) or hasattr(obj, '__iter__') and not isinstance(obj, dict): return [to_serializable_dict(item) for item in obj] # Handle frozensets, which are not iterable if isinstance(obj, frozenset): return [to_serializable_dict(item) for item in list(obj)] # Handle dictionaries - preserve them as-is if isinstance(obj, dict): return { "type": "dict", # Mark as plain dictionary "value": {str(k): to_serializable_dict(v) for k, v in obj.items()}, } _type = obj.__class__.__name__ # Handle class instances if hasattr(obj, "__class__"): # Get constructor signature sig = inspect.signature(obj.__class__.__init__) params = sig.parameters # Get current values current_values = {} for name, param in params.items(): if name == "self": continue value = getattr(obj, name, param.default) # Only include if different from default, considering empty values if not (is_empty_value(value) and is_empty_value(param.default)): if value != param.default and not ignore_default_value: current_values[name] = to_serializable_dict(value) # Don't serialize private __slots__ - they're internal implementation details # not constructor parameters. This was causing URLPatternFilter to fail # because _simple_suffixes was being serialized as 'simple_suffixes' # if hasattr(obj, '__slots__'): # for slot in obj.__slots__: # if slot.startswith('_'): # Handle private slots # attr_name = slot[1:] # Remove leading '_' # value = getattr(obj, slot, None) # if value is not None: # current_values[attr_name] = to_serializable_dict(value) return { "type": obj.__class__.__name__, "params": current_values } return str(obj) def from_serializable_dict(data: Any) -> Any: """ Recursively convert a serializable dictionary back to an object instance. """ if data is None: return None # Handle basic types if isinstance(data, (str, int, float, bool)): return data # Handle typed data if isinstance(data, dict) and "type" in data: # Handle plain dictionaries if data["type"] == "dict" and "value" in data: return {k: from_serializable_dict(v) for k, v in data["value"].items()} cls = None # If you are receiving an error while trying to convert a dict to an object: # Either add a module to `modules_paths` list, or add the `data["type"]` to the crawl4ai __init__.py file module_paths = ["crawl4ai"] for module_path in module_paths: try: mod = importlib.import_module(module_path) if hasattr(mod, data["type"]): cls = getattr(mod, data["type"]) break except (ImportError, AttributeError): continue if cls is not None: # Handle Enum if issubclass(cls, Enum): return cls(data["params"]) if "params" in data: # Handle class instances constructor_args = { k: from_serializable_dict(v) for k, v in data["params"].items() } return cls(**constructor_args) # Handle lists if isinstance(data, list): return [from_serializable_dict(item) for item in data] # Handle raw dictionaries (legacy support) if isinstance(data, dict): return {k: from_serializable_dict(v) for k, v in data.items()} return data def is_empty_value(value: Any) -> bool: """Check if a value is effectively empty/null.""" if value is None: return True if isinstance(value, (list, tuple, set, dict, str)) and len(value) == 0: return True return False class GeolocationConfig: def __init__( self, latitude: float, longitude: float, accuracy: Optional[float] = 0.0 ): """Configuration class for geolocation settings. Args: latitude: Latitude coordinate (e.g., 37.7749) longitude: Longitude coordinate (e.g., -122.4194) accuracy: Accuracy in meters. Default: 0.0 """ self.latitude = latitude self.longitude = longitude self.accuracy = accuracy @staticmethod def from_dict(geo_dict: Dict) -> "GeolocationConfig": """Create a GeolocationConfig from a dictionary.""" return GeolocationConfig( latitude=geo_dict.get("latitude"), longitude=geo_dict.get("longitude"), accuracy=geo_dict.get("accuracy", 0.0) ) def to_dict(self) -> Dict: """Convert to dictionary representation.""" return { "latitude": self.latitude, "longitude": self.longitude, "accuracy": self.accuracy } def clone(self, **kwargs) -> "GeolocationConfig": """Create a copy of this configuration with updated values. Args: **kwargs: Key-value pairs of configuration options to update Returns: GeolocationConfig: A new instance with the specified updates """ config_dict = self.to_dict() config_dict.update(kwargs) return GeolocationConfig.from_dict(config_dict) class ProxyConfig: def __init__( self, server: str, username: Optional[str] = None, password: Optional[str] = None, ip: Optional[str] = None, ): """Configuration class for a single proxy. Args: server: Proxy server URL (e.g., "http://127.0.0.1:8080") username: Optional username for proxy authentication password: Optional password for proxy authentication ip: Optional IP address for verification purposes """ self.server = server self.username = username self.password = password # Extract IP from server if not explicitly provided self.ip = ip or self._extract_ip_from_server() def _extract_ip_from_server(self) -> Optional[str]: """Extract IP address from server URL.""" try: # Simple extraction assuming http://ip:port format if "://" in self.server: parts = self.server.split("://")[1].split(":") return parts[0] else: parts = self.server.split(":") return parts[0] except Exception: return None @staticmethod def from_string(proxy_str: str) -> "ProxyConfig": """Create a ProxyConfig from a string. Supported formats: - 'http://username:password@ip:port' - 'http://ip:port' - 'socks5://ip:port' - 'ip:port:username:password' - 'ip:port' """ s = (proxy_str or "").strip() # URL with credentials if "@" in s and "://" in s: auth_part, server_part = s.split("@", 1) protocol, credentials = auth_part.split("://", 1) if ":" in credentials: username, password = credentials.split(":", 1) return ProxyConfig( server=f"{protocol}://{server_part}", username=username, password=password, ) # URL without credentials (keep scheme) if "://" in s and "@" not in s: return ProxyConfig(server=s) # Colon separated forms parts = s.split(":") if len(parts) == 4: ip, port, username, password = parts return ProxyConfig(server=f"http://{ip}:{port}", username=username, password=password) if len(parts) == 2: ip, port = parts return ProxyConfig(server=f"http://{ip}:{port}") raise ValueError(f"Invalid proxy string format: {proxy_str}") @staticmethod def from_dict(proxy_dict: Dict) -> "ProxyConfig": """Create a ProxyConfig from a dictionary.""" return ProxyConfig( server=proxy_dict.get("server"), username=proxy_dict.get("username"), password=proxy_dict.get("password"), ip=proxy_dict.get("ip") ) @staticmethod def from_env(env_var: str = "PROXIES") -> List["ProxyConfig"]: """Load proxies from environment variable. Args: env_var: Name of environment variable containing comma-separated proxy strings Returns: List of ProxyConfig objects """ proxies = [] try: proxy_list = os.getenv(env_var, "").split(",") for proxy in proxy_list: if not proxy: continue proxies.append(ProxyConfig.from_string(proxy)) except Exception as e: print(f"Error loading proxies from environment: {e}") return proxies def to_dict(self) -> Dict: """Convert to dictionary representation.""" return { "server": self.server, "username": self.username, "password": self.password, "ip": self.ip } def clone(self, **kwargs) -> "ProxyConfig": """Create a copy of this configuration with updated values. Args: **kwargs: Key-value pairs of configuration options to update Returns: ProxyConfig: A new instance with the specified updates """ config_dict = self.to_dict() config_dict.update(kwargs) return ProxyConfig.from_dict(config_dict) class BrowserConfig: """ Configuration class for setting up a browser instance and its context in AsyncPlaywrightCrawlerStrategy. This class centralizes all parameters that affect browser and context creation. Instead of passing scattered keyword arguments, users can instantiate and modify this configuration object. The crawler code will then reference these settings to initialize the browser in a consistent, documented manner. Attributes: browser_type (str): The type of browser to launch. Supported values: "chromium", "firefox", "webkit". Default: "chromium". headless (bool): Whether to run the browser in headless mode (no visible GUI). Default: True. browser_mode (str): Determines how the browser should be initialized: "builtin" - use the builtin CDP browser running in background "dedicated" - create a new dedicated browser instance each time "cdp" - use explicit CDP settings provided in cdp_url "docker" - run browser in Docker container with isolation Default: "dedicated" use_managed_browser (bool): Launch the browser using a managed approach (e.g., via CDP), allowing advanced manipulation. Default: False. cdp_url (str): URL for the Chrome DevTools Protocol (CDP) endpoint. Default: "ws://localhost:9222/devtools/browser/". debugging_port (int): Port for the browser debugging protocol. Default: 9222. use_persistent_context (bool): Use a persistent browser context (like a persistent profile). Automatically sets use_managed_browser=True. Default: False. user_data_dir (str or None): Path to a user data directory for persistent sessions. If None, a temporary directory may be used. Default: None. chrome_channel (str): The Chrome channel to launch (e.g., "chrome", "msedge"). Only applies if browser_type is "chromium". Default: "chromium". channel (str): The channel to launch (e.g., "chromium", "chrome", "msedge"). Only applies if browser_type is "chromium". Default: "chromium". proxy (Optional[str]): Proxy server URL (e.g., "http://username:password@proxy:port"). If None, no proxy is used. Default: None. proxy_config (ProxyConfig or dict or None): Detailed proxy configuration, e.g. {"server": "...", "username": "..."}. If None, no additional proxy config. Default: None. viewport_width (int): Default viewport width for pages. Default: 1080. viewport_height (int): Default viewport height for pages. Default: 600. viewport (dict): Default viewport dimensions for pages. If set, overrides viewport_width and viewport_height. Default: None. verbose (bool): Enable verbose logging. Default: True. accept_downloads (bool): Whether to allow file downloads. If True, requires a downloads_path. Default: False. downloads_path (str or None): Directory to store downloaded files. If None and accept_downloads is True, a default path will be created. Default: None. storage_state (str or dict or None): An in-memory storage state (cookies, localStorage). Default: None. ignore_https_errors (bool): Ignore HTTPS certificate errors. Default: True. java_script_enabled (bool): Enable JavaScript execution in pages. Default: True. cookies (list): List of cookies to add to the browser context. Each cookie is a dict with fields like {"name": "...", "value": "...", "url": "..."}. Default: []. headers (dict): Extra HTTP headers to apply to all requests in this context. Default: {}. user_agent (str): Custom User-Agent string to use. Default: "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) " "AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36". user_agent_mode (str or None): Mode for generating the user agent (e.g., "random"). If None, use the provided user_agent as-is. Default: None. user_agent_generator_config (dict or None): Configuration for user agent generation if user_agent_mode is set. Default: None. text_mode (bool): If True, disables images and other rich content for potentially faster load times. Default: False. light_mode (bool): Disables certain background features for performance gains. Default: False. extra_args (list): Additional command-line arguments passed to the browser. Default: []. enable_stealth (bool): If True, applies playwright-stealth to bypass basic bot detection. Cannot be used with use_undetected browser mode. Default: False. """ def __init__( self, browser_type: str = "chromium", headless: bool = True, browser_mode: str = "dedicated", use_managed_browser: bool = False, cdp_url: str = None, use_persistent_context: bool = False, user_data_dir: str = None, chrome_channel: str = "chromium", channel: str = "chromium", proxy: str = None, proxy_config: Union[ProxyConfig, dict, None] = None, viewport_width: int = 1080, viewport_height: int = 600, viewport: dict = None, accept_downloads: bool = False, downloads_path: str = None, storage_state: Union[str, dict, None] = None, ignore_https_errors: bool = True, java_script_enabled: bool = True, sleep_on_close: bool = False, verbose: bool = True, cookies: list = None, headers: dict = None, user_agent: str = ( # "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:109.0) AppleWebKit/537.36 " # "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 " # "(KHTML, like Gecko) Chrome/116.0.5845.187 Safari/604.1 Edg/117.0.2045.47" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 Chrome/116.0.0.0 Safari/537.36" ), user_agent_mode: str = "", user_agent_generator_config: dict = {}, text_mode: bool = False, light_mode: bool = False, extra_args: list = None, debugging_port: int = 9222, host: str = "localhost", enable_stealth: bool = False, ): self.browser_type = browser_type self.headless = headless self.browser_mode = browser_mode self.use_managed_browser = use_managed_browser self.cdp_url = cdp_url self.use_persistent_context = use_persistent_context self.user_data_dir = user_data_dir self.chrome_channel = chrome_channel or self.browser_type or "chromium" self.channel = channel or self.browser_type or "chromium" if self.browser_type in ["firefox", "webkit"]: self.channel = "" self.chrome_channel = "" if proxy: warnings.warn("The 'proxy' parameter is deprecated and will be removed in a future release. Use 'proxy_config' instead.", UserWarning) self.proxy = proxy self.proxy_config = proxy_config if isinstance(self.proxy_config, dict): self.proxy_config = ProxyConfig.from_dict(self.proxy_config) if isinstance(self.proxy_config, str): self.proxy_config = ProxyConfig.from_string(self.proxy_config) if self.proxy and self.proxy_config: warnings.warn("Both 'proxy' and 'proxy_config' are provided. 'proxy_config' will take precedence.", UserWarning) self.proxy = None elif self.proxy: # Convert proxy string to ProxyConfig if proxy_config is not provided self.proxy_config = ProxyConfig.from_string(self.proxy) self.proxy = None self.viewport_width = viewport_width self.viewport_height = viewport_height self.viewport = viewport if self.viewport is not None: self.viewport_width = self.viewport.get("width", 1080) self.viewport_height = self.viewport.get("height", 600) self.accept_downloads = accept_downloads self.downloads_path = downloads_path self.storage_state = storage_state self.ignore_https_errors = ignore_https_errors self.java_script_enabled = java_script_enabled self.cookies = cookies if cookies is not None else [] self.headers = headers if headers is not None else {} self.user_agent = user_agent self.user_agent_mode = user_agent_mode self.user_agent_generator_config = user_agent_generator_config self.text_mode = text_mode self.light_mode = light_mode self.extra_args = extra_args if extra_args is not None else [] self.sleep_on_close = sleep_on_close self.verbose = verbose self.debugging_port = debugging_port self.host = host self.enable_stealth = enable_stealth fa_user_agenr_generator = ValidUAGenerator() if self.user_agent_mode == "random": self.user_agent = fa_user_agenr_generator.generate( **(self.user_agent_generator_config or {}) ) else: pass self.browser_hint = UAGen.generate_client_hints(self.user_agent) self.headers.setdefault("sec-ch-ua", self.browser_hint) # Set appropriate browser management flags based on browser_mode if self.browser_mode == "builtin": # Builtin mode uses managed browser connecting to builtin CDP endpoint self.use_managed_browser = True # cdp_url will be set later by browser_manager elif self.browser_mode == "docker": # Docker mode uses managed browser with CDP to connect to browser in container self.use_managed_browser = True # cdp_url will be set later by docker browser strategy elif self.browser_mode == "custom" and self.cdp_url: # Custom mode with explicit CDP URL self.use_managed_browser = True elif self.browser_mode == "dedicated": # Dedicated mode uses a new browser instance each time pass # If persistent context is requested, ensure managed browser is enabled if self.use_persistent_context: self.use_managed_browser = True # Validate stealth configuration if self.enable_stealth and self.use_managed_browser and self.browser_mode == "builtin": raise ValueError( "enable_stealth cannot be used with browser_mode='builtin'. " "Stealth mode requires a dedicated browser instance." ) @staticmethod def from_kwargs(kwargs: dict) -> "BrowserConfig": return BrowserConfig( browser_type=kwargs.get("browser_type", "chromium"), headless=kwargs.get("headless", True), browser_mode=kwargs.get("browser_mode", "dedicated"), use_managed_browser=kwargs.get("use_managed_browser", False), cdp_url=kwargs.get("cdp_url"), use_persistent_context=kwargs.get("use_persistent_context", False), user_data_dir=kwargs.get("user_data_dir"), chrome_channel=kwargs.get("chrome_channel", "chromium"), channel=kwargs.get("channel", "chromium"), proxy=kwargs.get("proxy"), proxy_config=kwargs.get("proxy_config", None), viewport_width=kwargs.get("viewport_width", 1080), viewport_height=kwargs.get("viewport_height", 600), accept_downloads=kwargs.get("accept_downloads", False), downloads_path=kwargs.get("downloads_path"), storage_state=kwargs.get("storage_state"), ignore_https_errors=kwargs.get("ignore_https_errors", True), java_script_enabled=kwargs.get("java_script_enabled", True), cookies=kwargs.get("cookies", []), headers=kwargs.get("headers", {}), user_agent=kwargs.get( "user_agent", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) " "AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36", ), user_agent_mode=kwargs.get("user_agent_mode"), user_agent_generator_config=kwargs.get("user_agent_generator_config"), text_mode=kwargs.get("text_mode", False), light_mode=kwargs.get("light_mode", False), extra_args=kwargs.get("extra_args", []), debugging_port=kwargs.get("debugging_port", 9222), host=kwargs.get("host", "localhost"), enable_stealth=kwargs.get("enable_stealth", False), ) def to_dict(self): result = { "browser_type": self.browser_type, "headless": self.headless, "browser_mode": self.browser_mode, "use_managed_browser": self.use_managed_browser, "cdp_url": self.cdp_url, "use_persistent_context": self.use_persistent_context, "user_data_dir": self.user_data_dir, "chrome_channel": self.chrome_channel, "channel": self.channel, "proxy": self.proxy, "proxy_config": self.proxy_config.to_dict() if self.proxy_config else None, "viewport_width": self.viewport_width, "viewport_height": self.viewport_height, "accept_downloads": self.accept_downloads, "downloads_path": self.downloads_path, "storage_state": self.storage_state, "ignore_https_errors": self.ignore_https_errors, "java_script_enabled": self.java_script_enabled, "cookies": self.cookies, "headers": self.headers, "user_agent": self.user_agent, "user_agent_mode": self.user_agent_mode, "user_agent_generator_config": self.user_agent_generator_config, "text_mode": self.text_mode, "light_mode": self.light_mode, "extra_args": self.extra_args, "sleep_on_close": self.sleep_on_close, "verbose": self.verbose, "debugging_port": self.debugging_port, "host": self.host, "enable_stealth": self.enable_stealth, } return result def clone(self, **kwargs): """Create a copy of this configuration with updated values. Args: **kwargs: Key-value pairs of configuration options to update Returns: BrowserConfig: A new instance with the specified updates """ config_dict = self.to_dict() config_dict.update(kwargs) return BrowserConfig.from_kwargs(config_dict) # Create a funciton returns dict of the object def dump(self) -> dict: # Serialize the object to a dictionary return to_serializable_dict(self) @staticmethod def load(data: dict) -> "BrowserConfig": # Deserialize the object from a dictionary config = from_serializable_dict(data) if isinstance(config, BrowserConfig): return config return BrowserConfig.from_kwargs(config) def set_nstproxy( self, token: str, channel_id: str, country: str = "ANY", state: str = "", city: str = "", protocol: str = "http", session_duration: int = 10, ): """ Fetch a proxy from NSTProxy API and automatically assign it to proxy_config. Get your NSTProxy token from: https://app.nstproxy.com/profile Args: token (str): NSTProxy API token. channel_id (str): NSTProxy channel ID. country (str, optional): Country code (default: "ANY"). state (str, optional): State code (default: ""). city (str, optional): City name (default: ""). protocol (str, optional): Proxy protocol ("http" or "socks5"). Defaults to "http". session_duration (int, optional): Session duration in minutes (0 = rotate each request). Defaults to 10. Raises: ValueError: If the API response format is invalid. PermissionError: If the API returns an error message. """ # --- Validate input early --- if not token or not channel_id: raise ValueError("[NSTProxy] token and channel_id are required") if protocol not in ("http", "socks5"): raise ValueError(f"[NSTProxy] Invalid protocol: {protocol}") # --- Build NSTProxy API URL --- params = { "fType": 2, "count": 1, "channelId": channel_id, "country": country, "protocol": protocol, "sessionDuration": session_duration, "token": token, } if state: params["state"] = state if city: params["city"] = city url = "https://api.nstproxy.com/api/v1/generate/apiproxies" try: response = requests.get(url, params=params, timeout=10) response.raise_for_status() data = response.json() # --- Handle API error response --- if isinstance(data, dict) and data.get("err"): raise PermissionError(f"[NSTProxy] API Error: {data.get('msg', 'Unknown error')}") if not isinstance(data, list) or not data: raise ValueError("[NSTProxy] Invalid API response — expected a non-empty list") proxy_info = data[0] # --- Apply proxy config --- self.proxy_config = ProxyConfig( server=f"{protocol}://{proxy_info['ip']}:{proxy_info['port']}", username=proxy_info["username"], password=proxy_info["password"], ) except Exception as e: print(f"[NSTProxy] ❌ Failed to set proxy: {e}") raise class VirtualScrollConfig: """Configuration for virtual scroll handling. This config enables capturing content from pages with virtualized scrolling (like Twitter, Instagram feeds) where DOM elements are recycled as user scrolls. """ def __init__( self, container_selector: str, scroll_count: int = 10, scroll_by: Union[str, int] = "container_height", wait_after_scroll: float = 0.5, ): """ Initialize virtual scroll configuration. Args: container_selector: CSS selector for the scrollable container scroll_count: Maximum number of scrolls to perform scroll_by: Amount to scroll - can be: - "container_height": scroll by container's height - "page_height": scroll by viewport height - int: fixed pixel amount wait_after_scroll: Seconds to wait after each scroll for content to load """ self.container_selector = container_selector self.scroll_count = scroll_count self.scroll_by = scroll_by self.wait_after_scroll = wait_after_scroll def to_dict(self) -> dict: """Convert to dictionary for serialization.""" return { "container_selector": self.container_selector, "scroll_count": self.scroll_count, "scroll_by": self.scroll_by, "wait_after_scroll": self.wait_after_scroll, } @classmethod def from_dict(cls, data: dict) -> "VirtualScrollConfig": """Create instance from dictionary.""" return cls(**data) class LinkPreviewConfig: """Configuration for link head extraction and scoring.""" def __init__( self, include_internal: bool = True, include_external: bool = False, include_patterns: Optional[List[str]] = None, exclude_patterns: Optional[List[str]] = None, concurrency: int = 10, timeout: int = 5, max_links: int = 100, query: Optional[str] = None, score_threshold: Optional[float] = None, verbose: bool = False ): """ Initialize link extraction configuration.
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/content_filter_strategy.py
crawl4ai/content_filter_strategy.py
import inspect import re import time from bs4 import BeautifulSoup, Tag from typing import List, Tuple, Dict, Optional from rank_bm25 import BM25Okapi from collections import deque from bs4 import NavigableString, Comment from .utils import ( clean_tokens, perform_completion_with_backoff, escape_json_string, sanitize_html, get_home_folder, extract_xml_data, merge_chunks, ) from .types import LLMConfig from .config import DEFAULT_PROVIDER, OVERLAP_RATE, WORD_TOKEN_RATE from abc import ABC, abstractmethod import math from snowballstemmer import stemmer from .models import TokenUsage from .prompts import PROMPT_FILTER_CONTENT import json import hashlib from pathlib import Path from concurrent.futures import ThreadPoolExecutor from .async_logger import AsyncLogger, LogLevel, LogColor class RelevantContentFilter(ABC): """Abstract base class for content filtering strategies""" def __init__( self, user_query: str = None, verbose: bool = False, logger: Optional[AsyncLogger] = None, ): """ Initializes the RelevantContentFilter class with optional user query. Args: user_query (str): User query for filtering (optional). verbose (bool): Enable verbose logging (default: False). """ self.user_query = user_query self.included_tags = { # Primary structure "article", "main", "section", "div", # List structures "ul", "ol", "li", "dl", "dt", "dd", # Text content "p", "span", "blockquote", "pre", "code", # Headers "h1", "h2", "h3", "h4", "h5", "h6", # Tables "table", "thead", "tbody", "tr", "td", "th", # Other semantic elements "figure", "figcaption", "details", "summary", # Text formatting "em", "strong", "b", "i", "mark", "small", # Rich content "time", "address", "cite", "q", } self.excluded_tags = { "nav", "footer", "header", "aside", "script", "style", "form", "iframe", "noscript", } self.header_tags = {"h1", "h2", "h3", "h4", "h5", "h6"} self.negative_patterns = re.compile( r"nav|footer|header|sidebar|ads|comment|promo|advert|social|share", re.I ) self.min_word_count = 2 self.verbose = False self.logger = logger @abstractmethod def filter_content(self, html: str) -> List[str]: """Abstract method to be implemented by specific filtering strategies""" pass def extract_page_query(self, soup: BeautifulSoup, body: Tag) -> str: """Common method to extract page metadata with fallbacks""" if self.user_query: return self.user_query query_parts = [] # Title try: title = soup.title.string if title: query_parts.append(title) except Exception: pass if soup.find("h1"): query_parts.append(soup.find("h1").get_text()) # Meta tags temp = "" for meta_name in ["keywords", "description"]: meta = soup.find("meta", attrs={"name": meta_name}) if meta and meta.get("content"): query_parts.append(meta["content"]) temp += meta["content"] # If still empty, grab first significant paragraph if not temp: # Find the first tag P thatits text contains more than 50 characters for p in body.find_all("p"): if len(p.get_text()) > 150: query_parts.append(p.get_text()[:150]) break return " ".join(filter(None, query_parts)) def extract_text_chunks( self, body: Tag, min_word_threshold: int = None ) -> List[Tuple[str, str]]: """ Extracts text chunks from a BeautifulSoup body element while preserving order. Returns list of tuples (text, tag_name) for classification. Args: body: BeautifulSoup Tag object representing the body element Returns: List of (text, tag_name) tuples """ # Tags to ignore - inline elements that shouldn't break text flow INLINE_TAGS = { "a", "abbr", "acronym", "b", "bdo", "big", "br", "button", "cite", "code", "dfn", "em", "i", "img", "input", "kbd", "label", "map", "object", "q", "samp", "script", "select", "small", "span", "strong", "sub", "sup", "textarea", "time", "tt", "var", } # Tags that typically contain meaningful headers HEADER_TAGS = {"h1", "h2", "h3", "h4", "h5", "h6", "header"} chunks = [] current_text = [] chunk_index = 0 def should_break_chunk(tag: Tag) -> bool: """Determine if a tag should cause a break in the current text chunk""" return tag.name not in INLINE_TAGS and not ( tag.name == "p" and len(current_text) == 0 ) # Use deque for efficient push/pop operations stack = deque([(body, False)]) while stack: element, visited = stack.pop() if visited: # End of block element - flush accumulated text if current_text and should_break_chunk(element): text = " ".join("".join(current_text).split()) if text: tag_type = ( "header" if element.name in HEADER_TAGS else "content" ) chunks.append((chunk_index, text, tag_type, element)) chunk_index += 1 current_text = [] continue if isinstance(element, NavigableString): if str(element).strip(): current_text.append(str(element).strip()) continue # Pre-allocate children to avoid multiple list operations children = list(element.children) if not children: continue # Mark block for revisit after processing children stack.append((element, True)) # Add children in reverse order for correct processing for child in reversed(children): if isinstance(child, (Tag, NavigableString)): stack.append((child, False)) # Handle any remaining text if current_text: text = " ".join("".join(current_text).split()) if text: chunks.append((chunk_index, text, "content", body)) if min_word_threshold: chunks = [ chunk for chunk in chunks if len(chunk[1].split()) >= min_word_threshold ] return chunks def _deprecated_extract_text_chunks( self, soup: BeautifulSoup ) -> List[Tuple[int, str, Tag]]: """Common method for extracting text chunks""" _text_cache = {} def fast_text(element: Tag) -> str: elem_id = id(element) if elem_id in _text_cache: return _text_cache[elem_id] texts = [] for content in element.contents: if isinstance(content, str): text = content.strip() if text: texts.append(text) result = " ".join(texts) _text_cache[elem_id] = result return result candidates = [] index = 0 def dfs(element): nonlocal index if isinstance(element, Tag): if element.name in self.included_tags: if not self.is_excluded(element): text = fast_text(element) word_count = len(text.split()) # Headers pass through with adjusted minimum if element.name in self.header_tags: if word_count >= 3: # Minimal sanity check for headers candidates.append((index, text, element)) index += 1 # Regular content uses standard minimum elif word_count >= self.min_word_count: candidates.append((index, text, element)) index += 1 for child in element.children: dfs(child) dfs(soup.body if soup.body else soup) return candidates def is_excluded(self, tag: Tag) -> bool: """Common method for exclusion logic""" if tag.name in self.excluded_tags: return True class_id = " ".join( filter(None, [" ".join(tag.get("class", [])), tag.get("id", "")]) ) return bool(self.negative_patterns.search(class_id)) def clean_element(self, tag: Tag) -> str: """Common method for cleaning HTML elements with minimal overhead""" if not tag or not isinstance(tag, Tag): return "" unwanted_tags = {"script", "style", "aside", "form", "iframe", "noscript"} unwanted_attrs = { "style", "onclick", "onmouseover", "align", "bgcolor", "class", "id", } # Use string builder pattern for better performance builder = [] def render_tag(elem): if not isinstance(elem, Tag): if isinstance(elem, str): builder.append(elem.strip()) return if elem.name in unwanted_tags: return # Start tag builder.append(f"<{elem.name}") # Add cleaned attributes attrs = {k: v for k, v in elem.attrs.items() if k not in unwanted_attrs} for key, value in attrs.items(): builder.append(f' {key}="{value}"') builder.append(">") # Process children for child in elem.children: render_tag(child) # Close tag builder.append(f"</{elem.name}>") try: render_tag(tag) return "".join(builder) except Exception: return str(tag) # Fallback to original if anything fails class BM25ContentFilter(RelevantContentFilter): """ Content filtering using BM25 algorithm with priority tag handling. How it works: 1. Extracts page metadata with fallbacks. 2. Extracts text chunks from the body element. 3. Tokenizes the corpus and query. 4. Applies BM25 algorithm to calculate scores for each chunk. 5. Filters out chunks below the threshold. 6. Sorts chunks by score in descending order. 7. Returns the top N chunks. Attributes: user_query (str): User query for filtering (optional). bm25_threshold (float): BM25 threshold for filtering (default: 1.0). language (str): Language for stemming (default: 'english'). Methods: filter_content(self, html: str, min_word_threshold: int = None) """ def __init__( self, user_query: str = None, bm25_threshold: float = 1.0, language: str = "english", use_stemming: bool = True, ): """ Initializes the BM25ContentFilter class, if not provided, falls back to page metadata. Note: If no query is given and no page metadata is available, then it tries to pick up the first significant paragraph. Args: user_query (str): User query for filtering (optional). bm25_threshold (float): BM25 threshold for filtering (default: 1.0). language (str): Language for stemming (default: 'english'). use_stemming (bool): Whether to apply stemming (default: True). """ super().__init__(user_query=user_query) self.bm25_threshold = bm25_threshold self.use_stemming = use_stemming self.priority_tags = { "h1": 5.0, "h2": 4.0, "h3": 3.0, "title": 4.0, "strong": 2.0, "b": 1.5, "em": 1.5, "blockquote": 2.0, "code": 2.0, "pre": 1.5, "th": 1.5, # Table headers } self.stemmer = stemmer(language) if use_stemming else None def filter_content(self, html: str, min_word_threshold: int = None) -> List[str]: """ Implements content filtering using BM25 algorithm with priority tag handling. Note: This method implements the filtering logic for the BM25ContentFilter class. It takes HTML content as input and returns a list of filtered text chunks. Args: html (str): HTML content to be filtered. min_word_threshold (int): Minimum word threshold for filtering (optional). Returns: List[str]: List of filtered text chunks. """ if not html or not isinstance(html, str): return [] soup = BeautifulSoup(html, "lxml") # Check if body is present if not soup.body: # Wrap in body tag if missing soup = BeautifulSoup(f"<body>{html}</body>", "lxml") body = soup.find("body") query = self.extract_page_query(soup, body) if not query: return [] # return [self.clean_element(soup)] candidates = self.extract_text_chunks(body, min_word_threshold) if not candidates: return [] # Tokenize corpus # tokenized_corpus = [chunk.lower().split() for _, chunk, _, _ in candidates] # tokenized_query = query.lower().split() # tokenized_corpus = [[ps.stem(word) for word in chunk.lower().split()] # for _, chunk, _, _ in candidates] # tokenized_query = [ps.stem(word) for word in query.lower().split()] if self.use_stemming: tokenized_corpus = [ [self.stemmer.stemWord(word) for word in chunk.lower().split()] for _, chunk, _, _ in candidates ] tokenized_query = [ self.stemmer.stemWord(word) for word in query.lower().split() ] else: tokenized_corpus = [ chunk.lower().split() for _, chunk, _, _ in candidates ] tokenized_query = query.lower().split() # tokenized_corpus = [[self.stemmer.stemWord(word) for word in tokenize_text(chunk.lower())] # for _, chunk, _, _ in candidates] # tokenized_query = [self.stemmer.stemWord(word) for word in tokenize_text(query.lower())] # Clean from stop words and noise tokenized_corpus = [clean_tokens(tokens) for tokens in tokenized_corpus] tokenized_query = clean_tokens(tokenized_query) bm25 = BM25Okapi(tokenized_corpus) scores = bm25.get_scores(tokenized_query) # Adjust scores with tag weights adjusted_candidates = [] for score, (index, chunk, tag_type, tag) in zip(scores, candidates): tag_weight = self.priority_tags.get(tag.name, 1.0) adjusted_score = score * tag_weight adjusted_candidates.append((adjusted_score, index, chunk, tag)) # Filter candidates by threshold selected_candidates = [ (index, chunk, tag) for adjusted_score, index, chunk, tag in adjusted_candidates if adjusted_score >= self.bm25_threshold ] if not selected_candidates: return [] # Sort selected candidates by original document order selected_candidates.sort(key=lambda x: x[0]) return [self.clean_element(tag) for _, _, tag in selected_candidates] class PruningContentFilter(RelevantContentFilter): """ Content filtering using pruning algorithm with dynamic threshold. How it works: 1. Extracts page metadata with fallbacks. 2. Extracts text chunks from the body element. 3. Applies pruning algorithm to calculate scores for each chunk. 4. Filters out chunks below the threshold. 5. Sorts chunks by score in descending order. 6. Returns the top N chunks. Attributes: user_query (str): User query for filtering (optional), if not provided, falls back to page metadata. min_word_threshold (int): Minimum word threshold for filtering (optional). threshold_type (str): Threshold type for dynamic threshold (default: 'fixed'). threshold (float): Fixed threshold value (default: 0.48). Methods: filter_content(self, html: str, min_word_threshold: int = None): """ def __init__( self, user_query: str = None, min_word_threshold: int = None, threshold_type: str = "fixed", threshold: float = 0.48, ): """ Initializes the PruningContentFilter class, if not provided, falls back to page metadata. Note: If no query is given and no page metadata is available, then it tries to pick up the first significant paragraph. Args: user_query (str): User query for filtering (optional). min_word_threshold (int): Minimum word threshold for filtering (optional). threshold_type (str): Threshold type for dynamic threshold (default: 'fixed'). threshold (float): Fixed threshold value (default: 0.48). """ super().__init__(None) self.min_word_threshold = min_word_threshold self.threshold_type = threshold_type self.threshold = threshold # Add tag importance for dynamic threshold self.tag_importance = { "article": 1.5, "main": 1.4, "section": 1.3, "p": 1.2, "h1": 1.4, "h2": 1.3, "h3": 1.2, "div": 0.7, "span": 0.6, } # Metric configuration self.metric_config = { "text_density": True, "link_density": True, "tag_weight": True, "class_id_weight": True, "text_length": True, } self.metric_weights = { "text_density": 0.4, "link_density": 0.2, "tag_weight": 0.2, "class_id_weight": 0.1, "text_length": 0.1, } self.tag_weights = { "div": 0.5, "p": 1.0, "article": 1.5, "section": 1.0, "span": 0.3, "li": 0.5, "ul": 0.5, "ol": 0.5, "h1": 1.2, "h2": 1.1, "h3": 1.0, "h4": 0.9, "h5": 0.8, "h6": 0.7, } def filter_content(self, html: str, min_word_threshold: int = None) -> List[str]: """ Implements content filtering using pruning algorithm with dynamic threshold. Note: This method implements the filtering logic for the PruningContentFilter class. It takes HTML content as input and returns a list of filtered text chunks. Args: html (str): HTML content to be filtered. min_word_threshold (int): Minimum word threshold for filtering (optional). Returns: List[str]: List of filtered text chunks. """ if not html or not isinstance(html, str): return [] soup = BeautifulSoup(html, "lxml") if not soup.body: soup = BeautifulSoup(f"<body>{html}</body>", "lxml") # Remove comments and unwanted tags self._remove_comments(soup) self._remove_unwanted_tags(soup) # Prune tree starting from body body = soup.find("body") self._prune_tree(body) # Extract remaining content as list of HTML strings content_blocks = [] for element in body.children: if isinstance(element, str) or not hasattr(element, "name"): continue if len(element.get_text(strip=True)) > 0: content_blocks.append(str(element)) return content_blocks def _remove_comments(self, soup): """Removes HTML comments""" for element in soup(text=lambda text: isinstance(text, Comment)): element.extract() def _remove_unwanted_tags(self, soup): """Removes unwanted tags""" for tag in self.excluded_tags: for element in soup.find_all(tag): element.decompose() def _prune_tree(self, node): """ Prunes the tree starting from the given node. Args: node (Tag): The node from which the pruning starts. """ if not node or not hasattr(node, "name") or node.name is None: return text_len = len(node.get_text(strip=True)) tag_len = len(node.encode_contents().decode("utf-8")) link_text_len = sum( len(s.strip()) for s in (a.string for a in node.find_all("a", recursive=False)) if s ) metrics = { "node": node, "tag_name": node.name, "text_len": text_len, "tag_len": tag_len, "link_text_len": link_text_len, } score = self._compute_composite_score(metrics, text_len, tag_len, link_text_len) if self.threshold_type == "fixed": should_remove = score < self.threshold else: # dynamic tag_importance = self.tag_importance.get(node.name, 0.7) text_ratio = text_len / tag_len if tag_len > 0 else 0 link_ratio = link_text_len / text_len if text_len > 0 else 1 threshold = self.threshold # base threshold if tag_importance > 1: threshold *= 0.8 if text_ratio > 0.4: threshold *= 0.9 if link_ratio > 0.6: threshold *= 1.2 should_remove = score < threshold if should_remove: node.decompose() else: children = [child for child in node.children if hasattr(child, "name")] for child in children: self._prune_tree(child) def _compute_composite_score(self, metrics, text_len, tag_len, link_text_len): """Computes the composite score""" if self.min_word_threshold: # Get raw text from metrics node - avoid extra processing text = metrics["node"].get_text(strip=True) word_count = text.count(" ") + 1 if word_count < self.min_word_threshold: return -1.0 # Guaranteed removal score = 0.0 total_weight = 0.0 if self.metric_config["text_density"]: density = text_len / tag_len if tag_len > 0 else 0 score += self.metric_weights["text_density"] * density total_weight += self.metric_weights["text_density"] if self.metric_config["link_density"]: density = 1 - (link_text_len / text_len if text_len > 0 else 0) score += self.metric_weights["link_density"] * density total_weight += self.metric_weights["link_density"] if self.metric_config["tag_weight"]: tag_score = self.tag_weights.get(metrics["tag_name"], 0.5) score += self.metric_weights["tag_weight"] * tag_score total_weight += self.metric_weights["tag_weight"] if self.metric_config["class_id_weight"]: class_score = self._compute_class_id_weight(metrics["node"]) score += self.metric_weights["class_id_weight"] * max(0, class_score) total_weight += self.metric_weights["class_id_weight"] if self.metric_config["text_length"]: score += self.metric_weights["text_length"] * math.log(text_len + 1) total_weight += self.metric_weights["text_length"] return score / total_weight if total_weight > 0 else 0 def _compute_class_id_weight(self, node): """Computes the class ID weight""" class_id_score = 0 if "class" in node.attrs: classes = " ".join(node["class"]) if self.negative_patterns.match(classes): class_id_score -= 0.5 if "id" in node.attrs: element_id = node["id"] if self.negative_patterns.match(element_id): class_id_score -= 0.5 return class_id_score class LLMContentFilter(RelevantContentFilter): """Content filtering using LLMs to generate relevant markdown. How it works: 1. Extracts page metadata with fallbacks. 2. Extracts text chunks from the body element. 3. Applies LLMs to generate markdown for each chunk. 4. Filters out chunks below the threshold. 5. Sorts chunks by score in descending order. 6. Returns the top N chunks. Attributes: llm_config (LLMConfig): LLM configuration object. instruction (str): Instruction for LLM markdown generation chunk_token_threshold (int): Chunk token threshold for splitting (default: 1e9). overlap_rate (float): Overlap rate for chunking (default: 0.5). word_token_rate (float): Word token rate for chunking (default: 0.2). verbose (bool): Enable verbose logging (default: False). logger (AsyncLogger): Custom logger for LLM operations (optional). """ _UNWANTED_PROPS = { 'provider' : 'Instead, use llm_config=LLMConfig(provider="...")', 'api_token' : 'Instead, use llm_config=LlMConfig(api_token="...")', 'base_url' : 'Instead, use llm_config=LLMConfig(base_url="...")', 'api_base' : 'Instead, use llm_config=LLMConfig(base_url="...")', } def __init__( self, llm_config: "LLMConfig" = None, instruction: str = None, chunk_token_threshold: int = int(1e9), overlap_rate: float = OVERLAP_RATE, word_token_rate: float = WORD_TOKEN_RATE, # char_token_rate: float = WORD_TOKEN_RATE * 5, # chunk_mode: str = "char", verbose: bool = False, logger: Optional[AsyncLogger] = None, ignore_cache: bool = True, # Deprecated properties provider: str = DEFAULT_PROVIDER, api_token: Optional[str] = None, base_url: Optional[str] = None, api_base: Optional[str] = None, extra_args: Dict = None, ): super().__init__(None) self.provider = provider self.api_token = api_token self.base_url = base_url or api_base self.llm_config = llm_config self.instruction = instruction self.chunk_token_threshold = chunk_token_threshold self.overlap_rate = overlap_rate self.word_token_rate = word_token_rate or WORD_TOKEN_RATE # self.chunk_mode: str = chunk_mode # self.char_token_rate = char_token_rate or word_token_rate / 5 # self.token_rate = word_token_rate if chunk_mode == "word" else self.char_token_rate self.token_rate = word_token_rate or WORD_TOKEN_RATE self.extra_args = extra_args or {} self.ignore_cache = ignore_cache self.verbose = verbose # Setup logger with custom styling for LLM operations if logger: self.logger = logger elif verbose: self.logger = AsyncLogger( verbose=verbose, icons={ **AsyncLogger.DEFAULT_ICONS, "LLM": "★", # Star for LLM operations "CHUNK": "◈", # Diamond for chunks "CACHE": "⚡", # Lightning for cache operations }, colors={ **AsyncLogger.DEFAULT_COLORS, LogLevel.INFO: LogColor.DIM_MAGENTA # Dimmed purple for LLM ops }, ) else: self.logger = None self.usages = [] self.total_usage = TokenUsage() def __setattr__(self, name, value): """Handle attribute setting.""" # TODO: Planning to set properties dynamically based on the __init__ signature sig = inspect.signature(self.__init__) all_params = sig.parameters # Dictionary of parameter names and their details if name in self._UNWANTED_PROPS and value is not all_params[name].default: raise AttributeError(f"Setting '{name}' is deprecated. {self._UNWANTED_PROPS[name]}") super().__setattr__(name, value) def _get_cache_key(self, html: str, instruction: str) -> str: """Generate a unique cache key based on HTML and instruction""" content = f"{html}{instruction}" return hashlib.md5(content.encode()).hexdigest() def _merge_chunks(self, text: str) -> List[str]: """Split text into chunks with overlap using char or word mode.""" ov = int(self.chunk_token_threshold * self.overlap_rate) sections = merge_chunks( docs=[text], target_size=self.chunk_token_threshold, overlap=ov, word_token_ratio=self.word_token_rate, ) return sections def filter_content(self, html: str, ignore_cache: bool = True) -> List[str]: if not html or not isinstance(html, str): return [] if self.logger: self.logger.info( "Starting LLM markdown content filtering process", tag="LLM", params={"provider": self.llm_config.provider}, colors={"provider": LogColor.CYAN}, ) # Cache handling cache_dir = Path(get_home_folder()) / "llm_cache" / "content_filter" cache_dir.mkdir(parents=True, exist_ok=True) cache_key = self._get_cache_key(html, self.instruction or "") cache_file = cache_dir / f"{cache_key}.json" # if ignore_cache == None: ignore_cache = self.ignore_cache if not ignore_cache and cache_file.exists(): if self.logger: self.logger.info("Found cached markdown result", tag="CACHE") try: with cache_file.open("r") as f: cached_data = json.load(f) usage = TokenUsage(**cached_data["usage"]) self.usages.append(usage) self.total_usage.completion_tokens += usage.completion_tokens self.total_usage.prompt_tokens += usage.prompt_tokens self.total_usage.total_tokens += usage.total_tokens return cached_data["blocks"] except Exception as e: if self.logger: self.logger.error( f"LLM markdown: Cache read error: {str(e)}", tag="CACHE" ) # Split into chunks html_chunks = self._merge_chunks(html) if self.logger: self.logger.info( "LLM markdown: Split content into {chunk_count} chunks", tag="CHUNK", params={"chunk_count": len(html_chunks)}, colors={"chunk_count": LogColor.YELLOW}, ) start_time = time.time() # Process chunks in parallel with ThreadPoolExecutor(max_workers=4) as executor: futures = [] for i, chunk in enumerate(html_chunks): if self.logger: self.logger.debug( "LLM markdown: Processing chunk {chunk_num}/{total_chunks}", tag="CHUNK",
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true
unclecode/crawl4ai
https://github.com/unclecode/crawl4ai/blob/c85f56b085a1d5b62779774d48887345e566869b/crawl4ai/browser_profiler.py
crawl4ai/browser_profiler.py
""" Browser Profiler Module This module provides a dedicated class for managing browser profiles that can be used for identity-based crawling with Crawl4AI. """ import os import asyncio import signal import sys import datetime import uuid import shutil import json import subprocess import time from typing import List, Dict, Optional, Any from rich.console import Console from .async_configs import BrowserConfig from .browser_manager import ManagedBrowser from .async_logger import AsyncLogger, AsyncLoggerBase, LogColor from .utils import get_home_folder class BrowserProfiler: """ A dedicated class for managing browser profiles for Crawl4AI. The BrowserProfiler allows you to: - Create browser profiles interactively - List available profiles - Delete profiles when no longer needed - Get profile paths for use in BrowserConfig Profiles are stored by default in ~/.crawl4ai/profiles/ """ def __init__(self, logger: Optional[AsyncLoggerBase] = None): """ Initialize the BrowserProfiler. Args: logger (AsyncLoggerBase, optional): Logger for outputting messages. If None, a default AsyncLogger will be created. """ # Initialize rich console for colorful input prompts self.console = Console() # Create a logger if not provided if logger is None: self.logger = AsyncLogger(verbose=True) elif not isinstance(logger, AsyncLoggerBase): self.logger = AsyncLogger(verbose=True) else: self.logger = logger # Ensure profiles directory exists self.profiles_dir = os.path.join(get_home_folder(), "profiles") os.makedirs(self.profiles_dir, exist_ok=True) # Builtin browser config file self.builtin_browser_dir = os.path.join(get_home_folder(), "builtin-browser") self.builtin_config_file = os.path.join(self.builtin_browser_dir, "browser_config.json") os.makedirs(self.builtin_browser_dir, exist_ok=True) def _is_windows(self) -> bool: """Check if running on Windows platform.""" return sys.platform.startswith('win') or sys.platform == 'cygwin' def _is_macos(self) -> bool: """Check if running on macOS platform.""" return sys.platform == 'darwin' def _is_linux(self) -> bool: """Check if running on Linux platform.""" return sys.platform.startswith('linux') def _get_quit_message(self, tag: str) -> str: """Get appropriate quit message based on context.""" if tag == "PROFILE": return "Closing browser and saving profile..." elif tag == "CDP": return "Closing browser..." else: return "Closing browser..." async def _listen_windows(self, user_done_event, check_browser_process, tag: str): """Windows-specific keyboard listener using msvcrt.""" try: import msvcrt except ImportError: raise ImportError("msvcrt module not available on this platform") while True: try: # Check for keyboard input if msvcrt.kbhit(): raw = msvcrt.getch() # Handle Unicode decoding more robustly key = None try: key = raw.decode("utf-8") except UnicodeDecodeError: try: # Try different encodings key = raw.decode("latin1") except UnicodeDecodeError: # Skip if we can't decode continue # Validate key if not key or len(key) != 1: continue # Check for printable characters only if not key.isprintable(): continue # Check for quit command if key.lower() == "q": self.logger.info( self._get_quit_message(tag), tag=tag, base_color=LogColor.GREEN ) user_done_event.set() return # Check if browser process ended if await check_browser_process(): return # Small delay to prevent busy waiting await asyncio.sleep(0.1) except Exception as e: self.logger.warning(f"Error in Windows keyboard listener: {e}", tag=tag) # Continue trying instead of failing completely await asyncio.sleep(0.1) continue async def _listen_unix(self, user_done_event: asyncio.Event, check_browser_process, tag: str): """Unix/Linux/macOS keyboard listener using termios and select.""" try: import termios import tty import select except ImportError: raise ImportError("termios/tty/select modules not available on this platform") # Get stdin file descriptor try: fd = sys.stdin.fileno() except (AttributeError, OSError): raise ImportError("stdin is not a terminal") # Save original terminal settings old_settings = None try: old_settings = termios.tcgetattr(fd) except termios.error as e: raise ImportError(f"Cannot get terminal attributes: {e}") try: # Switch to non-canonical mode (cbreak mode) tty.setcbreak(fd) while True: try: # Use select to check if input is available (non-blocking) # Timeout of 0.5 seconds to periodically check browser process readable, _, _ = select.select([sys.stdin], [], [], 0.5) if readable: # Read one character key = sys.stdin.read(1) if key and key.lower() == "q": self.logger.info( self._get_quit_message(tag), tag=tag, base_color=LogColor.GREEN ) user_done_event.set() return # Check if browser process ended if await check_browser_process(): return # Small delay to prevent busy waiting await asyncio.sleep(0.1) except (KeyboardInterrupt, EOFError): # Handle Ctrl+C or EOF gracefully self.logger.info("Keyboard interrupt received", tag=tag) user_done_event.set() return except Exception as e: self.logger.warning(f"Error in Unix keyboard listener: {e}", tag=tag) await asyncio.sleep(0.1) continue finally: # Always restore terminal settings if old_settings is not None: try: termios.tcsetattr(fd, termios.TCSADRAIN, old_settings) except Exception as e: self.logger.error(f"Failed to restore terminal settings: {e}", tag=tag) async def _listen_fallback(self, user_done_event: asyncio.Event, check_browser_process, tag: str): """Fallback keyboard listener using simple input() method.""" self.logger.info("Using fallback input mode. Type 'q' and press Enter to quit.", tag=tag) # Run input in a separate thread to avoid blocking import threading import queue input_queue = queue.Queue() def input_thread(): """Thread function to handle input.""" try: while not user_done_event.is_set(): try: # Use input() with a prompt user_input = input("Press 'q' + Enter to quit: ").strip().lower() input_queue.put(user_input) if user_input == 'q': break except (EOFError, KeyboardInterrupt): input_queue.put('q') break except Exception as e: self.logger.warning(f"Error in input thread: {e}", tag=tag) break except Exception as e: self.logger.error(f"Input thread failed: {e}", tag=tag) # Start input thread thread = threading.Thread(target=input_thread, daemon=True) thread.start() try: while not user_done_event.is_set(): # Check for user input try: user_input = input_queue.get_nowait() if user_input == 'q': self.logger.info( self._get_quit_message(tag), tag=tag, base_color=LogColor.GREEN ) user_done_event.set() return except queue.Empty: pass # Check if browser process ended if await check_browser_process(): return # Small delay await asyncio.sleep(0.5) except Exception as e: self.logger.error(f"Fallback listener failed: {e}", tag=tag) user_done_event.set() async def create_profile(self, profile_name: Optional[str] = None, browser_config: Optional[BrowserConfig] = None) -> Optional[str]: """ Creates a browser profile by launching a browser for interactive user setup and waits until the user closes it. The profile is stored in a directory that can be used later with BrowserConfig.user_data_dir. Args: profile_name (str, optional): Name for the profile directory. If None, a name is generated based on timestamp. browser_config (BrowserConfig, optional): Configuration for the browser. If None, a default configuration is used with headless=False. Returns: str: Path to the created profile directory, or None if creation failed Example: ```python profiler = BrowserProfiler() # Create a profile interactively profile_path = await profiler.create_profile( profile_name="my-login-profile" ) # Use the profile in a crawler browser_config = BrowserConfig( headless=True, use_managed_browser=True, user_data_dir=profile_path ) async with AsyncWebCrawler(config=browser_config) as crawler: # The crawler will now use your profile with all your cookies and login state result = await crawler.arun("https://example.com/dashboard") ``` """ # Create default browser config if none provided if browser_config is None: from .async_configs import BrowserConfig browser_config = BrowserConfig( browser_type="chromium", headless=False, # Must be visible for user interaction verbose=True ) else: # Ensure headless is False for user interaction browser_config.headless = False # Generate profile name if not provided if not profile_name: timestamp = datetime.datetime.now().strftime("%Y%m%d_%H%M%S") profile_name = f"profile_{timestamp}_{uuid.uuid4().hex[:6]}" # Sanitize profile name (replace spaces and special chars) profile_name = "".join(c if c.isalnum() or c in "-_" else "_" for c in profile_name) # Set user data directory profile_path = os.path.join(self.profiles_dir, profile_name) os.makedirs(profile_path, exist_ok=True) # Print instructions for the user with rich formatting border = f"{'='*80}" self.logger.info("{border}", tag="PROFILE", params={"border": f"\n{border}"}, colors={"border": LogColor.CYAN}) self.logger.info("Creating browser profile: {profile_name}", tag="PROFILE", params={"profile_name": profile_name}, colors={"profile_name": LogColor.GREEN}) self.logger.info("Profile directory: {profile_path}", tag="PROFILE", params={"profile_path": profile_path}, colors={"profile_path": LogColor.YELLOW}) self.logger.info("\nInstructions:", tag="PROFILE") self.logger.info("1. A browser window will open for you to set up your profile.", tag="PROFILE") self.logger.info("{segment}, configure settings, etc. as needed.", tag="PROFILE", params={"segment": "2. Log in to websites"}, colors={"segment": LogColor.CYAN}) self.logger.info("3. When you're done, {segment} to close the browser.", tag="PROFILE", params={"segment": "press 'q' in this terminal"}, colors={"segment": LogColor.YELLOW}) self.logger.info("4. The profile will be saved and ready to use with Crawl4AI.", tag="PROFILE") self.logger.info("{border}", tag="PROFILE", params={"border": f"{border}\n"}, colors={"border": LogColor.CYAN}) browser_config.headless = False browser_config.user_data_dir = profile_path # Create managed browser instance managed_browser = ManagedBrowser( browser_config=browser_config, # user_data_dir=profile_path, # headless=False, # Must be visible logger=self.logger, # debugging_port=browser_config.debugging_port ) # Set up signal handlers to ensure cleanup on interrupt original_sigint = signal.getsignal(signal.SIGINT) original_sigterm = signal.getsignal(signal.SIGTERM) # Define cleanup handler for signals async def cleanup_handler(sig, frame): self.logger.warning("\nCleaning up browser process...", tag="PROFILE") await managed_browser.cleanup() # Restore original signal handlers signal.signal(signal.SIGINT, original_sigint) signal.signal(signal.SIGTERM, original_sigterm) if sig == signal.SIGINT: self.logger.error("Profile creation interrupted. Profile may be incomplete.", tag="PROFILE") sys.exit(1) # Set signal handlers def sigint_handler(sig, frame): asyncio.create_task(cleanup_handler(sig, frame)) signal.signal(signal.SIGINT, sigint_handler) signal.signal(signal.SIGTERM, sigint_handler) # Event to signal when user is done with the browser user_done_event = asyncio.Event() # Run keyboard input loop in a separate task async def listen_for_quit_command(): """Cross-platform keyboard listener that waits for 'q' key press.""" # First output the prompt self.logger.info( "Press {segment} when you've finished using the browser...", tag="PROFILE", params={"segment": "'q'"}, colors={"segment": LogColor.YELLOW}, base_color=LogColor.CYAN ) async def check_browser_process(): """Check if browser process is still running.""" if ( managed_browser.browser_process and managed_browser.browser_process.poll() is not None ): self.logger.info( "Browser already closed. Ending input listener.", tag="PROFILE" ) user_done_event.set() return True return False # Try platform-specific implementations with fallback try: if self._is_windows(): await self._listen_windows(user_done_event, check_browser_process, "PROFILE") else: await self._listen_unix(user_done_event, check_browser_process, "PROFILE") except Exception as e: self.logger.warning(f"Platform-specific keyboard listener failed: {e}", tag="PROFILE") self.logger.info("Falling back to simple input mode...", tag="PROFILE") await self._listen_fallback(user_done_event, check_browser_process, "PROFILE") try: from playwright.async_api import async_playwright # Start the browser # await managed_browser.start() # 1. ── Start the browser ───────────────────────────────────────── cdp_url = await managed_browser.start() # 2. ── Attach Playwright to that running Chrome ────────────────── pw = await async_playwright().start() browser = await pw.chromium.connect_over_cdp(cdp_url) # Grab the existing default context (there is always one) context = browser.contexts[0] # Check if browser started successfully browser_process = managed_browser.browser_process if not browser_process: self.logger.error("Failed to start browser process.", tag="PROFILE") return None self.logger.info("Browser launched. Waiting for you to finish...", tag="PROFILE") # Start listening for keyboard input listener_task = asyncio.create_task(listen_for_quit_command()) # Wait for either the user to press 'q' or for the browser process to exit naturally while not user_done_event.is_set() and browser_process.poll() is None: await asyncio.sleep(0.5) # Cancel the listener task if it's still running if not listener_task.done(): listener_task.cancel() try: await listener_task except asyncio.CancelledError: pass # 3. ── Persist storage state *before* we kill Chrome ───────────── state_file = os.path.join(profile_path, "storage_state.json") try: await context.storage_state(path=state_file) self.logger.info(f"[PROFILE].i storage_state saved → {state_file}", tag="PROFILE") except Exception as e: self.logger.warning(f"[PROFILE].w failed to save storage_state: {e}", tag="PROFILE") # 4. ── Close everything cleanly ────────────────────────────────── await browser.close() await pw.stop() # If the browser is still running and the user pressed 'q', terminate it if browser_process.poll() is None and user_done_event.is_set(): self.logger.info("Terminating browser process...", tag="PROFILE") await managed_browser.cleanup() self.logger.success(f"Browser closed. Profile saved at: {profile_path}", tag="PROFILE") except Exception as e: self.logger.error(f"Error creating profile: {e!s}", tag="PROFILE") await managed_browser.cleanup() return None finally: # Restore original signal handlers signal.signal(signal.SIGINT, original_sigint) signal.signal(signal.SIGTERM, original_sigterm) # Make sure browser is fully cleaned up await managed_browser.cleanup() # Return the profile path return profile_path def list_profiles(self) -> List[Dict[str, Any]]: """ Lists all available browser profiles in the Crawl4AI profiles directory. Returns: list: A list of dictionaries containing profile information: [{"name": "profile_name", "path": "/path/to/profile", "created": datetime, "type": "chromium|firefox"}] Example: ```python profiler = BrowserProfiler() # List all available profiles profiles = profiler.list_profiles() for profile in profiles: print(f"Profile: {profile['name']}") print(f" Path: {profile['path']}") print(f" Created: {profile['created']}") print(f" Browser type: {profile['type']}") ``` """ if not os.path.exists(self.profiles_dir): return [] profiles = [] for name in os.listdir(self.profiles_dir): profile_path = os.path.join(self.profiles_dir, name) # Skip if not a directory if not os.path.isdir(profile_path): continue # Check if this looks like a valid browser profile # For Chromium: Look for Preferences file # For Firefox: Look for prefs.js file is_valid = False if os.path.exists(os.path.join(profile_path, "Preferences")) or \ os.path.exists(os.path.join(profile_path, "Default", "Preferences")): is_valid = "chromium" elif os.path.exists(os.path.join(profile_path, "prefs.js")): is_valid = "firefox" if is_valid: # Get creation time created = datetime.datetime.fromtimestamp( os.path.getctime(profile_path) ) profiles.append({ "name": name, "path": profile_path, "created": created, "type": is_valid }) # Sort by creation time, newest first profiles.sort(key=lambda x: x["created"], reverse=True) return profiles def get_profile_path(self, profile_name: str) -> Optional[str]: """ Get the full path to a profile by name. Args: profile_name (str): Name of the profile (not the full path) Returns: str: Full path to the profile directory, or None if not found Example: ```python profiler = BrowserProfiler() path = profiler.get_profile_path("my-profile") if path: print(f"Profile path: {path}") else: print("Profile not found") ``` """ profile_path = os.path.join(self.profiles_dir, profile_name) # Check if path exists and is a valid profile if not os.path.isdir(profile_path): # Chrck if profile_name itself is full path if os.path.isabs(profile_name): profile_path = profile_name else: return None # Look for profile indicators is_profile = ( os.path.exists(os.path.join(profile_path, "Preferences")) or os.path.exists(os.path.join(profile_path, "Default", "Preferences")) or os.path.exists(os.path.join(profile_path, "prefs.js")) ) if not is_profile: return None # Not a valid browser profile return profile_path def delete_profile(self, profile_name_or_path: str) -> bool: """ Delete a browser profile by name or path. Args: profile_name_or_path (str): Name of the profile or full path to profile directory Returns: bool: True if the profile was deleted successfully, False otherwise Example: ```python profiler = BrowserProfiler() # Delete by name success = profiler.delete_profile("my-profile") # Delete by path success = profiler.delete_profile("/path/to/.crawl4ai/profiles/my-profile") ``` """ # Determine if input is a name or a path if os.path.isabs(profile_name_or_path): # Full path provided profile_path = profile_name_or_path else: # Just a name provided, construct path profile_path = os.path.join(self.profiles_dir, profile_name_or_path) # Check if path exists and is a valid profile if not os.path.isdir(profile_path): return False # Look for profile indicators is_profile = ( os.path.exists(os.path.join(profile_path, "Preferences")) or os.path.exists(os.path.join(profile_path, "Default", "Preferences")) or os.path.exists(os.path.join(profile_path, "prefs.js")) ) if not is_profile: return False # Not a valid browser profile # Delete the profile directory try: shutil.rmtree(profile_path) return True except Exception: return False async def interactive_manager(self, crawl_callback=None): """ Launch an interactive profile management console. Args: crawl_callback (callable, optional): Function to call when selecting option to use a profile for crawling. It will be called with (profile_path, url). Example: ```python profiler = BrowserProfiler() # Define a custom crawl function async def my_crawl_function(profile_path, url): print(f"Crawling {url} with profile {profile_path}") # Implement your crawling logic here # Start interactive manager await profiler.interactive_manager(crawl_callback=my_crawl_function) ``` """ while True: self.logger.info("\nProfile Management Options:", tag="MENU") self.logger.info("1. Create a new profile", tag="MENU", base_color=LogColor.GREEN) self.logger.info("2. List available profiles", tag="MENU", base_color=LogColor.YELLOW) self.logger.info("3. Delete a profile", tag="MENU", base_color=LogColor.RED) # Only show crawl option if callback provided if crawl_callback: self.logger.info("4. Use a profile to crawl a website", tag="MENU", base_color=LogColor.CYAN) self.logger.info("5. Exit", tag="MENU", base_color=LogColor.MAGENTA) exit_option = "5" else: self.logger.info("4. Exit", tag="MENU", base_color=LogColor.MAGENTA) exit_option = "4" self.logger.info(f"\n[cyan]Enter your choice (1-{exit_option}): [/cyan]", end="") choice = input() if choice == "1": # Create new profile self.console.print("[green]Enter a name for the new profile (or press Enter for auto-generated name): [/green]", end="") name = input() await self.create_profile(name or None) elif choice == "2": # List profiles profiles = self.list_profiles() if not profiles: self.logger.warning(" No profiles found. Create one first with option 1.", tag="PROFILES") continue # Print profile information self.logger.info("\nAvailable profiles:", tag="PROFILES") for i, profile in enumerate(profiles): self.logger.info(f"[{i+1}] {profile['name']}", tag="PROFILES") self.logger.info(f" Path: {profile['path']}", tag="PROFILES", base_color=LogColor.YELLOW) self.logger.info(f" Created: {profile['created'].strftime('%Y-%m-%d %H:%M:%S')}", tag="PROFILES") self.logger.info(f" Browser type: {profile['type']}", tag="PROFILES") self.logger.info("", tag="PROFILES") # Empty line for spacing elif choice == "3": # Delete profile profiles = self.list_profiles() if not profiles: self.logger.warning("No profiles found to delete", tag="PROFILES") continue # Display numbered list self.logger.info("\nAvailable profiles:", tag="PROFILES", base_color=LogColor.YELLOW) for i, profile in enumerate(profiles): self.logger.info(f"[{i+1}] {profile['name']}", tag="PROFILES") # Get profile to delete self.console.print("[red]Enter the number of the profile to delete (or 'c' to cancel): [/red]", end="") profile_idx = input() if profile_idx.lower() == 'c': continue try: idx = int(profile_idx) - 1 if 0 <= idx < len(profiles): profile_name = profiles[idx]["name"] self.logger.info(f"Deleting profile: [yellow]{profile_name}[/yellow]", tag="PROFILES") # Confirm deletion self.console.print("[red]Are you sure you want to delete this profile? (y/n): [/red]", end="") confirm = input() if confirm.lower() == 'y': success = self.delete_profile(profiles[idx]["path"]) if success: self.logger.success(f"Profile {profile_name} deleted successfully", tag="PROFILES") else: self.logger.error(f"Failed to delete profile {profile_name}", tag="PROFILES") else: self.logger.error("Invalid profile number", tag="PROFILES") except ValueError: self.logger.error("Please enter a valid number", tag="PROFILES") elif choice == "4" and crawl_callback: # Use profile to crawl a site profiles = self.list_profiles() if not profiles: self.logger.warning("No profiles found. Create one first.", tag="PROFILES") continue # Display numbered list self.logger.info("\nAvailable profiles:", tag="PROFILES", base_color=LogColor.YELLOW) for i, profile in enumerate(profiles): self.logger.info(f"[{i+1}] {profile['name']}", tag="PROFILES") # Get profile to use self.console.print("[cyan]Enter the number of the profile to use (or 'c' to cancel): [/cyan]", end="") profile_idx = input() if profile_idx.lower() == 'c':
python
Apache-2.0
c85f56b085a1d5b62779774d48887345e566869b
2026-01-04T14:38:51.943025Z
true