"""GPU utility functions for detecting and managing GPU availability.""" import logging from typing import Dict, Optional logger = logging.getLogger(__name__) def is_gpu_available() -> bool: """Check if GPU is available for deep learning models. Returns: True if GPU is available, False otherwise """ try: import torch if torch.cuda.is_available(): gpu_count = torch.cuda.device_count() gpu_name = torch.cuda.get_device_name(0) if gpu_count > 0 else "Unknown" logger.info(f"GPU detected: {gpu_name} (count: {gpu_count})") return True else: logger.info("No CUDA GPU available") return False except ImportError: logger.info("PyTorch not available, assuming no GPU") return False except Exception as e: logger.warning(f"Error checking GPU availability: {e}") return False def get_gpu_info() -> Dict: """Get detailed GPU information. Returns: Dictionary with GPU information """ info = { "available": False, "count": 0, "names": [], "memory": [] } try: import torch if torch.cuda.is_available(): info["available"] = True info["count"] = torch.cuda.device_count() info["names"] = [torch.cuda.get_device_name(i) for i in range(info["count"])] info["memory"] = [torch.cuda.get_device_properties(i).total_memory for i in range(info["count"])] except ImportError: pass except Exception as e: logger.warning(f"Error getting GPU info: {e}") return info def should_use_gpu_processor() -> bool: """Determine if GPU processor should be used based on GPU availability. Returns: True if GPU processor should be used, False otherwise """ return is_gpu_available() def get_processor_preference() -> str: """Get the preferred processor type based on system capabilities. Returns: 'gpu' if GPU is available Raises: RuntimeError: If GPU is not available """ if should_use_gpu_processor(): return 'gpu' else: raise RuntimeError( "GPU is not available. Please ensure CUDA is installed and a compatible GPU is present, " "or use cloud processing mode." )