""" Gradient Descent and Backpropagation Training System ================================================== This module provides a comprehensive implementation of gradient descent optimization algorithms and backpropagation for neural network training, specifically designed for the MangoMAS multi-agent system. Key Components: - Optimizers: SGD, Adam, AdamW with proper mathematical implementations - Backpropagation: Chain rule-based gradient computation - Training Loop: Complete training orchestration with monitoring - Loss Functions: Various loss implementations for different tasks - Monitoring: Comprehensive gradient and training metrics tracking Usage: from src.training.gradient_descent import GradientDescentTrainer trainer = GradientDescentTrainer() results = trainer.train_agent(agent_spec) """ from .optimizers import SGD, Adam, AdamW, Optimizer from .backpropagation import BackpropagationEngine from .training_loop import GradientDescentTrainer from .loss_functions import CrossEntropyLoss, KLDivergenceLoss, LossFunction from .monitoring import GradientMonitor, TrainingMonitor from .model_wrapper import ModelWrapper from .schedulers import LinearScheduler, CosineScheduler, StepScheduler __version__ = "1.0.0" __author__ = "MangoMAS Team" __all__ = [ "SGD", "Adam", "AdamW", "Optimizer", "BackpropagationEngine", "GradientDescentTrainer", "CrossEntropyLoss", "KLDivergenceLoss", "LossFunction", "GradientMonitor", "TrainingMonitor", "ModelWrapper", "LinearScheduler", "CosineScheduler", "StepScheduler" ]