{ "algorithm": "PPO", "total_timesteps": 150000, "learning_rate": 0.0003, "gamma": 0.99, "training_candles": 100000 }