{ "name": "root", "gauges": { "Pyramids.Policy.Entropy.mean": { "value": 0.4548112154006958, "min": 0.4548112154006958, "max": 1.4480611085891724, "count": 33 }, "Pyramids.Policy.Entropy.sum": { "value": 13775.322265625, "min": 13670.9609375, "max": 43928.3828125, "count": 33 }, "Pyramids.Step.mean": { "value": 989997.0, "min": 29952.0, "max": 989997.0, "count": 33 }, "Pyramids.Step.sum": { "value": 989997.0, "min": 29952.0, "max": 989997.0, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.mean": { "value": 0.36824795603752136, "min": -0.10178116708993912, "max": 0.42843231558799744, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.sum": { "value": 96.84921264648438, "min": -24.529260635375977, "max": 113.96299743652344, "count": 33 }, "Pyramids.Policy.RndValueEstimate.mean": { "value": 0.006934191100299358, "min": 0.0028177027124911547, "max": 0.34559914469718933, "count": 33 }, "Pyramids.Policy.RndValueEstimate.sum": { "value": 1.8236922025680542, "min": 0.7072433829307556, "max": 81.90699768066406, "count": 33 }, "Pyramids.Losses.PolicyLoss.mean": { "value": 0.06956796737377016, "min": 0.06379592067394758, "max": 0.07321309103909669, "count": 33 }, "Pyramids.Losses.PolicyLoss.sum": { "value": 0.9739515432327822, "min": 0.5015818004693098, "max": 1.070607882817664, "count": 33 }, "Pyramids.Losses.ValueLoss.mean": { "value": 0.014214756165829871, "min": 0.00025640409403211625, "max": 0.014862789750276578, "count": 33 }, "Pyramids.Losses.ValueLoss.sum": { "value": 0.1990065863216182, "min": 0.0028204450343532785, "max": 0.21029198639250052, "count": 33 }, "Pyramids.Policy.LearningRate.mean": { "value": 7.517518922764284e-06, "min": 7.517518922764284e-06, "max": 0.00029515063018788575, "count": 33 }, "Pyramids.Policy.LearningRate.sum": { "value": 0.00010524526491869997, "min": 0.00010524526491869997, "max": 0.0032559047146985, "count": 33 }, "Pyramids.Policy.Epsilon.mean": { "value": 0.10250580714285715, "min": 0.10250580714285715, "max": 0.19838354285714285, "count": 33 }, "Pyramids.Policy.Epsilon.sum": { "value": 1.4350813, "min": 1.3691136000000002, "max": 2.4021698000000002, "count": 33 }, "Pyramids.Policy.Beta.mean": { "value": 0.0002603301335714285, "min": 0.0002603301335714285, "max": 0.00983851593142857, "count": 33 }, "Pyramids.Policy.Beta.sum": { "value": 0.0036446218699999994, "min": 0.0036446218699999994, "max": 0.10855161985, "count": 33 }, "Pyramids.Losses.RNDLoss.mean": { "value": 0.009032952599227428, "min": 0.0086537916213274, "max": 0.43593263626098633, "count": 33 }, "Pyramids.Losses.RNDLoss.sum": { "value": 0.12646134197711945, "min": 0.1211530864238739, "max": 3.0515284538269043, "count": 33 }, "Pyramids.Environment.EpisodeLength.mean": { "value": 471.4032258064516, "min": 440.463768115942, "max": 999.0, "count": 33 }, "Pyramids.Environment.EpisodeLength.sum": { "value": 29227.0, "min": 15984.0, "max": 32924.0, "count": 33 }, "Pyramids.Environment.CumulativeReward.mean": { "value": 1.3011048914956265, "min": -1.0000000521540642, "max": 1.4725594002267588, "count": 33 }, "Pyramids.Environment.CumulativeReward.sum": { "value": 79.36739838123322, "min": -32.000001668930054, "max": 101.60659861564636, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.mean": { "value": 1.3011048914956265, "min": -1.0000000521540642, "max": 1.4725594002267588, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.sum": { "value": 79.36739838123322, "min": -32.000001668930054, "max": 101.60659861564636, "count": 33 }, "Pyramids.Policy.RndReward.mean": { "value": 0.04429158767719646, "min": 0.043352641953407, "max": 8.53546473942697, "count": 33 }, "Pyramids.Policy.RndReward.sum": { "value": 2.701786848308984, "min": 2.6599582975322846, "max": 136.56743583083153, "count": 33 }, "Pyramids.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 }, "Pyramids.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1686112265", "python_version": "3.10.11 (main, Apr 5 2023, 14:15:10) [GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1686114358" }, "total": 2092.9848652600003, "count": 1, "self": 0.575282798000444, "children": { "run_training.setup": { "total": 0.04292366000004222, "count": 1, "self": 0.04292366000004222 }, "TrainerController.start_learning": { "total": 2092.366658802, "count": 1, "self": 1.4404065180688121, "children": { "TrainerController._reset_env": { "total": 3.736506961000032, "count": 1, "self": 3.736506961000032 }, "TrainerController.advance": { "total": 2087.089941646931, "count": 63390, "self": 1.3712998078508463, "children": { "env_step": { "total": 1448.0797201420341, "count": 63390, "self": 1335.6835791899957, "children": { "SubprocessEnvManager._take_step": { "total": 111.57247465801902, "count": 63390, "self": 4.859888107013376, "children": { "TorchPolicy.evaluate": { "total": 106.71258655100564, "count": 62552, "self": 106.71258655100564 } } }, "workers": { "total": 0.8236662940194037, "count": 63390, "self": 0.0, "children": { "worker_root": { "total": 2087.4207112570193, "count": 63390, "is_parallel": true, "self": 866.3884227230128, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.001830280000149287, "count": 1, "is_parallel": true, "self": 0.0006319470003290917, "children": { "_process_rank_one_or_two_observation": { "total": 0.0011983329998201953, "count": 8, "is_parallel": true, "self": 0.0011983329998201953 } } }, "UnityEnvironment.step": { "total": 0.053344245000062074, "count": 1, "is_parallel": true, "self": 0.0005557460001455183, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0005415170001015213, "count": 1, "is_parallel": true, "self": 0.0005415170001015213 }, "communicator.exchange": { "total": 0.05046834399990985, "count": 1, "is_parallel": true, "self": 0.05046834399990985 }, "steps_from_proto": { "total": 0.001778637999905186, "count": 1, "is_parallel": true, "self": 0.0003668269996524032, "children": { "_process_rank_one_or_two_observation": { "total": 0.0014118110002527828, "count": 8, "is_parallel": true, "self": 0.0014118110002527828 } } } } } } }, "UnityEnvironment.step": { "total": 1221.0322885340065, "count": 63389, "is_parallel": true, "self": 32.81479635010237, "children": { "UnityEnvironment._generate_step_input": { "total": 23.436398587927215, "count": 63389, "is_parallel": true, "self": 23.436398587927215 }, "communicator.exchange": { "total": 1061.6281738520104, "count": 63389, "is_parallel": true, "self": 1061.6281738520104 }, "steps_from_proto": { "total": 103.15291974396655, "count": 63389, "is_parallel": true, "self": 20.69779507699195, "children": { "_process_rank_one_or_two_observation": { "total": 82.4551246669746, "count": 507112, "is_parallel": true, "self": 82.4551246669746 } } } } } } } } } } }, "trainer_advance": { "total": 637.6389216970458, "count": 63390, "self": 2.5722135250296105, "children": { "process_trajectory": { "total": 111.34335536401795, "count": 63390, "self": 111.04837539201799, "children": { "RLTrainer._checkpoint": { "total": 0.2949799719999646, "count": 2, "self": 0.2949799719999646 } } }, "_update_policy": { "total": 523.7233528079983, "count": 442, "self": 336.1868543580042, "children": { "TorchPPOOptimizer.update": { "total": 187.53649844999404, "count": 22818, "self": 187.53649844999404 } } } } } } }, "trainer_threads": { "total": 8.269998943433166e-07, "count": 1, "self": 8.269998943433166e-07 }, "TrainerController._save_models": { "total": 0.09980284900029801, "count": 1, "self": 0.0014069150001887465, "children": { "RLTrainer._checkpoint": { "total": 0.09839593400010926, "count": 1, "self": 0.09839593400010926 } } } } } } }