{ "name": "root", "gauges": { "Pyramids.Policy.Entropy.mean": { "value": 0.5021506547927856, "min": 0.5021506547927856, "max": 1.4920095205307007, "count": 33 }, "Pyramids.Policy.Entropy.sum": { "value": 14984.17578125, "min": 14984.17578125, "max": 45261.6015625, "count": 33 }, "Pyramids.Step.mean": { "value": 989965.0, "min": 29952.0, "max": 989965.0, "count": 33 }, "Pyramids.Step.sum": { "value": 989965.0, "min": 29952.0, "max": 989965.0, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.mean": { "value": 0.22611448168754578, "min": -0.10216780751943588, "max": 0.2841082513332367, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.sum": { "value": 58.3375358581543, "min": -24.724609375, "max": 73.86814880371094, "count": 33 }, "Pyramids.Policy.RndValueEstimate.mean": { "value": 0.9184799790382385, "min": -0.29537999629974365, "max": 1.2115176916122437, "count": 33 }, "Pyramids.Policy.RndValueEstimate.sum": { "value": 236.96783447265625, "min": -76.50341796875, "max": 314.9945983886719, "count": 33 }, "Pyramids.Losses.PolicyLoss.mean": { "value": 0.06821941310010429, "min": 0.06513066821757496, "max": 0.07527306955424806, "count": 33 }, "Pyramids.Losses.PolicyLoss.sum": { "value": 0.9550717834014602, "min": 0.492703045007837, "max": 1.0722641541015394, "count": 33 }, "Pyramids.Losses.ValueLoss.mean": { "value": 0.15624019664080283, "min": 0.001007773937901953, "max": 0.22454367401983794, "count": 33 }, "Pyramids.Losses.ValueLoss.sum": { "value": 2.1873627529712394, "min": 0.007902372321903637, "max": 3.368155110297569, "count": 33 }, "Pyramids.Policy.LearningRate.mean": { "value": 7.500118928564282e-06, "min": 7.500118928564282e-06, "max": 0.00029515063018788575, "count": 33 }, "Pyramids.Policy.LearningRate.sum": { "value": 0.00010500166499989995, "min": 0.00010500166499989995, "max": 0.0032572250142584, "count": 33 }, "Pyramids.Policy.Epsilon.mean": { "value": 0.10250000714285715, "min": 0.10250000714285715, "max": 0.19838354285714285, "count": 33 }, "Pyramids.Policy.Epsilon.sum": { "value": 1.4350001000000001, "min": 1.3691136000000002, "max": 2.4852578000000003, "count": 33 }, "Pyramids.Policy.Beta.mean": { "value": 0.0002597507135714285, "min": 0.0002597507135714285, "max": 0.00983851593142857, "count": 33 }, "Pyramids.Policy.Beta.sum": { "value": 0.003636509989999999, "min": 0.003636509989999999, "max": 0.10859558584, "count": 33 }, "Pyramids.Losses.RNDLoss.mean": { "value": 0.00944130215793848, "min": 0.008440539240837097, "max": 0.3676963448524475, "count": 33 }, "Pyramids.Losses.RNDLoss.sum": { "value": 0.13217823207378387, "min": 0.11816754937171936, "max": 2.5738744735717773, "count": 33 }, "Pyramids.Environment.EpisodeLength.mean": { "value": 549.0188679245283, "min": 500.7142857142857, "max": 999.0, "count": 33 }, "Pyramids.Environment.EpisodeLength.sum": { "value": 29098.0, "min": 15984.0, "max": 32366.0, "count": 33 }, "Pyramids.Environment.CumulativeReward.mean": { "value": 0.9609962642468788, "min": -1.0000000521540642, "max": 1.072667901246053, "count": 33 }, "Pyramids.Environment.CumulativeReward.sum": { "value": 51.893798269331455, "min": -32.000001668930054, "max": 64.23039814829826, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.mean": { "value": 0.9609962642468788, "min": -1.0000000521540642, "max": 1.072667901246053, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.sum": { "value": 51.893798269331455, "min": -32.000001668930054, "max": 64.23039814829826, "count": 33 }, "Pyramids.Policy.RndReward.mean": { "value": 0.05391255627856245, "min": 0.051680048903906525, "max": 8.046361776068807, "count": 33 }, "Pyramids.Policy.RndReward.sum": { "value": 2.9112780390423723, "min": 2.53232239629142, "max": 128.7417884171009, "count": 33 }, "Pyramids.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 }, "Pyramids.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1696244276", "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1696246570" }, "total": 2294.185716109, "count": 1, "self": 0.5305372720004016, "children": { "run_training.setup": { "total": 0.04022258200006945, "count": 1, "self": 0.04022258200006945 }, "TrainerController.start_learning": { "total": 2293.614956255, "count": 1, "self": 1.3827964190140847, "children": { "TrainerController._reset_env": { "total": 5.6247360479999315, "count": 1, "self": 5.6247360479999315 }, "TrainerController.advance": { "total": 2286.513123790987, "count": 63493, "self": 1.3462596818712882, "children": { "env_step": { "total": 1610.2826746760588, "count": 63493, "self": 1501.2480643811186, "children": { "SubprocessEnvManager._take_step": { "total": 108.19846278290652, "count": 63493, "self": 4.682967251962509, "children": { "TorchPolicy.evaluate": { "total": 103.51549553094401, "count": 62557, "self": 103.51549553094401 } } }, "workers": { "total": 0.8361475120336763, "count": 63493, "self": 0.0, "children": { "worker_root": { "total": 2288.418504687979, "count": 63493, "is_parallel": true, "self": 901.439076575052, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.002766390000033425, "count": 1, "is_parallel": true, "self": 0.0007953129997986252, "children": { "_process_rank_one_or_two_observation": { "total": 0.0019710770002348, "count": 8, "is_parallel": true, "self": 0.0019710770002348 } } }, "UnityEnvironment.step": { "total": 0.08629489800000556, "count": 1, "is_parallel": true, "self": 0.0006624949999149976, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0005493070000284206, "count": 1, "is_parallel": true, "self": 0.0005493070000284206 }, "communicator.exchange": { "total": 0.08214933399995061, "count": 1, "is_parallel": true, "self": 0.08214933399995061 }, "steps_from_proto": { "total": 0.0029337620001115283, "count": 1, "is_parallel": true, "self": 0.0004008160001376382, "children": { "_process_rank_one_or_two_observation": { "total": 0.00253294599997389, "count": 8, "is_parallel": true, "self": 0.00253294599997389 } } } } } } }, "UnityEnvironment.step": { "total": 1386.9794281129268, "count": 63492, "is_parallel": true, "self": 34.174127390876265, "children": { "UnityEnvironment._generate_step_input": { "total": 23.104340074013408, "count": 63492, "is_parallel": true, "self": 23.104340074013408 }, "communicator.exchange": { "total": 1224.1752410320287, "count": 63492, "is_parallel": true, "self": 1224.1752410320287 }, "steps_from_proto": { "total": 105.52571961600847, "count": 63492, "is_parallel": true, "self": 20.615875372899836, "children": { "_process_rank_one_or_two_observation": { "total": 84.90984424310864, "count": 507936, "is_parallel": true, "self": 84.90984424310864 } } } } } } } } } } }, "trainer_advance": { "total": 674.8841894330565, "count": 63493, "self": 2.637011567025638, "children": { "process_trajectory": { "total": 111.94504011603544, "count": 63493, "self": 111.73689922703511, "children": { "RLTrainer._checkpoint": { "total": 0.20814088900033312, "count": 2, "self": 0.20814088900033312 } } }, "_update_policy": { "total": 560.3021377499954, "count": 443, "self": 366.5897470269949, "children": { "TorchPPOOptimizer.update": { "total": 193.71239072300045, "count": 22818, "self": 193.71239072300045 } } } } } } }, "trainer_threads": { "total": 1.3089993444737047e-06, "count": 1, "self": 1.3089993444737047e-06 }, "TrainerController._save_models": { "total": 0.09429868799998076, "count": 1, "self": 0.0014279840006565792, "children": { "RLTrainer._checkpoint": { "total": 0.09287070399932418, "count": 1, "self": 0.09287070399932418 } } } } } } }