{ "name": "root", "gauges": { "Pyramids.Policy.Entropy.mean": { "value": 0.3751542866230011, "min": 0.3751542866230011, "max": 1.3917635679244995, "count": 33 }, "Pyramids.Policy.Entropy.sum": { "value": 11062.5498046875, "min": 11062.5498046875, "max": 42220.5390625, "count": 33 }, "Pyramids.Step.mean": { "value": 989875.0, "min": 29920.0, "max": 989875.0, "count": 33 }, "Pyramids.Step.sum": { "value": 989875.0, "min": 29920.0, "max": 989875.0, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.mean": { "value": 0.5894595384597778, "min": -0.08475109934806824, "max": 0.5894595384597778, "count": 33 }, "Pyramids.Policy.ExtrinsicValueEstimate.sum": { "value": 165.04867553710938, "min": -20.42501449584961, "max": 165.04867553710938, "count": 33 }, "Pyramids.Policy.RndValueEstimate.mean": { "value": 0.0038175820372998714, "min": -0.0664333924651146, "max": 0.5999228358268738, "count": 33 }, "Pyramids.Policy.RndValueEstimate.sum": { "value": 1.068922996520996, "min": -16.874082565307617, "max": 142.1817169189453, "count": 33 }, "Pyramids.Losses.PolicyLoss.mean": { "value": 0.06632594499489683, "min": 0.06457859685178846, "max": 0.07508671516487832, "count": 33 }, "Pyramids.Losses.PolicyLoss.sum": { "value": 0.9285632299285557, "min": 0.55559605451381, "max": 1.0403894486719532, "count": 33 }, "Pyramids.Losses.ValueLoss.mean": { "value": 0.015127578886498596, "min": 0.0004044822102349373, "max": 0.015362370851410862, "count": 33 }, "Pyramids.Losses.ValueLoss.sum": { "value": 0.21178610441098034, "min": 0.005258268733054185, "max": 0.21507319191975208, "count": 33 }, "Pyramids.Policy.LearningRate.mean": { "value": 7.460561798892855e-06, "min": 7.460561798892855e-06, "max": 0.0002948436017188, "count": 33 }, "Pyramids.Policy.LearningRate.sum": { "value": 0.00010444786518449997, "min": 0.00010444786518449997, "max": 0.0035090126303292, "count": 33 }, "Pyramids.Policy.Epsilon.mean": { "value": 0.10248682142857143, "min": 0.10248682142857143, "max": 0.1982812, "count": 33 }, "Pyramids.Policy.Epsilon.sum": { "value": 1.4348155, "min": 1.4348155, "max": 2.5696708, "count": 33 }, "Pyramids.Policy.Beta.mean": { "value": 0.00025843346071428565, "min": 0.00025843346071428565, "max": 0.00982829188, "count": 33 }, "Pyramids.Policy.Beta.sum": { "value": 0.0036180684499999994, "min": 0.0036180684499999994, "max": 0.11699011292000003, "count": 33 }, "Pyramids.Losses.RNDLoss.mean": { "value": 0.015708057209849358, "min": 0.01550370268523693, "max": 0.6156061887741089, "count": 33 }, "Pyramids.Losses.RNDLoss.sum": { "value": 0.2199127972126007, "min": 0.2199127972126007, "max": 4.924849510192871, "count": 33 }, "Pyramids.Environment.EpisodeLength.mean": { "value": 317.5, "min": 317.5, "max": 999.0, "count": 33 }, "Pyramids.Environment.EpisodeLength.sum": { "value": 31115.0, "min": 16463.0, "max": 32078.0, "count": 33 }, "Pyramids.Environment.CumulativeReward.mean": { "value": 1.6620836554437268, "min": -0.9999500517733395, "max": 1.6620836554437268, "count": 33 }, "Pyramids.Environment.CumulativeReward.sum": { "value": 162.88419823348522, "min": -31.998401656746864, "max": 162.88419823348522, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.mean": { "value": 1.6620836554437268, "min": -0.9999500517733395, "max": 1.6620836554437268, "count": 33 }, "Pyramids.Policy.ExtrinsicReward.sum": { "value": 162.88419823348522, "min": -31.998401656746864, "max": 162.88419823348522, "count": 33 }, "Pyramids.Policy.RndReward.mean": { "value": 0.05126715123645097, "min": 0.05126715123645097, "max": 11.714249651221667, "count": 33 }, "Pyramids.Policy.RndReward.sum": { "value": 5.0241808211721946, "min": 4.539654043619521, "max": 199.14224407076836, "count": 33 }, "Pyramids.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 }, "Pyramids.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 33 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1673457832", "python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1673460131" }, "total": 2298.869401232, "count": 1, "self": 0.42510950299993056, "children": { "run_training.setup": { "total": 0.1124471100001756, "count": 1, "self": 0.1124471100001756 }, "TrainerController.start_learning": { "total": 2298.331844619, "count": 1, "self": 1.4342928829846642, "children": { "TrainerController._reset_env": { "total": 6.848743022999997, "count": 1, "self": 6.848743022999997 }, "TrainerController.advance": { "total": 2289.953882238015, "count": 63932, "self": 1.4759382040551827, "children": { "env_step": { "total": 1585.152514205954, "count": 63932, "self": 1467.110062690917, "children": { "SubprocessEnvManager._take_step": { "total": 117.11747811096643, "count": 63932, "self": 4.8218035939748916, "children": { "TorchPolicy.evaluate": { "total": 112.29567451699154, "count": 62566, "self": 37.73733194707302, "children": { "TorchPolicy.sample_actions": { "total": 74.55834256991852, "count": 62566, "self": 74.55834256991852 } } } } }, "workers": { "total": 0.9249734040706699, "count": 63932, "self": 0.0, "children": { "worker_root": { "total": 2293.2333010749962, "count": 63932, "is_parallel": true, "self": 940.7484627540084, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0018015680000189604, "count": 1, "is_parallel": true, "self": 0.0006439239998599078, "children": { "_process_rank_one_or_two_observation": { "total": 0.0011576440001590527, "count": 8, "is_parallel": true, "self": 0.0011576440001590527 } } }, "UnityEnvironment.step": { "total": 0.04781888899992737, "count": 1, "is_parallel": true, "self": 0.0006469799998285453, "children": { "UnityEnvironment._generate_step_input": { "total": 0.000488424999957715, "count": 1, "is_parallel": true, "self": 0.000488424999957715 }, "communicator.exchange": { "total": 0.04503667700009828, "count": 1, "is_parallel": true, "self": 0.04503667700009828 }, "steps_from_proto": { "total": 0.001646807000042827, "count": 1, "is_parallel": true, "self": 0.00041805199998634635, "children": { "_process_rank_one_or_two_observation": { "total": 0.0012287550000564806, "count": 8, "is_parallel": true, "self": 0.0012287550000564806 } } } } } } }, "UnityEnvironment.step": { "total": 1352.4848383209878, "count": 63931, "is_parallel": true, "self": 30.534577841101054, "children": { "UnityEnvironment._generate_step_input": { "total": 24.230452593963037, "count": 63931, "is_parallel": true, "self": 24.230452593963037 }, "communicator.exchange": { "total": 1189.2243363159903, "count": 63931, "is_parallel": true, "self": 1189.2243363159903 }, "steps_from_proto": { "total": 108.49547156993344, "count": 63931, "is_parallel": true, "self": 24.068160102979164, "children": { "_process_rank_one_or_two_observation": { "total": 84.42731146695428, "count": 511448, "is_parallel": true, "self": 84.42731146695428 } } } } } } } } } } }, "trainer_advance": { "total": 703.3254298280058, "count": 63932, "self": 2.799658408024925, "children": { "process_trajectory": { "total": 151.71104306898496, "count": 63932, "self": 151.5093902829849, "children": { "RLTrainer._checkpoint": { "total": 0.20165278600006786, "count": 2, "self": 0.20165278600006786 } } }, "_update_policy": { "total": 548.8147283509959, "count": 456, "self": 211.4842232480146, "children": { "TorchPPOOptimizer.update": { "total": 337.3305051029813, "count": 22833, "self": 337.3305051029813 } } } } } } }, "trainer_threads": { "total": 1.1740003174054436e-06, "count": 1, "self": 1.1740003174054436e-06 }, "TrainerController._save_models": { "total": 0.09492530099987562, "count": 1, "self": 0.0014766799999961222, "children": { "RLTrainer._checkpoint": { "total": 0.0934486209998795, "count": 1, "self": 0.0934486209998795 } } } } } } }