| { | |
| "name": "root", | |
| "gauges": { | |
| "Pyramids.Policy.Entropy.mean": { | |
| "value": 0.30823203921318054, | |
| "min": 0.2590036988258362, | |
| "max": 1.3212605714797974, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Entropy.sum": { | |
| "value": 9316.0048828125, | |
| "min": 7811.55126953125, | |
| "max": 40081.76171875, | |
| "count": 33 | |
| }, | |
| "Pyramids.Step.mean": { | |
| "value": 989984.0, | |
| "min": 29989.0, | |
| "max": 989984.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Step.sum": { | |
| "value": 989984.0, | |
| "min": 29989.0, | |
| "max": 989984.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicValueEstimate.mean": { | |
| "value": 0.6038783192634583, | |
| "min": -0.10377731919288635, | |
| "max": 0.6722452640533447, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicValueEstimate.sum": { | |
| "value": 169.0859375, | |
| "min": -25.010334014892578, | |
| "max": 191.58990478515625, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndValueEstimate.mean": { | |
| "value": 0.027300408110022545, | |
| "min": 0.00039116060361266136, | |
| "max": 0.3165760338306427, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndValueEstimate.sum": { | |
| "value": 7.644114017486572, | |
| "min": 0.10913380980491638, | |
| "max": 75.3450927734375, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.PolicyLoss.mean": { | |
| "value": 0.06567606225767772, | |
| "min": 0.0656227797144156, | |
| "max": 0.0743524374070947, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.PolicyLoss.sum": { | |
| "value": 0.919464871607488, | |
| "min": 0.5204670618496628, | |
| "max": 1.0589628908686186, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.ValueLoss.mean": { | |
| "value": 0.014384692752390125, | |
| "min": 0.00022428097407098583, | |
| "max": 0.016094682862915634, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.ValueLoss.sum": { | |
| "value": 0.20138569853346175, | |
| "min": 0.00269137168885183, | |
| "max": 0.23286290983863486, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.LearningRate.mean": { | |
| "value": 7.289326141685715e-06, | |
| "min": 7.289326141685715e-06, | |
| "max": 0.00029515063018788575, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.LearningRate.sum": { | |
| "value": 0.00010205056598360001, | |
| "min": 0.00010205056598360001, | |
| "max": 0.003331593089469, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Epsilon.mean": { | |
| "value": 0.10242974285714286, | |
| "min": 0.10242974285714286, | |
| "max": 0.19838354285714285, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Epsilon.sum": { | |
| "value": 1.4340164, | |
| "min": 1.3886848, | |
| "max": 2.442952, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Beta.mean": { | |
| "value": 0.0002527313114285715, | |
| "min": 0.0002527313114285715, | |
| "max": 0.00983851593142857, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Beta.sum": { | |
| "value": 0.0035382383600000007, | |
| "min": 0.0035382383600000007, | |
| "max": 0.11106204690000002, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.RNDLoss.mean": { | |
| "value": 0.013543262146413326, | |
| "min": 0.012501879595220089, | |
| "max": 0.4942598342895508, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.RNDLoss.sum": { | |
| "value": 0.18960566818714142, | |
| "min": 0.1750263124704361, | |
| "max": 3.4598188400268555, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.EpisodeLength.mean": { | |
| "value": 300.3440860215054, | |
| "min": 287.7289719626168, | |
| "max": 999.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.EpisodeLength.sum": { | |
| "value": 27932.0, | |
| "min": 16788.0, | |
| "max": 32541.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.CumulativeReward.mean": { | |
| "value": 1.635126870165589, | |
| "min": -0.999837551265955, | |
| "max": 1.6748784882164447, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.CumulativeReward.sum": { | |
| "value": 152.06679892539978, | |
| "min": -31.99480164051056, | |
| "max": 179.21199823915958, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicReward.mean": { | |
| "value": 1.635126870165589, | |
| "min": -0.999837551265955, | |
| "max": 1.6748784882164447, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicReward.sum": { | |
| "value": 152.06679892539978, | |
| "min": -31.99480164051056, | |
| "max": 179.21199823915958, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndReward.mean": { | |
| "value": 0.04253368296501507, | |
| "min": 0.0420990171978795, | |
| "max": 8.939846154521494, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndReward.sum": { | |
| "value": 3.9556325157464016, | |
| "min": 3.9556325157464016, | |
| "max": 151.9773846268654, | |
| "count": 33 | |
| }, | |
| "Pyramids.IsTraining.mean": { | |
| "value": 1.0, | |
| "min": 1.0, | |
| "max": 1.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.IsTraining.sum": { | |
| "value": 1.0, | |
| "min": 1.0, | |
| "max": 1.0, | |
| "count": 33 | |
| } | |
| }, | |
| "metadata": { | |
| "timer_format_version": "0.1.0", | |
| "start_time_seconds": "1725455468", | |
| "python_version": "3.10.12 (main, Jul 29 2024, 16:56:48) [GCC 11.4.0]", | |
| "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", | |
| "mlagents_version": "1.1.0.dev0", | |
| "mlagents_envs_version": "1.1.0.dev0", | |
| "communication_protocol_version": "1.5.0", | |
| "pytorch_version": "2.4.0+cu121", | |
| "numpy_version": "1.23.5", | |
| "end_time_seconds": "1725458277" | |
| }, | |
| "total": 2808.690324873, | |
| "count": 1, | |
| "self": 0.5417087249998076, | |
| "children": { | |
| "run_training.setup": { | |
| "total": 0.07583376699994915, | |
| "count": 1, | |
| "self": 0.07583376699994915 | |
| }, | |
| "TrainerController.start_learning": { | |
| "total": 2808.0727823810003, | |
| "count": 1, | |
| "self": 1.8343723190228047, | |
| "children": { | |
| "TrainerController._reset_env": { | |
| "total": 1.872349722999843, | |
| "count": 1, | |
| "self": 1.872349722999843 | |
| }, | |
| "TrainerController.advance": { | |
| "total": 2804.293062076978, | |
| "count": 63878, | |
| "self": 1.8592885488969841, | |
| "children": { | |
| "env_step": { | |
| "total": 1861.1480081190616, | |
| "count": 63878, | |
| "self": 1729.754723312018, | |
| "children": { | |
| "SubprocessEnvManager._take_step": { | |
| "total": 130.27318121500457, | |
| "count": 63878, | |
| "self": 5.305852125036381, | |
| "children": { | |
| "TorchPolicy.evaluate": { | |
| "total": 124.96732908996819, | |
| "count": 62566, | |
| "self": 124.96732908996819 | |
| } | |
| } | |
| }, | |
| "workers": { | |
| "total": 1.1201035920389586, | |
| "count": 63878, | |
| "self": 0.0, | |
| "children": { | |
| "worker_root": { | |
| "total": 2802.454026264935, | |
| "count": 63878, | |
| "is_parallel": true, | |
| "self": 1220.0774426239313, | |
| "children": { | |
| "run_training.setup": { | |
| "total": 0.0, | |
| "count": 0, | |
| "is_parallel": true, | |
| "self": 0.0, | |
| "children": { | |
| "steps_from_proto": { | |
| "total": 0.006555874000241602, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.002710352000576677, | |
| "children": { | |
| "_process_rank_one_or_two_observation": { | |
| "total": 0.003845521999664925, | |
| "count": 8, | |
| "is_parallel": true, | |
| "self": 0.003845521999664925 | |
| } | |
| } | |
| }, | |
| "UnityEnvironment.step": { | |
| "total": 0.056553943999915646, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.000700233000316075, | |
| "children": { | |
| "UnityEnvironment._generate_step_input": { | |
| "total": 0.0005066229996373295, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.0005066229996373295 | |
| }, | |
| "communicator.exchange": { | |
| "total": 0.053332365000187565, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.053332365000187565 | |
| }, | |
| "steps_from_proto": { | |
| "total": 0.0020147229997746763, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.0005138159995112801, | |
| "children": { | |
| "_process_rank_one_or_two_observation": { | |
| "total": 0.0015009070002633962, | |
| "count": 8, | |
| "is_parallel": true, | |
| "self": 0.0015009070002633962 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "UnityEnvironment.step": { | |
| "total": 1582.3765836410034, | |
| "count": 63877, | |
| "is_parallel": true, | |
| "self": 41.761416157960866, | |
| "children": { | |
| "UnityEnvironment._generate_step_input": { | |
| "total": 27.802824746935585, | |
| "count": 63877, | |
| "is_parallel": true, | |
| "self": 27.802824746935585 | |
| }, | |
| "communicator.exchange": { | |
| "total": 1393.7710841601047, | |
| "count": 63877, | |
| "is_parallel": true, | |
| "self": 1393.7710841601047 | |
| }, | |
| "steps_from_proto": { | |
| "total": 119.04125857600229, | |
| "count": 63877, | |
| "is_parallel": true, | |
| "self": 25.022369727780642, | |
| "children": { | |
| "_process_rank_one_or_two_observation": { | |
| "total": 94.01888884822165, | |
| "count": 511016, | |
| "is_parallel": true, | |
| "self": 94.01888884822165 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "trainer_advance": { | |
| "total": 941.2857654090194, | |
| "count": 63878, | |
| "self": 3.304072214232292, | |
| "children": { | |
| "process_trajectory": { | |
| "total": 143.5407350457849, | |
| "count": 63878, | |
| "self": 143.38364553978454, | |
| "children": { | |
| "RLTrainer._checkpoint": { | |
| "total": 0.15708950600037497, | |
| "count": 2, | |
| "self": 0.15708950600037497 | |
| } | |
| } | |
| }, | |
| "_update_policy": { | |
| "total": 794.4409581490022, | |
| "count": 451, | |
| "self": 340.7522739279957, | |
| "children": { | |
| "TorchPPOOptimizer.update": { | |
| "total": 453.6886842210065, | |
| "count": 22842, | |
| "self": 453.6886842210065 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "trainer_threads": { | |
| "total": 1.0789999578264542e-06, | |
| "count": 1, | |
| "self": 1.0789999578264542e-06 | |
| }, | |
| "TrainerController._save_models": { | |
| "total": 0.07299718299964297, | |
| "count": 1, | |
| "self": 0.0017966440000236616, | |
| "children": { | |
| "RLTrainer._checkpoint": { | |
| "total": 0.07120053899961931, | |
| "count": 1, | |
| "self": 0.07120053899961931 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } |