Pyramids / run_logs /timers.json
stuti-srinath's picture
First Push
d1f0ab3 verified
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.45756399631500244,
"min": 0.45756399631500244,
"max": 1.5040336847305298,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 13609.783203125,
"min": 13609.783203125,
"max": 45626.3671875,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989889.0,
"min": 29995.0,
"max": 989889.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989889.0,
"min": 29995.0,
"max": 989889.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.4547107219696045,
"min": -0.17053431272506714,
"max": 0.6123825907707214,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 122.31718444824219,
"min": -40.58716583251953,
"max": 173.91665649414062,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.06040317192673683,
"min": -0.013297402299940586,
"max": 0.35190609097480774,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 16.24845314025879,
"min": -3.4706220626831055,
"max": 83.75364685058594,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.06661658520183338,
"min": 0.06348988306403737,
"max": 0.07438346972779232,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.9992487780275006,
"min": 0.5950677578223386,
"max": 1.043722482919168,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.014879163227434378,
"min": 5.819783857324592e-05,
"max": 0.01740750817615273,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.22318744841151567,
"min": 0.000698374062878951,
"max": 0.2437051144661382,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.4580375140200035e-06,
"min": 7.4580375140200035e-06,
"max": 0.0002948495642168125,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.00011187056271030005,
"min": 0.00011187056271030005,
"max": 0.0031405685531438994,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10248598,
"min": 0.10248598,
"max": 0.19828318749999999,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.5372897,
"min": 1.4777525,
"max": 2.3593829,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.00025834940200000013,
"min": 0.00025834940200000013,
"max": 0.00982849043125,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.003875241030000002,
"min": 0.003875241030000002,
"max": 0.10471092439,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.00784855056554079,
"min": 0.00784855056554079,
"max": 0.4417484700679779,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.11772826313972473,
"min": 0.11258344352245331,
"max": 3.5339877605438232,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 417.14666666666665,
"min": 318.6989247311828,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 31286.0,
"min": 16282.0,
"max": 32312.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.3482105018276918,
"min": -1.0000000521540642,
"max": 1.6559318952420925,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 102.46399813890457,
"min": -32.000001668930054,
"max": 155.6575981527567,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.3482105018276918,
"min": -1.0000000521540642,
"max": 1.6559318952420925,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 102.46399813890457,
"min": -32.000001668930054,
"max": 155.6575981527567,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.033934670875360576,
"min": 0.028049252265047567,
"max": 9.304248069139089,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 2.579034986527404,
"min": 2.329140003363136,
"max": 158.1722171753645,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1742812154",
"python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "1.2.0.dev0",
"mlagents_envs_version": "1.2.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.6.0+cu124",
"numpy_version": "1.23.5",
"end_time_seconds": "1742814391"
},
"total": 2236.349685046,
"count": 1,
"self": 0.5372856650001268,
"children": {
"run_training.setup": {
"total": 0.023198602999968898,
"count": 1,
"self": 0.023198602999968898
},
"TrainerController.start_learning": {
"total": 2235.789200778,
"count": 1,
"self": 1.3817650539635906,
"children": {
"TrainerController._reset_env": {
"total": 3.5499088220000203,
"count": 1,
"self": 3.5499088220000203
},
"TrainerController.advance": {
"total": 2230.7650272690366,
"count": 63770,
"self": 1.5002949470731437,
"children": {
"env_step": {
"total": 1548.9240599449613,
"count": 63770,
"self": 1390.443742678909,
"children": {
"SubprocessEnvManager._take_step": {
"total": 157.65761558704367,
"count": 63770,
"self": 4.900581816033423,
"children": {
"TorchPolicy.evaluate": {
"total": 152.75703377101024,
"count": 62567,
"self": 152.75703377101024
}
}
},
"workers": {
"total": 0.8227016790086736,
"count": 63770,
"self": 0.0,
"children": {
"worker_root": {
"total": 2230.734022660008,
"count": 63770,
"is_parallel": true,
"self": 954.912694901981,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.005917974999988473,
"count": 1,
"is_parallel": true,
"self": 0.004079378999790606,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0018385960001978674,
"count": 8,
"is_parallel": true,
"self": 0.0018385960001978674
}
}
},
"UnityEnvironment.step": {
"total": 0.0524213859999918,
"count": 1,
"is_parallel": true,
"self": 0.0005606560000614991,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0005038649999278277,
"count": 1,
"is_parallel": true,
"self": 0.0005038649999278277
},
"communicator.exchange": {
"total": 0.04960968499995033,
"count": 1,
"is_parallel": true,
"self": 0.04960968499995033
},
"steps_from_proto": {
"total": 0.0017471800000521398,
"count": 1,
"is_parallel": true,
"self": 0.0004916929999581043,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0012554870000940355,
"count": 8,
"is_parallel": true,
"self": 0.0012554870000940355
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1275.821327758027,
"count": 63769,
"is_parallel": true,
"self": 32.32314335805586,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 23.03877689896126,
"count": 63769,
"is_parallel": true,
"self": 23.03877689896126
},
"communicator.exchange": {
"total": 1123.395636481983,
"count": 63769,
"is_parallel": true,
"self": 1123.395636481983
},
"steps_from_proto": {
"total": 97.06377101902706,
"count": 63769,
"is_parallel": true,
"self": 19.806728485915414,
"children": {
"_process_rank_one_or_two_observation": {
"total": 77.25704253311164,
"count": 510152,
"is_parallel": true,
"self": 77.25704253311164
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 680.3406723770023,
"count": 63770,
"self": 2.591274365991012,
"children": {
"process_trajectory": {
"total": 130.16705984501073,
"count": 63770,
"self": 129.83959507201075,
"children": {
"RLTrainer._checkpoint": {
"total": 0.32746477299997423,
"count": 2,
"self": 0.32746477299997423
}
}
},
"_update_policy": {
"total": 547.5823381660006,
"count": 443,
"self": 297.358186570994,
"children": {
"TorchPPOOptimizer.update": {
"total": 250.22415159500656,
"count": 22827,
"self": 250.22415159500656
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.860003279347438e-07,
"count": 1,
"self": 7.860003279347438e-07
},
"TrainerController._save_models": {
"total": 0.0924988469996606,
"count": 1,
"self": 0.001290584999424027,
"children": {
"RLTrainer._checkpoint": {
"total": 0.09120826200023657,
"count": 1,
"self": 0.09120826200023657
}
}
}
}
}
}
}