ppo-PyramidsV1 / run_logs /timers.json
anilkumar2444's picture
First Push
b3120f1
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.31261855363845825,
"min": 0.31261855363845825,
"max": 1.401254653930664,
"count": 35
},
"Pyramids.Policy.Entropy.sum": {
"value": 9408.568359375,
"min": 9408.568359375,
"max": 42508.4609375,
"count": 35
},
"Pyramids.Step.mean": {
"value": 1049996.0,
"min": 29952.0,
"max": 1049996.0,
"count": 35
},
"Pyramids.Step.sum": {
"value": 1049996.0,
"min": 29952.0,
"max": 1049996.0,
"count": 35
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.5647602677345276,
"min": -0.24371707439422607,
"max": 0.6341997981071472,
"count": 35
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 158.13287353515625,
"min": -57.760948181152344,
"max": 173.1365509033203,
"count": 35
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.008768255822360516,
"min": -0.015624484978616238,
"max": 0.3907085657119751,
"count": 35
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 2.455111503601074,
"min": -4.26548433303833,
"max": 94.1607666015625,
"count": 35
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.0720819551657076,
"min": 0.06355605734613624,
"max": 0.07384525715523134,
"count": 35
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 1.0091473723199065,
"min": 0.5115841196227053,
"max": 1.0660511634293168,
"count": 35
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.014508242805147472,
"min": 0.0003372882433926354,
"max": 0.016382612570028175,
"count": 35
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.20311539927206462,
"min": 0.0037101706773189896,
"max": 0.22935657598039447,
"count": 35
},
"Pyramids.Policy.LearningRate.mean": {
"value": 0.0001965532701965405,
"min": 0.0001965532701965405,
"max": 0.00029838354339596195,
"count": 35
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.002751745782751567,
"min": 0.0020886848037717336,
"max": 0.003982106972631066,
"count": 35
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.1655177452380953,
"min": 0.1655177452380953,
"max": 0.19946118095238097,
"count": 35
},
"Pyramids.Policy.Epsilon.sum": {
"value": 2.317248433333334,
"min": 1.3962282666666668,
"max": 2.8273689333333336,
"count": 35
},
"Pyramids.Policy.Beta.mean": {
"value": 0.006555222749285715,
"min": 0.006555222749285715,
"max": 0.009946171977142856,
"count": 35
},
"Pyramids.Policy.Beta.sum": {
"value": 0.09177311849000001,
"min": 0.06962320384,
"max": 0.13275415644,
"count": 35
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.012526755221188068,
"min": 0.011322936043143272,
"max": 0.604117214679718,
"count": 35
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.1753745675086975,
"min": 0.15852110087871552,
"max": 4.228820323944092,
"count": 35
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 335.94444444444446,
"min": 321.1818181818182,
"max": 999.0,
"count": 35
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 30235.0,
"min": 15984.0,
"max": 32522.0,
"count": 35
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.5364921151587132,
"min": -1.0000000521540642,
"max": 1.6213792971838479,
"count": 35
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 136.74779824912548,
"min": -30.52500170469284,
"max": 141.05999885499477,
"count": 35
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.5364921151587132,
"min": -1.0000000521540642,
"max": 1.6213792971838479,
"count": 35
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 136.74779824912548,
"min": -30.52500170469284,
"max": 141.05999885499477,
"count": 35
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.042924876469534065,
"min": 0.04097906218425353,
"max": 11.786974547430873,
"count": 35
},
"Pyramids.Policy.RndReward.sum": {
"value": 3.820314005788532,
"min": 3.5651784100300574,
"max": 188.59159275889397,
"count": 35
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 35
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 35
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1682337170",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1682339641"
},
"total": 2471.0873518400003,
"count": 1,
"self": 1.2946393300007912,
"children": {
"run_training.setup": {
"total": 0.10985993600024813,
"count": 1,
"self": 0.10985993600024813
},
"TrainerController.start_learning": {
"total": 2469.6828525739993,
"count": 1,
"self": 1.8444322930663475,
"children": {
"TrainerController._reset_env": {
"total": 3.8707110739997006,
"count": 1,
"self": 3.8707110739997006
},
"TrainerController.advance": {
"total": 2463.9601103989326,
"count": 67915,
"self": 1.8636420658340285,
"children": {
"env_step": {
"total": 1784.5714731709913,
"count": 67915,
"self": 1649.8980007337468,
"children": {
"SubprocessEnvManager._take_step": {
"total": 133.48471894021804,
"count": 67915,
"self": 5.78576871018231,
"children": {
"TorchPolicy.evaluate": {
"total": 127.69895023003573,
"count": 66481,
"self": 127.69895023003573
}
}
},
"workers": {
"total": 1.1887534970264824,
"count": 67915,
"self": 0.0,
"children": {
"worker_root": {
"total": 2463.431883780002,
"count": 67915,
"is_parallel": true,
"self": 949.4841125458652,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0018205699998361524,
"count": 1,
"is_parallel": true,
"self": 0.0006347229982566205,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001185847001579532,
"count": 8,
"is_parallel": true,
"self": 0.001185847001579532
}
}
},
"UnityEnvironment.step": {
"total": 0.05111714899976505,
"count": 1,
"is_parallel": true,
"self": 0.000561658001061005,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0005661859995598206,
"count": 1,
"is_parallel": true,
"self": 0.0005661859995598206
},
"communicator.exchange": {
"total": 0.04814638199968613,
"count": 1,
"is_parallel": true,
"self": 0.04814638199968613
},
"steps_from_proto": {
"total": 0.0018429229994580965,
"count": 1,
"is_parallel": true,
"self": 0.00042151699926762376,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014214060001904727,
"count": 8,
"is_parallel": true,
"self": 0.0014214060001904727
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1513.9477712341368,
"count": 67914,
"is_parallel": true,
"self": 37.256896678773046,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 27.61560155029747,
"count": 67914,
"is_parallel": true,
"self": 27.61560155029747
},
"communicator.exchange": {
"total": 1334.3911262350884,
"count": 67914,
"is_parallel": true,
"self": 1334.3911262350884
},
"steps_from_proto": {
"total": 114.68414676997781,
"count": 67914,
"is_parallel": true,
"self": 25.220293256189507,
"children": {
"_process_rank_one_or_two_observation": {
"total": 89.46385351378831,
"count": 543312,
"is_parallel": true,
"self": 89.46385351378831
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 677.5249951621072,
"count": 67915,
"self": 3.3102578240841467,
"children": {
"process_trajectory": {
"total": 124.44356309001705,
"count": 67915,
"self": 124.15465131901692,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2889117710001301,
"count": 2,
"self": 0.2889117710001301
}
}
},
"_update_policy": {
"total": 549.771174248006,
"count": 478,
"self": 348.29546326792934,
"children": {
"TorchPPOOptimizer.update": {
"total": 201.47571098007666,
"count": 24211,
"self": 201.47571098007666
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.5000005078036338e-06,
"count": 1,
"self": 1.5000005078036338e-06
},
"TrainerController._save_models": {
"total": 0.007597308000185876,
"count": 1,
"self": 3.219700010959059e-05,
"children": {
"RLTrainer._checkpoint": {
"total": 0.0075651110000762856,
"count": 1,
"self": 0.0075651110000762856
}
}
}
}
}
}
}