PyramidsRND / run_logs /timers.json
parsasam's picture
initial commit
75941d1
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.16424055397510529,
"min": 0.1559770256280899,
"max": 1.3720276355743408,
"count": 92
},
"Pyramids.Policy.Entropy.sum": {
"value": 4919.3330078125,
"min": 4700.45947265625,
"max": 41621.83203125,
"count": 92
},
"Pyramids.Step.mean": {
"value": 2759977.0,
"min": 29947.0,
"max": 2759977.0,
"count": 92
},
"Pyramids.Step.sum": {
"value": 2759977.0,
"min": 29947.0,
"max": 2759977.0,
"count": 92
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.7984915971755981,
"min": -0.0801166221499443,
"max": 0.8205011487007141,
"count": 92
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 238.74899291992188,
"min": -19.30810546875,
"max": 249.43235778808594,
"count": 92
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.00798786897212267,
"min": -0.005292269866913557,
"max": 0.4261401295661926,
"count": 92
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 2.3883728981018066,
"min": -1.571804165840149,
"max": 100.99520874023438,
"count": 92
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.06836758276676082,
"min": 0.064009909468994,
"max": 0.07476346208187305,
"count": 92
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.9571461587346515,
"min": 0.5636522952911808,
"max": 1.0633553473208117,
"count": 92
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.013863306923866264,
"min": 0.0008922286392557729,
"max": 0.01644228222236658,
"count": 92
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.19408629693412768,
"min": 0.00892228639255773,
"max": 0.23312334825728884,
"count": 92
},
"Pyramids.Policy.LearningRate.mean": {
"value": 2.542401295393095e-05,
"min": 2.542401295393095e-05,
"max": 0.0002982801880732708,
"count": 92
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.0003559361813550333,
"min": 0.0003559361813550333,
"max": 0.004027672757442467,
"count": 92
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10847464047619049,
"min": 0.10847464047619049,
"max": 0.1994267291666667,
"count": 92
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.5186449666666668,
"min": 1.5186449666666668,
"max": 2.8425575333333333,
"count": 92
},
"Pyramids.Policy.Beta.mean": {
"value": 0.0008566165835714284,
"min": 0.0008566165835714284,
"max": 0.00994273024375,
"count": 92
},
"Pyramids.Policy.Beta.sum": {
"value": 0.011992632169999998,
"min": 0.011992632169999998,
"max": 0.13427149758,
"count": 92
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.006343779154121876,
"min": 0.006209437269717455,
"max": 0.3692438006401062,
"count": 92
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.08881291002035141,
"min": 0.08881291002035141,
"max": 2.9539504051208496,
"count": 92
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 237.8,
"min": 225.60162601626016,
"max": 999.0,
"count": 92
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 29725.0,
"min": 16362.0,
"max": 32903.0,
"count": 92
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.7611693438262708,
"min": -0.999987552408129,
"max": 1.7743983547377393,
"count": 92
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 218.3849986344576,
"min": -31.999601677060127,
"max": 229.4491978958249,
"count": 92
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.7611693438262708,
"min": -0.999987552408129,
"max": 1.7743983547377393,
"count": 92
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 218.3849986344576,
"min": -31.999601677060127,
"max": 229.4491978958249,
"count": 92
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.01570810448550131,
"min": 0.015573458049006067,
"max": 7.2554192700806786,
"count": 92
},
"Pyramids.Policy.RndReward.sum": {
"value": 1.9478049562021624,
"min": 1.9124590799037833,
"max": 123.34212759137154,
"count": 92
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 92
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 92
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1676907041",
"python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1676914233"
},
"total": 7191.705955924999,
"count": 1,
"self": 0.5351134530001218,
"children": {
"run_training.setup": {
"total": 0.10881939899991266,
"count": 1,
"self": 0.10881939899991266
},
"TrainerController.start_learning": {
"total": 7191.062023072999,
"count": 1,
"self": 3.9619973119642964,
"children": {
"TrainerController._reset_env": {
"total": 7.200940046999904,
"count": 1,
"self": 7.200940046999904
},
"TrainerController.advance": {
"total": 7179.758435156035,
"count": 179056,
"self": 4.158575470084543,
"children": {
"env_step": {
"total": 5062.768275863919,
"count": 179056,
"self": 4732.1622895467735,
"children": {
"SubprocessEnvManager._take_step": {
"total": 328.173371414995,
"count": 179056,
"self": 13.46620600000142,
"children": {
"TorchPolicy.evaluate": {
"total": 314.7071654149936,
"count": 172654,
"self": 106.06929664887957,
"children": {
"TorchPolicy.sample_actions": {
"total": 208.63786876611402,
"count": 172654,
"self": 208.63786876611402
}
}
}
}
},
"workers": {
"total": 2.43261490215059,
"count": 179056,
"self": 0.0,
"children": {
"worker_root": {
"total": 7173.588793095096,
"count": 179056,
"is_parallel": true,
"self": 2783.1632639899653,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0021503739999388927,
"count": 1,
"is_parallel": true,
"self": 0.0007818960000349762,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013684779999039165,
"count": 8,
"is_parallel": true,
"self": 0.0013684779999039165
}
}
},
"UnityEnvironment.step": {
"total": 0.053946387999985745,
"count": 1,
"is_parallel": true,
"self": 0.0006242360002488567,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00048189399990405946,
"count": 1,
"is_parallel": true,
"self": 0.00048189399990405946
},
"communicator.exchange": {
"total": 0.05097597999997561,
"count": 1,
"is_parallel": true,
"self": 0.05097597999997561
},
"steps_from_proto": {
"total": 0.0018642779998572223,
"count": 1,
"is_parallel": true,
"self": 0.0004928769997150084,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013714010001422139,
"count": 8,
"is_parallel": true,
"self": 0.0013714010001422139
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 4390.42552910513,
"count": 179055,
"is_parallel": true,
"self": 89.8525033401429,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 68.09506498397286,
"count": 179055,
"is_parallel": true,
"self": 68.09506498397286
},
"communicator.exchange": {
"total": 3949.774780879955,
"count": 179055,
"is_parallel": true,
"self": 3949.774780879955
},
"steps_from_proto": {
"total": 282.7031799010599,
"count": 179055,
"is_parallel": true,
"self": 68.02969105748502,
"children": {
"_process_rank_one_or_two_observation": {
"total": 214.6734888435749,
"count": 1432440,
"is_parallel": true,
"self": 214.6734888435749
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 2112.8315838220315,
"count": 179056,
"self": 7.897758254177916,
"children": {
"process_trajectory": {
"total": 483.43315452684624,
"count": 179056,
"self": 482.9762032428464,
"children": {
"RLTrainer._checkpoint": {
"total": 0.4569512839998424,
"count": 5,
"self": 0.4569512839998424
}
}
},
"_update_policy": {
"total": 1621.5006710410073,
"count": 1292,
"self": 633.9426676590533,
"children": {
"TorchPPOOptimizer.update": {
"total": 987.558003381954,
"count": 62954,
"self": 987.558003381954
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.2790005712304264e-06,
"count": 1,
"self": 1.2790005712304264e-06
},
"TrainerController._save_models": {
"total": 0.1406492789992626,
"count": 1,
"self": 0.0018860069994843798,
"children": {
"RLTrainer._checkpoint": {
"total": 0.13876327199977823,
"count": 1,
"self": 0.13876327199977823
}
}
}
}
}
}
}