poca-SoccerTwos / run_logs /timers.json
phonenix's picture
First Push
6ad8ea4
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.4606472253799438,
"min": 1.2818001508712769,
"max": 3.295718193054199,
"count": 5000
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 30287.98046875,
"min": 16307.234375,
"max": 109189.453125,
"count": 5000
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 91.50943396226415,
"min": 40.30833333333333,
"max": 999.0,
"count": 5000
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19400.0,
"min": 13036.0,
"max": 28132.0,
"count": 5000
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1638.1093737463184,
"min": 1198.4322416455373,
"max": 1715.266684285121,
"count": 4978
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 173639.59361710976,
"min": 2397.567130297888,
"max": 397417.51250759757,
"count": 4978
},
"SoccerTwos.Step.mean": {
"value": 49999972.0,
"min": 9086.0,
"max": 49999972.0,
"count": 5000
},
"SoccerTwos.Step.sum": {
"value": 49999972.0,
"min": 9086.0,
"max": 49999972.0,
"count": 5000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.035314299166202545,
"min": -0.1513925939798355,
"max": 0.261381059885025,
"count": 5000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -3.778630018234253,
"min": -33.76054763793945,
"max": 35.04165267944336,
"count": 5000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.03379722312092781,
"min": -0.15381675958633423,
"max": 0.26418569684028625,
"count": 5000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -3.616302728652954,
"min": -34.3011360168457,
"max": 35.56377410888672,
"count": 5000
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 5000
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 5000
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.1563420557530127,
"min": -0.5447555556893349,
"max": 0.6125354814913965,
"count": 5000
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 16.728599965572357,
"min": -87.84300005435944,
"max": 76.05240023136139,
"count": 5000
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.1563420557530127,
"min": -0.5447555556893349,
"max": 0.6125354814913965,
"count": 5000
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 16.728599965572357,
"min": -87.84300005435944,
"max": 76.05240023136139,
"count": 5000
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 5000
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 5000
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.020271552559764434,
"min": 0.009705064026638866,
"max": 0.02512652772832856,
"count": 2424
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.020271552559764434,
"min": 0.009705064026638866,
"max": 0.02512652772832856,
"count": 2424
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.08147014304995537,
"min": 0.00011633944911106179,
"max": 0.12695539717872936,
"count": 2424
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.08147014304995537,
"min": 0.00011633944911106179,
"max": 0.12695539717872936,
"count": 2424
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.0821972573796908,
"min": 0.00010153193531247477,
"max": 0.12931535094976426,
"count": 2424
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.0821972573796908,
"min": 0.00010153193531247477,
"max": 0.12931535094976426,
"count": 2424
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 2424
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 2424
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 2424
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 2424
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 2424
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 2424
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1678720696",
"python_version": "3.9.16 (main, Mar 8 2023, 14:00:05) \n[GCC 11.2.0]",
"command_line_arguments": "/home/ivan/miniconda3/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos --no-graphics --force",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1678864766"
},
"total": 144070.11821003904,
"count": 1,
"self": 0.2191282089916058,
"children": {
"run_training.setup": {
"total": 0.008097673009615391,
"count": 1,
"self": 0.008097673009615391
},
"TrainerController.start_learning": {
"total": 144069.89098415704,
"count": 1,
"self": 69.02466520911548,
"children": {
"TrainerController._reset_env": {
"total": 4.632106091885362,
"count": 250,
"self": 4.632106091885362
},
"TrainerController.advance": {
"total": 143996.0793338091,
"count": 3442990,
"self": 66.45647222403204,
"children": {
"env_step": {
"total": 46244.63626892376,
"count": 3442990,
"self": 37887.79250735021,
"children": {
"SubprocessEnvManager._take_step": {
"total": 8317.804537528718,
"count": 3442990,
"self": 336.71693324833177,
"children": {
"TorchPolicy.evaluate": {
"total": 7981.087604280387,
"count": 6283122,
"self": 7981.087604280387
}
}
},
"workers": {
"total": 39.03922404482728,
"count": 3442990,
"self": 0.0,
"children": {
"worker_root": {
"total": 143976.92156786704,
"count": 3442990,
"is_parallel": true,
"self": 113122.60298189969,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.001963631948456168,
"count": 2,
"is_parallel": true,
"self": 0.000467139994725585,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014964919537305832,
"count": 8,
"is_parallel": true,
"self": 0.0014964919537305832
}
}
},
"UnityEnvironment.step": {
"total": 0.019571140001062304,
"count": 1,
"is_parallel": true,
"self": 0.0004837249871343374,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0003792460192926228,
"count": 1,
"is_parallel": true,
"self": 0.0003792460192926228
},
"communicator.exchange": {
"total": 0.017254301987122744,
"count": 1,
"is_parallel": true,
"self": 0.017254301987122744
},
"steps_from_proto": {
"total": 0.0014538670075125992,
"count": 2,
"is_parallel": true,
"self": 0.0003151089767925441,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001138758030720055,
"count": 8,
"is_parallel": true,
"self": 0.001138758030720055
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 30853.926119588723,
"count": 3442989,
"is_parallel": true,
"self": 1769.8957734311116,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 1270.0130755080027,
"count": 3442989,
"is_parallel": true,
"self": 1270.0130755080027
},
"communicator.exchange": {
"total": 22603.56078915886,
"count": 3442989,
"is_parallel": true,
"self": 22603.56078915886
},
"steps_from_proto": {
"total": 5210.45648149075,
"count": 6885978,
"is_parallel": true,
"self": 1112.2107851236942,
"children": {
"_process_rank_one_or_two_observation": {
"total": 4098.245696367056,
"count": 27543912,
"is_parallel": true,
"self": 4098.245696367056
}
}
}
}
},
"steps_from_proto": {
"total": 0.39246637863107026,
"count": 498,
"is_parallel": true,
"self": 0.08386936254100874,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.3085970160900615,
"count": 1992,
"is_parallel": true,
"self": 0.3085970160900615
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 97684.9865926613,
"count": 3442990,
"self": 435.5247600186849,
"children": {
"process_trajectory": {
"total": 8790.586209100147,
"count": 3442990,
"self": 8773.4900656924,
"children": {
"RLTrainer._checkpoint": {
"total": 17.09614340774715,
"count": 100,
"self": 17.09614340774715
}
}
},
"_update_policy": {
"total": 88458.87562354247,
"count": 2424,
"self": 6828.4682076094905,
"children": {
"TorchPOCAOptimizer.update": {
"total": 81630.40741593298,
"count": 72720,
"self": 81630.40741593298
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.189810276031494e-07,
"count": 1,
"self": 7.189810276031494e-07
},
"TrainerController._save_models": {
"total": 0.15487832797225565,
"count": 1,
"self": 0.0010931319557130337,
"children": {
"RLTrainer._checkpoint": {
"total": 0.1537851960165426,
"count": 1,
"self": 0.1537851960165426
}
}
}
}
}
}
}