poca-SoccerTwo / run_logs /timers.json
lsaulier's picture
First Push
da69cf8
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.7185440063476562,
"min": 1.712857723236084,
"max": 3.295708656311035,
"count": 800
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 36790.58984375,
"min": 18034.27734375,
"max": 105462.6796875,
"count": 800
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 61.38461538461539,
"min": 42.839285714285715,
"max": 999.0,
"count": 800
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19152.0,
"min": 3996.0,
"max": 29320.0,
"count": 800
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1720.4272013985615,
"min": 1195.1268524246282,
"max": 1747.530858757312,
"count": 782
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 268386.6434181756,
"min": 2390.253704849257,
"max": 386028.50408218155,
"count": 782
},
"SoccerTwos.Step.mean": {
"value": 7999944.0,
"min": 9388.0,
"max": 7999944.0,
"count": 800
},
"SoccerTwos.Step.sum": {
"value": 7999944.0,
"min": 9388.0,
"max": 7999944.0,
"count": 800
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.04483793303370476,
"min": -0.11136051267385483,
"max": 0.32275745272636414,
"count": 800
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -7.039555549621582,
"min": -24.94475555419922,
"max": 42.92183303833008,
"count": 800
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.04256414249539375,
"min": -0.11507757753133774,
"max": 0.32423368096351624,
"count": 800
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -6.682570457458496,
"min": -25.77737808227539,
"max": 43.09465789794922,
"count": 800
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 800
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 800
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": -0.1414394894982599,
"min": -0.46153846153846156,
"max": 0.8060407765860697,
"count": 800
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": -22.205999851226807,
"min": -59.10460019111633,
"max": 83.02219998836517,
"count": 800
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": -0.1414394894982599,
"min": -0.46153846153846156,
"max": 0.8060407765860697,
"count": 800
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": -22.205999851226807,
"min": -59.10460019111633,
"max": 83.02219998836517,
"count": 800
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 800
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 800
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.016280188617793102,
"min": 0.011284240738799174,
"max": 0.025832497673885277,
"count": 385
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.016280188617793102,
"min": 0.011284240738799174,
"max": 0.025832497673885277,
"count": 385
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.10526493216554324,
"min": 1.2015580068691634e-05,
"max": 0.12397578631838163,
"count": 385
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.10526493216554324,
"min": 1.2015580068691634e-05,
"max": 0.12397578631838163,
"count": 385
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.10657266477743785,
"min": 1.1846101369883399e-05,
"max": 0.12565159474809964,
"count": 385
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.10657266477743785,
"min": 1.1846101369883399e-05,
"max": 0.12565159474809964,
"count": 385
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 385
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 385
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 385
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.19999999999999996,
"max": 0.20000000000000007,
"count": 385
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 385
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 385
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1675969539",
"python_version": "3.9.16 (main, Jan 11 2023, 16:05:54) \n[GCC 11.2.0]",
"command_line_arguments": "/home/lsaulier/anaconda3/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos_2 --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1675991702"
},
"total": 22163.11969683,
"count": 1,
"self": 0.21965920000366168,
"children": {
"run_training.setup": {
"total": 0.007608962994709145,
"count": 1,
"self": 0.007608962994709145
},
"TrainerController.start_learning": {
"total": 22162.892428667,
"count": 1,
"self": 11.635546017918386,
"children": {
"TrainerController._reset_env": {
"total": 0.852964419987984,
"count": 20,
"self": 0.852964419987984
},
"TrainerController.advance": {
"total": 22150.273212965105,
"count": 546024,
"self": 10.067874203523388,
"children": {
"env_step": {
"total": 6366.853094141094,
"count": 546024,
"self": 5136.617505596732,
"children": {
"SubprocessEnvManager._take_step": {
"total": 1223.4178501296628,
"count": 546024,
"self": 50.36454255526769,
"children": {
"TorchPolicy.evaluate": {
"total": 1173.0533075743951,
"count": 1003434,
"self": 1173.0533075743951
}
}
},
"workers": {
"total": 6.817738414698397,
"count": 546024,
"self": 0.0,
"children": {
"worker_root": {
"total": 22150.67067910044,
"count": 546024,
"is_parallel": true,
"self": 18097.142765594042,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.001882700998976361,
"count": 2,
"is_parallel": true,
"self": 0.0005679990063072182,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0013147019926691428,
"count": 8,
"is_parallel": true,
"self": 0.0013147019926691428
}
}
},
"UnityEnvironment.step": {
"total": 0.017189674996188842,
"count": 1,
"is_parallel": true,
"self": 0.00042573100654408336,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00039789799484424293,
"count": 1,
"is_parallel": true,
"self": 0.00039789799484424293
},
"communicator.exchange": {
"total": 0.0151750759978313,
"count": 1,
"is_parallel": true,
"self": 0.0151750759978313
},
"steps_from_proto": {
"total": 0.0011909699969692156,
"count": 2,
"is_parallel": true,
"self": 0.00025141499645542353,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.000939555000513792,
"count": 8,
"is_parallel": true,
"self": 0.000939555000513792
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 4053.5034054913704,
"count": 546023,
"is_parallel": true,
"self": 245.24890382892045,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 189.3249936548309,
"count": 546023,
"is_parallel": true,
"self": 189.3249936548309
},
"communicator.exchange": {
"total": 2939.566085178936,
"count": 546023,
"is_parallel": true,
"self": 2939.566085178936
},
"steps_from_proto": {
"total": 679.3634228286828,
"count": 1092046,
"is_parallel": true,
"self": 132.3677148117349,
"children": {
"_process_rank_one_or_two_observation": {
"total": 546.9957080169479,
"count": 4368184,
"is_parallel": true,
"self": 546.9957080169479
}
}
}
}
},
"steps_from_proto": {
"total": 0.024508015027095098,
"count": 38,
"is_parallel": true,
"self": 0.004874444020970259,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.01963357100612484,
"count": 152,
"is_parallel": true,
"self": 0.01963357100612484
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 15773.352244620488,
"count": 546024,
"self": 90.01150891112047,
"children": {
"process_trajectory": {
"total": 1849.1886234544581,
"count": 546024,
"self": 1847.0379783684475,
"children": {
"RLTrainer._checkpoint": {
"total": 2.150645086010627,
"count": 16,
"self": 2.150645086010627
}
}
},
"_update_policy": {
"total": 13834.15211225491,
"count": 385,
"self": 893.2984637601694,
"children": {
"TorchPOCAOptimizer.update": {
"total": 12940.85364849474,
"count": 11568,
"self": 12940.85364849474
}
}
}
}
}
}
},
"trainer_threads": {
"total": 6.809932528994977e-07,
"count": 1,
"self": 6.809932528994977e-07
},
"TrainerController._save_models": {
"total": 0.13070458299625898,
"count": 1,
"self": 0.0012228909981786273,
"children": {
"RLTrainer._checkpoint": {
"total": 0.12948169199808035,
"count": 1,
"self": 0.12948169199808035
}
}
}
}
}
}
}