SoccerTwos / run_logs /timers.json
jcramirezpr's picture
commit soccer
2abdf88
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 2.0481293201446533,
"min": 2.031398057937622,
"max": 3.2957351207733154,
"count": 603
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 40372.7265625,
"min": 13354.611328125,
"max": 135264.46875,
"count": 603
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 53.26881720430107,
"min": 41.42372881355932,
"max": 999.0,
"count": 603
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19816.0,
"min": 14988.0,
"max": 26244.0,
"count": 603
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1546.940166718161,
"min": 1197.094490819942,
"max": 1548.8583906123458,
"count": 559
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 287730.87100957794,
"min": 2394.188981639884,
"max": 356815.5596697034,
"count": 559
},
"SoccerTwos.Step.mean": {
"value": 6029984.0,
"min": 9140.0,
"max": 6029984.0,
"count": 603
},
"SoccerTwos.Step.sum": {
"value": 6029984.0,
"min": 9140.0,
"max": 6029984.0,
"count": 603
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.0018405868904665112,
"min": -0.10299093276262283,
"max": 0.169850692152977,
"count": 603
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -0.34234917163848877,
"min": -18.332386016845703,
"max": 21.55048370361328,
"count": 603
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.005029987543821335,
"min": -0.10474444925785065,
"max": 0.16812866926193237,
"count": 603
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": 0.9355776906013489,
"min": -18.644512176513672,
"max": 21.8997802734375,
"count": 603
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 603
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 603
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.02575698911502797,
"min": -0.73235789725655,
"max": 0.5598113199449936,
"count": 603
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 4.790799975395203,
"min": -49.42200016975403,
"max": 56.86640000343323,
"count": 603
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.02575698911502797,
"min": -0.73235789725655,
"max": 0.5598113199449936,
"count": 603
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 4.790799975395203,
"min": -49.42200016975403,
"max": 56.86640000343323,
"count": 603
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 603
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 603
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.0177919195984335,
"min": 0.010626989774755203,
"max": 0.024175661398718754,
"count": 286
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.0177919195984335,
"min": 0.010626989774755203,
"max": 0.024175661398718754,
"count": 286
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.11514378090699513,
"min": 2.682095528143691e-05,
"max": 0.12018098334471385,
"count": 286
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.11514378090699513,
"min": 2.682095528143691e-05,
"max": 0.12018098334471385,
"count": 286
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.11886790816982587,
"min": 2.758895795219966e-05,
"max": 0.12267898718516032,
"count": 286
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.11886790816982587,
"min": 2.758895795219966e-05,
"max": 0.12267898718516032,
"count": 286
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 286
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 286
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 286
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 286
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 286
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 286
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1680214312",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=/notebooks/ml-agents/training-envs-executables/SoccerTwos/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics --force",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1680227456"
},
"total": 13144.30611139501,
"count": 1,
"self": 0.22970467898994684,
"children": {
"run_training.setup": {
"total": 0.02708850195631385,
"count": 1,
"self": 0.02708850195631385
},
"TrainerController.start_learning": {
"total": 13144.049318214064,
"count": 1,
"self": 13.20497715624515,
"children": {
"TrainerController._reset_env": {
"total": 9.555639024241827,
"count": 31,
"self": 9.555639024241827
},
"TrainerController.advance": {
"total": 13120.880024859449,
"count": 403605,
"self": 12.91263545525726,
"children": {
"env_step": {
"total": 10199.374266911298,
"count": 403605,
"self": 7845.777121244697,
"children": {
"SubprocessEnvManager._take_step": {
"total": 2345.8237072145566,
"count": 403605,
"self": 67.34951957874,
"children": {
"TorchPolicy.evaluate": {
"total": 2278.4741876358166,
"count": 766598,
"self": 2278.4741876358166
}
}
},
"workers": {
"total": 7.773438452044502,
"count": 403604,
"self": 0.0,
"children": {
"worker_root": {
"total": 13121.212236626307,
"count": 403604,
"is_parallel": true,
"self": 6702.960799925611,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0028533160220831633,
"count": 2,
"is_parallel": true,
"self": 0.0007448650430887938,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0021084509789943695,
"count": 8,
"is_parallel": true,
"self": 0.0021084509789943695
}
}
},
"UnityEnvironment.step": {
"total": 0.03437218500766903,
"count": 1,
"is_parallel": true,
"self": 0.0008932790951803327,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0007292130030691624,
"count": 1,
"is_parallel": true,
"self": 0.0007292130030691624
},
"communicator.exchange": {
"total": 0.03057268902193755,
"count": 1,
"is_parallel": true,
"self": 0.03057268902193755
},
"steps_from_proto": {
"total": 0.0021770038874819875,
"count": 2,
"is_parallel": true,
"self": 0.0005012680776417255,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001675735809840262,
"count": 8,
"is_parallel": true,
"self": 0.001675735809840262
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 6418.185005472973,
"count": 403603,
"is_parallel": true,
"self": 281.9115276389057,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 186.6211402276531,
"count": 403603,
"is_parallel": true,
"self": 186.6211402276531
},
"communicator.exchange": {
"total": 5090.0244737637695,
"count": 403603,
"is_parallel": true,
"self": 5090.0244737637695
},
"steps_from_proto": {
"total": 859.6278638426447,
"count": 807206,
"is_parallel": true,
"self": 183.71781118237413,
"children": {
"_process_rank_one_or_two_observation": {
"total": 675.9100526602706,
"count": 3228824,
"is_parallel": true,
"self": 675.9100526602706
}
}
}
}
},
"steps_from_proto": {
"total": 0.06643122772220522,
"count": 60,
"is_parallel": true,
"self": 0.014306416152976453,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.05212481156922877,
"count": 240,
"is_parallel": true,
"self": 0.05212481156922877
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 2908.5931224928936,
"count": 403604,
"self": 101.2361810276052,
"children": {
"process_trajectory": {
"total": 1005.2014897217741,
"count": 403604,
"self": 999.0768546346808,
"children": {
"RLTrainer._checkpoint": {
"total": 6.124635087093338,
"count": 12,
"self": 6.124635087093338
}
}
},
"_update_policy": {
"total": 1802.1554517435143,
"count": 286,
"self": 1126.1437411573716,
"children": {
"TorchPOCAOptimizer.update": {
"total": 676.0117105861427,
"count": 8580,
"self": 676.0117105861427
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.8300488591194153e-06,
"count": 1,
"self": 1.8300488591194153e-06
},
"TrainerController._save_models": {
"total": 0.4086753440788016,
"count": 1,
"self": 0.0021712761372327805,
"children": {
"RLTrainer._checkpoint": {
"total": 0.4065040679415688,
"count": 1,
"self": 0.4065040679415688
}
}
}
}
}
}
}