poca-SoccerTwos / run_logs /timers.json
Issacwong
First Push
d0ac5d3
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.207340717315674,
"min": 3.2038371562957764,
"max": 3.2957329750061035,
"count": 49
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 73691.859375,
"min": 33190.0234375,
"max": 105463.453125,
"count": 49
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 999.0,
"min": 548.2222222222222,
"max": 999.0,
"count": 49
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19980.0,
"min": 16436.0,
"max": 31968.0,
"count": 49
},
"SoccerTwos.Step.mean": {
"value": 499508.0,
"min": 9000.0,
"max": 499508.0,
"count": 50
},
"SoccerTwos.Step.sum": {
"value": 499508.0,
"min": 9000.0,
"max": 499508.0,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.000544462469406426,
"min": -0.08753174543380737,
"max": -0.0005048831808380783,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -0.005444624461233616,
"min": -1.1300594806671143,
"max": -0.005444624461233616,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.0009487472707405686,
"min": -0.0875873863697052,
"max": -0.0007544748368673027,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -0.009487472474575043,
"min": -1.1205412149429321,
"max": -0.008299223147332668,
"count": 50
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 50
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.0,
"min": -0.6423733333746592,
"max": 0.15703333417574564,
"count": 50
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 0.0,
"min": -9.635600000619888,
"max": 1.8844000101089478,
"count": 50
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.0,
"min": -0.6423733333746592,
"max": 0.15703333417574564,
"count": 50
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 0.0,
"min": -9.635600000619888,
"max": 1.8844000101089478,
"count": 50
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 50
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 50
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1198.2180175999972,
"min": 1194.7392473541395,
"max": 1202.0026651031105,
"count": 40
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 2396.4360351999944,
"min": 2390.7043647929877,
"max": 14398.022971881543,
"count": 40
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.015105669860107203,
"min": 0.012382276609908633,
"max": 0.02145515911009473,
"count": 22
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.015105669860107203,
"min": 0.012382276609908633,
"max": 0.02145515911009473,
"count": 22
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.001043486234266311,
"min": 0.00018685251537438793,
"max": 0.006256534438580275,
"count": 22
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.001043486234266311,
"min": 0.00018685251537438793,
"max": 0.006256534438580275,
"count": 22
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.0010441043404474234,
"min": 0.00018883000793721295,
"max": 0.006314269049714009,
"count": 22
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.0010441043404474234,
"min": 0.00018883000793721295,
"max": 0.006314269049714009,
"count": 22
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 22
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 22
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.19999999999999993,
"max": 0.20000000000000007,
"count": 22
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.19999999999999993,
"max": 0.20000000000000007,
"count": 22
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 22
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 22
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1681539755",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/linux/SoccerTwos --run-id=SoccerTwos --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1681540874"
},
"total": 1119.3354234619999,
"count": 1,
"self": 0.3908143159997053,
"children": {
"run_training.setup": {
"total": 0.11485964199999898,
"count": 1,
"self": 0.11485964199999898
},
"TrainerController.start_learning": {
"total": 1118.8297495040001,
"count": 1,
"self": 0.7752854250072687,
"children": {
"TrainerController._reset_env": {
"total": 5.116547662999949,
"count": 3,
"self": 5.116547662999949
},
"TrainerController.advance": {
"total": 1112.6760890209928,
"count": 32599,
"self": 0.8786119010062521,
"children": {
"env_step": {
"total": 901.8980586439775,
"count": 32599,
"self": 710.6199166659889,
"children": {
"SubprocessEnvManager._take_step": {
"total": 190.81615588399188,
"count": 32599,
"self": 5.67221772500551,
"children": {
"TorchPolicy.evaluate": {
"total": 185.14393815898637,
"count": 64780,
"self": 185.14393815898637
}
}
},
"workers": {
"total": 0.4619860939967566,
"count": 32599,
"self": 0.0,
"children": {
"worker_root": {
"total": 1116.181876154984,
"count": 32599,
"is_parallel": true,
"self": 514.2796645149695,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.005971231999978954,
"count": 2,
"is_parallel": true,
"self": 0.0036164189999681184,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.002354813000010836,
"count": 8,
"is_parallel": true,
"self": 0.002354813000010836
}
}
},
"UnityEnvironment.step": {
"total": 0.06992805999999518,
"count": 1,
"is_parallel": true,
"self": 0.0010409269999627213,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0009196529999826453,
"count": 1,
"is_parallel": true,
"self": 0.0009196529999826453
},
"communicator.exchange": {
"total": 0.06453340400003071,
"count": 1,
"is_parallel": true,
"self": 0.06453340400003071
},
"steps_from_proto": {
"total": 0.003434076000019104,
"count": 2,
"is_parallel": true,
"self": 0.0007099779999748534,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0027240980000442505,
"count": 8,
"is_parallel": true,
"self": 0.0027240980000442505
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 601.8962270070145,
"count": 32598,
"is_parallel": true,
"self": 35.38102434600012,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 26.43470972297007,
"count": 32598,
"is_parallel": true,
"self": 26.43470972297007
},
"communicator.exchange": {
"total": 428.3626060020186,
"count": 32598,
"is_parallel": true,
"self": 428.3626060020186
},
"steps_from_proto": {
"total": 111.7178869360257,
"count": 65196,
"is_parallel": true,
"self": 20.731826283994224,
"children": {
"_process_rank_one_or_two_observation": {
"total": 90.98606065203148,
"count": 260784,
"is_parallel": true,
"self": 90.98606065203148
}
}
}
}
},
"steps_from_proto": {
"total": 0.00598463299991181,
"count": 4,
"is_parallel": true,
"self": 0.0011972630001082507,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.004787369999803559,
"count": 16,
"is_parallel": true,
"self": 0.004787369999803559
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 209.89941847600903,
"count": 32599,
"self": 6.894574033016454,
"children": {
"process_trajectory": {
"total": 54.78363244999235,
"count": 32599,
"self": 54.528717558992355,
"children": {
"RLTrainer._checkpoint": {
"total": 0.25491489099999853,
"count": 1,
"self": 0.25491489099999853
}
}
},
"_update_policy": {
"total": 148.22121199300022,
"count": 22,
"self": 94.37949303799633,
"children": {
"TorchPOCAOptimizer.update": {
"total": 53.8417189550039,
"count": 675,
"self": 53.8417189550039
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.630000477045542e-07,
"count": 1,
"self": 9.630000477045542e-07
},
"TrainerController._save_models": {
"total": 0.26182643200013445,
"count": 1,
"self": 0.0019213699999909295,
"children": {
"RLTrainer._checkpoint": {
"total": 0.2599050620001435,
"count": 1,
"self": 0.2599050620001435
}
}
}
}
}
}
}