poca-SoccerTwos / run_logs /timers.json
ninja21's picture
First Push
181dbd5 verified
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 3.170539379119873,
"min": 3.144883632659912,
"max": 3.295755386352539,
"count": 100
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 70208.421875,
"min": 13924.3701171875,
"max": 111513.0703125,
"count": 100
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 992.6,
"min": 421.0,
"max": 999.0,
"count": 100
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19852.0,
"min": 12372.0,
"max": 27972.0,
"count": 100
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1197.2151870767996,
"min": 1191.6544841157602,
"max": 1203.8147289055166,
"count": 71
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 2394.430374153599,
"min": 2384.2969477556185,
"max": 12023.919422083303,
"count": 71
},
"SoccerTwos.Step.mean": {
"value": 999244.0,
"min": 9740.0,
"max": 999244.0,
"count": 100
},
"SoccerTwos.Step.sum": {
"value": 999244.0,
"min": 9740.0,
"max": 999244.0,
"count": 100
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.0031801178120076656,
"min": -0.03964674472808838,
"max": 0.0138687239959836,
"count": 100
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -0.03180117905139923,
"min": -0.47576093673706055,
"max": 0.18029341101646423,
"count": 100
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.002169694285839796,
"min": -0.036696907132864,
"max": 0.01457876805216074,
"count": 100
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -0.021696943789720535,
"min": -0.47621333599090576,
"max": 0.18952398002147675,
"count": 100
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 100
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.025139999389648438,
"min": -0.5,
"max": 0.3055999938930784,
"count": 100
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 0.2513999938964844,
"min": -6.0,
"max": 4.2783999145030975,
"count": 100
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.025139999389648438,
"min": -0.5,
"max": 0.3055999938930784,
"count": 100
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 0.2513999938964844,
"min": -6.0,
"max": 4.2783999145030975,
"count": 100
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 100
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.016157496247130136,
"min": 0.011717753964088235,
"max": 0.024546924280002713,
"count": 45
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.016157496247130136,
"min": 0.011717753964088235,
"max": 0.024546924280002713,
"count": 45
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.004855411833462616,
"min": 1.1203015541620213e-06,
"max": 0.004855411833462616,
"count": 45
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.004855411833462616,
"min": 1.1203015541620213e-06,
"max": 0.004855411833462616,
"count": 45
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.004853233058626453,
"min": 1.4627920412143188e-06,
"max": 0.0049429924537738165,
"count": 45
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.004853233058626453,
"min": 1.4627920412143188e-06,
"max": 0.0049429924537738165,
"count": 45
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 45
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 45
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 45
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.2,
"max": 0.20000000000000007,
"count": 45
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 45
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005,
"max": 0.005000000000000001,
"count": 45
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1711373504",
"python_version": "3.10.12 | packaged by conda-forge | (main, Jun 23 2023, 22:40:32) [GCC 12.3.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.x86_64 --run-id=SoccerTwos --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.2.1+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1711378953"
},
"total": 5449.153454893,
"count": 1,
"self": 0.24403187800089654,
"children": {
"run_training.setup": {
"total": 0.09170745699998406,
"count": 1,
"self": 0.09170745699998406
},
"TrainerController.start_learning": {
"total": 5448.817715558,
"count": 1,
"self": 2.2251287779818085,
"children": {
"TrainerController._reset_env": {
"total": 2.813229217999151,
"count": 5,
"self": 2.813229217999151
},
"TrainerController.advance": {
"total": 5443.22722290102,
"count": 65172,
"self": 2.7172877041502943,
"children": {
"env_step": {
"total": 2201.5783943737842,
"count": 65172,
"self": 1767.4962277078612,
"children": {
"SubprocessEnvManager._take_step": {
"total": 432.54229496884045,
"count": 65172,
"self": 15.269354169766075,
"children": {
"TorchPolicy.evaluate": {
"total": 417.2729407990744,
"count": 129404,
"self": 417.2729407990744
}
}
},
"workers": {
"total": 1.5398716970826172,
"count": 65172,
"self": 0.0,
"children": {
"worker_root": {
"total": 5381.059578365979,
"count": 65172,
"is_parallel": true,
"self": 3920.0858749181225,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.007013139000036972,
"count": 2,
"is_parallel": true,
"self": 0.004208731999597148,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0028044070004398236,
"count": 8,
"is_parallel": true,
"self": 0.0028044070004398236
}
}
},
"UnityEnvironment.step": {
"total": 0.04630665800004863,
"count": 1,
"is_parallel": true,
"self": 0.001330119000385821,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0009984280000026047,
"count": 1,
"is_parallel": true,
"self": 0.0009984280000026047
},
"communicator.exchange": {
"total": 0.04002135699988685,
"count": 1,
"is_parallel": true,
"self": 0.04002135699988685
},
"steps_from_proto": {
"total": 0.003956753999773355,
"count": 2,
"is_parallel": true,
"self": 0.0007377499996437109,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.003219004000129644,
"count": 8,
"is_parallel": true,
"self": 0.003219004000129644
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1460.9627264038581,
"count": 65171,
"is_parallel": true,
"self": 89.00598279105566,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 54.840749249899545,
"count": 65171,
"is_parallel": true,
"self": 54.840749249899545
},
"communicator.exchange": {
"total": 1041.0566117079454,
"count": 65171,
"is_parallel": true,
"self": 1041.0566117079454
},
"steps_from_proto": {
"total": 276.0593826549575,
"count": 130342,
"is_parallel": true,
"self": 48.61762786653708,
"children": {
"_process_rank_one_or_two_observation": {
"total": 227.44175478842044,
"count": 521368,
"is_parallel": true,
"self": 227.44175478842044
}
}
}
}
},
"steps_from_proto": {
"total": 0.010977043998536828,
"count": 8,
"is_parallel": true,
"self": 0.00219157800256653,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.008785465995970299,
"count": 32,
"is_parallel": true,
"self": 0.008785465995970299
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 3238.931540823085,
"count": 65172,
"self": 20.58632396802932,
"children": {
"process_trajectory": {
"total": 403.5471596110549,
"count": 65172,
"self": 403.0601233310547,
"children": {
"RLTrainer._checkpoint": {
"total": 0.48703628000021126,
"count": 2,
"self": 0.48703628000021126
}
}
},
"_update_policy": {
"total": 2814.7980572440006,
"count": 46,
"self": 200.79373283101677,
"children": {
"TorchPOCAOptimizer.update": {
"total": 2614.004324412984,
"count": 1382,
"self": 2614.004324412984
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.9929993868572637e-06,
"count": 1,
"self": 1.9929993868572637e-06
},
"TrainerController._save_models": {
"total": 0.5521326680000129,
"count": 1,
"self": 0.005468391999784217,
"children": {
"RLTrainer._checkpoint": {
"total": 0.5466642760002287,
"count": 1,
"self": 0.5466642760002287
}
}
}
}
}
}
}