poca-SoccerTwos / run_logs /timers.json
CodyKilpatrick's picture
First Push
afa3065
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.9132435321807861,
"min": 1.9059504270553589,
"max": 3.295731544494629,
"count": 580
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 35326.12890625,
"min": 20676.146484375,
"max": 133449.828125,
"count": 580
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 64.41558441558442,
"min": 40.714285714285715,
"max": 999.0,
"count": 580
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19840.0,
"min": 9192.0,
"max": 31252.0,
"count": 580
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1573.7548044006687,
"min": 1194.4113977825145,
"max": 1598.15703327479,
"count": 566
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 242358.239877703,
"min": 2388.822795565029,
"max": 355263.8604044648,
"count": 566
},
"SoccerTwos.Step.mean": {
"value": 5799879.0,
"min": 9642.0,
"max": 5799879.0,
"count": 580
},
"SoccerTwos.Step.sum": {
"value": 5799879.0,
"min": 9642.0,
"max": 5799879.0,
"count": 580
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.02948528155684471,
"min": -0.07843954116106033,
"max": 0.20543165504932404,
"count": 580
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -4.540733337402344,
"min": -13.80535888671875,
"max": 30.754785537719727,
"count": 580
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.03074849769473076,
"min": -0.0813608393073082,
"max": 0.20297856628894806,
"count": 580
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -4.735268592834473,
"min": -14.319507598876953,
"max": 29.838438034057617,
"count": 580
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 580
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 580
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.08035065059538012,
"min": -0.5370666682720184,
"max": 0.3993285766669682,
"count": 580
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 12.374000191688538,
"min": -59.418399930000305,
"max": 47.898799657821655,
"count": 580
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.08035065059538012,
"min": -0.5370666682720184,
"max": 0.3993285766669682,
"count": 580
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 12.374000191688538,
"min": -59.418399930000305,
"max": 47.898799657821655,
"count": 580
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 580
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 580
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.015443711108916128,
"min": 0.0105148810641064,
"max": 0.02422826134522135,
"count": 278
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.015443711108916128,
"min": 0.0105148810641064,
"max": 0.02422826134522135,
"count": 278
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.1008040428161621,
"min": 1.1924418231501477e-05,
"max": 0.12162800828615825,
"count": 278
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.1008040428161621,
"min": 1.1924418231501477e-05,
"max": 0.12162800828615825,
"count": 278
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.10219638074437777,
"min": 1.4110462204068124e-05,
"max": 0.12444661979873975,
"count": 278
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.10219638074437777,
"min": 1.4110462204068124e-05,
"max": 0.12444661979873975,
"count": 278
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 278
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 278
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 278
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000004,
"max": 0.20000000000000007,
"count": 278
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 278
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 278
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1687198843",
"python_version": "3.9.16 | packaged by conda-forge | (main, Feb 1 2023, 21:38:11) \n[Clang 14.0.6 ]",
"command_line_arguments": "/opt/homebrew/Caskroom/miniforge/base/envs/rl/bin/mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos/SoccerTwos.app --run-id=SoccerTwos --no-graphics --force",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0",
"numpy_version": "1.21.2",
"end_time_seconds": "1687220678"
},
"total": 21834.080426458,
"count": 1,
"self": 0.284593832999235,
"children": {
"run_training.setup": {
"total": 0.013124083000000009,
"count": 1,
"self": 0.013124083000000009
},
"TrainerController.start_learning": {
"total": 21833.782708542,
"count": 1,
"self": 4.813739496825292,
"children": {
"TrainerController._reset_env": {
"total": 3.6249906230106754,
"count": 29,
"self": 3.6249906230106754
},
"TrainerController.advance": {
"total": 21825.216123297167,
"count": 396203,
"self": 4.189913268706732,
"children": {
"env_step": {
"total": 17045.965847778807,
"count": 396203,
"self": 16351.30863075149,
"children": {
"SubprocessEnvManager._take_step": {
"total": 691.5932731516705,
"count": 396203,
"self": 19.00876508560441,
"children": {
"TorchPolicy.evaluate": {
"total": 672.5845080660661,
"count": 732874,
"self": 672.5845080660661
}
}
},
"workers": {
"total": 3.0639438756473028,
"count": 396202,
"self": 0.0,
"children": {
"worker_root": {
"total": 21823.5592402453,
"count": 396202,
"is_parallel": true,
"self": 6074.188804953585,
"children": {
"steps_from_proto": {
"total": 0.03814041300750404,
"count": 58,
"is_parallel": true,
"self": 0.005507164023321742,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.032633248984182295,
"count": 232,
"is_parallel": true,
"self": 0.032633248984182295
}
}
},
"UnityEnvironment.step": {
"total": 15749.332294878708,
"count": 396202,
"is_parallel": true,
"self": 43.57100942318357,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 248.96811793657102,
"count": 396202,
"is_parallel": true,
"self": 248.96811793657102
},
"communicator.exchange": {
"total": 14954.632041105906,
"count": 396202,
"is_parallel": true,
"self": 14954.632041105906
},
"steps_from_proto": {
"total": 502.1611264130481,
"count": 792404,
"is_parallel": true,
"self": 69.03948763145127,
"children": {
"_process_rank_one_or_two_observation": {
"total": 433.12163878159686,
"count": 3169616,
"is_parallel": true,
"self": 433.12163878159686
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 4775.060362249652,
"count": 396202,
"self": 37.79991290281305,
"children": {
"process_trajectory": {
"total": 973.1014304748356,
"count": 396202,
"self": 971.6940835168338,
"children": {
"RLTrainer._checkpoint": {
"total": 1.407346958001881,
"count": 11,
"self": 1.407346958001881
}
}
},
"_update_policy": {
"total": 3764.1590188720033,
"count": 278,
"self": 511.6840756118004,
"children": {
"TorchPOCAOptimizer.update": {
"total": 3252.474943260203,
"count": 8349,
"self": 3252.474943260203
}
}
}
}
}
}
},
"trainer_threads": {
"total": 4.999965312890708e-07,
"count": 1,
"self": 4.999965312890708e-07
},
"TrainerController._save_models": {
"total": 0.12785462499959976,
"count": 1,
"self": 0.002245249997940846,
"children": {
"RLTrainer._checkpoint": {
"total": 0.12560937500165892,
"count": 1,
"self": 0.12560937500165892
}
}
}
}
}
}
}