ppo-Huggy / run_logs /timers.json
sohm's picture
Huggy
bb4dec3
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4029500484466553,
"min": 1.4029500484466553,
"max": 1.427134394645691,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 68240.890625,
"min": 67824.8515625,
"max": 79353.578125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 117.40995260663507,
"min": 84.11945392491468,
"max": 381.4503816793893,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49547.0,
"min": 48940.0,
"max": 50368.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999946.0,
"min": 49950.0,
"max": 1999946.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999946.0,
"min": 49950.0,
"max": 1999946.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.306443691253662,
"min": 0.07948636263608932,
"max": 2.4858205318450928,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 973.3192138671875,
"min": 10.333227157592773,
"max": 1438.078125,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.4616845073575657,
"min": 1.8123217739737951,
"max": 4.105002243518829,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1460.8308621048927,
"min": 235.60183061659336,
"max": 2225.898113846779,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.4616845073575657,
"min": 1.8123217739737951,
"max": 4.105002243518829,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1460.8308621048927,
"min": 235.60183061659336,
"max": 2225.898113846779,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.016154403954796077,
"min": 0.013882778619558342,
"max": 0.02106150973607631,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.048463211864388234,
"min": 0.027858269662829116,
"max": 0.05514129457429438,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04534555743965838,
"min": 0.022708302984635036,
"max": 0.06060358677059412,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.13603667231897515,
"min": 0.04541660596927007,
"max": 0.17140753430624803,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.3400488866833246e-06,
"min": 3.3400488866833246e-06,
"max": 0.00029532255155915007,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.0020146660049974e-05,
"min": 1.0020146660049974e-05,
"max": 0.0008440612686462497,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10111331666666669,
"min": 0.10111331666666669,
"max": 0.19844085,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.3033399500000001,
"min": 0.2073955,
"max": 0.58135375,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.555450166666654e-05,
"min": 6.555450166666654e-05,
"max": 0.0049221984150000005,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00019666350499999963,
"min": 0.00019666350499999963,
"max": 0.014069552125000004,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1670899706",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1670901918"
},
"total": 2211.9882652610004,
"count": 1,
"self": 0.4348890110004504,
"children": {
"run_training.setup": {
"total": 0.10711237100008475,
"count": 1,
"self": 0.10711237100008475
},
"TrainerController.start_learning": {
"total": 2211.446263879,
"count": 1,
"self": 3.9073637871310893,
"children": {
"TrainerController._reset_env": {
"total": 10.172514633999981,
"count": 1,
"self": 10.172514633999981
},
"TrainerController.advance": {
"total": 2197.250240443869,
"count": 232421,
"self": 3.9935037146760806,
"children": {
"env_step": {
"total": 1722.4658394911257,
"count": 232421,
"self": 1445.503340941234,
"children": {
"SubprocessEnvManager._take_step": {
"total": 274.3513423609877,
"count": 232421,
"self": 14.228028341023105,
"children": {
"TorchPolicy.evaluate": {
"total": 260.1233140199646,
"count": 222993,
"self": 65.34839569591259,
"children": {
"TorchPolicy.sample_actions": {
"total": 194.774918324052,
"count": 222993,
"self": 194.774918324052
}
}
}
}
},
"workers": {
"total": 2.611156188904147,
"count": 232421,
"self": 0.0,
"children": {
"worker_root": {
"total": 2203.685773803074,
"count": 232421,
"is_parallel": true,
"self": 1018.9052431620112,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0019430820000252425,
"count": 1,
"is_parallel": true,
"self": 0.000311869000029219,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0016312129999960234,
"count": 2,
"is_parallel": true,
"self": 0.0016312129999960234
}
}
},
"UnityEnvironment.step": {
"total": 0.02809839400003966,
"count": 1,
"is_parallel": true,
"self": 0.0002744500000062544,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0001773080000475602,
"count": 1,
"is_parallel": true,
"self": 0.0001773080000475602
},
"communicator.exchange": {
"total": 0.0269492289999107,
"count": 1,
"is_parallel": true,
"self": 0.0269492289999107
},
"steps_from_proto": {
"total": 0.0006974070000751453,
"count": 1,
"is_parallel": true,
"self": 0.00023462100000415376,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0004627860000709916,
"count": 2,
"is_parallel": true,
"self": 0.0004627860000709916
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1184.7805306410628,
"count": 232420,
"is_parallel": true,
"self": 34.19085954609341,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 74.99743670295618,
"count": 232420,
"is_parallel": true,
"self": 74.99743670295618
},
"communicator.exchange": {
"total": 982.9845625260191,
"count": 232420,
"is_parallel": true,
"self": 982.9845625260191
},
"steps_from_proto": {
"total": 92.607671865994,
"count": 232420,
"is_parallel": true,
"self": 37.830659432086236,
"children": {
"_process_rank_one_or_two_observation": {
"total": 54.77701243390777,
"count": 464840,
"is_parallel": true,
"self": 54.77701243390777
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 470.7908972380675,
"count": 232421,
"self": 6.257566197091251,
"children": {
"process_trajectory": {
"total": 149.21605765297465,
"count": 232421,
"self": 148.74709216497433,
"children": {
"RLTrainer._checkpoint": {
"total": 0.4689654880003218,
"count": 4,
"self": 0.4689654880003218
}
}
},
"_update_policy": {
"total": 315.3172733880016,
"count": 97,
"self": 260.94258537999417,
"children": {
"TorchPPOOptimizer.update": {
"total": 54.37468800800741,
"count": 2910,
"self": 54.37468800800741
}
}
}
}
}
}
},
"trainer_threads": {
"total": 8.729998626222368e-07,
"count": 1,
"self": 8.729998626222368e-07
},
"TrainerController._save_models": {
"total": 0.11614414099994974,
"count": 1,
"self": 0.0021073560001241276,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11403678499982561,
"count": 1,
"self": 0.11403678499982561
}
}
}
}
}
}
}