ppo-Huggy / run_logs /timers.json
LOGQS's picture
Huggy
34320d0
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4070566892623901,
"min": 1.4070566892623901,
"max": 1.4276663064956665,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 71325.109375,
"min": 68204.21875,
"max": 78327.03125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 81.09031198686371,
"min": 77.65408805031447,
"max": 395.4803149606299,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49384.0,
"min": 48840.0,
"max": 50226.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999992.0,
"min": 49672.0,
"max": 1999992.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999992.0,
"min": 49672.0,
"max": 1999992.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.5277159214019775,
"min": 0.052642323076725006,
"max": 2.5277159214019775,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1539.3790283203125,
"min": 6.632932662963867,
"max": 1539.577880859375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.9085409951523213,
"min": 1.9222564030261267,
"max": 4.019346476408223,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2380.301466047764,
"min": 242.20430678129196,
"max": 2400.582337141037,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.9085409951523213,
"min": 1.9222564030261267,
"max": 4.019346476408223,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2380.301466047764,
"min": 242.20430678129196,
"max": 2400.582337141037,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.017903978318079477,
"min": 0.013629258640058754,
"max": 0.02068094924131098,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.053711934954238434,
"min": 0.02747704444724756,
"max": 0.05910704522781696,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.062020447933011584,
"min": 0.021423754965265593,
"max": 0.06304539125412703,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.18606134379903475,
"min": 0.042847509930531186,
"max": 0.18606134379903475,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.4901488366500003e-06,
"min": 3.4901488366500003e-06,
"max": 0.0002953350015549999,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.047044650995e-05,
"min": 1.047044650995e-05,
"max": 0.0008443254185581999,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10116335000000003,
"min": 0.10116335000000003,
"max": 0.198445,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30349005000000007,
"min": 0.20746989999999998,
"max": 0.5814418000000001,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.805116500000003e-05,
"min": 6.805116500000003e-05,
"max": 0.004922405500000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0002041534950000001,
"min": 0.0002041534950000001,
"max": 0.01407394582,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1685430000",
"python_version": "3.10.11 (main, Apr 5 2023, 14:15:10) [GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1685432323"
},
"total": 2323.243148009,
"count": 1,
"self": 0.7380630469997413,
"children": {
"run_training.setup": {
"total": 0.04502905699996518,
"count": 1,
"self": 0.04502905699996518
},
"TrainerController.start_learning": {
"total": 2322.460055905,
"count": 1,
"self": 4.173163498053327,
"children": {
"TrainerController._reset_env": {
"total": 3.91665989899991,
"count": 1,
"self": 3.91665989899991
},
"TrainerController.advance": {
"total": 2314.1972209809464,
"count": 232695,
"self": 4.28697782285326,
"children": {
"env_step": {
"total": 1805.4118888730907,
"count": 232695,
"self": 1525.877000536141,
"children": {
"SubprocessEnvManager._take_step": {
"total": 276.763194763078,
"count": 232695,
"self": 17.04014405109706,
"children": {
"TorchPolicy.evaluate": {
"total": 259.72305071198093,
"count": 222970,
"self": 259.72305071198093
}
}
},
"workers": {
"total": 2.771693573871744,
"count": 232695,
"self": 0.0,
"children": {
"worker_root": {
"total": 2314.957366741983,
"count": 232695,
"is_parallel": true,
"self": 1070.6848906030095,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0009920280000415005,
"count": 1,
"is_parallel": true,
"self": 0.00029201300003478536,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0007000150000067151,
"count": 2,
"is_parallel": true,
"self": 0.0007000150000067151
}
}
},
"UnityEnvironment.step": {
"total": 0.028909755999961817,
"count": 1,
"is_parallel": true,
"self": 0.0003476929999806089,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00026331100002607855,
"count": 1,
"is_parallel": true,
"self": 0.00026331100002607855
},
"communicator.exchange": {
"total": 0.027576537999948414,
"count": 1,
"is_parallel": true,
"self": 0.027576537999948414
},
"steps_from_proto": {
"total": 0.0007222140000067157,
"count": 1,
"is_parallel": true,
"self": 0.00019744200005789025,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005247719999488254,
"count": 2,
"is_parallel": true,
"self": 0.0005247719999488254
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1244.2724761389736,
"count": 232694,
"is_parallel": true,
"self": 37.57182651410494,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 77.39394426889692,
"count": 232694,
"is_parallel": true,
"self": 77.39394426889692
},
"communicator.exchange": {
"total": 1038.3484595979653,
"count": 232694,
"is_parallel": true,
"self": 1038.3484595979653
},
"steps_from_proto": {
"total": 90.95824575800634,
"count": 232694,
"is_parallel": true,
"self": 33.30263193496614,
"children": {
"_process_rank_one_or_two_observation": {
"total": 57.6556138230402,
"count": 465388,
"is_parallel": true,
"self": 57.6556138230402
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 504.4983542850025,
"count": 232695,
"self": 6.447900051954321,
"children": {
"process_trajectory": {
"total": 133.2132649750472,
"count": 232695,
"self": 131.78404895304743,
"children": {
"RLTrainer._checkpoint": {
"total": 1.429216021999764,
"count": 10,
"self": 1.429216021999764
}
}
},
"_update_policy": {
"total": 364.837189258001,
"count": 97,
"self": 305.5086788620024,
"children": {
"TorchPPOOptimizer.update": {
"total": 59.32851039599859,
"count": 2910,
"self": 59.32851039599859
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.2069999684172217e-06,
"count": 1,
"self": 1.2069999684172217e-06
},
"TrainerController._save_models": {
"total": 0.17301032000023042,
"count": 1,
"self": 0.0026289220004400704,
"children": {
"RLTrainer._checkpoint": {
"total": 0.17038139799979035,
"count": 1,
"self": 0.17038139799979035
}
}
}
}
}
}
}