ppo-Huggy / run_logs /timers.json
S1X3L4's picture
Huggy
11f402a
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4046322107315063,
"min": 1.4046322107315063,
"max": 1.4275102615356445,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 69284.890625,
"min": 67695.8046875,
"max": 75528.328125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 82.55183946488295,
"min": 73.91154422788605,
"max": 386.8914728682171,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49366.0,
"min": 48697.0,
"max": 50039.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999941.0,
"min": 49361.0,
"max": 1999941.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999941.0,
"min": 49361.0,
"max": 1999941.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4990994930267334,
"min": 0.14088945090770721,
"max": 2.533768653869629,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1494.46142578125,
"min": 18.033849716186523,
"max": 1639.181884765625,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.867014236474117,
"min": 1.769575375597924,
"max": 4.088328120693109,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2312.474513411522,
"min": 226.50564807653427,
"max": 2634.363876402378,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.867014236474117,
"min": 1.769575375597924,
"max": 4.088328120693109,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2312.474513411522,
"min": 226.50564807653427,
"max": 2634.363876402378,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.016210231348587616,
"min": 0.013600344116760728,
"max": 0.020433036771525317,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.04863069404576285,
"min": 0.027200688233521456,
"max": 0.05943559415221292,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.0605986467252175,
"min": 0.0209213190091153,
"max": 0.06299416803651386,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.1817959401756525,
"min": 0.0418426380182306,
"max": 0.1889825041095416,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.557898814066667e-06,
"min": 3.557898814066667e-06,
"max": 0.00029534070155309996,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.0673696442200001e-05,
"min": 1.0673696442200001e-05,
"max": 0.0008441149686283498,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10118593333333331,
"min": 0.10118593333333331,
"max": 0.1984469,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30355779999999993,
"min": 0.20750295000000005,
"max": 0.5813716499999999,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.917807333333335e-05,
"min": 6.917807333333335e-05,
"max": 0.004922500310000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00020753422000000004,
"min": 0.00020753422000000004,
"max": 0.014070445335000003,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1688679441",
"python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1688681775"
},
"total": 2333.3614991249997,
"count": 1,
"self": 0.7481942959998378,
"children": {
"run_training.setup": {
"total": 0.041812861999915185,
"count": 1,
"self": 0.041812861999915185
},
"TrainerController.start_learning": {
"total": 2332.571491967,
"count": 1,
"self": 4.101789002914757,
"children": {
"TrainerController._reset_env": {
"total": 4.190612704000046,
"count": 1,
"self": 4.190612704000046
},
"TrainerController.advance": {
"total": 2324.0884653440853,
"count": 233099,
"self": 4.398315180113968,
"children": {
"env_step": {
"total": 1805.3281935290197,
"count": 233099,
"self": 1521.0547212189172,
"children": {
"SubprocessEnvManager._take_step": {
"total": 281.524146038975,
"count": 233099,
"self": 16.194496782988836,
"children": {
"TorchPolicy.evaluate": {
"total": 265.32964925598617,
"count": 222932,
"self": 265.32964925598617
}
}
},
"workers": {
"total": 2.7493262711274156,
"count": 233099,
"self": 0.0,
"children": {
"worker_root": {
"total": 2325.0230581070655,
"count": 233099,
"is_parallel": true,
"self": 1081.5180724180095,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.001005941000016719,
"count": 1,
"is_parallel": true,
"self": 0.0003312149999601388,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006747260000565802,
"count": 2,
"is_parallel": true,
"self": 0.0006747260000565802
}
}
},
"UnityEnvironment.step": {
"total": 0.028931767000017317,
"count": 1,
"is_parallel": true,
"self": 0.0003418969999984256,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00021507700000711338,
"count": 1,
"is_parallel": true,
"self": 0.00021507700000711338
},
"communicator.exchange": {
"total": 0.0276285119999784,
"count": 1,
"is_parallel": true,
"self": 0.0276285119999784
},
"steps_from_proto": {
"total": 0.0007462810000333775,
"count": 1,
"is_parallel": true,
"self": 0.00020322900013525214,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005430519998981254,
"count": 2,
"is_parallel": true,
"self": 0.0005430519998981254
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1243.504985689056,
"count": 233098,
"is_parallel": true,
"self": 38.94969981816166,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 77.68511629398381,
"count": 233098,
"is_parallel": true,
"self": 77.68511629398381
},
"communicator.exchange": {
"total": 1034.1394593119524,
"count": 233098,
"is_parallel": true,
"self": 1034.1394593119524
},
"steps_from_proto": {
"total": 92.73071026495802,
"count": 233098,
"is_parallel": true,
"self": 33.115929958926245,
"children": {
"_process_rank_one_or_two_observation": {
"total": 59.614780306031776,
"count": 466196,
"is_parallel": true,
"self": 59.614780306031776
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 514.3619566349516,
"count": 233099,
"self": 6.321204007998745,
"children": {
"process_trajectory": {
"total": 134.94143800995334,
"count": 233099,
"self": 133.68916281295333,
"children": {
"RLTrainer._checkpoint": {
"total": 1.2522751970000172,
"count": 10,
"self": 1.2522751970000172
}
}
},
"_update_policy": {
"total": 373.09931461699955,
"count": 97,
"self": 313.2894940229986,
"children": {
"TorchPPOOptimizer.update": {
"total": 59.80982059400094,
"count": 2910,
"self": 59.80982059400094
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.0579997251625173e-06,
"count": 1,
"self": 1.0579997251625173e-06
},
"TrainerController._save_models": {
"total": 0.19062385799998083,
"count": 1,
"self": 0.002648765999765601,
"children": {
"RLTrainer._checkpoint": {
"total": 0.18797509200021523,
"count": 1,
"self": 0.18797509200021523
}
}
}
}
}
}
}