ppo-Huggy / run_logs /timers.json
jemal's picture
Pushed Shadow Huggy to the Hub
2e6d7a4
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4087244272232056,
"min": 1.4086841344833374,
"max": 1.4273147583007812,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 69762.8515625,
"min": 68227.3828125,
"max": 77970.4453125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 107.48156182212581,
"min": 105.02991452991454,
"max": 407.0813008130081,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49549.0,
"min": 49024.0,
"max": 50246.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999995.0,
"min": 49978.0,
"max": 1999995.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999995.0,
"min": 49978.0,
"max": 1999995.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4268085956573486,
"min": 0.07700655609369278,
"max": 2.615288257598877,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1118.7587890625,
"min": 9.394800186157227,
"max": 1154.527587890625,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.543316109320088,
"min": 1.9420151962120025,
"max": 3.78355118707957,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1633.4687263965607,
"min": 236.9258539378643,
"max": 1679.539180636406,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.543316109320088,
"min": 1.9420151962120025,
"max": 3.78355118707957,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1633.4687263965607,
"min": 236.9258539378643,
"max": 1679.539180636406,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.020278186908884284,
"min": 0.013358998819926962,
"max": 0.020278186908884284,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.04055637381776857,
"min": 0.027825396285334134,
"max": 0.05756960379318722,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.0606825872634848,
"min": 0.02423858071366946,
"max": 0.0650582895303766,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.1213651745269696,
"min": 0.04847716142733892,
"max": 0.1862375574807326,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 4.481498506200002e-06,
"min": 4.481498506200002e-06,
"max": 0.0002953469265510251,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 8.962997012400004e-06,
"min": 8.962997012400004e-06,
"max": 0.00084411796862735,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10149379999999997,
"min": 0.10149379999999997,
"max": 0.19844897500000006,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20298759999999993,
"min": 0.20298759999999993,
"max": 0.5813726499999999,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 8.454062000000004e-05,
"min": 8.454062000000004e-05,
"max": 0.0049226038525000005,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0001690812400000001,
"min": 0.0001690812400000001,
"max": 0.014070495235000001,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1702438312",
"python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.1.1+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1702440924"
},
"total": 2612.603338897,
"count": 1,
"self": 0.7368781799996214,
"children": {
"run_training.setup": {
"total": 0.09116736300006778,
"count": 1,
"self": 0.09116736300006778
},
"TrainerController.start_learning": {
"total": 2611.7752933540005,
"count": 1,
"self": 4.740716538874949,
"children": {
"TrainerController._reset_env": {
"total": 3.977682601999959,
"count": 1,
"self": 3.977682601999959
},
"TrainerController.advance": {
"total": 2602.905934902125,
"count": 230656,
"self": 5.229445637906792,
"children": {
"env_step": {
"total": 2083.4408426331056,
"count": 230656,
"self": 1721.1658823111802,
"children": {
"SubprocessEnvManager._take_step": {
"total": 359.04400718896954,
"count": 230656,
"self": 19.03705946700984,
"children": {
"TorchPolicy.evaluate": {
"total": 340.0069477219597,
"count": 222871,
"self": 340.0069477219597
}
}
},
"workers": {
"total": 3.2309531329558467,
"count": 230656,
"self": 0.0,
"children": {
"worker_root": {
"total": 2603.7090671048027,
"count": 230656,
"is_parallel": true,
"self": 1210.1772545888623,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0009723309999571939,
"count": 1,
"is_parallel": true,
"self": 0.00030612499995186226,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006662060000053316,
"count": 2,
"is_parallel": true,
"self": 0.0006662060000053316
}
}
},
"UnityEnvironment.step": {
"total": 0.03190657900017868,
"count": 1,
"is_parallel": true,
"self": 0.00031564300024911063,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00024819000009301817,
"count": 1,
"is_parallel": true,
"self": 0.00024819000009301817
},
"communicator.exchange": {
"total": 0.030605272000002515,
"count": 1,
"is_parallel": true,
"self": 0.030605272000002515
},
"steps_from_proto": {
"total": 0.0007374739998340374,
"count": 1,
"is_parallel": true,
"self": 0.00022347099979924678,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005140030000347906,
"count": 2,
"is_parallel": true,
"self": 0.0005140030000347906
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1393.5318125159404,
"count": 230655,
"is_parallel": true,
"self": 42.7707668779492,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 91.96311139116688,
"count": 230655,
"is_parallel": true,
"self": 91.96311139116688
},
"communicator.exchange": {
"total": 1161.3889356698687,
"count": 230655,
"is_parallel": true,
"self": 1161.3889356698687
},
"steps_from_proto": {
"total": 97.40899857695558,
"count": 230655,
"is_parallel": true,
"self": 36.31927286580617,
"children": {
"_process_rank_one_or_two_observation": {
"total": 61.08972571114941,
"count": 461310,
"is_parallel": true,
"self": 61.08972571114941
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 514.2356466311126,
"count": 230656,
"self": 7.466302198188259,
"children": {
"process_trajectory": {
"total": 159.28887198292523,
"count": 230656,
"self": 158.02148155892633,
"children": {
"RLTrainer._checkpoint": {
"total": 1.2673904239989042,
"count": 10,
"self": 1.2673904239989042
}
}
},
"_update_policy": {
"total": 347.48047244999907,
"count": 96,
"self": 283.5140256899965,
"children": {
"TorchPPOOptimizer.update": {
"total": 63.966446760002555,
"count": 2880,
"self": 63.966446760002555
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.7620004655327648e-06,
"count": 1,
"self": 1.7620004655327648e-06
},
"TrainerController._save_models": {
"total": 0.15095754899994063,
"count": 1,
"self": 0.0024163070002032327,
"children": {
"RLTrainer._checkpoint": {
"total": 0.1485412419997374,
"count": 1,
"self": 0.1485412419997374
}
}
}
}
}
}
}