ppo-Huggy / run_logs /timers.json
Winmodel's picture
Huggy
9d1f49f
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4063383340835571,
"min": 1.4063383340835571,
"max": 1.4274674654006958,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70115.8125,
"min": 68829.8125,
"max": 76945.5390625,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 113.33638443935926,
"min": 103.2041237113402,
"max": 393.244094488189,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49528.0,
"min": 49046.0,
"max": 50176.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999601.0,
"min": 49908.0,
"max": 1999601.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999601.0,
"min": 49908.0,
"max": 1999601.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.280457019805908,
"min": 0.05170527473092079,
"max": 2.3714523315429688,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 996.5597534179688,
"min": 6.514864444732666,
"max": 1126.9261474609375,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.4671294563695003,
"min": 1.725784605930722,
"max": 3.7377596497535706,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1515.1355724334717,
"min": 217.44886034727097,
"max": 1778.473171889782,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.4671294563695003,
"min": 1.725784605930722,
"max": 3.7377596497535706,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1515.1355724334717,
"min": 217.44886034727097,
"max": 1778.473171889782,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.013718903138457486,
"min": 0.01359782126164646,
"max": 0.019467268939479258,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.027437806276914972,
"min": 0.02719564252329292,
"max": 0.05840180681843777,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.05024488866329194,
"min": 0.02394623619814714,
"max": 0.060948111489415174,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.10048977732658387,
"min": 0.04789247239629428,
"max": 0.17487719717125097,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 4.461248512950004e-06,
"min": 4.461248512950004e-06,
"max": 0.00029534992655002494,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 8.922497025900007e-06,
"min": 8.922497025900007e-06,
"max": 0.000844233018589,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10148705000000001,
"min": 0.10148705000000001,
"max": 0.198449975,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.20297410000000002,
"min": 0.20297410000000002,
"max": 0.5814110000000001,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 8.420379500000007e-05,
"min": 8.420379500000007e-05,
"max": 0.004922653752500001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00016840759000000014,
"min": 0.00016840759000000014,
"max": 0.0140724089,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1688876345",
"python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1688878972"
},
"total": 2627.368784034,
"count": 1,
"self": 0.44872559099985665,
"children": {
"run_training.setup": {
"total": 0.04266644799997721,
"count": 1,
"self": 0.04266644799997721
},
"TrainerController.start_learning": {
"total": 2626.877391995,
"count": 1,
"self": 4.710315692051154,
"children": {
"TrainerController._reset_env": {
"total": 4.635545810999986,
"count": 1,
"self": 4.635545810999986
},
"TrainerController.advance": {
"total": 2617.4042393939485,
"count": 230604,
"self": 5.07132117209585,
"children": {
"env_step": {
"total": 2048.0084776488593,
"count": 230604,
"self": 1723.4556343177996,
"children": {
"SubprocessEnvManager._take_step": {
"total": 321.45976891205277,
"count": 230604,
"self": 18.50148270404071,
"children": {
"TorchPolicy.evaluate": {
"total": 302.95828620801205,
"count": 223028,
"self": 302.95828620801205
}
}
},
"workers": {
"total": 3.0930744190067685,
"count": 230604,
"self": 0.0,
"children": {
"worker_root": {
"total": 2618.7641466749387,
"count": 230604,
"is_parallel": true,
"self": 1214.2767743089412,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0011336989999790603,
"count": 1,
"is_parallel": true,
"self": 0.00034282899997606364,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0007908700000029967,
"count": 2,
"is_parallel": true,
"self": 0.0007908700000029967
}
}
},
"UnityEnvironment.step": {
"total": 0.07920485400001098,
"count": 1,
"is_parallel": true,
"self": 0.000362783000014133,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00025121499999158914,
"count": 1,
"is_parallel": true,
"self": 0.00025121499999158914
},
"communicator.exchange": {
"total": 0.07777482900002042,
"count": 1,
"is_parallel": true,
"self": 0.07777482900002042
},
"steps_from_proto": {
"total": 0.0008160269999848424,
"count": 1,
"is_parallel": true,
"self": 0.000212560999955258,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006034660000295844,
"count": 2,
"is_parallel": true,
"self": 0.0006034660000295844
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1404.4873723659975,
"count": 230603,
"is_parallel": true,
"self": 41.70111107501384,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 89.6970849329553,
"count": 230603,
"is_parallel": true,
"self": 89.6970849329553
},
"communicator.exchange": {
"total": 1168.757293306043,
"count": 230603,
"is_parallel": true,
"self": 1168.757293306043
},
"steps_from_proto": {
"total": 104.33188305198536,
"count": 230603,
"is_parallel": true,
"self": 39.000494890083814,
"children": {
"_process_rank_one_or_two_observation": {
"total": 65.33138816190154,
"count": 461206,
"is_parallel": true,
"self": 65.33138816190154
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 564.3244405729935,
"count": 230604,
"self": 6.974379765095819,
"children": {
"process_trajectory": {
"total": 144.72723658089728,
"count": 230604,
"self": 143.02619409889678,
"children": {
"RLTrainer._checkpoint": {
"total": 1.7010424820005028,
"count": 10,
"self": 1.7010424820005028
}
}
},
"_update_policy": {
"total": 412.62282422700036,
"count": 96,
"self": 349.7311670670134,
"children": {
"TorchPPOOptimizer.update": {
"total": 62.89165715998695,
"count": 2880,
"self": 62.89165715998695
}
}
}
}
}
}
},
"trainer_threads": {
"total": 9.4600000011269e-07,
"count": 1,
"self": 9.4600000011269e-07
},
"TrainerController._save_models": {
"total": 0.12729015200011418,
"count": 1,
"self": 0.002255745000184106,
"children": {
"RLTrainer._checkpoint": {
"total": 0.12503440699993007,
"count": 1,
"self": 0.12503440699993007
}
}
}
}
}
}
}