Huggy / run_logs /timers.json
NathanS-HuggingFace's picture
Huggy
a605afb
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4059518575668335,
"min": 1.4059518575668335,
"max": 1.4209702014923096,
"count": 28
},
"Huggy.Policy.Entropy.sum": {
"value": 70488.8046875,
"min": 48166.62890625,
"max": 71700.53125,
"count": 28
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 100.39148073022312,
"min": 79.0624,
"max": 101.63406940063092,
"count": 28
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49493.0,
"min": 32218.0,
"max": 49589.0,
"count": 28
},
"Huggy.Step.mean": {
"value": 1999966.0,
"min": 649938.0,
"max": 1999966.0,
"count": 28
},
"Huggy.Step.sum": {
"value": 1999966.0,
"min": 649938.0,
"max": 1999966.0,
"count": 28
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.3824172019958496,
"min": 2.2781317234039307,
"max": 2.46649432182312,
"count": 28
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1174.53173828125,
"min": 721.60302734375,
"max": 1541.5589599609375,
"count": 28
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.708418772380202,
"min": 3.641531162929041,
"max": 4.018090721829843,
"count": 28
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1828.2504547834396,
"min": 1195.6019991636276,
"max": 2428.5858787298203,
"count": 28
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.708418772380202,
"min": 3.641531162929041,
"max": 4.018090721829843,
"count": 28
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1828.2504547834396,
"min": 1195.6019991636276,
"max": 2428.5858787298203,
"count": 28
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.017469912778574324,
"min": 0.014146781922803105,
"max": 0.020454756249900025,
"count": 28
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.052409738335722975,
"min": 0.017522567624109798,
"max": 0.06136426874970008,
"count": 28
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04911581894589795,
"min": 0.04135653575261434,
"max": 0.06160926582912604,
"count": 28
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.14734745683769385,
"min": 0.04135653575261434,
"max": 0.1798206618676583,
"count": 28
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.824998725033331e-06,
"min": 3.824998725033331e-06,
"max": 0.00020429133190289994,
"count": 28
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.1474996175099993e-05,
"min": 1.1474996175099993e-05,
"max": 0.0005943349018884,
"count": 28
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.1012749666666667,
"min": 0.1012749666666667,
"max": 0.16809709999999994,
"count": 28
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.3038249000000001,
"min": 0.16809709999999994,
"max": 0.4981115999999999,
"count": 28
},
"Huggy.Policy.Beta.mean": {
"value": 7.362083666666664e-05,
"min": 7.362083666666664e-05,
"max": 0.0034080452899999994,
"count": 28
},
"Huggy.Policy.Beta.sum": {
"value": 0.00022086250999999991,
"min": 0.00022086250999999991,
"max": 0.009915768839999999,
"count": 28
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 28
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 28
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1678316608",
"python_version": "3.8.16 (default, Feb 14 2023, 21:20:30) \n[GCC 11.3.0]",
"command_line_arguments": "/home/thekeymaster/Server/Projects/Hugging-Face-Course/Unit-1-Bonus-Huggy/Huggy/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --resume --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.19.0",
"end_time_seconds": "1678319203"
},
"total": 2594.804821166,
"count": 1,
"self": 1.223000597999544,
"children": {
"run_training.setup": {
"total": 0.17241783500003294,
"count": 1,
"self": 0.17241783500003294
},
"TrainerController.start_learning": {
"total": 2593.409402733,
"count": 1,
"self": 4.433360074037864,
"children": {
"TrainerController._reset_env": {
"total": 48.10232846600002,
"count": 1,
"self": 48.10232846600002
},
"TrainerController.advance": {
"total": 2537.800454396963,
"count": 162096,
"self": 4.379974758770459,
"children": {
"env_step": {
"total": 1871.7011423120941,
"count": 162096,
"self": 1611.256111434021,
"children": {
"SubprocessEnvManager._take_step": {
"total": 257.3448072810812,
"count": 162096,
"self": 13.216748077183752,
"children": {
"TorchPolicy.evaluate": {
"total": 244.12805920389746,
"count": 154310,
"self": 72.79456038686067,
"children": {
"TorchPolicy.sample_actions": {
"total": 171.3334988170368,
"count": 154310,
"self": 171.3334988170368
}
}
}
}
},
"workers": {
"total": 3.1002235969920093,
"count": 162096,
"self": 0.0,
"children": {
"worker_root": {
"total": 2578.066867394974,
"count": 162096,
"is_parallel": true,
"self": 1273.0393655169635,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0016456859999607332,
"count": 1,
"is_parallel": true,
"self": 0.00019698999994943733,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014486960000112958,
"count": 2,
"is_parallel": true,
"self": 0.0014486960000112958
}
}
},
"UnityEnvironment.step": {
"total": 0.02414632700003949,
"count": 1,
"is_parallel": true,
"self": 0.0002958750001198496,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00027042699991852714,
"count": 1,
"is_parallel": true,
"self": 0.00027042699991852714
},
"communicator.exchange": {
"total": 0.022637408000036885,
"count": 1,
"is_parallel": true,
"self": 0.022637408000036885
},
"steps_from_proto": {
"total": 0.0009426169999642298,
"count": 1,
"is_parallel": true,
"self": 0.00021727699993334681,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.000725340000030883,
"count": 2,
"is_parallel": true,
"self": 0.000725340000030883
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1305.0275018780105,
"count": 162095,
"is_parallel": true,
"self": 48.64382054904968,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 116.33505832498281,
"count": 162095,
"is_parallel": true,
"self": 116.33505832498281
},
"communicator.exchange": {
"total": 969.4295607468749,
"count": 162095,
"is_parallel": true,
"self": 969.4295607468749
},
"steps_from_proto": {
"total": 170.619062257103,
"count": 162095,
"is_parallel": true,
"self": 52.02597234808684,
"children": {
"_process_rank_one_or_two_observation": {
"total": 118.59308990901616,
"count": 324190,
"is_parallel": true,
"self": 118.59308990901616
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 661.7193373260983,
"count": 162096,
"self": 6.161762898047755,
"children": {
"process_trajectory": {
"total": 190.16286054004968,
"count": 162096,
"self": 171.06188684304948,
"children": {
"RLTrainer._checkpoint": {
"total": 19.1009736970002,
"count": 7,
"self": 19.1009736970002
}
}
},
"_update_policy": {
"total": 465.39471388800087,
"count": 67,
"self": 427.02999190299624,
"children": {
"TorchPPOOptimizer.update": {
"total": 38.36472198500462,
"count": 2010,
"self": 38.36472198500462
}
}
}
}
}
}
},
"trainer_threads": {
"total": 2.7149999368702993e-06,
"count": 1,
"self": 2.7149999368702993e-06
},
"TrainerController._save_models": {
"total": 3.0732570809996105,
"count": 1,
"self": 0.2834825939999064,
"children": {
"RLTrainer._checkpoint": {
"total": 2.789774486999704,
"count": 1,
"self": 2.789774486999704
}
}
}
}
}
}
}