ppo-Huggy / run_logs /timers.json
amarmol's picture
Huggy
8d2b25a
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4058184623718262,
"min": 1.4058184623718262,
"max": 1.4290870428085327,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70632.5390625,
"min": 68658.9296875,
"max": 75939.203125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 105.94017094017094,
"min": 88.89818181818181,
"max": 383.14503816793894,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49580.0,
"min": 48894.0,
"max": 50339.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999977.0,
"min": 49635.0,
"max": 1999977.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999977.0,
"min": 49635.0,
"max": 1999977.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.37117075920105,
"min": -0.04457467794418335,
"max": 2.445763349533081,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1109.7078857421875,
"min": -5.794708251953125,
"max": 1330.786376953125,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.692634591816837,
"min": 1.7467674154501696,
"max": 3.8797175172116427,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1728.1529889702797,
"min": 227.07976400852203,
"max": 2092.3500990867615,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.692634591816837,
"min": 1.7467674154501696,
"max": 3.8797175172116427,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1728.1529889702797,
"min": 227.07976400852203,
"max": 2092.3500990867615,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.01710728253496604,
"min": 0.01379818477629063,
"max": 0.01964733062874681,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.05132184760489812,
"min": 0.02759636955258126,
"max": 0.05555038007829959,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.04693340725368924,
"min": 0.022942741525669892,
"max": 0.05569313185082542,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.1408002217610677,
"min": 0.045885483051339784,
"max": 0.16707939555247625,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.347098884333334e-06,
"min": 3.347098884333334e-06,
"max": 0.0002953068015644,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.0041296653000002e-05,
"min": 1.0041296653000002e-05,
"max": 0.0008440389186537001,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10111566666666667,
"min": 0.10111566666666667,
"max": 0.19843560000000007,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30334700000000003,
"min": 0.20740139999999996,
"max": 0.5813463,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.567176666666666e-05,
"min": 6.567176666666666e-05,
"max": 0.00492193644,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0001970153,
"min": 0.0001970153,
"max": 0.01406918037,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1676117667",
"python_version": "3.8.10 (default, Nov 14 2022, 12:59:47) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1676120092"
},
"total": 2425.6845173379998,
"count": 1,
"self": 0.7426066230000288,
"children": {
"run_training.setup": {
"total": 0.1949968080000417,
"count": 1,
"self": 0.1949968080000417
},
"TrainerController.start_learning": {
"total": 2424.746913907,
"count": 1,
"self": 4.344736706982985,
"children": {
"TrainerController._reset_env": {
"total": 10.82148600000005,
"count": 1,
"self": 10.82148600000005
},
"TrainerController.advance": {
"total": 2409.400278068017,
"count": 231566,
"self": 4.5924286511422,
"children": {
"env_step": {
"total": 1877.671654658011,
"count": 231566,
"self": 1565.4042092851741,
"children": {
"SubprocessEnvManager._take_step": {
"total": 309.44215483997186,
"count": 231566,
"self": 16.135154984104247,
"children": {
"TorchPolicy.evaluate": {
"total": 293.3069998558676,
"count": 223011,
"self": 73.40504580277684,
"children": {
"TorchPolicy.sample_actions": {
"total": 219.90195405309078,
"count": 223011,
"self": 219.90195405309078
}
}
}
}
},
"workers": {
"total": 2.825290532864983,
"count": 231566,
"self": 0.0,
"children": {
"worker_root": {
"total": 2415.7817741148933,
"count": 231566,
"is_parallel": true,
"self": 1146.3215572759696,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.002436761000126353,
"count": 1,
"is_parallel": true,
"self": 0.0003431870004533266,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0020935739996730263,
"count": 2,
"is_parallel": true,
"self": 0.0020935739996730263
}
}
},
"UnityEnvironment.step": {
"total": 0.029797623999911593,
"count": 1,
"is_parallel": true,
"self": 0.0003065350001634215,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00021883699992031325,
"count": 1,
"is_parallel": true,
"self": 0.00021883699992031325
},
"communicator.exchange": {
"total": 0.028522949000034714,
"count": 1,
"is_parallel": true,
"self": 0.028522949000034714
},
"steps_from_proto": {
"total": 0.0007493029997931444,
"count": 1,
"is_parallel": true,
"self": 0.000260231000083877,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0004890719997092674,
"count": 2,
"is_parallel": true,
"self": 0.0004890719997092674
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1269.4602168389238,
"count": 231565,
"is_parallel": true,
"self": 39.64054002514149,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 80.47207621294274,
"count": 231565,
"is_parallel": true,
"self": 80.47207621294274
},
"communicator.exchange": {
"total": 1054.9072886059744,
"count": 231565,
"is_parallel": true,
"self": 1054.9072886059744
},
"steps_from_proto": {
"total": 94.44031199486517,
"count": 231565,
"is_parallel": true,
"self": 38.38573004109139,
"children": {
"_process_rank_one_or_two_observation": {
"total": 56.05458195377378,
"count": 463130,
"is_parallel": true,
"self": 56.05458195377378
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 527.1361947588639,
"count": 231566,
"self": 6.823154037911081,
"children": {
"process_trajectory": {
"total": 161.26529488895403,
"count": 231566,
"self": 160.05118810095473,
"children": {
"RLTrainer._checkpoint": {
"total": 1.2141067879992988,
"count": 10,
"self": 1.2141067879992988
}
}
},
"_update_policy": {
"total": 359.0477458319988,
"count": 97,
"self": 300.2139922780134,
"children": {
"TorchPPOOptimizer.update": {
"total": 58.83375355398539,
"count": 2910,
"self": 58.83375355398539
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.4469997040578164e-06,
"count": 1,
"self": 1.4469997040578164e-06
},
"TrainerController._save_models": {
"total": 0.1804116849998536,
"count": 1,
"self": 0.0028779539998140535,
"children": {
"RLTrainer._checkpoint": {
"total": 0.17753373100003955,
"count": 1,
"self": 0.17753373100003955
}
}
}
}
}
}
}