ppo-Pyramids / run_logs /timers.json
rvukasin's picture
First Push
54236c8 verified
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.3191472291946411,
"min": 0.3057059645652771,
"max": 1.3908493518829346,
"count": 33
},
"Pyramids.Policy.Entropy.sum": {
"value": 9615.267578125,
"min": 9127.1572265625,
"max": 42192.8046875,
"count": 33
},
"Pyramids.Step.mean": {
"value": 989979.0,
"min": 29952.0,
"max": 989979.0,
"count": 33
},
"Pyramids.Step.sum": {
"value": 989979.0,
"min": 29952.0,
"max": 989979.0,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.5525098443031311,
"min": -0.09002811461687088,
"max": 0.6763673424720764,
"count": 33
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 159.12283325195312,
"min": -21.78680419921875,
"max": 191.41195678710938,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.022005505859851837,
"min": -0.004445615224540234,
"max": 0.47642505168914795,
"count": 33
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 6.33758544921875,
"min": -1.2581090927124023,
"max": 112.91273498535156,
"count": 33
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.0682436155994747,
"min": 0.06557597851900662,
"max": 0.07430821960790131,
"count": 33
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 0.9554106183926457,
"min": 0.49749007896911934,
"max": 1.0575463627014243,
"count": 33
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.020275005826754112,
"min": 0.0005918263905836543,
"max": 0.020275005826754112,
"count": 33
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.28385008157455754,
"min": 0.004142784734085581,
"max": 0.28385008157455754,
"count": 33
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.53061177554286e-06,
"min": 7.53061177554286e-06,
"max": 0.00029515063018788575,
"count": 33
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.00010542856485760003,
"min": 0.00010542856485760003,
"max": 0.0033835298721568,
"count": 33
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.10251017142857144,
"min": 0.10251017142857144,
"max": 0.19838354285714285,
"count": 33
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.4351424000000002,
"min": 1.3691136000000002,
"max": 2.6173256,
"count": 33
},
"Pyramids.Policy.Beta.mean": {
"value": 0.0002607661257142858,
"min": 0.0002607661257142858,
"max": 0.00983851593142857,
"count": 33
},
"Pyramids.Policy.Beta.sum": {
"value": 0.003650725760000001,
"min": 0.003650725760000001,
"max": 0.11281153567999999,
"count": 33
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.01308396551758051,
"min": 0.01308396551758051,
"max": 0.46969836950302124,
"count": 33
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.18317551910877228,
"min": 0.18317551910877228,
"max": 3.287888526916504,
"count": 33
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 326.7448979591837,
"min": 284.1730769230769,
"max": 999.0,
"count": 33
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 32021.0,
"min": 15984.0,
"max": 34282.0,
"count": 33
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.6114484676809022,
"min": -1.0000000521540642,
"max": 1.705459985435009,
"count": 33
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 159.53339830040932,
"min": -32.000001668930054,
"max": 176.44499846547842,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.6114484676809022,
"min": -1.0000000521540642,
"max": 1.705459985435009,
"count": 33
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 159.53339830040932,
"min": -32.000001668930054,
"max": 176.44499846547842,
"count": 33
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.04434726943174904,
"min": 0.03904319512282829,
"max": 8.690058313310146,
"count": 33
},
"Pyramids.Policy.RndReward.sum": {
"value": 4.390379673743155,
"min": 3.8511583506769966,
"max": 139.04093301296234,
"count": 33
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 33
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1715733974",
"python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.2.1+cu121",
"numpy_version": "1.23.5",
"end_time_seconds": "1715737400"
},
"total": 3425.8004379920003,
"count": 1,
"self": 0.754805729000509,
"children": {
"run_training.setup": {
"total": 0.09334687099999428,
"count": 1,
"self": 0.09334687099999428
},
"TrainerController.start_learning": {
"total": 3424.952285392,
"count": 1,
"self": 2.165698326058191,
"children": {
"TrainerController._reset_env": {
"total": 3.861868382999944,
"count": 1,
"self": 3.861868382999944
},
"TrainerController.advance": {
"total": 3418.7870484169416,
"count": 64015,
"self": 2.4450644949261005,
"children": {
"env_step": {
"total": 2335.8827663210245,
"count": 64015,
"self": 2176.385351018057,
"children": {
"SubprocessEnvManager._take_step": {
"total": 158.09781304698936,
"count": 64015,
"self": 7.238873927956547,
"children": {
"TorchPolicy.evaluate": {
"total": 150.8589391190328,
"count": 62570,
"self": 150.8589391190328
}
}
},
"workers": {
"total": 1.3996022559781522,
"count": 64015,
"self": 0.0,
"children": {
"worker_root": {
"total": 3418.0171755469846,
"count": 64015,
"is_parallel": true,
"self": 1428.4135461949995,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.006838068000035946,
"count": 1,
"is_parallel": true,
"self": 0.004324192000012772,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.002513876000023174,
"count": 8,
"is_parallel": true,
"self": 0.002513876000023174
}
}
},
"UnityEnvironment.step": {
"total": 0.06349043700004131,
"count": 1,
"is_parallel": true,
"self": 0.000732551999988118,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.0005718710000337524,
"count": 1,
"is_parallel": true,
"self": 0.0005718710000337524
},
"communicator.exchange": {
"total": 0.060193959000002906,
"count": 1,
"is_parallel": true,
"self": 0.060193959000002906
},
"steps_from_proto": {
"total": 0.001992055000016535,
"count": 1,
"is_parallel": true,
"self": 0.000408656999923096,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001583398000093439,
"count": 8,
"is_parallel": true,
"self": 0.001583398000093439
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1989.603629351985,
"count": 64014,
"is_parallel": true,
"self": 50.72619947687076,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 28.634496125036947,
"count": 64014,
"is_parallel": true,
"self": 28.634496125036947
},
"communicator.exchange": {
"total": 1778.3041234640268,
"count": 64014,
"is_parallel": true,
"self": 1778.3041234640268
},
"steps_from_proto": {
"total": 131.93881028605045,
"count": 64014,
"is_parallel": true,
"self": 28.05096226509022,
"children": {
"_process_rank_one_or_two_observation": {
"total": 103.88784802096023,
"count": 512112,
"is_parallel": true,
"self": 103.88784802096023
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 1080.459217600991,
"count": 64015,
"self": 4.469252293973113,
"children": {
"process_trajectory": {
"total": 165.56075755001746,
"count": 64015,
"self": 165.23695537501737,
"children": {
"RLTrainer._checkpoint": {
"total": 0.32380217500008257,
"count": 2,
"self": 0.32380217500008257
}
}
},
"_update_policy": {
"total": 910.4292077570005,
"count": 449,
"self": 374.89932245699515,
"children": {
"TorchPPOOptimizer.update": {
"total": 535.5298853000054,
"count": 22833,
"self": 535.5298853000054
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.568999778100988e-06,
"count": 1,
"self": 1.568999778100988e-06
},
"TrainerController._save_models": {
"total": 0.13766869700020834,
"count": 1,
"self": 0.002987692000260722,
"children": {
"RLTrainer._checkpoint": {
"total": 0.13468100499994762,
"count": 1,
"self": 0.13468100499994762
}
}
}
}
}
}
}