| { | |
| "name": "root", | |
| "gauges": { | |
| "Pyramids.Policy.Entropy.mean": { | |
| "value": 0.285113662481308, | |
| "min": 0.28019753098487854, | |
| "max": 1.3373417854309082, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Entropy.sum": { | |
| "value": 8553.41015625, | |
| "min": 8320.74609375, | |
| "max": 40569.6015625, | |
| "count": 33 | |
| }, | |
| "Pyramids.Step.mean": { | |
| "value": 989968.0, | |
| "min": 29977.0, | |
| "max": 989968.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Step.sum": { | |
| "value": 989968.0, | |
| "min": 29977.0, | |
| "max": 989968.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicValueEstimate.mean": { | |
| "value": 0.742624044418335, | |
| "min": -0.07317552715539932, | |
| "max": 0.7502455711364746, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicValueEstimate.sum": { | |
| "value": 217.58883666992188, | |
| "min": -17.708477020263672, | |
| "max": 222.07269287109375, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndValueEstimate.mean": { | |
| "value": 0.01739008165895939, | |
| "min": 0.00175029959063977, | |
| "max": 0.38777002692222595, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndValueEstimate.sum": { | |
| "value": 5.095293998718262, | |
| "min": 0.5075868964195251, | |
| "max": 93.06480407714844, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.PolicyLoss.mean": { | |
| "value": 0.06926023140355635, | |
| "min": 0.06414183754840795, | |
| "max": 0.07220410677733038, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.PolicyLoss.sum": { | |
| "value": 0.9696432396497888, | |
| "min": 0.5552216753219285, | |
| "max": 1.0607346540200524, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.ValueLoss.mean": { | |
| "value": 0.01638117568689354, | |
| "min": 0.001232118952374212, | |
| "max": 0.017923484782971042, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.ValueLoss.sum": { | |
| "value": 0.22933645961650956, | |
| "min": 0.012321189523742119, | |
| "max": 0.26189641419235465, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.LearningRate.mean": { | |
| "value": 7.532154632171433e-06, | |
| "min": 7.532154632171433e-06, | |
| "max": 0.00029522051409316247, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.LearningRate.sum": { | |
| "value": 0.00010545016485040006, | |
| "min": 0.00010545016485040006, | |
| "max": 0.0036337453887515998, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Epsilon.mean": { | |
| "value": 0.10251068571428572, | |
| "min": 0.10251068571428572, | |
| "max": 0.1984068375, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Epsilon.sum": { | |
| "value": 1.4351496000000001, | |
| "min": 1.4351496000000001, | |
| "max": 2.6173617, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Beta.mean": { | |
| "value": 0.000260817502857143, | |
| "min": 0.000260817502857143, | |
| "max": 0.00984084306625, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.Beta.sum": { | |
| "value": 0.003651445040000002, | |
| "min": 0.003651445040000002, | |
| "max": 0.12114371516, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.RNDLoss.mean": { | |
| "value": 0.012743408791720867, | |
| "min": 0.012743408791720867, | |
| "max": 0.6665088534355164, | |
| "count": 33 | |
| }, | |
| "Pyramids.Losses.RNDLoss.sum": { | |
| "value": 0.1784077286720276, | |
| "min": 0.1784077286720276, | |
| "max": 5.332070827484131, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.EpisodeLength.mean": { | |
| "value": 248.86614173228347, | |
| "min": 247.01612903225808, | |
| "max": 999.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.EpisodeLength.sum": { | |
| "value": 31606.0, | |
| "min": 16776.0, | |
| "max": 33431.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.CumulativeReward.mean": { | |
| "value": 1.7346190318820023, | |
| "min": -1.0000000521540642, | |
| "max": 1.7472543688981157, | |
| "count": 33 | |
| }, | |
| "Pyramids.Environment.CumulativeReward.sum": { | |
| "value": 218.56199801713228, | |
| "min": -31.00000161677599, | |
| "max": 218.56199801713228, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicReward.mean": { | |
| "value": 1.7346190318820023, | |
| "min": -1.0000000521540642, | |
| "max": 1.7472543688981157, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.ExtrinsicReward.sum": { | |
| "value": 218.56199801713228, | |
| "min": -31.00000161677599, | |
| "max": 218.56199801713228, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndReward.mean": { | |
| "value": 0.0327601416957184, | |
| "min": 0.0327601416957184, | |
| "max": 14.381607474649654, | |
| "count": 33 | |
| }, | |
| "Pyramids.Policy.RndReward.sum": { | |
| "value": 4.127777853660518, | |
| "min": 3.966959278041031, | |
| "max": 244.4873270690441, | |
| "count": 33 | |
| }, | |
| "Pyramids.IsTraining.mean": { | |
| "value": 1.0, | |
| "min": 1.0, | |
| "max": 1.0, | |
| "count": 33 | |
| }, | |
| "Pyramids.IsTraining.sum": { | |
| "value": 1.0, | |
| "min": 1.0, | |
| "max": 1.0, | |
| "count": 33 | |
| } | |
| }, | |
| "metadata": { | |
| "timer_format_version": "0.1.0", | |
| "start_time_seconds": "1756378652", | |
| "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]", | |
| "command_line_arguments": "/usr/local/bin/mlagents-learn ../config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", | |
| "mlagents_version": "1.1.0", | |
| "mlagents_envs_version": "1.1.0", | |
| "communication_protocol_version": "1.5.0", | |
| "pytorch_version": "2.8.0+cu128", | |
| "numpy_version": "1.23.5", | |
| "end_time_seconds": "1756380981" | |
| }, | |
| "total": 2328.278660369, | |
| "count": 1, | |
| "self": 0.4762808940004106, | |
| "children": { | |
| "run_training.setup": { | |
| "total": 0.023940635999679216, | |
| "count": 1, | |
| "self": 0.023940635999679216 | |
| }, | |
| "TrainerController.start_learning": { | |
| "total": 2327.778438839, | |
| "count": 1, | |
| "self": 1.339578540047114, | |
| "children": { | |
| "TrainerController._reset_env": { | |
| "total": 2.0610387349997836, | |
| "count": 1, | |
| "self": 2.0610387349997836 | |
| }, | |
| "TrainerController.advance": { | |
| "total": 2324.3017410129532, | |
| "count": 64495, | |
| "self": 1.3548746797541753, | |
| "children": { | |
| "env_step": { | |
| "total": 1651.3001822041106, | |
| "count": 64495, | |
| "self": 1501.537429327977, | |
| "children": { | |
| "SubprocessEnvManager._take_step": { | |
| "total": 148.98576050510655, | |
| "count": 64495, | |
| "self": 4.51671116405214, | |
| "children": { | |
| "TorchPolicy.evaluate": { | |
| "total": 144.46904934105441, | |
| "count": 62558, | |
| "self": 144.46904934105441 | |
| } | |
| } | |
| }, | |
| "workers": { | |
| "total": 0.7769923710270632, | |
| "count": 64495, | |
| "self": 0.0, | |
| "children": { | |
| "worker_root": { | |
| "total": 2323.150059353868, | |
| "count": 64495, | |
| "is_parallel": true, | |
| "self": 933.2224880878898, | |
| "children": { | |
| "run_training.setup": { | |
| "total": 0.0, | |
| "count": 0, | |
| "is_parallel": true, | |
| "self": 0.0, | |
| "children": { | |
| "steps_from_proto": { | |
| "total": 0.002264876000026561, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.0007253339999806485, | |
| "children": { | |
| "_process_rank_one_or_two_observation": { | |
| "total": 0.0015395420000459126, | |
| "count": 8, | |
| "is_parallel": true, | |
| "self": 0.0015395420000459126 | |
| } | |
| } | |
| }, | |
| "UnityEnvironment.step": { | |
| "total": 0.04912001799993959, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.0005158320000191452, | |
| "children": { | |
| "UnityEnvironment._generate_step_input": { | |
| "total": 0.0004982979999113013, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.0004982979999113013 | |
| }, | |
| "communicator.exchange": { | |
| "total": 0.04642198200008352, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.04642198200008352 | |
| }, | |
| "steps_from_proto": { | |
| "total": 0.0016839059999256278, | |
| "count": 1, | |
| "is_parallel": true, | |
| "self": 0.0003655319997051265, | |
| "children": { | |
| "_process_rank_one_or_two_observation": { | |
| "total": 0.0013183740002205013, | |
| "count": 8, | |
| "is_parallel": true, | |
| "self": 0.0013183740002205013 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "UnityEnvironment.step": { | |
| "total": 1389.9275712659783, | |
| "count": 64494, | |
| "is_parallel": true, | |
| "self": 31.355824202882104, | |
| "children": { | |
| "UnityEnvironment._generate_step_input": { | |
| "total": 23.015810393014817, | |
| "count": 64494, | |
| "is_parallel": true, | |
| "self": 23.015810393014817 | |
| }, | |
| "communicator.exchange": { | |
| "total": 1240.005566531072, | |
| "count": 64494, | |
| "is_parallel": true, | |
| "self": 1240.005566531072 | |
| }, | |
| "steps_from_proto": { | |
| "total": 95.55037013900937, | |
| "count": 64494, | |
| "is_parallel": true, | |
| "self": 19.16597075935124, | |
| "children": { | |
| "_process_rank_one_or_two_observation": { | |
| "total": 76.38439937965813, | |
| "count": 515952, | |
| "is_parallel": true, | |
| "self": 76.38439937965813 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "trainer_advance": { | |
| "total": 671.6466841290885, | |
| "count": 64495, | |
| "self": 2.6764614840049035, | |
| "children": { | |
| "process_trajectory": { | |
| "total": 128.1584318350774, | |
| "count": 64495, | |
| "self": 127.9594408980779, | |
| "children": { | |
| "RLTrainer._checkpoint": { | |
| "total": 0.19899093699950754, | |
| "count": 2, | |
| "self": 0.19899093699950754 | |
| } | |
| } | |
| }, | |
| "_update_policy": { | |
| "total": 540.8117908100062, | |
| "count": 457, | |
| "self": 299.36157837494466, | |
| "children": { | |
| "TorchPPOOptimizer.update": { | |
| "total": 241.45021243506153, | |
| "count": 22803, | |
| "self": 241.45021243506153 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| }, | |
| "trainer_threads": { | |
| "total": 8.240003808168694e-07, | |
| "count": 1, | |
| "self": 8.240003808168694e-07 | |
| }, | |
| "TrainerController._save_models": { | |
| "total": 0.0760797269995237, | |
| "count": 1, | |
| "self": 0.0014419099989027018, | |
| "children": { | |
| "RLTrainer._checkpoint": { | |
| "total": 0.074637817000621, | |
| "count": 1, | |
| "self": 0.074637817000621 | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } | |
| } |