{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4006353616714478, "min": 1.4006353616714478, "max": 1.426529884338379, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 70044.375, "min": 66284.75, "max": 78918.109375, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 78.86871961102106, "min": 73.21216617210682, "max": 422.33613445378154, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 48662.0, "min": 48662.0, "max": 50258.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999962.0, "min": 49654.0, "max": 1999962.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999962.0, "min": 49654.0, "max": 1999962.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.5036091804504395, "min": 0.027692269533872604, "max": 2.5407638549804688, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1544.726806640625, "min": 3.2676877975463867, "max": 1671.862060546875, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.8966730767362896, "min": 1.7619207895408242, "max": 4.022127675334203, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2404.2472883462906, "min": 207.90665316581726, "max": 2657.6173388957977, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.8966730767362896, "min": 1.7619207895408242, "max": 4.022127675334203, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2404.2472883462906, "min": 207.90665316581726, "max": 2657.6173388957977, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.01501045446352186, "min": 0.014234450271275515, "max": 0.019678100145635348, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.04503136339056558, "min": 0.02846890054255103, "max": 0.058819624052072556, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.05640256123410331, "min": 0.0228787323149542, "max": 0.06099907575796047, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.16920768370230993, "min": 0.0457574646299084, "max": 0.1791786236067613, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 4.013348662250002e-06, "min": 4.013348662250002e-06, "max": 0.00029538562653812493, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.2040045986750005e-05, "min": 1.2040045986750005e-05, "max": 0.0008445489184837, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10133774999999999, "min": 0.10133774999999999, "max": 0.198461875, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30401324999999996, "min": 0.2078612, "max": 0.5815163, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 7.675372500000003e-05, "min": 7.675372500000003e-05, "max": 0.004923247562500002, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.0002302611750000001, "min": 0.0002302611750000001, "max": 0.01407766337, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1694269078", "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Doggy.yaml --env=./trained-envs-executables/linux/Doggy/Doggy/Doggy.x86_64 --run-id=Doggy9 --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1694273278" }, "total": 4200.038466457999, "count": 1, "self": 0.37259262399857107, "children": { "run_training.setup": { "total": 0.0354097740000725, "count": 1, "self": 0.0354097740000725 }, "TrainerController.start_learning": { "total": 4199.63046406, "count": 1, "self": 5.756110109900874, "children": { "TrainerController._reset_env": { "total": 1.0806191629999375, "count": 1, "self": 1.0806191629999375 }, "TrainerController.advance": { "total": 4192.707246960101, "count": 233350, "self": 5.6354263535668, "children": { "env_step": { "total": 1863.9672194068344, "count": 233350, "self": 1524.0976070025408, "children": { "SubprocessEnvManager._take_step": { "total": 336.24224466556916, "count": 233350, "self": 16.735262777801836, "children": { "TorchPolicy.evaluate": { "total": 319.5069818877673, "count": 223000, "self": 319.5069818877673 } } }, "workers": { "total": 3.627367738724388, "count": 233350, "self": 0.0, "children": { "worker_root": { "total": 4190.9048125101335, "count": 233350, "is_parallel": true, "self": 2992.3644590519652, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0005514399999810848, "count": 1, "is_parallel": true, "self": 0.0001406499995937338, "children": { "_process_rank_one_or_two_observation": { "total": 0.00041079000038735103, "count": 2, "is_parallel": true, "self": 0.00041079000038735103 } } }, "UnityEnvironment.step": { "total": 0.03259481500026595, "count": 1, "is_parallel": true, "self": 0.0002628699994602357, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0001746200005072751, "count": 1, "is_parallel": true, "self": 0.0001746200005072751 }, "communicator.exchange": { "total": 0.0315045450006437, "count": 1, "is_parallel": true, "self": 0.0315045450006437 }, "steps_from_proto": { "total": 0.0006527799996547401, "count": 1, "is_parallel": true, "self": 0.00018219000048702583, "children": { "_process_rank_one_or_two_observation": { "total": 0.00047058999916771427, "count": 2, "is_parallel": true, "self": 0.00047058999916771427 } } } } } } }, "UnityEnvironment.step": { "total": 1198.5403534581683, "count": 233349, "is_parallel": true, "self": 37.48062159698566, "children": { "UnityEnvironment._generate_step_input": { "total": 58.51944908810492, "count": 233349, "is_parallel": true, "self": 58.51944908810492 }, "communicator.exchange": { "total": 1009.4007290808431, "count": 233349, "is_parallel": true, "self": 1009.4007290808431 }, "steps_from_proto": { "total": 93.13955369223459, "count": 233349, "is_parallel": true, "self": 33.0212697277002, "children": { "_process_rank_one_or_two_observation": { "total": 60.118283964534385, "count": 466698, "is_parallel": true, "self": 60.118283964534385 } } } } } } } } } } }, "trainer_advance": { "total": 2323.1046011996996, "count": 233350, "self": 8.521025031101999, "children": { "process_trajectory": { "total": 231.67016222059738, "count": 233350, "self": 230.6670259675975, "children": { "RLTrainer._checkpoint": { "total": 1.003136252999866, "count": 10, "self": 1.003136252999866 } } }, "_update_policy": { "total": 2082.913413948, "count": 97, "self": 276.4319280220552, "children": { "TorchPPOOptimizer.update": { "total": 1806.481485925945, "count": 2910, "self": 1806.481485925945 } } } } } } }, "trainer_threads": { "total": 1.2700002116616815e-06, "count": 1, "self": 1.2700002116616815e-06 }, "TrainerController._save_models": { "total": 0.08648655699835217, "count": 1, "self": 0.0014056799991521984, "children": { "RLTrainer._checkpoint": { "total": 0.08508087699919997, "count": 1, "self": 0.08508087699919997 } } } } } } }