{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4077200889587402, "min": 1.4077200889587402, "max": 1.4286668300628662, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 71233.453125, "min": 69112.6953125, "max": 76022.734375, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 76.12345679012346, "min": 73.0727002967359, "max": 381.90909090909093, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49328.0, "min": 49002.0, "max": 50412.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999961.0, "min": 49794.0, "max": 1999961.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999961.0, "min": 49794.0, "max": 1999961.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4532899856567383, "min": 0.09209198504686356, "max": 2.5125558376312256, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1589.73193359375, "min": 12.06404972076416, "max": 1675.418212890625, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.7230075182002267, "min": 1.7334371086295324, "max": 4.056929603615464, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2412.508871793747, "min": 227.08026123046875, "max": 2660.16019821167, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.7230075182002267, "min": 1.7334371086295324, "max": 4.056929603615464, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2412.508871793747, "min": 227.08026123046875, "max": 2660.16019821167, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.018714435778545323, "min": 0.014477003315793505, "max": 0.02151310969719715, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.05614330733563597, "min": 0.029132395891550308, "max": 0.06453932909159145, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.056048711720440125, "min": 0.02177607814470927, "max": 0.06738110147416591, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.16814613516132038, "min": 0.04355215628941854, "max": 0.19305108276506266, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.6048987983999974e-06, "min": 3.6048987983999974e-06, "max": 0.0002953089015637, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.0814696395199992e-05, "min": 1.0814696395199992e-05, "max": 0.0008437728187423998, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10120160000000004, "min": 0.10120160000000004, "max": 0.1984363, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.3036048000000001, "min": 0.20759775000000003, "max": 0.5812576, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 6.995983999999995e-05, "min": 6.995983999999995e-05, "max": 0.00492197137, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00020987951999999985, "min": 0.00020987951999999985, "max": 0.014064754239999998, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1679709032", "python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1679711446" }, "total": 2414.186469364, "count": 1, "self": 0.44104908000008436, "children": { "run_training.setup": { "total": 0.11294381199991221, "count": 1, "self": 0.11294381199991221 }, "TrainerController.start_learning": { "total": 2413.632476472, "count": 1, "self": 4.447008245995676, "children": { "TrainerController._reset_env": { "total": 8.729362329000082, "count": 1, "self": 8.729362329000082 }, "TrainerController.advance": { "total": 2400.3384790920036, "count": 233257, "self": 4.710391653898114, "children": { "env_step": { "total": 1860.108357730009, "count": 233257, "self": 1573.0615503780905, "children": { "SubprocessEnvManager._take_step": { "total": 284.2685999408782, "count": 233257, "self": 17.3057520478568, "children": { "TorchPolicy.evaluate": { "total": 266.9628478930214, "count": 222970, "self": 266.9628478930214 } } }, "workers": { "total": 2.7782074110403983, "count": 233257, "self": 0.0, "children": { "worker_root": { "total": 2405.5697015600063, "count": 233257, "is_parallel": true, "self": 1121.988483189033, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0009692200000017692, "count": 1, "is_parallel": true, "self": 0.00028154199992513895, "children": { "_process_rank_one_or_two_observation": { "total": 0.0006876780000766303, "count": 2, "is_parallel": true, "self": 0.0006876780000766303 } } }, "UnityEnvironment.step": { "total": 0.0288610410000274, "count": 1, "is_parallel": true, "self": 0.0003459340000517841, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00020189299993944587, "count": 1, "is_parallel": true, "self": 0.00020189299993944587 }, "communicator.exchange": { "total": 0.027622740000083468, "count": 1, "is_parallel": true, "self": 0.027622740000083468 }, "steps_from_proto": { "total": 0.000690473999952701, "count": 1, "is_parallel": true, "self": 0.000212631999943369, "children": { "_process_rank_one_or_two_observation": { "total": 0.00047784200000933197, "count": 2, "is_parallel": true, "self": 0.00047784200000933197 } } } } } } }, "UnityEnvironment.step": { "total": 1283.5812183709731, "count": 233256, "is_parallel": true, "self": 38.501591318934516, "children": { "UnityEnvironment._generate_step_input": { "total": 77.37988996102399, "count": 233256, "is_parallel": true, "self": 77.37988996102399 }, "communicator.exchange": { "total": 1078.6279335439897, "count": 233256, "is_parallel": true, "self": 1078.6279335439897 }, "steps_from_proto": { "total": 89.07180354702484, "count": 233256, "is_parallel": true, "self": 33.50142663503857, "children": { "_process_rank_one_or_two_observation": { "total": 55.57037691198627, "count": 466512, "is_parallel": true, "self": 55.57037691198627 } } } } } } } } } } }, "trainer_advance": { "total": 535.5197297080964, "count": 233257, "self": 6.936320227143256, "children": { "process_trajectory": { "total": 149.39268992895313, "count": 233257, "self": 147.9431576389518, "children": { "RLTrainer._checkpoint": { "total": 1.4495322900013434, "count": 10, "self": 1.4495322900013434 } } }, "_update_policy": { "total": 379.190719552, "count": 97, "self": 319.8809438560097, "children": { "TorchPPOOptimizer.update": { "total": 59.30977569599031, "count": 2910, "self": 59.30977569599031 } } } } } } }, "trainer_threads": { "total": 8.71000338520389e-07, "count": 1, "self": 8.71000338520389e-07 }, "TrainerController._save_models": { "total": 0.11762593399998877, "count": 1, "self": 0.002767821999896114, "children": { "RLTrainer._checkpoint": { "total": 0.11485811200009266, "count": 1, "self": 0.11485811200009266 } } } } } } }