{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4097583293914795, "min": 1.4097416400909424, "max": 1.4283286333084106, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 69963.484375, "min": 69624.296875, "max": 76190.828125, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 88.3989266547406, "min": 79.20161290322581, "max": 410.74590163934425, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49415.0, "min": 48836.0, "max": 50225.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999987.0, "min": 49597.0, "max": 1999987.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999987.0, "min": 49597.0, "max": 1999987.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.436450481414795, "min": 0.1388910710811615, "max": 2.5286405086517334, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1361.975830078125, "min": 16.80582046508789, "max": 1554.177490234375, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.6963064767593563, "min": 1.9325981111812198, "max": 4.064635463171108, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2066.23532050848, "min": 233.8443714529276, "max": 2416.915520429611, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.6963064767593563, "min": 1.9325981111812198, "max": 4.064635463171108, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2066.23532050848, "min": 233.8443714529276, "max": 2416.915520429611, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.01728263283633472, "min": 0.013700785408309456, "max": 0.020520886103622616, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.05184789850900416, "min": 0.02740157081661891, "max": 0.058425409940052006, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.05113179658850034, "min": 0.021534893475472924, "max": 0.06375581758717695, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.15339538976550102, "min": 0.04306978695094585, "max": 0.19126745276153087, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.6928487690833334e-06, "min": 3.6928487690833334e-06, "max": 0.00029528175157274994, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.107854630725e-05, "min": 1.107854630725e-05, "max": 0.0008439489186837, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10123091666666668, "min": 0.10123091666666668, "max": 0.19842725, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30369275000000007, "min": 0.20759110000000003, "max": 0.5813163, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 7.142274166666667e-05, "min": 7.142274166666667e-05, "max": 0.004921519774999999, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00021426822500000004, "min": 0.00021426822500000004, "max": 0.014067683370000004, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1670509917", "python_version": "3.8.15 (default, Oct 12 2022, 19:14:39) \n[GCC 7.5.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1670514261" }, "total": 4343.952967973, "count": 1, "self": 0.43901156399988395, "children": { "run_training.setup": { "total": 0.11272517399993376, "count": 1, "self": 0.11272517399993376 }, "TrainerController.start_learning": { "total": 4343.4012312350005, "count": 1, "self": 6.534613204193192, "children": { "TrainerController._reset_env": { "total": 9.657168056000046, "count": 1, "self": 9.657168056000046 }, "TrainerController.advance": { "total": 4327.106404997808, "count": 232947, "self": 6.307936196845731, "children": { "env_step": { "total": 2156.6456573250007, "count": 232947, "self": 1773.286168771683, "children": { "SubprocessEnvManager._take_step": { "total": 378.893346499136, "count": 232947, "self": 19.23982730914986, "children": { "TorchPolicy.evaluate": { "total": 359.6535191899861, "count": 223012, "self": 44.3396424688508, "children": { "TorchPolicy.sample_actions": { "total": 315.31387672113533, "count": 223012, "self": 315.31387672113533 } } } } }, "workers": { "total": 4.466142054181773, "count": 232947, "self": 0.0, "children": { "worker_root": { "total": 4331.168089047204, "count": 232947, "is_parallel": true, "self": 2920.6352740472967, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.007966459000044779, "count": 1, "is_parallel": true, "self": 0.0006023400001140544, "children": { "_process_rank_one_or_two_observation": { "total": 0.007364118999930724, "count": 2, "is_parallel": true, "self": 0.007364118999930724 } } }, "UnityEnvironment.step": { "total": 0.03193897599999218, "count": 1, "is_parallel": true, "self": 0.00027658000021801854, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00024113999984365364, "count": 1, "is_parallel": true, "self": 0.00024113999984365364 }, "communicator.exchange": { "total": 0.030662465999967026, "count": 1, "is_parallel": true, "self": 0.030662465999967026 }, "steps_from_proto": { "total": 0.0007587899999634828, "count": 1, "is_parallel": true, "self": 0.0002375500000653119, "children": { "_process_rank_one_or_two_observation": { "total": 0.0005212399998981709, "count": 2, "is_parallel": true, "self": 0.0005212399998981709 } } } } } } }, "UnityEnvironment.step": { "total": 1410.5328149999075, "count": 232946, "is_parallel": true, "self": 43.272642082017, "children": { "UnityEnvironment._generate_step_input": { "total": 64.71605120582262, "count": 232946, "is_parallel": true, "self": 64.71605120582262 }, "communicator.exchange": { "total": 1190.5913320730428, "count": 232946, "is_parallel": true, "self": 1190.5913320730428 }, "steps_from_proto": { "total": 111.95278963902501, "count": 232946, "is_parallel": true, "self": 43.351653754931704, "children": { "_process_rank_one_or_two_observation": { "total": 68.60113588409331, "count": 465892, "is_parallel": true, "self": 68.60113588409331 } } } } } } } } } } }, "trainer_advance": { "total": 2164.1528114759612, "count": 232947, "self": 11.066469340214098, "children": { "process_trajectory": { "total": 252.35740804274815, "count": 232947, "self": 251.97641377974855, "children": { "RLTrainer._checkpoint": { "total": 0.38099426299959305, "count": 4, "self": 0.38099426299959305 } } }, "_update_policy": { "total": 1900.7289340929988, "count": 97, "self": 243.58899946900647, "children": { "TorchPPOOptimizer.update": { "total": 1657.1399346239923, "count": 2910, "self": 1657.1399346239923 } } } } } } }, "trainer_threads": { "total": 8.110000635497272e-07, "count": 1, "self": 8.110000635497272e-07 }, "TrainerController._save_models": { "total": 0.10304416599956312, "count": 1, "self": 0.0018346179995205603, "children": { "RLTrainer._checkpoint": { "total": 0.10120954800004256, "count": 1, "self": 0.10120954800004256 } } } } } } }