{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4083466529846191, "min": 1.4083466529846191, "max": 1.4314814805984497, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 70544.0859375, "min": 69563.8125, "max": 76569.4296875, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 77.08477237048666, "min": 68.77513966480447, "max": 405.01612903225805, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49103.0, "min": 48982.0, "max": 50222.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999946.0, "min": 49985.0, "max": 1999946.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999946.0, "min": 49985.0, "max": 1999946.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4982240200042725, "min": 0.10753491520881653, "max": 2.574455976486206, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1591.36865234375, "min": 13.226794242858887, "max": 1785.5816650390625, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.888633198509785, "min": 1.692333216831936, "max": 4.074271710525299, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2477.059347450733, "min": 208.15698567032814, "max": 2837.310325920582, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.888633198509785, "min": 1.692333216831936, "max": 4.074271710525299, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2477.059347450733, "min": 208.15698567032814, "max": 2837.310325920582, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.016377053786224376, "min": 0.013268326189669704, "max": 0.02091339319013059, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.049131161358673126, "min": 0.026536652379339408, "max": 0.057837815184029744, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.06304530021217135, "min": 0.021909340471029284, "max": 0.06304530021217135, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.18913590063651403, "min": 0.04381868094205857, "max": 0.18913590063651403, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.921248692950007e-06, "min": 3.921248692950007e-06, "max": 0.0002953233015588999, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.1763746078850019e-05, "min": 1.1763746078850019e-05, "max": 0.0008440482186505999, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10130705000000002, "min": 0.10130705000000002, "max": 0.19844110000000004, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.3039211500000001, "min": 0.20776455, "max": 0.5813494, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 7.52217950000001e-05, "min": 7.52217950000001e-05, "max": 0.0049222108900000006, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.0002256653850000003, "min": 0.0002256653850000003, "max": 0.01406933506, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1671645495", "python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.29.0.dev0", "mlagents_envs_version": "0.29.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.8.1+cu102", "numpy_version": "1.21.6", "end_time_seconds": "1671647691" }, "total": 2196.225475872, "count": 1, "self": 0.4397386030004782, "children": { "run_training.setup": { "total": 0.11159942899996622, "count": 1, "self": 0.11159942899996622 }, "TrainerController.start_learning": { "total": 2195.6741378399997, "count": 1, "self": 3.708219253050629, "children": { "TrainerController._reset_env": { "total": 8.988482865000037, "count": 1, "self": 8.988482865000037 }, "TrainerController.advance": { "total": 2182.855264748949, "count": 233643, "self": 3.966572111979076, "children": { "env_step": { "total": 1713.0355876469928, "count": 233643, "self": 1438.136168033997, "children": { "SubprocessEnvManager._take_step": { "total": 272.38369722102595, "count": 233643, "self": 13.970420735975779, "children": { "TorchPolicy.evaluate": { "total": 258.4132764850502, "count": 222953, "self": 64.89727219606681, "children": { "TorchPolicy.sample_actions": { "total": 193.51600428898337, "count": 222953, "self": 193.51600428898337 } } } } }, "workers": { "total": 2.5157223919698595, "count": 233643, "self": 0.0, "children": { "worker_root": { "total": 2187.988831673913, "count": 233643, "is_parallel": true, "self": 1004.7927639689392, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0020018260000256305, "count": 1, "is_parallel": true, "self": 0.0003796230000148171, "children": { "_process_rank_one_or_two_observation": { "total": 0.0016222030000108134, "count": 2, "is_parallel": true, "self": 0.0016222030000108134 } } }, "UnityEnvironment.step": { "total": 0.026786828000012974, "count": 1, "is_parallel": true, "self": 0.0002639350000208651, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00018746599999985847, "count": 1, "is_parallel": true, "self": 0.00018746599999985847 }, "communicator.exchange": { "total": 0.025611377000018365, "count": 1, "is_parallel": true, "self": 0.025611377000018365 }, "steps_from_proto": { "total": 0.0007240499999738859, "count": 1, "is_parallel": true, "self": 0.0002725889999624087, "children": { "_process_rank_one_or_two_observation": { "total": 0.00045146100001147715, "count": 2, "is_parallel": true, "self": 0.00045146100001147715 } } } } } } }, "UnityEnvironment.step": { "total": 1183.1960677049738, "count": 233642, "is_parallel": true, "self": 34.634428519060066, "children": { "UnityEnvironment._generate_step_input": { "total": 74.43702776298193, "count": 233642, "is_parallel": true, "self": 74.43702776298193 }, "communicator.exchange": { "total": 982.283763170914, "count": 233642, "is_parallel": true, "self": 982.283763170914 }, "steps_from_proto": { "total": 91.8408482520179, "count": 233642, "is_parallel": true, "self": 37.45487470209309, "children": { "_process_rank_one_or_two_observation": { "total": 54.38597354992481, "count": 467284, "is_parallel": true, "self": 54.38597354992481 } } } } } } } } } } }, "trainer_advance": { "total": 465.8531049899774, "count": 233643, "self": 5.821555946000444, "children": { "process_trajectory": { "total": 150.30675127497545, "count": 233643, "self": 149.1481316949753, "children": { "RLTrainer._checkpoint": { "total": 1.1586195800001633, "count": 10, "self": 1.1586195800001633 } } }, "_update_policy": { "total": 309.7247977690015, "count": 97, "self": 256.8394524859996, "children": { "TorchPPOOptimizer.update": { "total": 52.885345283001925, "count": 2910, "self": 52.885345283001925 } } } } } } }, "trainer_threads": { "total": 7.929997991595883e-07, "count": 1, "self": 7.929997991595883e-07 }, "TrainerController._save_models": { "total": 0.1221701800000119, "count": 1, "self": 0.0018218730001535732, "children": { "RLTrainer._checkpoint": { "total": 0.12034830699985832, "count": 1, "self": 0.12034830699985832 } } } } } } }