ppo-Huggy / run_logs /timers.json
Krystof Saml
Huggy
9082512
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4067423343658447,
"min": 1.4067423343658447,
"max": 1.429145336151123,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 71228.9921875,
"min": 69858.078125,
"max": 76519.03125,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 76.51782945736434,
"min": 71.5421511627907,
"max": 413.8512396694215,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49354.0,
"min": 48815.0,
"max": 50076.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999941.0,
"min": 49519.0,
"max": 1999941.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999941.0,
"min": 49519.0,
"max": 1999941.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.5491013526916504,
"min": 0.12725305557250977,
"max": 2.551847457885742,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1644.17041015625,
"min": 15.270366668701172,
"max": 1752.016357421875,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 4.056073940447135,
"min": 1.9481734133015076,
"max": 4.08722722599034,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2616.167691588402,
"min": 233.78080959618092,
"max": 2726.0639117360115,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 4.056073940447135,
"min": 1.9481734133015076,
"max": 4.08722722599034,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2616.167691588402,
"min": 233.78080959618092,
"max": 2726.0639117360115,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.018508391798241066,
"min": 0.01363709384992641,
"max": 0.018941554454310484,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.055525175394723195,
"min": 0.02727418769985282,
"max": 0.055525175394723195,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.059904573733607934,
"min": 0.021432141525050005,
"max": 0.062334409231940904,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.1797137212008238,
"min": 0.04286428305010001,
"max": 0.18529379330575466,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.820648726483345e-06,
"min": 3.820648726483345e-06,
"max": 0.00029532360155879985,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.1461946179450036e-05,
"min": 1.1461946179450036e-05,
"max": 0.0008440023186659,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10127351666666669,
"min": 0.10127351666666669,
"max": 0.1984412,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.3038205500000001,
"min": 0.20767385000000008,
"max": 0.5813341000000001,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 7.354848166666685e-05,
"min": 7.354848166666685e-05,
"max": 0.004922215880000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00022064544500000055,
"min": 0.00022064544500000055,
"max": 0.014068571589999999,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1670535220",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1670537461"
},
"total": 2241.303073763,
"count": 1,
"self": 0.38720015899980353,
"children": {
"run_training.setup": {
"total": 0.10242823600003703,
"count": 1,
"self": 0.10242823600003703
},
"TrainerController.start_learning": {
"total": 2240.8134453680004,
"count": 1,
"self": 3.9688945259458706,
"children": {
"TrainerController._reset_env": {
"total": 10.937759443000004,
"count": 1,
"self": 10.937759443000004
},
"TrainerController.advance": {
"total": 2225.7926708510545,
"count": 233461,
"self": 4.333425060242462,
"children": {
"env_step": {
"total": 1733.668599052856,
"count": 233461,
"self": 1449.9240029810135,
"children": {
"SubprocessEnvManager._take_step": {
"total": 281.09571312377636,
"count": 233461,
"self": 14.572254073750628,
"children": {
"TorchPolicy.evaluate": {
"total": 266.52345905002574,
"count": 222988,
"self": 66.28917391414905,
"children": {
"TorchPolicy.sample_actions": {
"total": 200.2342851358767,
"count": 222988,
"self": 200.2342851358767
}
}
}
}
},
"workers": {
"total": 2.64888294806633,
"count": 233461,
"self": 0.0,
"children": {
"worker_root": {
"total": 2232.8217278410093,
"count": 233461,
"is_parallel": true,
"self": 1048.013605719942,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.002090160999955515,
"count": 1,
"is_parallel": true,
"self": 0.0003536819999681029,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.001736478999987412,
"count": 2,
"is_parallel": true,
"self": 0.001736478999987412
}
}
},
"UnityEnvironment.step": {
"total": 0.026574137000011433,
"count": 1,
"is_parallel": true,
"self": 0.0002489730001116186,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00020912999991651304,
"count": 1,
"is_parallel": true,
"self": 0.00020912999991651304
},
"communicator.exchange": {
"total": 0.025416790000008405,
"count": 1,
"is_parallel": true,
"self": 0.025416790000008405
},
"steps_from_proto": {
"total": 0.0006992439999748967,
"count": 1,
"is_parallel": true,
"self": 0.00024293599994962278,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00045630800002527394,
"count": 2,
"is_parallel": true,
"self": 0.00045630800002527394
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1184.8081221210673,
"count": 233460,
"is_parallel": true,
"self": 34.0457946558231,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 72.82173735210824,
"count": 233460,
"is_parallel": true,
"self": 72.82173735210824
},
"communicator.exchange": {
"total": 987.7483152691191,
"count": 233460,
"is_parallel": true,
"self": 987.7483152691191
},
"steps_from_proto": {
"total": 90.19227484401699,
"count": 233460,
"is_parallel": true,
"self": 37.296962618159796,
"children": {
"_process_rank_one_or_two_observation": {
"total": 52.8953122258572,
"count": 466920,
"is_parallel": true,
"self": 52.8953122258572
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 487.7906467379561,
"count": 233461,
"self": 6.114717760957319,
"children": {
"process_trajectory": {
"total": 155.95641447999787,
"count": 233461,
"self": 155.3248487729977,
"children": {
"RLTrainer._checkpoint": {
"total": 0.6315657070001635,
"count": 4,
"self": 0.6315657070001635
}
}
},
"_update_policy": {
"total": 325.7195144970009,
"count": 97,
"self": 270.97082095600945,
"children": {
"TorchPPOOptimizer.update": {
"total": 54.74869354099144,
"count": 2910,
"self": 54.74869354099144
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.2129999049648177e-06,
"count": 1,
"self": 1.2129999049648177e-06
},
"TrainerController._save_models": {
"total": 0.1141193349999412,
"count": 1,
"self": 0.0019587869996939844,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11216054800024722,
"count": 1,
"self": 0.11216054800024722
}
}
}
}
}
}
}