ppo-Huggy / run_logs /timers.json
moodlep's picture
Huggy
73b088b
raw
history blame
17.9 kB
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4083466529846191,
"min": 1.4083466529846191,
"max": 1.4314814805984497,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 70544.0859375,
"min": 69563.8125,
"max": 76569.4296875,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 77.08477237048666,
"min": 68.77513966480447,
"max": 405.01612903225805,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49103.0,
"min": 48982.0,
"max": 50222.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999946.0,
"min": 49985.0,
"max": 1999946.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999946.0,
"min": 49985.0,
"max": 1999946.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4982240200042725,
"min": 0.10753491520881653,
"max": 2.574455976486206,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1591.36865234375,
"min": 13.226794242858887,
"max": 1785.5816650390625,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.888633198509785,
"min": 1.692333216831936,
"max": 4.074271710525299,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2477.059347450733,
"min": 208.15698567032814,
"max": 2837.310325920582,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.888633198509785,
"min": 1.692333216831936,
"max": 4.074271710525299,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2477.059347450733,
"min": 208.15698567032814,
"max": 2837.310325920582,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.016377053786224376,
"min": 0.013268326189669704,
"max": 0.02091339319013059,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.049131161358673126,
"min": 0.026536652379339408,
"max": 0.057837815184029744,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.06304530021217135,
"min": 0.021909340471029284,
"max": 0.06304530021217135,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.18913590063651403,
"min": 0.04381868094205857,
"max": 0.18913590063651403,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.921248692950007e-06,
"min": 3.921248692950007e-06,
"max": 0.0002953233015588999,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.1763746078850019e-05,
"min": 1.1763746078850019e-05,
"max": 0.0008440482186505999,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10130705000000002,
"min": 0.10130705000000002,
"max": 0.19844110000000004,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.3039211500000001,
"min": 0.20776455,
"max": 0.5813494,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 7.52217950000001e-05,
"min": 7.52217950000001e-05,
"max": 0.0049222108900000006,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0002256653850000003,
"min": 0.0002256653850000003,
"max": 0.01406933506,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1671645495",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1671647691"
},
"total": 2196.225475872,
"count": 1,
"self": 0.4397386030004782,
"children": {
"run_training.setup": {
"total": 0.11159942899996622,
"count": 1,
"self": 0.11159942899996622
},
"TrainerController.start_learning": {
"total": 2195.6741378399997,
"count": 1,
"self": 3.708219253050629,
"children": {
"TrainerController._reset_env": {
"total": 8.988482865000037,
"count": 1,
"self": 8.988482865000037
},
"TrainerController.advance": {
"total": 2182.855264748949,
"count": 233643,
"self": 3.966572111979076,
"children": {
"env_step": {
"total": 1713.0355876469928,
"count": 233643,
"self": 1438.136168033997,
"children": {
"SubprocessEnvManager._take_step": {
"total": 272.38369722102595,
"count": 233643,
"self": 13.970420735975779,
"children": {
"TorchPolicy.evaluate": {
"total": 258.4132764850502,
"count": 222953,
"self": 64.89727219606681,
"children": {
"TorchPolicy.sample_actions": {
"total": 193.51600428898337,
"count": 222953,
"self": 193.51600428898337
}
}
}
}
},
"workers": {
"total": 2.5157223919698595,
"count": 233643,
"self": 0.0,
"children": {
"worker_root": {
"total": 2187.988831673913,
"count": 233643,
"is_parallel": true,
"self": 1004.7927639689392,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0020018260000256305,
"count": 1,
"is_parallel": true,
"self": 0.0003796230000148171,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0016222030000108134,
"count": 2,
"is_parallel": true,
"self": 0.0016222030000108134
}
}
},
"UnityEnvironment.step": {
"total": 0.026786828000012974,
"count": 1,
"is_parallel": true,
"self": 0.0002639350000208651,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00018746599999985847,
"count": 1,
"is_parallel": true,
"self": 0.00018746599999985847
},
"communicator.exchange": {
"total": 0.025611377000018365,
"count": 1,
"is_parallel": true,
"self": 0.025611377000018365
},
"steps_from_proto": {
"total": 0.0007240499999738859,
"count": 1,
"is_parallel": true,
"self": 0.0002725889999624087,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00045146100001147715,
"count": 2,
"is_parallel": true,
"self": 0.00045146100001147715
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1183.1960677049738,
"count": 233642,
"is_parallel": true,
"self": 34.634428519060066,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 74.43702776298193,
"count": 233642,
"is_parallel": true,
"self": 74.43702776298193
},
"communicator.exchange": {
"total": 982.283763170914,
"count": 233642,
"is_parallel": true,
"self": 982.283763170914
},
"steps_from_proto": {
"total": 91.8408482520179,
"count": 233642,
"is_parallel": true,
"self": 37.45487470209309,
"children": {
"_process_rank_one_or_two_observation": {
"total": 54.38597354992481,
"count": 467284,
"is_parallel": true,
"self": 54.38597354992481
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 465.8531049899774,
"count": 233643,
"self": 5.821555946000444,
"children": {
"process_trajectory": {
"total": 150.30675127497545,
"count": 233643,
"self": 149.1481316949753,
"children": {
"RLTrainer._checkpoint": {
"total": 1.1586195800001633,
"count": 10,
"self": 1.1586195800001633
}
}
},
"_update_policy": {
"total": 309.7247977690015,
"count": 97,
"self": 256.8394524859996,
"children": {
"TorchPPOOptimizer.update": {
"total": 52.885345283001925,
"count": 2910,
"self": 52.885345283001925
}
}
}
}
}
}
},
"trainer_threads": {
"total": 7.929997991595883e-07,
"count": 1,
"self": 7.929997991595883e-07
},
"TrainerController._save_models": {
"total": 0.1221701800000119,
"count": 1,
"self": 0.0018218730001535732,
"children": {
"RLTrainer._checkpoint": {
"total": 0.12034830699985832,
"count": 1,
"self": 0.12034830699985832
}
}
}
}
}
}
}