ppo-Huggy / run_logs /timers.json
kumasura's picture
Huggy
0e81c35
raw
history blame
17.4 kB
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.3655809164047241,
"min": 1.3655809164047241,
"max": 1.4220852851867676,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 68448.375,
"min": 67092.578125,
"max": 77067.484375,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 85.46434782608695,
"min": 75.79846153846154,
"max": 402.256,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49142.0,
"min": 49142.0,
"max": 50282.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999981.0,
"min": 49891.0,
"max": 1999981.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999981.0,
"min": 49891.0,
"max": 1999981.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.3977816104888916,
"min": 0.22666120529174805,
"max": 2.481987714767456,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1378.7244873046875,
"min": 28.105989456176758,
"max": 1591.406494140625,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.7594304469357365,
"min": 1.7458883824367677,
"max": 4.119779796903706,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 2161.6725069880486,
"min": 216.4901594221592,
"max": 2568.3164764642715,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.7594304469357365,
"min": 1.7458883824367677,
"max": 4.119779796903706,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 2161.6725069880486,
"min": 216.4901594221592,
"max": 2568.3164764642715,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.016941861639206763,
"min": 0.014954924264930013,
"max": 0.020531737734891066,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.05082558491762029,
"min": 0.029909848529860027,
"max": 0.05499477257500985,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.0973258104423682,
"min": 0.04163783516734838,
"max": 0.0973258104423682,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.2919774313271046,
"min": 0.08327567033469677,
"max": 0.2919774313271046,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.601448799550003e-06,
"min": 3.601448799550003e-06,
"max": 0.0002953024515658499,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.0804346398650009e-05,
"min": 1.0804346398650009e-05,
"max": 0.0008439864186711998,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10120045,
"min": 0.10120045,
"max": 0.19843415000000003,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30360135,
"min": 0.20755525000000002,
"max": 0.5813288000000001,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 6.990245500000006e-05,
"min": 6.990245500000006e-05,
"max": 0.004921864085,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.00020970736500000017,
"min": 0.00020970736500000017,
"max": 0.014068307120000002,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1680506821",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1680509786"
},
"total": 2965.886232971,
"count": 1,
"self": 0.4441644130001805,
"children": {
"run_training.setup": {
"total": 0.11025033699996811,
"count": 1,
"self": 0.11025033699996811
},
"TrainerController.start_learning": {
"total": 2965.331818221,
"count": 1,
"self": 5.767511819916308,
"children": {
"TrainerController._reset_env": {
"total": 10.006586082000013,
"count": 1,
"self": 10.006586082000013
},
"TrainerController.advance": {
"total": 2949.4364998800834,
"count": 231987,
"self": 6.578978099892993,
"children": {
"env_step": {
"total": 2146.411861281002,
"count": 231987,
"self": 1800.469472629019,
"children": {
"SubprocessEnvManager._take_step": {
"total": 342.1497806129756,
"count": 231987,
"self": 19.984126717934657,
"children": {
"TorchPolicy.evaluate": {
"total": 322.1656538950409,
"count": 222896,
"self": 322.1656538950409
}
}
},
"workers": {
"total": 3.792608039007405,
"count": 231987,
"self": 0.0,
"children": {
"worker_root": {
"total": 2955.256965436892,
"count": 231987,
"is_parallel": true,
"self": 1505.659248661711,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0009571260000029724,
"count": 1,
"is_parallel": true,
"self": 0.00033080400004337207,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0006263219999596004,
"count": 2,
"is_parallel": true,
"self": 0.0006263219999596004
}
}
},
"UnityEnvironment.step": {
"total": 0.031385136999972474,
"count": 1,
"is_parallel": true,
"self": 0.00034825799991722306,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00020871399999577989,
"count": 1,
"is_parallel": true,
"self": 0.00020871399999577989
},
"communicator.exchange": {
"total": 0.030034461000013835,
"count": 1,
"is_parallel": true,
"self": 0.030034461000013835
},
"steps_from_proto": {
"total": 0.0007937040000456363,
"count": 1,
"is_parallel": true,
"self": 0.00023842199999535296,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0005552820000502834,
"count": 2,
"is_parallel": true,
"self": 0.0005552820000502834
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1449.5977167751812,
"count": 231986,
"is_parallel": true,
"self": 43.010625543113065,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 87.21803803706126,
"count": 231986,
"is_parallel": true,
"self": 87.21803803706126
},
"communicator.exchange": {
"total": 1216.3331294120517,
"count": 231986,
"is_parallel": true,
"self": 1216.3331294120517
},
"steps_from_proto": {
"total": 103.0359237829552,
"count": 231986,
"is_parallel": true,
"self": 40.199435320785426,
"children": {
"_process_rank_one_or_two_observation": {
"total": 62.83648846216977,
"count": 463972,
"is_parallel": true,
"self": 62.83648846216977
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 796.4456604991883,
"count": 231987,
"self": 8.688414288302283,
"children": {
"process_trajectory": {
"total": 161.46791912388522,
"count": 231987,
"self": 160.04566630088465,
"children": {
"RLTrainer._checkpoint": {
"total": 1.4222528230005764,
"count": 10,
"self": 1.4222528230005764
}
}
},
"_update_policy": {
"total": 626.2893270870009,
"count": 97,
"self": 526.7230702639991,
"children": {
"TorchPPOOptimizer.update": {
"total": 99.56625682300171,
"count": 4850,
"self": 99.56625682300171
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.032999989547534e-06,
"count": 1,
"self": 1.032999989547534e-06
},
"TrainerController._save_models": {
"total": 0.12121940600036396,
"count": 1,
"self": 0.0021338890005608846,
"children": {
"RLTrainer._checkpoint": {
"total": 0.11908551699980308,
"count": 1,
"self": 0.11908551699980308
}
}
}
}
}
}
}