ppo-Huggy / run_logs /timers.json
wooihen's picture
Huggy - Initial Commit
4a77e59
{
"name": "root",
"gauges": {
"Huggy.Policy.Entropy.mean": {
"value": 1.4065192937850952,
"min": 1.4065192937850952,
"max": 1.4261252880096436,
"count": 40
},
"Huggy.Policy.Entropy.sum": {
"value": 69991.2109375,
"min": 67456.6640625,
"max": 76484.796875,
"count": 40
},
"Huggy.Environment.EpisodeLength.mean": {
"value": 93.30131826741996,
"min": 83.41722972972973,
"max": 390.968992248062,
"count": 40
},
"Huggy.Environment.EpisodeLength.sum": {
"value": 49543.0,
"min": 48970.0,
"max": 50435.0,
"count": 40
},
"Huggy.Step.mean": {
"value": 1999986.0,
"min": 49814.0,
"max": 1999986.0,
"count": 40
},
"Huggy.Step.sum": {
"value": 1999986.0,
"min": 49814.0,
"max": 1999986.0,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.mean": {
"value": 2.4391233921051025,
"min": -0.017253749072551727,
"max": 2.4846079349517822,
"count": 40
},
"Huggy.Policy.ExtrinsicValueEstimate.sum": {
"value": 1295.174560546875,
"min": -2.208479881286621,
"max": 1430.368408203125,
"count": 40
},
"Huggy.Environment.CumulativeReward.mean": {
"value": 3.762664524736602,
"min": 1.7149685039184988,
"max": 4.074545225575423,
"count": 40
},
"Huggy.Environment.CumulativeReward.sum": {
"value": 1997.9748626351357,
"min": 219.51596850156784,
"max": 2223.750473201275,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.mean": {
"value": 3.762664524736602,
"min": 1.7149685039184988,
"max": 4.074545225575423,
"count": 40
},
"Huggy.Policy.ExtrinsicReward.sum": {
"value": 1997.9748626351357,
"min": 219.51596850156784,
"max": 2223.750473201275,
"count": 40
},
"Huggy.Losses.PolicyLoss.mean": {
"value": 0.016829739866726517,
"min": 0.012623462902694074,
"max": 0.019379492297108905,
"count": 40
},
"Huggy.Losses.PolicyLoss.sum": {
"value": 0.05048921960017955,
"min": 0.02546476586236774,
"max": 0.058138476891326715,
"count": 40
},
"Huggy.Losses.ValueLoss.mean": {
"value": 0.05689984754555755,
"min": 0.024054072424769403,
"max": 0.0626257926846544,
"count": 40
},
"Huggy.Losses.ValueLoss.sum": {
"value": 0.17069954263667264,
"min": 0.048108144849538806,
"max": 0.17803839867313706,
"count": 40
},
"Huggy.Policy.LearningRate.mean": {
"value": 3.7041987653000107e-06,
"min": 3.7041987653000107e-06,
"max": 0.00029532930155690005,
"count": 40
},
"Huggy.Policy.LearningRate.sum": {
"value": 1.1112596295900032e-05,
"min": 1.1112596295900032e-05,
"max": 0.0008442843185718998,
"count": 40
},
"Huggy.Policy.Epsilon.mean": {
"value": 0.10123470000000001,
"min": 0.10123470000000001,
"max": 0.1984431,
"count": 40
},
"Huggy.Policy.Epsilon.sum": {
"value": 0.30370410000000003,
"min": 0.20761614999999997,
"max": 0.5814281,
"count": 40
},
"Huggy.Policy.Beta.mean": {
"value": 7.16115300000002e-05,
"min": 7.16115300000002e-05,
"max": 0.004922310690000001,
"count": 40
},
"Huggy.Policy.Beta.sum": {
"value": 0.0002148345900000006,
"min": 0.0002148345900000006,
"max": 0.014073262189999998,
"count": 40
},
"Huggy.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
},
"Huggy.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 40
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1671545038",
"python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics",
"mlagents_version": "0.29.0.dev0",
"mlagents_envs_version": "0.29.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.8.1+cu102",
"numpy_version": "1.21.6",
"end_time_seconds": "1671547176"
},
"total": 2138.214693779,
"count": 1,
"self": 0.43907510799999727,
"children": {
"run_training.setup": {
"total": 0.12022019799996997,
"count": 1,
"self": 0.12022019799996997
},
"TrainerController.start_learning": {
"total": 2137.655398473,
"count": 1,
"self": 3.536953458905373,
"children": {
"TrainerController._reset_env": {
"total": 7.305738372000178,
"count": 1,
"self": 7.305738372000178
},
"TrainerController.advance": {
"total": 2126.7020190370945,
"count": 232384,
"self": 3.7464575921599135,
"children": {
"env_step": {
"total": 1665.6109306439062,
"count": 232384,
"self": 1400.3316483257004,
"children": {
"SubprocessEnvManager._take_step": {
"total": 262.81666863521605,
"count": 232384,
"self": 13.655521117301305,
"children": {
"TorchPolicy.evaluate": {
"total": 249.16114751791474,
"count": 222974,
"self": 62.26937671593487,
"children": {
"TorchPolicy.sample_actions": {
"total": 186.89177080197987,
"count": 222974,
"self": 186.89177080197987
}
}
}
}
},
"workers": {
"total": 2.4626136829897405,
"count": 232384,
"self": 0.0,
"children": {
"worker_root": {
"total": 2130.492201104026,
"count": 232384,
"is_parallel": true,
"self": 974.4160901560481,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.0018358750003244495,
"count": 1,
"is_parallel": true,
"self": 0.00029716500012000324,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0015387100002044463,
"count": 2,
"is_parallel": true,
"self": 0.0015387100002044463
}
}
},
"UnityEnvironment.step": {
"total": 0.0294132020003417,
"count": 1,
"is_parallel": true,
"self": 0.0001794350005184242,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00017176699975607335,
"count": 1,
"is_parallel": true,
"self": 0.00017176699975607335
},
"communicator.exchange": {
"total": 0.02862963600000512,
"count": 1,
"is_parallel": true,
"self": 0.02862963600000512
},
"steps_from_proto": {
"total": 0.00043236400006207987,
"count": 1,
"is_parallel": true,
"self": 0.0001494100001764309,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00028295399988564895,
"count": 2,
"is_parallel": true,
"self": 0.00028295399988564895
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 1156.0761109479781,
"count": 232383,
"is_parallel": true,
"self": 33.87417085877496,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 74.25427492912058,
"count": 232383,
"is_parallel": true,
"self": 74.25427492912058
},
"communicator.exchange": {
"total": 957.9073844999848,
"count": 232383,
"is_parallel": true,
"self": 957.9073844999848
},
"steps_from_proto": {
"total": 90.04028066009778,
"count": 232383,
"is_parallel": true,
"self": 36.770761030148606,
"children": {
"_process_rank_one_or_two_observation": {
"total": 53.26951962994917,
"count": 464766,
"is_parallel": true,
"self": 53.26951962994917
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 457.3446308010284,
"count": 232384,
"self": 5.567366483992373,
"children": {
"process_trajectory": {
"total": 142.7584857310335,
"count": 232384,
"self": 141.63048059303355,
"children": {
"RLTrainer._checkpoint": {
"total": 1.128005137999935,
"count": 10,
"self": 1.128005137999935
}
}
},
"_update_policy": {
"total": 309.01877858600255,
"count": 97,
"self": 256.62899227703247,
"children": {
"TorchPPOOptimizer.update": {
"total": 52.38978630897009,
"count": 2910,
"self": 52.38978630897009
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.2210002751089633e-06,
"count": 1,
"self": 1.2210002751089633e-06
},
"TrainerController._save_models": {
"total": 0.11068638399956399,
"count": 1,
"self": 0.0019273449997854186,
"children": {
"RLTrainer._checkpoint": {
"total": 0.10875903899977857,
"count": 1,
"self": 0.10875903899977857
}
}
}
}
}
}
}