inkasaras's picture
Second Push
cc8d7bc
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": 1.8565446138381958,
"min": 1.8300139904022217,
"max": 3.2036256790161133,
"count": 401
},
"SoccerTwos.Policy.Entropy.sum": {
"value": 34992.15234375,
"min": 32595.626953125,
"max": 102866.40625,
"count": 401
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 80.8360655737705,
"min": 48.53465346534654,
"max": 884.8333333333334,
"count": 401
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 19724.0,
"min": 1484.0,
"max": 25944.0,
"count": 401
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1594.5005268827101,
"min": 1199.8062420984254,
"max": 1615.6039021747852,
"count": 401
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 194529.06427969065,
"min": 2408.5595086931794,
"max": 310671.0313586862,
"count": 401
},
"SoccerTwos.Step.mean": {
"value": 4499688.0,
"min": 499928.0,
"max": 4499688.0,
"count": 401
},
"SoccerTwos.Step.sum": {
"value": 4499688.0,
"min": 499928.0,
"max": 4499688.0,
"count": 401
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -0.007189917378127575,
"min": -0.10055205225944519,
"max": 0.21586117148399353,
"count": 401
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -0.8771699070930481,
"min": -16.395347595214844,
"max": 29.357118606567383,
"count": 401
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -0.00981832854449749,
"min": -0.10455340147018433,
"max": 0.22598282992839813,
"count": 401
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -1.1978360414505005,
"min": -16.937650680541992,
"max": 30.733665466308594,
"count": 401
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 401
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 401
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.01649508466486071,
"min": -1.0,
"max": 0.5655735284090042,
"count": 401
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 2.0124003291130066,
"min": -57.84639984369278,
"max": 76.91799986362457,
"count": 401
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.01649508466486071,
"min": -1.0,
"max": 0.5655735284090042,
"count": 401
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 2.0124003291130066,
"min": -57.84639984369278,
"max": 76.91799986362457,
"count": 401
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 401
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 401
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.01713785919030973,
"min": 0.010751689630463563,
"max": 0.02467565069285532,
"count": 193
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.01713785919030973,
"min": 0.010751689630463563,
"max": 0.02467565069285532,
"count": 193
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 0.07816493560870488,
"min": 0.001869278926945602,
"max": 0.11182952572902044,
"count": 193
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 0.07816493560870488,
"min": 0.001869278926945602,
"max": 0.11182952572902044,
"count": 193
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 0.07915815537174543,
"min": 0.0018843572567372273,
"max": 0.11378700186808904,
"count": 193
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 0.07915815537174543,
"min": 0.0018843572567372273,
"max": 0.11378700186808904,
"count": 193
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 193
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 193
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 193
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 193
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 193
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 193
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1679669529",
"python_version": "3.9.16 (main, Mar 8 2023, 10:39:24) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "C:\\Users\\joaop\\miniconda3\\envs\\rl\\Scripts\\mlagents-learn ./config/poca/SoccerTwos.yaml --env=./training-envs-executables/SoccerTwos.exe --run-id=SoccerTwos --no-graphics --resume",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.0.0+cpu",
"numpy_version": "1.21.2",
"end_time_seconds": "1679692907"
},
"total": 23378.376714600003,
"count": 1,
"self": 0.8634517000027699,
"children": {
"run_training.setup": {
"total": 0.19488839999999996,
"count": 1,
"self": 0.19488839999999996
},
"TrainerController.start_learning": {
"total": 23377.3183745,
"count": 1,
"self": 10.698057600424363,
"children": {
"TrainerController._reset_env": {
"total": 6.217875199999011,
"count": 21,
"self": 6.217875199999011
},
"TrainerController.advance": {
"total": 23360.136206399577,
"count": 271861,
"self": 10.819512099318672,
"children": {
"env_step": {
"total": 8126.954050799219,
"count": 271861,
"self": 6263.979860599473,
"children": {
"SubprocessEnvManager._take_step": {
"total": 1856.6918990994225,
"count": 271861,
"self": 67.9788764986547,
"children": {
"TorchPolicy.evaluate": {
"total": 1788.7130226007678,
"count": 502870,
"self": 1788.7130226007678
}
}
},
"workers": {
"total": 6.282291100323192,
"count": 271861,
"self": 0.0,
"children": {
"worker_root": {
"total": 23356.25566779996,
"count": 271861,
"is_parallel": true,
"self": 18305.991184700397,
"children": {
"steps_from_proto": {
"total": 0.08139189999524987,
"count": 42,
"is_parallel": true,
"self": 0.016833400005478616,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.06455849998977126,
"count": 168,
"is_parallel": true,
"self": 0.06455849998977126
}
}
},
"UnityEnvironment.step": {
"total": 5050.183091199568,
"count": 271861,
"is_parallel": true,
"self": 272.647700600568,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 244.5002960992781,
"count": 271861,
"is_parallel": true,
"self": 244.5002960992781
},
"communicator.exchange": {
"total": 3645.9851571996655,
"count": 271861,
"is_parallel": true,
"self": 3645.9851571996655
},
"steps_from_proto": {
"total": 887.0499373000555,
"count": 543722,
"is_parallel": true,
"self": 174.81389119825144,
"children": {
"_process_rank_one_or_two_observation": {
"total": 712.236046101804,
"count": 2174888,
"is_parallel": true,
"self": 712.236046101804
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 15222.362643501041,
"count": 271861,
"self": 75.56815480070145,
"children": {
"process_trajectory": {
"total": 2167.6675626003457,
"count": 271861,
"self": 2164.528453300348,
"children": {
"RLTrainer._checkpoint": {
"total": 3.139109299997987,
"count": 9,
"self": 3.139109299997987
}
}
},
"_update_policy": {
"total": 12979.126926099994,
"count": 193,
"self": 1164.9475176999513,
"children": {
"TorchPOCAOptimizer.update": {
"total": 11814.179408400043,
"count": 5790,
"self": 11814.179408400043
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.9999970390927047e-06,
"count": 1,
"self": 1.9999970390927047e-06
},
"TrainerController._save_models": {
"total": 0.2662333000007493,
"count": 1,
"self": 0.017151000000012573,
"children": {
"RLTrainer._checkpoint": {
"total": 0.24908230000073672,
"count": 1,
"self": 0.24908230000073672
}
}
}
}
}
}
}