poca-SoccerTwos / run_logs /timers.json
jroblesgomez's picture
Initial push
160bc63 verified
raw
history blame
16 kB
{
"name": "root",
"gauges": {
"SoccerTwos.Policy.Entropy.mean": {
"value": -3.576278402306343e-07,
"min": -25.074377059936523,
"max": 1.8537969589233398,
"count": 1000
},
"SoccerTwos.Policy.Entropy.sum": {
"value": -0.008468626998364925,
"min": -458159.03125,
"max": 51451.98046875,
"count": 1000
},
"SoccerTwos.Environment.EpisodeLength.mean": {
"value": 527.6,
"min": 192.4814814814815,
"max": 999.0,
"count": 1000
},
"SoccerTwos.Environment.EpisodeLength.sum": {
"value": 21104.0,
"min": 16076.0,
"max": 24840.0,
"count": 1000
},
"SoccerTwos.Self-play.ELO.mean": {
"value": 1155.711411124665,
"min": 1150.598487782326,
"max": 1207.8826292429792,
"count": 816
},
"SoccerTwos.Self-play.ELO.sum": {
"value": 11557.114111246648,
"min": 2301.196975564652,
"max": 59476.97958428653,
"count": 816
},
"SoccerTwos.Step.mean": {
"value": 9999874.0,
"min": 9436.0,
"max": 9999874.0,
"count": 1000
},
"SoccerTwos.Step.sum": {
"value": 9999874.0,
"min": 9436.0,
"max": 9999874.0,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.mean": {
"value": -1.4355356693267822,
"min": -86.131591796875,
"max": 142.8604278564453,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicBaselineEstimate.sum": {
"value": -28.71071434020996,
"min": -1188.624267578125,
"max": 2597.698486328125,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.mean": {
"value": -1.4908798933029175,
"min": -84.45001983642578,
"max": 139.96893310546875,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicValueEstimate.sum": {
"value": -29.817598342895508,
"min": -1232.548583984375,
"max": 2549.23486328125,
"count": 1000
},
"SoccerTwos.Environment.CumulativeReward.mean": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1000
},
"SoccerTwos.Environment.CumulativeReward.sum": {
"value": 0.0,
"min": 0.0,
"max": 0.0,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicReward.mean": {
"value": 0.08212000131607056,
"min": -0.6242260881092238,
"max": 0.5161636363376271,
"count": 1000
},
"SoccerTwos.Policy.ExtrinsicReward.sum": {
"value": 1.6424000263214111,
"min": -18.172400057315826,
"max": 12.67320004105568,
"count": 1000
},
"SoccerTwos.Environment.GroupCumulativeReward.mean": {
"value": 0.08212000131607056,
"min": -0.6242260881092238,
"max": 0.5161636363376271,
"count": 1000
},
"SoccerTwos.Environment.GroupCumulativeReward.sum": {
"value": 1.6424000263214111,
"min": -18.172400057315826,
"max": 12.67320004105568,
"count": 1000
},
"SoccerTwos.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1000
},
"SoccerTwos.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 1000
},
"SoccerTwos.Losses.PolicyLoss.mean": {
"value": 0.014513091626577079,
"min": 0.010575032234191895,
"max": 0.5904294659694036,
"count": 465
},
"SoccerTwos.Losses.PolicyLoss.sum": {
"value": 0.014513091626577079,
"min": 0.010575032234191895,
"max": 0.5904294659694036,
"count": 465
},
"SoccerTwos.Losses.ValueLoss.mean": {
"value": 332.52110525468987,
"min": 0.011835511432339747,
"max": 79919.56179827452,
"count": 465
},
"SoccerTwos.Losses.ValueLoss.sum": {
"value": 332.52110525468987,
"min": 0.011835511432339747,
"max": 79919.56179827452,
"count": 465
},
"SoccerTwos.Losses.BaselineLoss.mean": {
"value": 300.86022165815035,
"min": 0.011721353543301423,
"max": 94309.54581700961,
"count": 465
},
"SoccerTwos.Losses.BaselineLoss.sum": {
"value": 300.86022165815035,
"min": 0.011721353543301423,
"max": 94309.54581700961,
"count": 465
},
"SoccerTwos.Policy.LearningRate.mean": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 465
},
"SoccerTwos.Policy.LearningRate.sum": {
"value": 0.0003,
"min": 0.0003,
"max": 0.0003,
"count": 465
},
"SoccerTwos.Policy.Epsilon.mean": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 465
},
"SoccerTwos.Policy.Epsilon.sum": {
"value": 0.20000000000000007,
"min": 0.20000000000000007,
"max": 0.20000000000000007,
"count": 465
},
"SoccerTwos.Policy.Beta.mean": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 465
},
"SoccerTwos.Policy.Beta.sum": {
"value": 0.005000000000000001,
"min": 0.005000000000000001,
"max": 0.005000000000000001,
"count": 465
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1723573992",
"python_version": "3.10.12 | packaged by Anaconda, Inc. | (main, Jul 5 2023, 19:01:18) [MSC v.1916 64 bit (AMD64)]",
"command_line_arguments": "\\\\?\\C:\\Users\\Juan\\anaconda3\\envs\\rl\\Scripts\\mlagents-learn ./config/poca/SoccerTwos.yaml --env=./ml-agents/training-envs-executables/SoccerTwos/SoccerTwos.exe --run-id=SoccerTwos --no-graphics",
"mlagents_version": "1.1.0.dev0",
"mlagents_envs_version": "1.1.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "2.4.0+cpu",
"numpy_version": "1.23.5",
"end_time_seconds": "1723648498"
},
"total": 74505.75136419991,
"count": 1,
"self": 2.1500184000469744,
"children": {
"run_training.setup": {
"total": 0.16744569991715252,
"count": 1,
"self": 0.16744569991715252
},
"TrainerController.start_learning": {
"total": 74503.43390009995,
"count": 1,
"self": 25.694641928654164,
"children": {
"TrainerController._reset_env": {
"total": 23.986774299759418,
"count": 50,
"self": 23.986774299759418
},
"TrainerController.advance": {
"total": 74453.58467907156,
"count": 653111,
"self": 26.129078436410055,
"children": {
"env_step": {
"total": 20470.51052446314,
"count": 653111,
"self": 15894.306698830565,
"children": {
"SubprocessEnvManager._take_step": {
"total": 4560.4840015205555,
"count": 653111,
"self": 177.3953197451774,
"children": {
"TorchPolicy.evaluate": {
"total": 4383.088681775378,
"count": 1293826,
"self": 4383.088681775378
}
}
},
"workers": {
"total": 15.719824112020433,
"count": 653111,
"self": 0.0,
"children": {
"worker_root": {
"total": 74448.80314426776,
"count": 653111,
"is_parallel": true,
"self": 61870.764370877296,
"children": {
"steps_from_proto": {
"total": 0.11197260022163391,
"count": 100,
"is_parallel": true,
"self": 0.02306770207360387,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.08890489814803004,
"count": 400,
"is_parallel": true,
"self": 0.08890489814803004
}
}
},
"UnityEnvironment.step": {
"total": 12577.926800790243,
"count": 653111,
"is_parallel": true,
"self": 690.2568437336013,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 588.6797559948172,
"count": 653111,
"is_parallel": true,
"self": 588.6797559948172
},
"communicator.exchange": {
"total": 9108.145537183154,
"count": 653111,
"is_parallel": true,
"self": 9108.145537183154
},
"steps_from_proto": {
"total": 2190.84466387867,
"count": 1306222,
"is_parallel": true,
"self": 415.1400192868896,
"children": {
"_process_rank_one_or_two_observation": {
"total": 1775.7046445917804,
"count": 5224888,
"is_parallel": true,
"self": 1775.7046445917804
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 53956.94507617201,
"count": 653111,
"self": 187.72696588002145,
"children": {
"process_trajectory": {
"total": 5933.103623187635,
"count": 653111,
"self": 5929.990689388011,
"children": {
"RLTrainer._checkpoint": {
"total": 3.112933799624443,
"count": 20,
"self": 3.112933799624443
}
}
},
"_update_policy": {
"total": 47836.11448710435,
"count": 465,
"self": 1850.7883133892901,
"children": {
"TorchPOCAOptimizer.update": {
"total": 45985.32617371506,
"count": 13950,
"self": 45985.32617371506
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.700129359960556e-06,
"count": 1,
"self": 1.700129359960556e-06
},
"TrainerController._save_models": {
"total": 0.16780309984460473,
"count": 1,
"self": 0.009353499859571457,
"children": {
"RLTrainer._checkpoint": {
"total": 0.15844959998503327,
"count": 1,
"self": 0.15844959998503327
}
}
}
}
}
}
}