PPO-LunarLander-v2 / results.json
asuzuki's picture
first commit - model PPO performing good
71e2525
raw
history blame
165 Bytes
{"mean_reward": 239.92169866976738, "std_reward": 20.791617039010347, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-01-06T08:49:26.545362"}