EpicPinkPenguin commited on
Commit
0bd16df
1 Parent(s): 6321375

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -649,7 +649,7 @@ Each data instance represents a single step consisting of tuples of the form (ob
649
  The dataset is divided into a `train` (90%) and `test` (10%) split. Each environment-dataset has in sum 10M steps (data points).
650
 
651
  ## Dataset Creation
652
- The dataset was created by training an RL agent with [PPO](https://arxiv.org/abs/1707.06347) for 50M steps in each environment. The trajectories where generated by sampling from the predicted action distribution at each step (not taking the argmax). The environments were created on `distribution_mode=easy` and with unlimited levels.
653
 
654
  ## Procgen Benchmark
655
  The [Procgen Benchmark](https://openai.com/index/procgen-benchmark/), released by OpenAI, consists of 16 procedurally-generated environments designed to measure how quickly reinforcement learning (RL) agents learn generalizable skills. It emphasizes experimental convenience, high diversity within and across environments, and is ideal for evaluating both sample efficiency and generalization. The benchmark allows for distinct training and test sets in each environment, making it a standard research platform for the OpenAI RL team. It aims to address the need for more diverse RL benchmarks compared to complex environments like Dota and StarCraft.
 
649
  The dataset is divided into a `train` (90%) and `test` (10%) split. Each environment-dataset has in sum 10M steps (data points).
650
 
651
  ## Dataset Creation
652
+ The dataset was created by training an RL agent with [PPO](https://arxiv.org/abs/1707.06347) for 25M steps in each environment. The trajectories where generated by sampling from the predicted action distribution at each step (not taking the argmax). The environments were created on `distribution_mode=easy` and with unlimited levels.
653
 
654
  ## Procgen Benchmark
655
  The [Procgen Benchmark](https://openai.com/index/procgen-benchmark/), released by OpenAI, consists of 16 procedurally-generated environments designed to measure how quickly reinforcement learning (RL) agents learn generalizable skills. It emphasizes experimental convenience, high diversity within and across environments, and is ideal for evaluating both sample efficiency and generalization. The benchmark allows for distinct training and test sets in each environment, making it a standard research platform for the OpenAI RL team. It aims to address the need for more diverse RL benchmarks compared to complex environments like Dota and StarCraft.