EpicPinkPenguin commited on
Commit
c806698
1 Parent(s): f85f87b

Update README.md

Browse files

Add description about environments

Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -99,7 +99,7 @@ tags:
99
 
100
  <video controls autoplay loop src="https://cdn-uploads.huggingface.co/production/uploads/633c1daf31c06121a58f2df9/brMaX1xgew7ulqkMU0Ahi.mp4"></video>
101
 
102
- This dataset contains expert trajectories generated by a [PPO](https://arxiv.org/abs/1707.06347) reinforcement learning agent trained on each of the 16 procedurally-generated gym environments from the [Procgen Benchmark](https://openai.com/index/procgen-benchmark/).
103
 
104
  Disclaimer: This is not an official repository from OpenAI.
105
 
@@ -190,7 +190,7 @@ Each data instance represents a single step consisting of tuples of the form (ob
190
  The dataset is divided into a `train` (90%) and `test` (10%) split
191
 
192
  ## Dataset Creation
193
- The dataset was created by training an RL agent with [PPO](https://arxiv.org/abs/1707.06347) for 50M steps in each environment. The trajectories where generated by taking the argmax action at each step, corresponding to taking the mode of the action distribution. Consequently the rollout policy is deterministic.
194
 
195
  ## Procgen Benchmark
196
  The [Procgen Benchmark](https://openai.com/index/procgen-benchmark/), released by OpenAI, consists of 16 procedurally-generated environments designed to measure how quickly reinforcement learning (RL) agents learn generalizable skills. It emphasizes experimental convenience, high diversity within and across environments, and is ideal for evaluating both sample efficiency and generalization. The benchmark allows for distinct training and test sets in each environment, making it a standard research platform for the OpenAI RL team. It aims to address the need for more diverse RL benchmarks compared to complex environments like Dota and StarCraft.
 
99
 
100
  <video controls autoplay loop src="https://cdn-uploads.huggingface.co/production/uploads/633c1daf31c06121a58f2df9/brMaX1xgew7ulqkMU0Ahi.mp4"></video>
101
 
102
+ This dataset contains expert trajectories generated by a [PPO](https://arxiv.org/abs/1707.06347) reinforcement learning agent trained on each of the 16 procedurally-generated gym environments from the [Procgen Benchmark](https://openai.com/index/procgen-benchmark/). The environments were created on `distribution_mode=easy` and with unlimited levels.
103
 
104
  Disclaimer: This is not an official repository from OpenAI.
105
 
 
190
  The dataset is divided into a `train` (90%) and `test` (10%) split
191
 
192
  ## Dataset Creation
193
+ The dataset was created by training an RL agent with [PPO](https://arxiv.org/abs/1707.06347) for 50M steps in each environment. The trajectories where generated by taking the argmax action at each step, corresponding to taking the mode of the action distribution. Consequently the rollout policy is deterministic. The environments were created on `distribution_mode=easy` and with unlimited levels.
194
 
195
  ## Procgen Benchmark
196
  The [Procgen Benchmark](https://openai.com/index/procgen-benchmark/), released by OpenAI, consists of 16 procedurally-generated environments designed to measure how quickly reinforcement learning (RL) agents learn generalizable skills. It emphasizes experimental convenience, high diversity within and across environments, and is ideal for evaluating both sample efficiency and generalization. The benchmark allows for distinct training and test sets in each environment, making it a standard research platform for the OpenAI RL team. It aims to address the need for more diverse RL benchmarks compared to complex environments like Dota and StarCraft.