|
@@ -148,11 +148,16 @@ Actions in the basic environment:
|
|
|
- Toggle (open doors, interact with objects)
|
|
|
- Done (task completed, optional)
|
|
|
|
|
|
+Default tile/observation encoding:
|
|
|
+- Each tile is encoded as a 3 dimensional tuple: (OBJECT_IDX, COLOR_IDX, STATE)
|
|
|
+- OBJECT_TO_IDX and COLOR_TO_IDX mapping can be found in [gym_minigrid/minigrid.py](gym_minigrid/minigrid.py)
|
|
|
+- e.g. door STATE -> 0: open, 1: closed, 2: locked
|
|
|
+
|
|
|
By default, sparse rewards are given for reaching a green goal tile. A
|
|
|
reward of 1 is given for success, and zero for failure. There is also an
|
|
|
environment-specific time step limit for completing the task.
|
|
|
You can define your own reward function by creating a class derived
|
|
|
-from `MiniGridEnv`. Extending the environment with new object types or action
|
|
|
+from `MiniGridEnv`. Extending the environment with new object types or new actions
|
|
|
should be very easy. If you wish to do this, you should take a look at the
|
|
|
[gym_minigrid/minigrid.py](gym_minigrid/minigrid.py) source file.
|
|
|
|