I have added a new agent -- PPO + LSTM, together with the new EpisodicRolloutBuffer, which is similar to VanillaRolloutBuffer but samples entire trajectories instead of random transitions in order to train the LSTM appropriately.
I have also added an example notebook to train it on Atari - Space Invaders, which achieves the following results:
In this case, it performs very similarly to vanilla PPO:
Motivation and Context
PPO LSTM can achieve better performance than PPO in partially observed environments.
[x] I have raised an issue to propose this change (required for new features and bug fixes)
Types of changes
[ ] Bug fix (non-breaking change which fixes an issue)
[x] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Description
I have added a new agent -- PPO + LSTM, together with the new EpisodicRolloutBuffer, which is similar to VanillaRolloutBuffer but samples entire trajectories instead of random transitions in order to train the LSTM appropriately.
I have also added an example notebook to train it on Atari - Space Invaders, which achieves the following results:
In this case, it performs very similarly to vanilla PPO:
Motivation and Context
PPO LSTM can achieve better performance than PPO in partially observed environments.
Types of changes
Checklist
make format
(required)make check-codestyle
andmake lint
(required)make pytest
andmake type
both pass. (required)make doc
(required)Note: You can run most of the checks using
make commit-checks
.Note: we are using a maximum length of 127 characters per line