This is the official implementation of "Learning from Visual Observation via Offline Pretrained State-to-Go Transformer" which has been accepted by NeurIPS 2023. A two-stage method, named STG, is proposed for reinforcement learning from visual observation.
Experiments are conducted in two video game environments: Atari and Minecraft. Codes & instructions about Atari can refer to atari and minecraft directory respectively.
pip clone https://github.com/zhoubohan0/STG-Transformer
Our paper is available on arxiv and you can access our website for more information. If you find our code useful or want to conduct further research based on STG, please consider citing us!
@article{zhou2023learning,
title={Learning from Visual Observation via Offline Pretrained State-to-Go Transformer},
author={Zhou, Bohan and Li, Ke and Jiang, Jiechuan and Lu, Zongqing},
journal={arXiv preprint arXiv:2306.12860},
year={2023}
}
We acknowledge cleanrl to provide a clear and simple implementation of PPO and SAC algorithm for our project. And the codebase from Plan4MC enables us to collect enough expert trajectories and to tackle Minecraft tasks.
MIT License