Closed rudrapoudel closed 1 year ago
Hey, thanks for the kind words.
We decided to focus on developing a world model architecture for learning in imagination. What you suggest has more to do with representation learning and would imply entangling the design of the world model with that of the policy. It would indeed be interesting to investigate architectures where the policy operates over latent states of the autoencoder and/or hidden states of the Transformer. At the moment, we have not run such experiments.
Nice work and thanks for the code!