hakuhodo-technologies / scope-rl

SCOPE-RL: A python library for offline reinforcement learning, off-policy evaluation, and selection
https://scope-rl.readthedocs.io/en/latest/
Apache License 2.0
106 stars 10 forks source link

[Questions] how can we generate offline dataset like D4RL? #24

Open return-sleep opened 7 months ago

return-sleep commented 7 months ago

Thanks for your outstanding work. I would like to ask: how should we generate offline datasets, such as medium or medium-expert version, like D4RL. Also is it possible to render states into images to support learning offline policy from visual observations?

aiueola commented 7 months ago

@return-sleep

Thank you for the question. For the dataset generation, please refer to this page in the documentation. Though we do not provide pre-trained agents for the data collection, one can train a policy online and use it as a data collection policy.

For the rendering of visual inputs, we do not have functions specialized for the visualization of the image observations. However, the offline reinforcement learning module should be able to handle it, taking advantage of PixelEncoderFactory provided in d3rlpy. For example, when training CQL, you can simply pass the instance as actor_encode_factory=pixel_encoder_factory, critic_encoder_factory=pixel_encoder_factory, as described in d3rlpy's documentation.