Open pseudo-rnd-thoughts opened 2 years ago
Will there be a required format for the tutorials?
Isn't the first tutorial on Implementation of a custom environment already done in: https://gymnasium.farama.org/content/environment_creation/ ?
We are using the sphinx-gallery project, therefore, we expect single python files with figures stored in a folder with the same name. We are in the process of transferring our current tutorials to this style but pytorch use this style for their tutorials
Since I'm a beginner, I'll take DQN for Atari tutorial.
I'm working on the Frozenlake one, still work in progress but I hope to have it ready soonish.
I am working on "Training a deep RL agent with pytorch (with the new step API)".
Could someone make an update on this proposal? For instance, which tasks are done, which ones are work in progress, so that one may know what to work on.
Could someone make an update on this proposal? For instance, which tasks are done, which ones are work in progress, so that one may know what to work on.
I've have updated
@pseudo-rnd-thoughts also the Frozenlake training results with different map sizes tutorial is done.
Thanks!
I am thinking to work on Training a deep RL agent with pytorch fromscratch
. However, doesn't the REINFORCE implementation fit that requirement?
Good point @hahas94 I think we are always happy for more tutorials. You can work on another tutorial if you want
Okay, if similar tutorials is a good thing I will stick to my aforementioned choice. I am thinking of writing a tutorial on implementing the C51 agent in PyTorch.
We need this. For practical use case. Need external environment tutorials not another memory environment
I can add a tutorial for "Train an agent with tianshou". The exact agent and env is free to choose right, as long as env is from Gymnasium?
@dantp-ai we have chosen to link tutorials, who purpose is to show how to use external libraries, to
@Kallinteris-Andreas I see, so for "Train an agent with tianshou" it would simply mean listing a link to the library with a brief explanation of its focus (similar to AgileRL in the section).
Regarding, the tutorial "How to use the action sample masking, with example from Taxi", I can tackle this one. Do I understand correctly that for this tutorial we want to show the impact of the RL agent learning, when we use at each state the info["action_mask"]
to pick up an action that results in a state change versus when we are not using the action_mask
info?
@dantp-ai yes that is general idea. Both a discussion of what action masking is, how it is impacted in gymnasium with space.sample(mask) and a small training implementation
Proposal
To encourage the use of Gymnasium and build up the RL community, I would propose that a large range of tutorials are created.
This is a list of tutorials that could be made
gym.make_vec
)hardcore
parameter on agent performance