Open TimotheeMathieu opened 1 year ago
An I suggest we use rundoc or something similar to verify that the code in the user guide actually does something and have exit code 0.
I think this should go into the long tests because the user guide will contain some code to train agents and it would be too heavy for azure.
An example of a user guide section from pr #276 : https://rlberry--276.org.readthedocs.build/en/276/basics/comparison.html
We can try Jupytext to edit markdown in jupyter.
I'm adding notes concerning Philippe's remarks (check your mailbox):
eval([eval_horizon, n_simulations, gamma])'':
Monte-Carlo policy evaluation [1] of an agent to estimate the value at the initial state.''
Basically, we should pass on each function/methods, and write the documentation in a better way (if needed), so that everything is documented and explicit.
I propose we do a user guide for rlberry. The outline of which would be something like this:
Feel free to suggest any change to this outline. Once we all agree to the outline, we can distribute the work among us.