Closed kosmylo closed 4 years ago
Duplicate of #791 and #509
To sum up, I did what is mentioned in #509, but the learning rate doesn't take the values that I want. Specifically, I use the following commands:
1) sched_LR = LinearSchedule(params.time_steps, 0.005, 0.00025)
for establishment of linear schedule from 0.005 to 0.00025
2) learning_rate = sched_LR.value
as an argument in PPO2
But what I am getting as a learning rate schedule according to tensorboard is the following:
The plot shows that the learning rate starts from 0.00025 and stays at this value.
Please look at how it is done in the rl zoo, and as mentioned in the doc, we recommend to use the rl zoo for best practices ;)
I also recommend you to give Stable-Baselines3 a try (as SB2 is in maintenance now), it has also a rl zoo: https://github.com/DLR-RM/rl-baselines3-zoo
Describe the bug I want to use learning rate schedule for training a SAC agent, but I cannot find the proper way to inform the algorithm about that. I am doing exactly the same as in PPO2 as follows:
Code example
The problem is that the training starts with the lowest value for the learning rate, namely 0.0001 in this case.
What am I doing wrong?