Open binbinyouli12 opened 2 months ago
@binbinyouli12 Thank you for using my code - I think it's normal, the most important metric is the success rate.
Hello, I have a few more questions. If I only use DDPG in your code, do I just need to remove the HER replacement in the memory part of the code? Another question is, will using HER lead to issues with reward overestimation and failure to correctly guide the agent to the desired location?
Hello,
First of all, thank you for providing the DDPG+HER code; it has been a great help. However, I have some basic questions as I am just starting to learn about reinforcement learning. After adapting your application to my custom environment, I noticed that during the initial stages of training, the printed actor loss is very small, typically around 0.000-something, and the critic loss is usually about 0.0000-something. I am not sure if this is normal or if there is a problem somewhere?