deligentfool / dqn_zoo

The implement of all kinds of dqn reinforcement learning with Pytorch
86 stars 19 forks source link

MoG-DQN not performing well #1

Open ege-kaya opened 1 month ago

ege-kaya commented 1 month ago

Hello, thanks for all the cool implementations.

I was specifically interested in the MoG-DQN, however, running your implementation, it seems that it does not manage to learn the simplest CartPole problem, even after thousands of episodes, whereas the standard DQN algorithm manages to consistently score the maximum possible points after about 450 episodes.

Is there something wrong with the implementation? I've been trying to implement the MoG-DQN for about a month myself, now, and all my attempts were unsuccessful. That's why I wanted to try your implementation out to see what I'm doing wrong, but it appears that your implementation does not work either. This is very surprising to me because in the paper, MoG-DQN appears to be able to learn even Atari games, whereas in my experience it does not even work on the CartPole problem. How come?

deligentfool commented 1 month ago

Hello, I'm glad you are interested in my code. The code I implemented occasionally works, but it is extremely unstable and shows a phenomenon where the return increases after training but then decreases again. I am actually quite curious about this too. Previously, I wanted to apply the characteristics of MoG-DQN to related research, but I found that the results did not meet expectations.

The authors of the paper said they used a mixture density network, and I also calculated the loss according to the formulas they provided. Do you think there is anything wrong with my implementation?