philtabor / Deep-Q-Learning-Paper-To-Code

MIT License
342 stars 145 forks source link

Fix typo, use "rewards" instead of "reward" in cartpole_naive_dqn.py #4

Closed lydxlx1 closed 4 years ago

lydxlx1 commented 4 years ago

According to Phil's comment, we should ideally use parameter rewards although the code still runs fine with parameter reward.

Ah, good point. Yes, this is a typo. Fortunately, it's a typo that doesn't affect the outcome. The reward (the numpy array) isn't connected to the graph, so apparently PyTorch is just as happy dealing with a numpy array as it is a tensor.