Open Achronferry opened 4 years ago
I notice that the model drops _weighthh in forward() and never recover them. The dropped neural unit will keep to 0 and never be recovered. After a series of turns, Every element of the weight matrix will be 0
I also have the same question. The code looks wrong to me if dropoutw > 0.
dropoutw > 0
+1 Implementation is blatantly incorrect
I notice that the model drops _weighthh in forward() and never recover them. The dropped neural unit will keep to 0 and never be recovered. After a series of turns, Every element of the weight matrix will be 0