EasonApolo / mstn

PyTorch reimplementation of Moving Semantic Transfer Network
35 stars 10 forks source link

is the Softmax function right? #5

Open lynnhhh opened 4 years ago

lynnhhh commented 4 years ago

Thanks for your nice work. I load the pre-train model and get the 78.6% acc on A->W. But I'm a little confused about the Softmax function in your code. As the input is N*C, why the Softmax function is applied to the 0-dimension? self.softmax = nn.Softmax(dim=0), I try to apply it to the 1-dimension and also get the 78.6% acc on A->W, but the model converges earlier.

datar001 commented 3 years ago

I have the same question, Why is dim=0 not the dim =1? a better performance?

jingzhengli commented 3 years ago

Thanks for your nice work. I load the pre-train model and get the 78.6% acc on A->W. But I'm a little confused about the Softmax function in your code. As the input is N*C, why the Softmax function is applied to the 0-dimension? self.softmax = nn.Softmax(dim=0), I try to apply it to the 1-dimension and also get the 78.6% acc on A->W, but the model converges earlier.

Hi, I don't know why I didn't reproduce the result of 78.6%. this is my result: validation: 183.0, 0.23018867924528302

jingzhengli commented 3 years ago

Thanks for your nice work. I load the pre-train model and get the 78.6% acc on A->W. But I'm a little confused about the Softmax function in your code. As the input is N*C, why the Softmax function is applied to the 0-dimension? self.softmax = nn.Softmax(dim=0), I try to apply it to the 1-dimension and also get the 78.6% acc on A->W, but the model converges earlier.

I didn't reproduce the result Even without the domain adaptation module. I didn't change the code except for the part about load the pre-trained AlexNet model. I download the pre-trained model from here: https://github.com/jiecaoyu/XNOR-Net-PyTorch.

jingzhengli commented 3 years ago

Thanks for your nice work. I load the pre-train model and get the 78.6% acc on A->W. But I'm a little confused about the Softmax function in your code. As the input is N*C, why the Softmax function is applied to the 0-dimension? self.softmax = nn.Softmax(dim=0), I try to apply it to the 1-dimension and also get the 78.6% acc on A->W, but the model converges earlier.

I didn't reproduce the result Even without the domain adaptation module. I didn't change the code except for the part about load the pre-trained AlexNet model. I download the pre-trained model from here: https://github.com/jiecaoyu/XNOR-Net-PyTorch.

I switched to the pre-trained model, then the result is largely improved. Namely, the algorithm heavily relies on a Pre-trained model.