Closed Varghese-Kuruvilla closed 2 years ago
Hi @Varghese-Kuruvilla After normalization (in data augmentation part), the value of the input tensor is among [-2, 2]. If you want to visualise the tensor, you need to reverse the normalisation.
Hi @layumi , Thanks for your quick response. However, I wanted to know why a single image contains 9 tiled images of the same person (as shown in the screenshot) . I would expect it to be a single image right? Thanks in advance!
Hi @layumi , Sorry my bad. Looks like numpy.reshape isn't the right way to visualise the tensor. Using the permute function from pytorch gave me the correct result. Would you mind commenting on the second point in my initial post? Thanks a lot!
Hi @Varghese-Kuruvilla
Actually, I found that the cross entropy loss is stable and easy to train in most cases. So generally I will keep cross entropy in most my attempts.
You may refer to my another paper on such phenomenon https://arxiv.org/abs/1711.05535
Thanks @layumi. Closing the issue.
Hi Layumi, Thanks a lot for the effort that you have put into this repository. It is really helpful for me. I have a couple of issues that I wanted to clarify
I tried to retrain ft_net with the Market-1501 dataset. When I visualise each image from the input tensor passed to the neural network, I get a tiled image of the same person. I have attached a screenshot of the same below. Is this correct? I am sure that I am missing something, I just am not able to figure out what.
The codebase seems to be computing the classification loss + triplet loss. However, the paper 'In defense of Triplet Loss for Person Reidentification' seems to advocate using the triplet loss alone. Could you comment on this?
To Reproduce Steps to reproduce the behavior:
Visualize input tensor with the following code snippet:
Screenshots
Thanks a lot for your help in advance!