Closed WangDeyu closed 2 years ago
Is the gradient of x just for calculating gradient penalty? How does the loss of l_gradient_penalty * gp backpropagate?
Yes, that's right. It's added to the loss, which is followed by loss.backward()
Could you please explain the process of model.update_embeddings ? What’s the meaning of self.N and self.m?
model.update_embeddings
is called after each model update step. self.N
is the number of points of a particular class in a minibatch, self.m
is the sum of the embeddings per class. Both are exponentially averaged.
This is explained below equation 6 in the paper (see 1. above).
I’m confused about the method of updating centroids, could you please explain it?
The paper mentioned:
I only found the paper, but didn't find the Appendix of van den Oord et al. (2017), Can you provide the link of the Appendix?
Is the gradient of
x
just for calculating gradient penalty? How does the loss ofl_gradient_penalty * gp
backpropagate?Could you please explain the process of
model.update_embeddings
? What’s the meaning ofself.N
andself.m
?Thank you so much!