Open ChineseYjh opened 1 year ago
hello, thanks for your interest @ChineseYjh
Equation(7) is about the Euclidean norm of the gradient, the norm should be a scalar rather than a vector.
Note that M_wi represents the projection matrix of w_i. For detail, pls refer the Equation (11) of paper "PR Product: A Substitute for Inner Product in Neural Networks". https://openaccess.thecvf.com/content_ICCV_2019/papers/Wang_PR_Product_A_Substitute_for_Inner_Product_in_Neural_Networks_ICCV_2019_paper.pdf
Thank you very much
hello, thanks for your interest @ChineseYjh
Equation(7) is about the Euclidean norm of the gradient, the norm should be a scalar rather than a vector.
Note that M_wi represents the projection matrix of w_i. For detail, pls refer the Equation (11) of paper "PR Product: A Substitute for Inner Product in Neural Networks". https://openaccess.thecvf.com/content_ICCV_2019/papers/Wang_PR_Product_A_Substitute_for_Inner_Product_in_Neural_Networks_ICCV_2019_paper.pdf
Thank you very much
Thank you very much~
Hello~ sorry for bothering you!
How is the sin derived from the norm in Eq(7)? Moreover, the gradient (LHS of Eq(7)) should be a vector, but the result seems to be a scalar (RHS of Eq(7)).
Looking forward for your reply~