Thanks for this awesome resource! May I ask if there is a neat way to clip the gradient norm of the inner loop (the fast adaptation steps) of MAML? Currently, we can pass max_grad_grad_norm into the meta-model to control the gradient norm of the outer-loop of MAML. I am wondering if there is a neat way to also clip the gradient norm for the inner-loop, that is, the gradient used in fast adaptations. In particular, I am wondering if this can be applied to IngraphRMSProp.
Hi!
Thanks for this awesome resource! May I ask if there is a neat way to clip the gradient norm of the inner loop (the fast adaptation steps) of MAML? Currently, we can pass
max_grad_grad_norm
into the meta-model to control the gradient norm of the outer-loop of MAML. I am wondering if there is a neat way to also clip the gradient norm for the inner-loop, that is, the gradient used in fast adaptations. In particular, I am wondering if this can be applied toIngraphRMSProp
.Thanks so much! ;-)
Cheers, Tianlin