gortizji / linearized-networks

Source code of "What can linearized neural networks actually say about generalization?
https://arxiv.org/abs/2106.06770
MIT License
18 stars 1 forks source link

Question about linearized NNs #1

Closed bwnjnOEI closed 2 years ago

bwnjnOEI commented 2 years ago

Hi, nice work! You expanded and delved the work Deep learning versus kernel learning: an empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. But some problem puzzle me:

What is the linearized training? This seem like different from training NNs using tiny learning rate(gradient flow). And i guess it just update parameters by SGD( or using tiny learning rate)?

The question I have is not directly from your paper, but I want to solve my trouble.

gortizji commented 2 years ago

Linearized training refers indeed to updating the parameters of the linearized model (i.e., first-order Taylor decomposition). This is in general not equivalent to training the full neural network with gradient flow, which would also update higher order terms in the decomposition.