taoyang1122 / adapt-image-models

[ICLR'23] AIM: Adapting Image Models for Efficient Video Action Recognition
Apache License 2.0
276 stars 21 forks source link

Training effiency #27

Closed Fanye12 closed 1 year ago

Fanye12 commented 1 year ago

I found that in practical applications, your method did not significantly improve training efficiency. Although the number of parameters to train is much smaller, it still requires calculating gradients for almost all layers due to backpropagation. As a result, there was not a significant reduction in memory usage or training time.

taoyang1122 commented 1 year ago

Hi, we had some discussion about memory usage and training time in Table 6. Yes, we still need the gradients for almost all layers, and the reduction in training time and memory may not be as significant as number of parameters. In fact. this is a problem of existing common PEFT methods such as Adapter, Prompt Tuning, LoRA, etc. There has been a work discussing this problem https://arxiv.org/abs/2206.06522. Further improving the efficiency could be a future direction.

Saving the number of parameters also has other benefits in applications such as communication-efficient distributed learning, privacy preserving federated learning. It also makes it easier to save multiple large models for different tasks.

Fanye12 commented 1 year ago

thank u for your reply