ChufanSuki / read-paper-and-code

0 stars 0 forks source link

ICML 2023 | Parameter-Level Soft-Masking for Continual Learning #75

Open ChufanSuki opened 5 months ago

ChufanSuki commented 5 months ago

https://arxiv.org/abs/2306.14775

https://github.com/UIC-Liu-Lab/SPG

ChufanSuki commented 5 months ago

Introduction

CF

Regularization-based:computes importance values of either parameters or their gradients on previous tasks, and adds a regularization in the loss to restrict changes to those important parameters

Regularization-based methods have difficulty to prevent CF


Memory-based: a small memory buffer to store data of previous tasks and replay them in learning a new task to prevent CF


Parameter isolation: learn to mask a sub-network for each task in a shared network. E.g. HAT, SupSup

poor KT

ChufanSuki commented 5 months ago

Soft-masking of Parameter-level Gradient flow

the importance of a parameter to a task is computed based on its gradient