Open sobir-git opened 3 years ago
I was wondering about this too. I was confused between 2 ways of doing the bias correction layer:
If anyone has any insight into which is the correct version that'd be helpful.
@bwolfson97 , I believe that there should be as many alphas and betas as incremental states
@EdenBelouadah Thanks!
@bwolfson97 For each increment, you need to learn a new set of bias paremeters for the new added classes for sure. For the old bias parameters, you can keep it (this repo) or discard it after distillation, depanding on how you do the distillation. I feel that the results from the two ways should be similar.
@wuyuebupt Awesome, thanks!
In the original paper, bias correction layer consists of only two parameters "alpha" and "beta", which are only applied to the logits corresponding to the last trained group of classes. Here it seems like you have multiple such layers for each group of classes. https://github.com/sairin1202/BIC/blob/56a34c47f4cce673aa646ecdb91030289e19fc84/trainer.py#L191-L202