arthurdouillard / dytox

Dynamic Token Expansion with Continual Transformers, accepted at CVPR 2022
https://arxiv.org/abs/2111.11326
Apache License 2.0
132 stars 18 forks source link

Why joint accuracy differs in different incremental settings? #21

Closed ChenJunzhi-buaa closed 1 year ago

ChenJunzhi-buaa commented 1 year ago

Thanky for your good work! I have some question about the upper bound. As below picture from your papper, joint accuracy differs in different incremental settings. However, joint accuracy has nothing to do with incremental setting? What's more, it seems that joint accuracy with resnet18 on CIFAR100 is about 70. In some other papers such as DER, when saving all old data, the average of all learning steps is about 80, so the acc of the last step is lower than 80. Can you give some explains? Thank you very much! image

arthurdouillard commented 1 year ago

For the joint, I'm reporting DER's metrics. Which indeed should have been the same, but i choose to report those for consistency.

I'm not sure about you mean in the second sentence. There is no "avg" for joint model, as it does only one phase. DER puts it as "avg" but it's a mistake from them.