RobustBench / robustbench

RobustBench: a standardized adversarial robustness benchmark [NeurIPS 2021 Benchmarks and Datasets Track]
https://robustbench.github.io
Other
669 stars 99 forks source link

[New Model] Chen2024Data #172

Closed CNOCycle closed 8 months ago

CNOCycle commented 8 months ago

Paper Information

Leaderboard Claim(s)

Model 1

Model 2

Model 3

Model 4

Model Zoo:

fra31 commented 8 months ago

Hi,

thanks for the submission. I was adding the models (see https://github.com/RobustBench/robustbench/commit/31b7aa166b9d41eac9a42915e0fcfdeda383e8a7), and for the ImageNet one, it seems that the checkpoint corresponds to a WideResNet-50-2, is that right? Also, with the Res256Crop224 pre-processing I get 68.76% accuracy: are you maybe using another one?

CNOCycle commented 8 months ago

Hi.

Thank for your reminder. After checking the code for imagenet evaluation, the model architecture I used was the ResNet architecture provided by Robustbench but the width was set to width_per_group=64*2 manually. It should be WideResNet-50-2. Also, the pre-processing was Res256Crop224.

Thank you for clarifying that for me.

fra31 commented 8 months ago

I still get 1% lower clean accuracy that reported. It might also have to do with different torchvision versions.

CNOCycle commented 8 months ago

I think that the subset I used for the evaluation might be different than the standard set provided by RobustBench. Besides, the evaluation was conducted almost two years ago, the version of AA or RobustBench I used seems to be older version. Variations (2%) in accuracy are acceptable.

fra31 commented 8 months ago

Added the models with https://github.com/RobustBench/robustbench/pull/176, please let me know if there's anything to change.

CNOCycle commented 8 months ago

I greatly appreciate your assistance. Those commits appear satisfactory to me.