Open Phd-man opened 5 days ago
Thank you for your kind words! I’m really glad you’re finding the repository helpful. 😊
Regarding your concern about clean accuracy during adversarial training, here are a few suggestions that might help strike a better balance between clean and robust accuracy:
Let me know if you need further clarification or assistance with implementation details—I’d be happy to help!
I’ll definitely give them a try. Specifically, I’m using the ViT-B network—it’s a bit time-consuming, but I’ll work through it.
I’ll definitely give them a try. Specifically, I’m using the ViT-B network—it’s a bit time-consuming, but I’ll work through it.
You can reference this paper: When Adversarial Training Meets Vision Transformers: Recipes from Training to Architecture[NeurIPS-22]
First off, I just want to say thank you for your amazing work on this repository! I've been learning a lot from it.
I’ve noticed that no matter which method of adversarial training I use, the clean accuracy tends to be lower compared to non-adversarial scenarios. Since my project requires high clean accuracy, I was wondering if you have any tips or techniques for improving this, or if there are specific hyperparameters I should focus on tuning.