-
Notifications
You must be signed in to change notification settings - Fork 113
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fixed evaluation of models with random defenses #105
base: master
Are you sure you want to change the base?
Conversation
@fra31 Could you please review this pr? |
您好,您的邮件已收到!
|
In Appendix L of our paper, we provide a detailed report on our fix for AutoAttack and its impact. We encourage future research to adopt this updated version when evaluating models with randomness, as it effectively reduces the risk of overestimating robustness. If you find our work useful for your research, please consider citing it: @Article{liu2024towards, |
您好,您的邮件已收到!
|
Good news! Our paper "Towards Understanding the Robustness of Diffusion-Based Purification: A Stochastic Perspective" has been accepted by ICLR2025. If it is helpful to your research, welcome to cite our paper! @inproceedings{ |
Thank you for your outstanding contributions.
@LYMDLUT and I put forward this PR to improve the evaluation of models with random defenses.
We've noticed that AutoAttack's current strategy for selecting the final output (clean/APGD etc) based on one time evaluation, regardless of whether the target models implement random defenses or not. This overlooks the variability of outputs in models with random defenses.
Relying on a single evaluation to filter samples for subsequent attacks leads to inflated success rate and hinders the exploration of attack methods that could potentially yield superior outcomes.
To address this, we propose to perform multiple time evaluations for models with random defenses and chose the adversarial example with the highest robustness as final output.