Balance, Imbalance, and Rebalance: Understanding Robust Overfitting from a Minimax Game Perspective

October 30, 2023 ยท Entered Twilight ยท ๐Ÿ› Neural Information Processing Systems

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: README.md, attacks.py, generate_validation.py, networks, train_cifar_wa.py, utils.py

Authors Yifei Wang, Liangchen Li, Jiansheng Yang, Zhouchen Lin, Yisen Wang arXiv ID 2310.19360 Category cs.LG: Machine Learning Cross-listed cs.AI, cs.CV, stat.ML Citations 19 Venue Neural Information Processing Systems Repository https://github.com/PKU-ML/ReBAT โญ 7 Last Checked 2 months ago
Abstract
Adversarial Training (AT) has become arguably the state-of-the-art algorithm for extracting robust features. However, researchers recently notice that AT suffers from severe robust overfitting problems, particularly after learning rate (LR) decay. In this paper, we explain this phenomenon by viewing adversarial training as a dynamic minimax game between the model trainer and the attacker. Specifically, we analyze how LR decay breaks the balance between the minimax game by empowering the trainer with a stronger memorization ability, and show such imbalance induces robust overfitting as a result of memorizing non-robust features. We validate this understanding with extensive experiments, and provide a holistic view of robust overfitting from the dynamics of both the two game players. This understanding further inspires us to alleviate robust overfitting by rebalancing the two players by either regularizing the trainer's capacity or improving the attack strength. Experiments show that the proposed ReBalanced Adversarial Training (ReBAT) can attain good robustness and does not suffer from robust overfitting even after very long training. Code is available at https://github.com/PKU-ML/ReBAT.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning