Published April 1, 2023 | Version v1
Journal article Restricted

Mitigating Robust Overfitting via Self-Residual-Calibration Regularization

  • 1. National Institute of Informatics, Tokyo, 101-8430, Japan
  • 2. University of Trento, Italy

Description

Overfitting in adversarial training has attracted the interest of researchers in the community of artificial intelligence and machine learning in recent years. To address this issue, in this paper we begin by evaluating the defense performances of several calibration methods on various robust models. Our analysis and experiments reveal two intriguing properties: 1) a well-calibrated robust model is decreasing the confidence of robust model; 2) there is a trade-off between the confidences of natural and adversarial images. These new properties offer a straightforward insight into designing a simple but effective regularization, called Self-Residual-Calibration (SRC). The proposed SRC calculates the absolute residual between adversarial and natural logit features corresponding to the ground-truth labels. Furthermore, we utilize the pinball loss to minimize the quantile residual between them, resulting in more robust regularization. Extensive experiments indicate that our SRC can effectively mitigate the overfitting problem while improving the robustness of state-of-the-art models. Importantly, SRC is complementary to various regularization methods. When combined with them, we are capable of achieving the top-rank performance on the AutoAttack benchmark leaderboard.

Files

Restricted

The record is publicly accessible, but files are restricted to users with access.

Additional details

Funding

AI4Media – A European Excellence Centre for Media, Society and Democracy 951911
European Commission