Adversarial robustness for latent models: Revisiting the robust-standard accuracies tradeoff

From MaRDI portal
Publication:6381065

arXiv2110.11950MaRDI QIDQ6381065

Author name not available (Why is that?)

Publication date: 22 October 2021

Abstract: Over the past few years, several adversarial training methods have been proposed to improve the robustness of machine learning models against adversarial perturbations in the input. Despite remarkable progress in this regard, adversarial training is often observed to drop the standard test accuracy. This phenomenon has intrigued the research community to investigate the potential tradeoff between standard accuracy (a.k.a generalization) and robust accuracy (a.k.a robust generalization) as two performance measures. In this paper, we revisit this tradeoff for latent models and argue that this tradeoff is mitigated when the data enjoys a low-dimensional structure. In particular, we consider binary classification under two data generative models, namely Gaussian mixture model and generalized linear model, where the features data lie on a low-dimensional manifold. We develop a theory to show that the low-dimensional manifold structure allows one to obtain models that are nearly optimal with respect to both, the standard accuracy and the robust accuracy measures. We further corroborate our theory with several numerical experiments, including Mixture of Factor Analyzers (MFA) model trained on the MNIST dataset.




Has companion code repository: https://github.com/cleverhans-lab/cleverhans








This page was built for publication: Adversarial robustness for latent models: Revisiting the robust-standard accuracies tradeoff

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6381065)