4.5 Article

On the Relationship between Generalization and Robustness to Adversarial Examples

Journal

SYMMETRY-BASEL
Volume 13, Issue 5, Pages -

Publisher

MDPI
DOI: 10.3390/sym13050817

Keywords

machine learning; computer vision; deep learning; adversarial examples; adversarial robustness; overfitting

Funding

  1. Spanish Ministry of Economy and Business [TIN2017-82113-C2-2-R]
  2. Autonomous Government of Castilla-La Mancha [SBPLY/17/180501/000543]
  3. Spanish Ministry of Science, Innovation, and Universities [FPU17/04758]

Ask authors/readers for more resources

Adversarial examples are visually equivalent to normal inputs but lead neural networks to output incorrect results, researchers are working on methods for generating and defending against them. The study focuses on characterizing this phenomenon and finds that there is greater robustness to adversarial examples in an overfitting regime. This loss of generalization and gain in robustness is explained as a manifestation of the fitting-generalization trade-off.
One of the most intriguing phenomenons related to deep learning is the so-called adversarial examples. These samples are visually equivalent to normal inputs, undetectable for humans, yet they cause the networks to output wrong results. The phenomenon can be framed as a symmetry/asymmetry problem, whereby inputs to a neural network with a similar/symmetric appearance to regular images, produce an opposite/asymmetric output. Some researchers are focused on developing methods for generating adversarial examples, while others propose defense methods. In parallel, there is a growing interest in characterizing the phenomenon, which is also the focus of this paper. From some well known datasets of common images, like CIFAR-10 and STL-10, a neural network architecture is first trained in a normal regime, where training and validation performances increase, reaching generalization. Additionally, the same architectures and datasets are trained in an overfitting regime, where there is a growing disparity in training and validation performances. The behaviour of these two regimes against adversarial examples is then compared. From the results, we observe greater robustness to adversarial examples in the overfitting regime. We explain this simultaneous loss of generalization and gain in robustness to adversarial examples as another manifestation of the well-known fitting-generalization trade-off.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available