Cargando…
Approaching Adversarial Example Classification with Chaos Theory
Adversarial examples are one of the most intriguing topics in modern deep learning. Imperceptible perturbations to the input can fool robust models. In relation to this problem, attack and defense methods are being developed almost on a daily basis. In parallel, efforts are being made to simply poin...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7712112/ https://www.ncbi.nlm.nih.gov/pubmed/33286969 http://dx.doi.org/10.3390/e22111201 |
_version_ | 1783618298112376832 |
---|---|
author | Pedraza, Anibal Deniz, Oscar Bueno, Gloria |
author_facet | Pedraza, Anibal Deniz, Oscar Bueno, Gloria |
author_sort | Pedraza, Anibal |
collection | PubMed |
description | Adversarial examples are one of the most intriguing topics in modern deep learning. Imperceptible perturbations to the input can fool robust models. In relation to this problem, attack and defense methods are being developed almost on a daily basis. In parallel, efforts are being made to simply pointing out when an input image is an adversarial example. This can help prevent potential issues, as the failure cases are easily recognizable by humans. The proposal in this work is to study how chaos theory methods can help distinguish adversarial examples from regular images. Our work is based on the assumption that deep networks behave as chaotic systems, and adversarial examples are the main manifestation of it (in the sense that a slight input variation produces a totally different output). In our experiments, we show that the Lyapunov exponents (an established measure of chaoticity), which have been recently proposed for classification of adversarial examples, are not robust to image processing transformations that alter image entropy. Furthermore, we show that entropy can complement Lyapunov exponents in such a way that the discriminating power is significantly enhanced. The proposed method achieves 65% to 100% accuracy detecting adversarials with a wide range of attacks (for example: CW, PGD, Spatial, HopSkip) for the MNIST dataset, with similar results when entropy-changing image processing methods (such as Equalization, Speckle and Gaussian noise) are applied. This is also corroborated with two other datasets, Fashion-MNIST and CIFAR 19. These results indicate that classifiers can enhance their robustness against the adversarial phenomenon, being applied in a wide variety of conditions that potentially matches real world cases and also other threatening scenarios. |
format | Online Article Text |
id | pubmed-7712112 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-77121122021-02-24 Approaching Adversarial Example Classification with Chaos Theory Pedraza, Anibal Deniz, Oscar Bueno, Gloria Entropy (Basel) Article Adversarial examples are one of the most intriguing topics in modern deep learning. Imperceptible perturbations to the input can fool robust models. In relation to this problem, attack and defense methods are being developed almost on a daily basis. In parallel, efforts are being made to simply pointing out when an input image is an adversarial example. This can help prevent potential issues, as the failure cases are easily recognizable by humans. The proposal in this work is to study how chaos theory methods can help distinguish adversarial examples from regular images. Our work is based on the assumption that deep networks behave as chaotic systems, and adversarial examples are the main manifestation of it (in the sense that a slight input variation produces a totally different output). In our experiments, we show that the Lyapunov exponents (an established measure of chaoticity), which have been recently proposed for classification of adversarial examples, are not robust to image processing transformations that alter image entropy. Furthermore, we show that entropy can complement Lyapunov exponents in such a way that the discriminating power is significantly enhanced. The proposed method achieves 65% to 100% accuracy detecting adversarials with a wide range of attacks (for example: CW, PGD, Spatial, HopSkip) for the MNIST dataset, with similar results when entropy-changing image processing methods (such as Equalization, Speckle and Gaussian noise) are applied. This is also corroborated with two other datasets, Fashion-MNIST and CIFAR 19. These results indicate that classifiers can enhance their robustness against the adversarial phenomenon, being applied in a wide variety of conditions that potentially matches real world cases and also other threatening scenarios. MDPI 2020-10-24 /pmc/articles/PMC7712112/ /pubmed/33286969 http://dx.doi.org/10.3390/e22111201 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Pedraza, Anibal Deniz, Oscar Bueno, Gloria Approaching Adversarial Example Classification with Chaos Theory |
title | Approaching Adversarial Example Classification with Chaos Theory |
title_full | Approaching Adversarial Example Classification with Chaos Theory |
title_fullStr | Approaching Adversarial Example Classification with Chaos Theory |
title_full_unstemmed | Approaching Adversarial Example Classification with Chaos Theory |
title_short | Approaching Adversarial Example Classification with Chaos Theory |
title_sort | approaching adversarial example classification with chaos theory |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7712112/ https://www.ncbi.nlm.nih.gov/pubmed/33286969 http://dx.doi.org/10.3390/e22111201 |
work_keys_str_mv | AT pedrazaanibal approachingadversarialexampleclassificationwithchaostheory AT denizoscar approachingadversarialexampleclassificationwithchaostheory AT buenogloria approachingadversarialexampleclassificationwithchaostheory |