Cargando…

Approaching Adversarial Example Classification with Chaos Theory

Adversarial examples are one of the most intriguing topics in modern deep learning. Imperceptible perturbations to the input can fool robust models. In relation to this problem, attack and defense methods are being developed almost on a daily basis. In parallel, efforts are being made to simply poin...

Descripción completa

Detalles Bibliográficos
Autores principales: Pedraza, Anibal, Deniz, Oscar, Bueno, Gloria
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7712112/
https://www.ncbi.nlm.nih.gov/pubmed/33286969
http://dx.doi.org/10.3390/e22111201
_version_ 1783618298112376832
author Pedraza, Anibal
Deniz, Oscar
Bueno, Gloria
author_facet Pedraza, Anibal
Deniz, Oscar
Bueno, Gloria
author_sort Pedraza, Anibal
collection PubMed
description Adversarial examples are one of the most intriguing topics in modern deep learning. Imperceptible perturbations to the input can fool robust models. In relation to this problem, attack and defense methods are being developed almost on a daily basis. In parallel, efforts are being made to simply pointing out when an input image is an adversarial example. This can help prevent potential issues, as the failure cases are easily recognizable by humans. The proposal in this work is to study how chaos theory methods can help distinguish adversarial examples from regular images. Our work is based on the assumption that deep networks behave as chaotic systems, and adversarial examples are the main manifestation of it (in the sense that a slight input variation produces a totally different output). In our experiments, we show that the Lyapunov exponents (an established measure of chaoticity), which have been recently proposed for classification of adversarial examples, are not robust to image processing transformations that alter image entropy. Furthermore, we show that entropy can complement Lyapunov exponents in such a way that the discriminating power is significantly enhanced. The proposed method achieves 65% to 100% accuracy detecting adversarials with a wide range of attacks (for example: CW, PGD, Spatial, HopSkip) for the MNIST dataset, with similar results when entropy-changing image processing methods (such as Equalization, Speckle and Gaussian noise) are applied. This is also corroborated with two other datasets, Fashion-MNIST and CIFAR 19. These results indicate that classifiers can enhance their robustness against the adversarial phenomenon, being applied in a wide variety of conditions that potentially matches real world cases and also other threatening scenarios.
format Online
Article
Text
id pubmed-7712112
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-77121122021-02-24 Approaching Adversarial Example Classification with Chaos Theory Pedraza, Anibal Deniz, Oscar Bueno, Gloria Entropy (Basel) Article Adversarial examples are one of the most intriguing topics in modern deep learning. Imperceptible perturbations to the input can fool robust models. In relation to this problem, attack and defense methods are being developed almost on a daily basis. In parallel, efforts are being made to simply pointing out when an input image is an adversarial example. This can help prevent potential issues, as the failure cases are easily recognizable by humans. The proposal in this work is to study how chaos theory methods can help distinguish adversarial examples from regular images. Our work is based on the assumption that deep networks behave as chaotic systems, and adversarial examples are the main manifestation of it (in the sense that a slight input variation produces a totally different output). In our experiments, we show that the Lyapunov exponents (an established measure of chaoticity), which have been recently proposed for classification of adversarial examples, are not robust to image processing transformations that alter image entropy. Furthermore, we show that entropy can complement Lyapunov exponents in such a way that the discriminating power is significantly enhanced. The proposed method achieves 65% to 100% accuracy detecting adversarials with a wide range of attacks (for example: CW, PGD, Spatial, HopSkip) for the MNIST dataset, with similar results when entropy-changing image processing methods (such as Equalization, Speckle and Gaussian noise) are applied. This is also corroborated with two other datasets, Fashion-MNIST and CIFAR 19. These results indicate that classifiers can enhance their robustness against the adversarial phenomenon, being applied in a wide variety of conditions that potentially matches real world cases and also other threatening scenarios. MDPI 2020-10-24 /pmc/articles/PMC7712112/ /pubmed/33286969 http://dx.doi.org/10.3390/e22111201 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Pedraza, Anibal
Deniz, Oscar
Bueno, Gloria
Approaching Adversarial Example Classification with Chaos Theory
title Approaching Adversarial Example Classification with Chaos Theory
title_full Approaching Adversarial Example Classification with Chaos Theory
title_fullStr Approaching Adversarial Example Classification with Chaos Theory
title_full_unstemmed Approaching Adversarial Example Classification with Chaos Theory
title_short Approaching Adversarial Example Classification with Chaos Theory
title_sort approaching adversarial example classification with chaos theory
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7712112/
https://www.ncbi.nlm.nih.gov/pubmed/33286969
http://dx.doi.org/10.3390/e22111201
work_keys_str_mv AT pedrazaanibal approachingadversarialexampleclassificationwithchaostheory
AT denizoscar approachingadversarialexampleclassificationwithchaostheory
AT buenogloria approachingadversarialexampleclassificationwithchaostheory