Cargando…

Minimum Adversarial Examples

Deep neural networks in the area of information security are facing a severe threat from adversarial examples (AEs). Existing methods of AE generation use two optimization models: (1) taking the successful attack as the objective function and limiting perturbations as the constraint; (2) taking the...

Descripción completa

Detalles Bibliográficos
Autores principales: Du, Zhenyu, Liu, Fangzheng, Yan, Xuehu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8947511/
https://www.ncbi.nlm.nih.gov/pubmed/35327907
http://dx.doi.org/10.3390/e24030396
_version_ 1784674456846729216
author Du, Zhenyu
Liu, Fangzheng
Yan, Xuehu
author_facet Du, Zhenyu
Liu, Fangzheng
Yan, Xuehu
author_sort Du, Zhenyu
collection PubMed
description Deep neural networks in the area of information security are facing a severe threat from adversarial examples (AEs). Existing methods of AE generation use two optimization models: (1) taking the successful attack as the objective function and limiting perturbations as the constraint; (2) taking the minimum of adversarial perturbations as the target and the successful attack as the constraint. These all involve two fundamental problems of AEs: the minimum boundary of constructing the AEs and whether that boundary is reachable. The reachability means whether the AEs of successful attack models exist equal to that boundary. Previous optimization models have no complete answer to the problems. Therefore, in this paper, for the first problem, we propose the definition of the minimum AEs and give the theoretical lower bound of the amplitude of the minimum AEs. For the second problem, we prove that solving the generation of the minimum AEs is an NPC problem, and then based on its computational inaccessibility, we establish a new third optimization model. This model is general and can adapt to any constraint. To verify the model, we devise two specific methods for generating controllable AEs under the widely used distance evaluation standard of adversarial perturbations, namely [Formula: see text] constraint and [Formula: see text] constraint (structural similarity). This model limits the amplitude of the AEs, reduces the solution space’s search cost, and is further improved in efficiency. In theory, those AEs generated by the new model which are closer to the actual minimum adversarial boundary overcome the blindness of the adversarial amplitude setting of the existing methods and further improve the attack success rate. In addition, this model can generate accurate AEs with controllable amplitude under different constraints, which is suitable for different application scenarios. In addition, through extensive experiments, they demonstrate a better attack ability under the same constraints as other baseline attacks. For all the datasets we test in the experiment, compared with other baseline methods, the attack success rate of our method is improved by approximately 10%.
format Online
Article
Text
id pubmed-8947511
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89475112022-03-25 Minimum Adversarial Examples Du, Zhenyu Liu, Fangzheng Yan, Xuehu Entropy (Basel) Article Deep neural networks in the area of information security are facing a severe threat from adversarial examples (AEs). Existing methods of AE generation use two optimization models: (1) taking the successful attack as the objective function and limiting perturbations as the constraint; (2) taking the minimum of adversarial perturbations as the target and the successful attack as the constraint. These all involve two fundamental problems of AEs: the minimum boundary of constructing the AEs and whether that boundary is reachable. The reachability means whether the AEs of successful attack models exist equal to that boundary. Previous optimization models have no complete answer to the problems. Therefore, in this paper, for the first problem, we propose the definition of the minimum AEs and give the theoretical lower bound of the amplitude of the minimum AEs. For the second problem, we prove that solving the generation of the minimum AEs is an NPC problem, and then based on its computational inaccessibility, we establish a new third optimization model. This model is general and can adapt to any constraint. To verify the model, we devise two specific methods for generating controllable AEs under the widely used distance evaluation standard of adversarial perturbations, namely [Formula: see text] constraint and [Formula: see text] constraint (structural similarity). This model limits the amplitude of the AEs, reduces the solution space’s search cost, and is further improved in efficiency. In theory, those AEs generated by the new model which are closer to the actual minimum adversarial boundary overcome the blindness of the adversarial amplitude setting of the existing methods and further improve the attack success rate. In addition, this model can generate accurate AEs with controllable amplitude under different constraints, which is suitable for different application scenarios. In addition, through extensive experiments, they demonstrate a better attack ability under the same constraints as other baseline attacks. For all the datasets we test in the experiment, compared with other baseline methods, the attack success rate of our method is improved by approximately 10%. MDPI 2022-03-12 /pmc/articles/PMC8947511/ /pubmed/35327907 http://dx.doi.org/10.3390/e24030396 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Du, Zhenyu
Liu, Fangzheng
Yan, Xuehu
Minimum Adversarial Examples
title Minimum Adversarial Examples
title_full Minimum Adversarial Examples
title_fullStr Minimum Adversarial Examples
title_full_unstemmed Minimum Adversarial Examples
title_short Minimum Adversarial Examples
title_sort minimum adversarial examples
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8947511/
https://www.ncbi.nlm.nih.gov/pubmed/35327907
http://dx.doi.org/10.3390/e24030396
work_keys_str_mv AT duzhenyu minimumadversarialexamples
AT liufangzheng minimumadversarialexamples
AT yanxuehu minimumadversarialexamples