Cargando…

Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model

Deep neural networks (DNNs) have been shown to be susceptible to critical vulnerabilities when attacked by adversarial samples. This has prompted the development of attack and defense strategies similar to those used in cyberspace security. The dependence of such strategies on attack and defense mec...

Descripción completa

Detalles Bibliográficos
Autores principales: Qin, Ruoxi, Wang, Linyuan, Du, Xuehui, Xie, Pengfei, Chen, Xingyuan, Yan, Bin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10442534/
https://www.ncbi.nlm.nih.gov/pubmed/37614968
http://dx.doi.org/10.3389/fnbot.2023.1205370
_version_ 1785093621294301184
author Qin, Ruoxi
Wang, Linyuan
Du, Xuehui
Xie, Pengfei
Chen, Xingyuan
Yan, Bin
author_facet Qin, Ruoxi
Wang, Linyuan
Du, Xuehui
Xie, Pengfei
Chen, Xingyuan
Yan, Bin
author_sort Qin, Ruoxi
collection PubMed
description Deep neural networks (DNNs) have been shown to be susceptible to critical vulnerabilities when attacked by adversarial samples. This has prompted the development of attack and defense strategies similar to those used in cyberspace security. The dependence of such strategies on attack and defense mechanisms makes the associated algorithms on both sides appear as closely processes, with the defense method being particularly passive in these processes. Inspired by the dynamic defense approach proposed in cyberspace to address endless arm races, this article defines ensemble quantity, network structure, and smoothing parameters as variable ensemble attributes and proposes a stochastic ensemble strategy based on heterogeneous and redundant sub-models. The proposed method introduces the diversity and randomness characteristic of deep neural networks to alter the fixed correspondence gradient between input and output. The unpredictability and diversity of the gradients make it more difficult for attackers to directly implement white-box attacks, helping to address the extreme transferability and vulnerability of ensemble models under white-box attacks. Experimental comparison of ASR-vs.-distortion curves with different attack scenarios under CIFAR10 preliminarily demonstrates the effectiveness of the proposed method that even the highest-capacity attacker cannot easily outperform the attack success rate associated with the ensemble smoothed model, especially for untargeted attacks.
format Online
Article
Text
id pubmed-10442534
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-104425342023-08-23 Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model Qin, Ruoxi Wang, Linyuan Du, Xuehui Xie, Pengfei Chen, Xingyuan Yan, Bin Front Neurorobot Neuroscience Deep neural networks (DNNs) have been shown to be susceptible to critical vulnerabilities when attacked by adversarial samples. This has prompted the development of attack and defense strategies similar to those used in cyberspace security. The dependence of such strategies on attack and defense mechanisms makes the associated algorithms on both sides appear as closely processes, with the defense method being particularly passive in these processes. Inspired by the dynamic defense approach proposed in cyberspace to address endless arm races, this article defines ensemble quantity, network structure, and smoothing parameters as variable ensemble attributes and proposes a stochastic ensemble strategy based on heterogeneous and redundant sub-models. The proposed method introduces the diversity and randomness characteristic of deep neural networks to alter the fixed correspondence gradient between input and output. The unpredictability and diversity of the gradients make it more difficult for attackers to directly implement white-box attacks, helping to address the extreme transferability and vulnerability of ensemble models under white-box attacks. Experimental comparison of ASR-vs.-distortion curves with different attack scenarios under CIFAR10 preliminarily demonstrates the effectiveness of the proposed method that even the highest-capacity attacker cannot easily outperform the attack success rate associated with the ensemble smoothed model, especially for untargeted attacks. Frontiers Media S.A. 2023-08-08 /pmc/articles/PMC10442534/ /pubmed/37614968 http://dx.doi.org/10.3389/fnbot.2023.1205370 Text en Copyright © 2023 Qin, Wang, Du, Xie, Chen and Yan. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Qin, Ruoxi
Wang, Linyuan
Du, Xuehui
Xie, Pengfei
Chen, Xingyuan
Yan, Bin
Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
title Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
title_full Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
title_fullStr Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
title_full_unstemmed Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
title_short Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
title_sort adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10442534/
https://www.ncbi.nlm.nih.gov/pubmed/37614968
http://dx.doi.org/10.3389/fnbot.2023.1205370
work_keys_str_mv AT qinruoxi adversarialrobustnessindeepneuralnetworksbasedonvariableattributesofthestochasticensemblemodel
AT wanglinyuan adversarialrobustnessindeepneuralnetworksbasedonvariableattributesofthestochasticensemblemodel
AT duxuehui adversarialrobustnessindeepneuralnetworksbasedonvariableattributesofthestochasticensemblemodel
AT xiepengfei adversarialrobustnessindeepneuralnetworksbasedonvariableattributesofthestochasticensemblemodel
AT chenxingyuan adversarialrobustnessindeepneuralnetworksbasedonvariableattributesofthestochasticensemblemodel
AT yanbin adversarialrobustnessindeepneuralnetworksbasedonvariableattributesofthestochasticensemblemodel