Cargando…

Clustering Approach for Detecting Multiple Types of Adversarial Examples

With intentional feature perturbations to a deep learning model, the adversary generates an adversarial example to deceive the deep learning model. As an adversarial example has recently been considered in the most severe problem of deep learning technology, its defense methods have been actively st...

Descripción completa

Detalles Bibliográficos
Autores principales: Choi, Seok-Hwan, Bahk, Tae-u, Ahn, Sungyong, Choi, Yoon-Ho
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9146128/
https://www.ncbi.nlm.nih.gov/pubmed/35632235
http://dx.doi.org/10.3390/s22103826
_version_ 1784716484353720320
author Choi, Seok-Hwan
Bahk, Tae-u
Ahn, Sungyong
Choi, Yoon-Ho
author_facet Choi, Seok-Hwan
Bahk, Tae-u
Ahn, Sungyong
Choi, Yoon-Ho
author_sort Choi, Seok-Hwan
collection PubMed
description With intentional feature perturbations to a deep learning model, the adversary generates an adversarial example to deceive the deep learning model. As an adversarial example has recently been considered in the most severe problem of deep learning technology, its defense methods have been actively studied. Such effective defense methods against adversarial examples are categorized into one of the three architectures: (1) model retraining architecture; (2) input transformation architecture; and (3) adversarial example detection architecture. Especially, defense methods using adversarial example detection architecture have been actively studied. This is because defense methods using adversarial example detection architecture do not make wrong decisions for the legitimate input data while others do. In this paper, we note that current defense methods using adversarial example detection architecture can classify the input data into only either a legitimate one or an adversarial one. That is, the current defense methods using adversarial example detection architecture can only detect the adversarial examples and cannot classify the input data into multiple classes of data, i.e., legitimate input data and various types of adversarial examples. To classify the input data into multiple classes of data while increasing the accuracy of the clustering model, we propose an advanced defense method using adversarial example detection architecture, which extracts the key features from the input data and feeds the extracted features into a clustering model. From the experimental results under various application datasets, we show that the proposed method can detect the adversarial examples while classifying the types of adversarial examples. We also show that the accuracy of the proposed method outperforms the accuracy of recent defense methods using adversarial example detection architecture.
format Online
Article
Text
id pubmed-9146128
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-91461282022-05-29 Clustering Approach for Detecting Multiple Types of Adversarial Examples Choi, Seok-Hwan Bahk, Tae-u Ahn, Sungyong Choi, Yoon-Ho Sensors (Basel) Article With intentional feature perturbations to a deep learning model, the adversary generates an adversarial example to deceive the deep learning model. As an adversarial example has recently been considered in the most severe problem of deep learning technology, its defense methods have been actively studied. Such effective defense methods against adversarial examples are categorized into one of the three architectures: (1) model retraining architecture; (2) input transformation architecture; and (3) adversarial example detection architecture. Especially, defense methods using adversarial example detection architecture have been actively studied. This is because defense methods using adversarial example detection architecture do not make wrong decisions for the legitimate input data while others do. In this paper, we note that current defense methods using adversarial example detection architecture can classify the input data into only either a legitimate one or an adversarial one. That is, the current defense methods using adversarial example detection architecture can only detect the adversarial examples and cannot classify the input data into multiple classes of data, i.e., legitimate input data and various types of adversarial examples. To classify the input data into multiple classes of data while increasing the accuracy of the clustering model, we propose an advanced defense method using adversarial example detection architecture, which extracts the key features from the input data and feeds the extracted features into a clustering model. From the experimental results under various application datasets, we show that the proposed method can detect the adversarial examples while classifying the types of adversarial examples. We also show that the accuracy of the proposed method outperforms the accuracy of recent defense methods using adversarial example detection architecture. MDPI 2022-05-18 /pmc/articles/PMC9146128/ /pubmed/35632235 http://dx.doi.org/10.3390/s22103826 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Choi, Seok-Hwan
Bahk, Tae-u
Ahn, Sungyong
Choi, Yoon-Ho
Clustering Approach for Detecting Multiple Types of Adversarial Examples
title Clustering Approach for Detecting Multiple Types of Adversarial Examples
title_full Clustering Approach for Detecting Multiple Types of Adversarial Examples
title_fullStr Clustering Approach for Detecting Multiple Types of Adversarial Examples
title_full_unstemmed Clustering Approach for Detecting Multiple Types of Adversarial Examples
title_short Clustering Approach for Detecting Multiple Types of Adversarial Examples
title_sort clustering approach for detecting multiple types of adversarial examples
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9146128/
https://www.ncbi.nlm.nih.gov/pubmed/35632235
http://dx.doi.org/10.3390/s22103826
work_keys_str_mv AT choiseokhwan clusteringapproachfordetectingmultipletypesofadversarialexamples
AT bahktaeu clusteringapproachfordetectingmultipletypesofadversarialexamples
AT ahnsungyong clusteringapproachfordetectingmultipletypesofadversarialexamples
AT choiyoonho clusteringapproachfordetectingmultipletypesofadversarialexamples