Cargando…
Clustering Approach for Detecting Multiple Types of Adversarial Examples
With intentional feature perturbations to a deep learning model, the adversary generates an adversarial example to deceive the deep learning model. As an adversarial example has recently been considered in the most severe problem of deep learning technology, its defense methods have been actively st...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9146128/ https://www.ncbi.nlm.nih.gov/pubmed/35632235 http://dx.doi.org/10.3390/s22103826 |
_version_ | 1784716484353720320 |
---|---|
author | Choi, Seok-Hwan Bahk, Tae-u Ahn, Sungyong Choi, Yoon-Ho |
author_facet | Choi, Seok-Hwan Bahk, Tae-u Ahn, Sungyong Choi, Yoon-Ho |
author_sort | Choi, Seok-Hwan |
collection | PubMed |
description | With intentional feature perturbations to a deep learning model, the adversary generates an adversarial example to deceive the deep learning model. As an adversarial example has recently been considered in the most severe problem of deep learning technology, its defense methods have been actively studied. Such effective defense methods against adversarial examples are categorized into one of the three architectures: (1) model retraining architecture; (2) input transformation architecture; and (3) adversarial example detection architecture. Especially, defense methods using adversarial example detection architecture have been actively studied. This is because defense methods using adversarial example detection architecture do not make wrong decisions for the legitimate input data while others do. In this paper, we note that current defense methods using adversarial example detection architecture can classify the input data into only either a legitimate one or an adversarial one. That is, the current defense methods using adversarial example detection architecture can only detect the adversarial examples and cannot classify the input data into multiple classes of data, i.e., legitimate input data and various types of adversarial examples. To classify the input data into multiple classes of data while increasing the accuracy of the clustering model, we propose an advanced defense method using adversarial example detection architecture, which extracts the key features from the input data and feeds the extracted features into a clustering model. From the experimental results under various application datasets, we show that the proposed method can detect the adversarial examples while classifying the types of adversarial examples. We also show that the accuracy of the proposed method outperforms the accuracy of recent defense methods using adversarial example detection architecture. |
format | Online Article Text |
id | pubmed-9146128 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-91461282022-05-29 Clustering Approach for Detecting Multiple Types of Adversarial Examples Choi, Seok-Hwan Bahk, Tae-u Ahn, Sungyong Choi, Yoon-Ho Sensors (Basel) Article With intentional feature perturbations to a deep learning model, the adversary generates an adversarial example to deceive the deep learning model. As an adversarial example has recently been considered in the most severe problem of deep learning technology, its defense methods have been actively studied. Such effective defense methods against adversarial examples are categorized into one of the three architectures: (1) model retraining architecture; (2) input transformation architecture; and (3) adversarial example detection architecture. Especially, defense methods using adversarial example detection architecture have been actively studied. This is because defense methods using adversarial example detection architecture do not make wrong decisions for the legitimate input data while others do. In this paper, we note that current defense methods using adversarial example detection architecture can classify the input data into only either a legitimate one or an adversarial one. That is, the current defense methods using adversarial example detection architecture can only detect the adversarial examples and cannot classify the input data into multiple classes of data, i.e., legitimate input data and various types of adversarial examples. To classify the input data into multiple classes of data while increasing the accuracy of the clustering model, we propose an advanced defense method using adversarial example detection architecture, which extracts the key features from the input data and feeds the extracted features into a clustering model. From the experimental results under various application datasets, we show that the proposed method can detect the adversarial examples while classifying the types of adversarial examples. We also show that the accuracy of the proposed method outperforms the accuracy of recent defense methods using adversarial example detection architecture. MDPI 2022-05-18 /pmc/articles/PMC9146128/ /pubmed/35632235 http://dx.doi.org/10.3390/s22103826 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Choi, Seok-Hwan Bahk, Tae-u Ahn, Sungyong Choi, Yoon-Ho Clustering Approach for Detecting Multiple Types of Adversarial Examples |
title | Clustering Approach for Detecting Multiple Types of Adversarial Examples |
title_full | Clustering Approach for Detecting Multiple Types of Adversarial Examples |
title_fullStr | Clustering Approach for Detecting Multiple Types of Adversarial Examples |
title_full_unstemmed | Clustering Approach for Detecting Multiple Types of Adversarial Examples |
title_short | Clustering Approach for Detecting Multiple Types of Adversarial Examples |
title_sort | clustering approach for detecting multiple types of adversarial examples |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9146128/ https://www.ncbi.nlm.nih.gov/pubmed/35632235 http://dx.doi.org/10.3390/s22103826 |
work_keys_str_mv | AT choiseokhwan clusteringapproachfordetectingmultipletypesofadversarialexamples AT bahktaeu clusteringapproachfordetectingmultipletypesofadversarialexamples AT ahnsungyong clusteringapproachfordetectingmultipletypesofadversarialexamples AT choiyoonho clusteringapproachfordetectingmultipletypesofadversarialexamples |