Cargando…
Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
Deep neural networks (DNNs) have been widely adopted in many fields, and they greatly promote the Internet of Health Things (IoHT) systems by mining health-related information. However, recent studies have shown the serious threat to DNN-based systems posed by adversarial attacks, which has raised w...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9977516/ https://www.ncbi.nlm.nih.gov/pubmed/36875749 http://dx.doi.org/10.1155/2023/1177635 |
_version_ | 1784899308834783232 |
---|---|
author | Zhang, Huan Tan, Hao Zhu, Bin Wang, Le Shafiq, Muhammad Gu, Zhaoquan |
author_facet | Zhang, Huan Tan, Hao Zhu, Bin Wang, Le Shafiq, Muhammad Gu, Zhaoquan |
author_sort | Zhang, Huan |
collection | PubMed |
description | Deep neural networks (DNNs) have been widely adopted in many fields, and they greatly promote the Internet of Health Things (IoHT) systems by mining health-related information. However, recent studies have shown the serious threat to DNN-based systems posed by adversarial attacks, which has raised widespread concerns. Attackers maliciously craft adversarial examples (AEs) and blend them into the normal examples (NEs) to fool the DNN models, which seriously affects the analysis results of the IoHT systems. Text data is a common form in such systems, such as the patients' medical records and prescriptions, and we study the security concerns of the DNNs for textural analysis. As identifying and correcting AEs in discrete textual representations is extremely challenging, the available detection techniques are still limited in performance and generalizability, especially in IoHT systems. In this paper, we propose an efficient and structure-free adversarial detection method, which detects AEs even in attack-unknown and model-agnostic circumstances. We reveal that sensitivity inconsistency prevails between AEs and NEs, leading them to react differently when important words in the text are perturbed. This discovery motivates us to design an adversarial detector based on adversarial features, which are extracted based on sensitivity inconsistency. Since the proposed detector is structure-free, it can be directly deployed in off-the-shelf applications without modifying the target models. Compared to the state-of-the-art detection methods, our proposed method improves adversarial detection performance, with an adversarial recall of up to 99.7% and an F1-score of up to 97.8%. In addition, extensive experiments have shown that our method achieves superior generalizability as it can be generalized across different attackers, models, and tasks. |
format | Online Article Text |
id | pubmed-9977516 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-99775162023-03-02 Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems Zhang, Huan Tan, Hao Zhu, Bin Wang, Le Shafiq, Muhammad Gu, Zhaoquan J Healthc Eng Research Article Deep neural networks (DNNs) have been widely adopted in many fields, and they greatly promote the Internet of Health Things (IoHT) systems by mining health-related information. However, recent studies have shown the serious threat to DNN-based systems posed by adversarial attacks, which has raised widespread concerns. Attackers maliciously craft adversarial examples (AEs) and blend them into the normal examples (NEs) to fool the DNN models, which seriously affects the analysis results of the IoHT systems. Text data is a common form in such systems, such as the patients' medical records and prescriptions, and we study the security concerns of the DNNs for textural analysis. As identifying and correcting AEs in discrete textual representations is extremely challenging, the available detection techniques are still limited in performance and generalizability, especially in IoHT systems. In this paper, we propose an efficient and structure-free adversarial detection method, which detects AEs even in attack-unknown and model-agnostic circumstances. We reveal that sensitivity inconsistency prevails between AEs and NEs, leading them to react differently when important words in the text are perturbed. This discovery motivates us to design an adversarial detector based on adversarial features, which are extracted based on sensitivity inconsistency. Since the proposed detector is structure-free, it can be directly deployed in off-the-shelf applications without modifying the target models. Compared to the state-of-the-art detection methods, our proposed method improves adversarial detection performance, with an adversarial recall of up to 99.7% and an F1-score of up to 97.8%. In addition, extensive experiments have shown that our method achieves superior generalizability as it can be generalized across different attackers, models, and tasks. Hindawi 2023-02-22 /pmc/articles/PMC9977516/ /pubmed/36875749 http://dx.doi.org/10.1155/2023/1177635 Text en Copyright © 2023 Huan Zhang et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Zhang, Huan Tan, Hao Zhu, Bin Wang, Le Shafiq, Muhammad Gu, Zhaoquan Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems |
title | Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems |
title_full | Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems |
title_fullStr | Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems |
title_full_unstemmed | Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems |
title_short | Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems |
title_sort | learning to discriminate adversarial examples by sensitivity inconsistency in ioht systems |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9977516/ https://www.ncbi.nlm.nih.gov/pubmed/36875749 http://dx.doi.org/10.1155/2023/1177635 |
work_keys_str_mv | AT zhanghuan learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems AT tanhao learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems AT zhubin learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems AT wangle learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems AT shafiqmuhammad learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems AT guzhaoquan learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems |