Cargando…

Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems

Deep neural networks (DNNs) have been widely adopted in many fields, and they greatly promote the Internet of Health Things (IoHT) systems by mining health-related information. However, recent studies have shown the serious threat to DNN-based systems posed by adversarial attacks, which has raised w...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Huan, Tan, Hao, Zhu, Bin, Wang, Le, Shafiq, Muhammad, Gu, Zhaoquan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9977516/
https://www.ncbi.nlm.nih.gov/pubmed/36875749
http://dx.doi.org/10.1155/2023/1177635
_version_ 1784899308834783232
author Zhang, Huan
Tan, Hao
Zhu, Bin
Wang, Le
Shafiq, Muhammad
Gu, Zhaoquan
author_facet Zhang, Huan
Tan, Hao
Zhu, Bin
Wang, Le
Shafiq, Muhammad
Gu, Zhaoquan
author_sort Zhang, Huan
collection PubMed
description Deep neural networks (DNNs) have been widely adopted in many fields, and they greatly promote the Internet of Health Things (IoHT) systems by mining health-related information. However, recent studies have shown the serious threat to DNN-based systems posed by adversarial attacks, which has raised widespread concerns. Attackers maliciously craft adversarial examples (AEs) and blend them into the normal examples (NEs) to fool the DNN models, which seriously affects the analysis results of the IoHT systems. Text data is a common form in such systems, such as the patients' medical records and prescriptions, and we study the security concerns of the DNNs for textural analysis. As identifying and correcting AEs in discrete textual representations is extremely challenging, the available detection techniques are still limited in performance and generalizability, especially in IoHT systems. In this paper, we propose an efficient and structure-free adversarial detection method, which detects AEs even in attack-unknown and model-agnostic circumstances. We reveal that sensitivity inconsistency prevails between AEs and NEs, leading them to react differently when important words in the text are perturbed. This discovery motivates us to design an adversarial detector based on adversarial features, which are extracted based on sensitivity inconsistency. Since the proposed detector is structure-free, it can be directly deployed in off-the-shelf applications without modifying the target models. Compared to the state-of-the-art detection methods, our proposed method improves adversarial detection performance, with an adversarial recall of up to 99.7% and an F1-score of up to 97.8%. In addition, extensive experiments have shown that our method achieves superior generalizability as it can be generalized across different attackers, models, and tasks.
format Online
Article
Text
id pubmed-9977516
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-99775162023-03-02 Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems Zhang, Huan Tan, Hao Zhu, Bin Wang, Le Shafiq, Muhammad Gu, Zhaoquan J Healthc Eng Research Article Deep neural networks (DNNs) have been widely adopted in many fields, and they greatly promote the Internet of Health Things (IoHT) systems by mining health-related information. However, recent studies have shown the serious threat to DNN-based systems posed by adversarial attacks, which has raised widespread concerns. Attackers maliciously craft adversarial examples (AEs) and blend them into the normal examples (NEs) to fool the DNN models, which seriously affects the analysis results of the IoHT systems. Text data is a common form in such systems, such as the patients' medical records and prescriptions, and we study the security concerns of the DNNs for textural analysis. As identifying and correcting AEs in discrete textual representations is extremely challenging, the available detection techniques are still limited in performance and generalizability, especially in IoHT systems. In this paper, we propose an efficient and structure-free adversarial detection method, which detects AEs even in attack-unknown and model-agnostic circumstances. We reveal that sensitivity inconsistency prevails between AEs and NEs, leading them to react differently when important words in the text are perturbed. This discovery motivates us to design an adversarial detector based on adversarial features, which are extracted based on sensitivity inconsistency. Since the proposed detector is structure-free, it can be directly deployed in off-the-shelf applications without modifying the target models. Compared to the state-of-the-art detection methods, our proposed method improves adversarial detection performance, with an adversarial recall of up to 99.7% and an F1-score of up to 97.8%. In addition, extensive experiments have shown that our method achieves superior generalizability as it can be generalized across different attackers, models, and tasks. Hindawi 2023-02-22 /pmc/articles/PMC9977516/ /pubmed/36875749 http://dx.doi.org/10.1155/2023/1177635 Text en Copyright © 2023 Huan Zhang et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Zhang, Huan
Tan, Hao
Zhu, Bin
Wang, Le
Shafiq, Muhammad
Gu, Zhaoquan
Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
title Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
title_full Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
title_fullStr Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
title_full_unstemmed Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
title_short Learning to Discriminate Adversarial Examples by Sensitivity Inconsistency in IoHT Systems
title_sort learning to discriminate adversarial examples by sensitivity inconsistency in ioht systems
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9977516/
https://www.ncbi.nlm.nih.gov/pubmed/36875749
http://dx.doi.org/10.1155/2023/1177635
work_keys_str_mv AT zhanghuan learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems
AT tanhao learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems
AT zhubin learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems
AT wangle learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems
AT shafiqmuhammad learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems
AT guzhaoquan learningtodiscriminateadversarialexamplesbysensitivityinconsistencyiniohtsystems