Cargando…

Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study

BACKGROUND: A backdoor attack controls the output of a machine learning model in 2 stages. First, the attacker poisons the training data set, introducing a back door into the victim’s trained model. Second, during test time, the attacker adds an imperceptible pattern called a trigger to the input va...

Descripción completa

Detalles Bibliográficos
Autores principales: Joe, Byunggill, Park, Yonghyeon, Hamm, Jihun, Shin, Insik, Lee, Jiyeon
Formato: Online Artículo Texto
Lenguaje:English
Publicado: JMIR Publications 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9440413/
https://www.ncbi.nlm.nih.gov/pubmed/35984701
http://dx.doi.org/10.2196/38440
_version_ 1784782345354608640
author Joe, Byunggill
Park, Yonghyeon
Hamm, Jihun
Shin, Insik
Lee, Jiyeon
author_facet Joe, Byunggill
Park, Yonghyeon
Hamm, Jihun
Shin, Insik
Lee, Jiyeon
author_sort Joe, Byunggill
collection PubMed
description BACKGROUND: A backdoor attack controls the output of a machine learning model in 2 stages. First, the attacker poisons the training data set, introducing a back door into the victim’s trained model. Second, during test time, the attacker adds an imperceptible pattern called a trigger to the input values, which forces the victim’s model to output the attacker’s intended values instead of true predictions or decisions. While backdoor attacks pose a serious threat to the reliability of machine learning–based medical diagnostics, existing backdoor attacks that directly change the input values are detectable relatively easily. OBJECTIVE: The goal of this study was to propose and study a robust backdoor attack on mortality-prediction machine learning models that use electronic health records. We showed that our backdoor attack grants attackers full control over classification outcomes for safety-critical tasks such as mortality prediction, highlighting the importance of undertaking safe artificial intelligence research in the medical field. METHODS: We present a trigger generation method based on missing patterns in electronic health record data. Compared to existing approaches, which introduce noise into the medical record, the proposed backdoor attack makes it simple to construct backdoor triggers without prior knowledge. To effectively avoid detection by manual inspectors, we employ variational autoencoders to learn the missing patterns in normal electronic health record data and produce trigger data that appears similar to this data. RESULTS: We experimented with the proposed backdoor attack on 4 machine learning models (linear regression, multilayer perceptron, long short-term memory, and gated recurrent units) that predict in-hospital mortality using a public electronic health record data set. The results showed that the proposed technique achieved a significant drop in the victim’s discrimination performance (reducing the area under the precision-recall curve by at most 0.45), with a low poisoning rate (2%) in the training data set. In addition, the impact of the attack on general classification performance was negligible (it reduced the area under the precision-recall curve by an average of 0.01025), which makes it difficult to detect the presence of poison. CONCLUSIONS: To the best of our knowledge, this is the first study to propose a backdoor attack that uses missing information from tabular data as a trigger. Through extensive experiments, we demonstrated that our backdoor attack can inflict severe damage on medical machine learning classifiers in practice.
format Online
Article
Text
id pubmed-9440413
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher JMIR Publications
record_format MEDLINE/PubMed
spelling pubmed-94404132022-09-04 Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study Joe, Byunggill Park, Yonghyeon Hamm, Jihun Shin, Insik Lee, Jiyeon JMIR Med Inform Original Paper BACKGROUND: A backdoor attack controls the output of a machine learning model in 2 stages. First, the attacker poisons the training data set, introducing a back door into the victim’s trained model. Second, during test time, the attacker adds an imperceptible pattern called a trigger to the input values, which forces the victim’s model to output the attacker’s intended values instead of true predictions or decisions. While backdoor attacks pose a serious threat to the reliability of machine learning–based medical diagnostics, existing backdoor attacks that directly change the input values are detectable relatively easily. OBJECTIVE: The goal of this study was to propose and study a robust backdoor attack on mortality-prediction machine learning models that use electronic health records. We showed that our backdoor attack grants attackers full control over classification outcomes for safety-critical tasks such as mortality prediction, highlighting the importance of undertaking safe artificial intelligence research in the medical field. METHODS: We present a trigger generation method based on missing patterns in electronic health record data. Compared to existing approaches, which introduce noise into the medical record, the proposed backdoor attack makes it simple to construct backdoor triggers without prior knowledge. To effectively avoid detection by manual inspectors, we employ variational autoencoders to learn the missing patterns in normal electronic health record data and produce trigger data that appears similar to this data. RESULTS: We experimented with the proposed backdoor attack on 4 machine learning models (linear regression, multilayer perceptron, long short-term memory, and gated recurrent units) that predict in-hospital mortality using a public electronic health record data set. The results showed that the proposed technique achieved a significant drop in the victim’s discrimination performance (reducing the area under the precision-recall curve by at most 0.45), with a low poisoning rate (2%) in the training data set. In addition, the impact of the attack on general classification performance was negligible (it reduced the area under the precision-recall curve by an average of 0.01025), which makes it difficult to detect the presence of poison. CONCLUSIONS: To the best of our knowledge, this is the first study to propose a backdoor attack that uses missing information from tabular data as a trigger. Through extensive experiments, we demonstrated that our backdoor attack can inflict severe damage on medical machine learning classifiers in practice. JMIR Publications 2022-08-19 /pmc/articles/PMC9440413/ /pubmed/35984701 http://dx.doi.org/10.2196/38440 Text en ©Byunggill Joe, Yonghyeon Park, Jihun Hamm, Insik Shin, Jiyeon Lee. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 19.08.2022. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.
spellingShingle Original Paper
Joe, Byunggill
Park, Yonghyeon
Hamm, Jihun
Shin, Insik
Lee, Jiyeon
Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study
title Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study
title_full Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study
title_fullStr Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study
title_full_unstemmed Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study
title_short Exploiting Missing Value Patterns for a Backdoor Attack on Machine Learning Models of Electronic Health Records: Development and Validation Study
title_sort exploiting missing value patterns for a backdoor attack on machine learning models of electronic health records: development and validation study
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9440413/
https://www.ncbi.nlm.nih.gov/pubmed/35984701
http://dx.doi.org/10.2196/38440
work_keys_str_mv AT joebyunggill exploitingmissingvaluepatternsforabackdoorattackonmachinelearningmodelsofelectronichealthrecordsdevelopmentandvalidationstudy
AT parkyonghyeon exploitingmissingvaluepatternsforabackdoorattackonmachinelearningmodelsofelectronichealthrecordsdevelopmentandvalidationstudy
AT hammjihun exploitingmissingvaluepatternsforabackdoorattackonmachinelearningmodelsofelectronichealthrecordsdevelopmentandvalidationstudy
AT shininsik exploitingmissingvaluepatternsforabackdoorattackonmachinelearningmodelsofelectronichealthrecordsdevelopmentandvalidationstudy
AT leejiyeon exploitingmissingvaluepatternsforabackdoorattackonmachinelearningmodelsofelectronichealthrecordsdevelopmentandvalidationstudy