Cargando…
A flexible symbolic regression method for constructing interpretable clinical prediction models
Machine learning (ML) models trained for triggering clinical decision support (CDS) are typically either accurate or interpretable but not both. Scaling CDS to the panoply of clinical use cases while mitigating risks to patients will require many ML models be intuitively interpretable for clinicians...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10241925/ https://www.ncbi.nlm.nih.gov/pubmed/37277550 http://dx.doi.org/10.1038/s41746-023-00833-8 |
_version_ | 1785054099753926656 |
---|---|
author | La Cava, William G. Lee, Paul C. Ajmal, Imran Ding, Xiruo Solanki, Priyanka Cohen, Jordana B. Moore, Jason H. Herman, Daniel S. |
author_facet | La Cava, William G. Lee, Paul C. Ajmal, Imran Ding, Xiruo Solanki, Priyanka Cohen, Jordana B. Moore, Jason H. Herman, Daniel S. |
author_sort | La Cava, William G. |
collection | PubMed |
description | Machine learning (ML) models trained for triggering clinical decision support (CDS) are typically either accurate or interpretable but not both. Scaling CDS to the panoply of clinical use cases while mitigating risks to patients will require many ML models be intuitively interpretable for clinicians. To this end, we adapted a symbolic regression method, coined the feature engineering automation tool (FEAT), to train concise and accurate models from high-dimensional electronic health record (EHR) data. We first present an in-depth application of FEAT to classify hypertension, hypertension with unexplained hypokalemia, and apparent treatment-resistant hypertension (aTRH) using EHR data for 1200 subjects receiving longitudinal care in a large healthcare system. FEAT models trained to predict phenotypes adjudicated by chart review had equivalent or higher discriminative performance (p < 0.001) and were at least three times smaller (p < 1 × 10(−6)) than other potentially interpretable models. For aTRH, FEAT generated a six-feature, highly discriminative (positive predictive value = 0.70, sensitivity = 0.62), and clinically intuitive model. To assess the generalizability of the approach, we tested FEAT on 25 benchmark clinical phenotyping tasks using the MIMIC-III critical care database. Under comparable dimensionality constraints, FEAT’s models exhibited higher area under the receiver-operating curve scores than penalized linear models across tasks (p < 6 × 10(−6)). In summary, FEAT can train EHR prediction models that are both intuitively interpretable and accurate, which should facilitate safe and effective scaling of ML-triggered CDS to the panoply of potential clinical use cases and healthcare practices. |
format | Online Article Text |
id | pubmed-10241925 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-102419252023-06-07 A flexible symbolic regression method for constructing interpretable clinical prediction models La Cava, William G. Lee, Paul C. Ajmal, Imran Ding, Xiruo Solanki, Priyanka Cohen, Jordana B. Moore, Jason H. Herman, Daniel S. NPJ Digit Med Article Machine learning (ML) models trained for triggering clinical decision support (CDS) are typically either accurate or interpretable but not both. Scaling CDS to the panoply of clinical use cases while mitigating risks to patients will require many ML models be intuitively interpretable for clinicians. To this end, we adapted a symbolic regression method, coined the feature engineering automation tool (FEAT), to train concise and accurate models from high-dimensional electronic health record (EHR) data. We first present an in-depth application of FEAT to classify hypertension, hypertension with unexplained hypokalemia, and apparent treatment-resistant hypertension (aTRH) using EHR data for 1200 subjects receiving longitudinal care in a large healthcare system. FEAT models trained to predict phenotypes adjudicated by chart review had equivalent or higher discriminative performance (p < 0.001) and were at least three times smaller (p < 1 × 10(−6)) than other potentially interpretable models. For aTRH, FEAT generated a six-feature, highly discriminative (positive predictive value = 0.70, sensitivity = 0.62), and clinically intuitive model. To assess the generalizability of the approach, we tested FEAT on 25 benchmark clinical phenotyping tasks using the MIMIC-III critical care database. Under comparable dimensionality constraints, FEAT’s models exhibited higher area under the receiver-operating curve scores than penalized linear models across tasks (p < 6 × 10(−6)). In summary, FEAT can train EHR prediction models that are both intuitively interpretable and accurate, which should facilitate safe and effective scaling of ML-triggered CDS to the panoply of potential clinical use cases and healthcare practices. Nature Publishing Group UK 2023-06-05 /pmc/articles/PMC10241925/ /pubmed/37277550 http://dx.doi.org/10.1038/s41746-023-00833-8 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article La Cava, William G. Lee, Paul C. Ajmal, Imran Ding, Xiruo Solanki, Priyanka Cohen, Jordana B. Moore, Jason H. Herman, Daniel S. A flexible symbolic regression method for constructing interpretable clinical prediction models |
title | A flexible symbolic regression method for constructing interpretable clinical prediction models |
title_full | A flexible symbolic regression method for constructing interpretable clinical prediction models |
title_fullStr | A flexible symbolic regression method for constructing interpretable clinical prediction models |
title_full_unstemmed | A flexible symbolic regression method for constructing interpretable clinical prediction models |
title_short | A flexible symbolic regression method for constructing interpretable clinical prediction models |
title_sort | flexible symbolic regression method for constructing interpretable clinical prediction models |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10241925/ https://www.ncbi.nlm.nih.gov/pubmed/37277550 http://dx.doi.org/10.1038/s41746-023-00833-8 |
work_keys_str_mv | AT lacavawilliamg aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT leepaulc aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT ajmalimran aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT dingxiruo aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT solankipriyanka aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT cohenjordanab aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT moorejasonh aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT hermandaniels aflexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT lacavawilliamg flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT leepaulc flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT ajmalimran flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT dingxiruo flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT solankipriyanka flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT cohenjordanab flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT moorejasonh flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels AT hermandaniels flexiblesymbolicregressionmethodforconstructinginterpretableclinicalpredictionmodels |