Cargando…

Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review

IMPORTANCE: Neuroimaging-based artificial intelligence (AI) diagnostic models have proliferated in psychiatry. However, their clinical applicability and reporting quality (ie, feasibility) for clinical practice have not been systematically evaluated. OBJECTIVE: To systematically assess the risk of b...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Zhiyi, Liu, Xuerong, Yang, Qingwu, Wang, Yan-Jiang, Miao, Kuan, Gong, Zheng, Yu, Yang, Leonov, Artemiy, Liu, Chunlei, Feng, Zhengzhi, Chuan-Peng, Hu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: American Medical Association 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9989906/
https://www.ncbi.nlm.nih.gov/pubmed/36877519
http://dx.doi.org/10.1001/jamanetworkopen.2023.1671
_version_ 1784901849461030912
author Chen, Zhiyi
Liu, Xuerong
Yang, Qingwu
Wang, Yan-Jiang
Miao, Kuan
Gong, Zheng
Yu, Yang
Leonov, Artemiy
Liu, Chunlei
Feng, Zhengzhi
Chuan-Peng, Hu
author_facet Chen, Zhiyi
Liu, Xuerong
Yang, Qingwu
Wang, Yan-Jiang
Miao, Kuan
Gong, Zheng
Yu, Yang
Leonov, Artemiy
Liu, Chunlei
Feng, Zhengzhi
Chuan-Peng, Hu
author_sort Chen, Zhiyi
collection PubMed
description IMPORTANCE: Neuroimaging-based artificial intelligence (AI) diagnostic models have proliferated in psychiatry. However, their clinical applicability and reporting quality (ie, feasibility) for clinical practice have not been systematically evaluated. OBJECTIVE: To systematically assess the risk of bias (ROB) and reporting quality of neuroimaging-based AI models for psychiatric diagnosis. EVIDENCE REVIEW: PubMed was searched for peer-reviewed, full-length articles published between January 1, 1990, and March 16, 2022. Studies aimed at developing or validating neuroimaging-based AI models for clinical diagnosis of psychiatric disorders were included. Reference lists were further searched for suitable original studies. Data extraction followed the CHARMS (Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modeling Studies) and PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-analyses) guidelines. A closed-loop cross-sequential design was used for quality control. The PROBAST (Prediction Model Risk of Bias Assessment Tool) and modified CLEAR (Checklist for Evaluation of Image-Based Artificial Intelligence Reports) benchmarks were used to systematically evaluate ROB and reporting quality. FINDINGS: A total of 517 studies presenting 555 AI models were included and evaluated. Of these models, 461 (83.1%; 95% CI, 80.0%-86.2%) were rated as having a high overall ROB based on the PROBAST. The ROB was particular high in the analysis domain, including inadequate sample size (398 of 555 models [71.7%; 95% CI, 68.0%-75.6%]), poor model performance examination (with 100% of models lacking calibration examination), and lack of handling data complexity (550 of 555 models [99.1%; 95% CI, 98.3%-99.9%]). None of the AI models was perceived to be applicable to clinical practices. Overall reporting completeness (ie, number of reported items/number of total items) for the AI models was 61.2% (95% CI, 60.6%-61.8%), and the completeness was poorest for the technical assessment domain with 39.9% (95% CI, 38.8%-41.1%). CONCLUSIONS AND RELEVANCE: This systematic review found that the clinical applicability and feasibility of neuroimaging-based AI models for psychiatric diagnosis were challenged by a high ROB and poor reporting quality. Particularly in the analysis domain, ROB in AI diagnostic models should be addressed before clinical application.
format Online
Article
Text
id pubmed-9989906
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher American Medical Association
record_format MEDLINE/PubMed
spelling pubmed-99899062023-03-08 Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review Chen, Zhiyi Liu, Xuerong Yang, Qingwu Wang, Yan-Jiang Miao, Kuan Gong, Zheng Yu, Yang Leonov, Artemiy Liu, Chunlei Feng, Zhengzhi Chuan-Peng, Hu JAMA Netw Open Original Investigation IMPORTANCE: Neuroimaging-based artificial intelligence (AI) diagnostic models have proliferated in psychiatry. However, their clinical applicability and reporting quality (ie, feasibility) for clinical practice have not been systematically evaluated. OBJECTIVE: To systematically assess the risk of bias (ROB) and reporting quality of neuroimaging-based AI models for psychiatric diagnosis. EVIDENCE REVIEW: PubMed was searched for peer-reviewed, full-length articles published between January 1, 1990, and March 16, 2022. Studies aimed at developing or validating neuroimaging-based AI models for clinical diagnosis of psychiatric disorders were included. Reference lists were further searched for suitable original studies. Data extraction followed the CHARMS (Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modeling Studies) and PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-analyses) guidelines. A closed-loop cross-sequential design was used for quality control. The PROBAST (Prediction Model Risk of Bias Assessment Tool) and modified CLEAR (Checklist for Evaluation of Image-Based Artificial Intelligence Reports) benchmarks were used to systematically evaluate ROB and reporting quality. FINDINGS: A total of 517 studies presenting 555 AI models were included and evaluated. Of these models, 461 (83.1%; 95% CI, 80.0%-86.2%) were rated as having a high overall ROB based on the PROBAST. The ROB was particular high in the analysis domain, including inadequate sample size (398 of 555 models [71.7%; 95% CI, 68.0%-75.6%]), poor model performance examination (with 100% of models lacking calibration examination), and lack of handling data complexity (550 of 555 models [99.1%; 95% CI, 98.3%-99.9%]). None of the AI models was perceived to be applicable to clinical practices. Overall reporting completeness (ie, number of reported items/number of total items) for the AI models was 61.2% (95% CI, 60.6%-61.8%), and the completeness was poorest for the technical assessment domain with 39.9% (95% CI, 38.8%-41.1%). CONCLUSIONS AND RELEVANCE: This systematic review found that the clinical applicability and feasibility of neuroimaging-based AI models for psychiatric diagnosis were challenged by a high ROB and poor reporting quality. Particularly in the analysis domain, ROB in AI diagnostic models should be addressed before clinical application. American Medical Association 2023-03-06 /pmc/articles/PMC9989906/ /pubmed/36877519 http://dx.doi.org/10.1001/jamanetworkopen.2023.1671 Text en Copyright 2023 Chen Z et al. JAMA Network Open. https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article distributed under the terms of the CC-BY-NC-ND License.
spellingShingle Original Investigation
Chen, Zhiyi
Liu, Xuerong
Yang, Qingwu
Wang, Yan-Jiang
Miao, Kuan
Gong, Zheng
Yu, Yang
Leonov, Artemiy
Liu, Chunlei
Feng, Zhengzhi
Chuan-Peng, Hu
Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review
title Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review
title_full Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review
title_fullStr Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review
title_full_unstemmed Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review
title_short Evaluation of Risk of Bias in Neuroimaging-Based Artificial Intelligence Models for Psychiatric Diagnosis: A Systematic Review
title_sort evaluation of risk of bias in neuroimaging-based artificial intelligence models for psychiatric diagnosis: a systematic review
topic Original Investigation
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9989906/
https://www.ncbi.nlm.nih.gov/pubmed/36877519
http://dx.doi.org/10.1001/jamanetworkopen.2023.1671
work_keys_str_mv AT chenzhiyi evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT liuxuerong evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT yangqingwu evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT wangyanjiang evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT miaokuan evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT gongzheng evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT yuyang evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT leonovartemiy evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT liuchunlei evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT fengzhengzhi evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview
AT chuanpenghu evaluationofriskofbiasinneuroimagingbasedartificialintelligencemodelsforpsychiatricdiagnosisasystematicreview