Cargando…

Should artificial intelligence have lower acceptable error rates than humans?

The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the...

Descripción completa

Detalles Bibliográficos
Autores principales: Lenskjold, Anders, Nybing, Janus Uhd, Trampedach, Charlotte, Galsgaard, Astrid, Brejnebøl, Mathias Willadsen, Raaschou, Henriette, Rose, Martin Høyer, Boesen, Mikael
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The British Institute of Radiology. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10301708/
https://www.ncbi.nlm.nih.gov/pubmed/37389001
http://dx.doi.org/10.1259/bjro.20220053
_version_ 1785064876470697984
author Lenskjold, Anders
Nybing, Janus Uhd
Trampedach, Charlotte
Galsgaard, Astrid
Brejnebøl, Mathias Willadsen
Raaschou, Henriette
Rose, Martin Høyer
Boesen, Mikael
author_facet Lenskjold, Anders
Nybing, Janus Uhd
Trampedach, Charlotte
Galsgaard, Astrid
Brejnebøl, Mathias Willadsen
Raaschou, Henriette
Rose, Martin Høyer
Boesen, Mikael
author_sort Lenskjold, Anders
collection PubMed
description The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI’s unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance.
format Online
Article
Text
id pubmed-10301708
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher The British Institute of Radiology.
record_format MEDLINE/PubMed
spelling pubmed-103017082023-06-29 Should artificial intelligence have lower acceptable error rates than humans? Lenskjold, Anders Nybing, Janus Uhd Trampedach, Charlotte Galsgaard, Astrid Brejnebøl, Mathias Willadsen Raaschou, Henriette Rose, Martin Høyer Boesen, Mikael BJR Open Opinion The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI’s unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance. The British Institute of Radiology. 2023-04-13 /pmc/articles/PMC10301708/ /pubmed/37389001 http://dx.doi.org/10.1259/bjro.20220053 Text en © 2023 The Authors. Published by the British Institute of Radiology https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited.
spellingShingle Opinion
Lenskjold, Anders
Nybing, Janus Uhd
Trampedach, Charlotte
Galsgaard, Astrid
Brejnebøl, Mathias Willadsen
Raaschou, Henriette
Rose, Martin Høyer
Boesen, Mikael
Should artificial intelligence have lower acceptable error rates than humans?
title Should artificial intelligence have lower acceptable error rates than humans?
title_full Should artificial intelligence have lower acceptable error rates than humans?
title_fullStr Should artificial intelligence have lower acceptable error rates than humans?
title_full_unstemmed Should artificial intelligence have lower acceptable error rates than humans?
title_short Should artificial intelligence have lower acceptable error rates than humans?
title_sort should artificial intelligence have lower acceptable error rates than humans?
topic Opinion
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10301708/
https://www.ncbi.nlm.nih.gov/pubmed/37389001
http://dx.doi.org/10.1259/bjro.20220053
work_keys_str_mv AT lenskjoldanders shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT nybingjanusuhd shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT trampedachcharlotte shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT galsgaardastrid shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT brejnebølmathiaswilladsen shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT raaschouhenriette shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT rosemartinhøyer shouldartificialintelligencehaveloweracceptableerrorratesthanhumans
AT boesenmikael shouldartificialintelligencehaveloweracceptableerrorratesthanhumans