Cargando…
Should artificial intelligence have lower acceptable error rates than humans?
The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The British Institute of Radiology.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10301708/ https://www.ncbi.nlm.nih.gov/pubmed/37389001 http://dx.doi.org/10.1259/bjro.20220053 |
_version_ | 1785064876470697984 |
---|---|
author | Lenskjold, Anders Nybing, Janus Uhd Trampedach, Charlotte Galsgaard, Astrid Brejnebøl, Mathias Willadsen Raaschou, Henriette Rose, Martin Høyer Boesen, Mikael |
author_facet | Lenskjold, Anders Nybing, Janus Uhd Trampedach, Charlotte Galsgaard, Astrid Brejnebøl, Mathias Willadsen Raaschou, Henriette Rose, Martin Høyer Boesen, Mikael |
author_sort | Lenskjold, Anders |
collection | PubMed |
description | The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI’s unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance. |
format | Online Article Text |
id | pubmed-10301708 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | The British Institute of Radiology. |
record_format | MEDLINE/PubMed |
spelling | pubmed-103017082023-06-29 Should artificial intelligence have lower acceptable error rates than humans? Lenskjold, Anders Nybing, Janus Uhd Trampedach, Charlotte Galsgaard, Astrid Brejnebøl, Mathias Willadsen Raaschou, Henriette Rose, Martin Høyer Boesen, Mikael BJR Open Opinion The first patient was misclassified in the diagnostic conclusion according to a local clinical expert opinion in a new clinical implementation of a knee osteoarthritis artificial intelligence (AI) algorithm at Bispebjerg-Frederiksberg University Hospital, Copenhagen, Denmark. In preparation for the evaluation of the AI algorithm, the implementation team collaborated with internal and external partners to plan workflows, and the algorithm was externally validated. After the misclassification, the team was left wondering: what is an acceptable error rate for a low-risk AI diagnostic algorithm? A survey among employees at the Department of Radiology showed significantly lower acceptable error rates for AI (6.8 %) than humans (11.3 %). A general mistrust of AI could cause the discrepancy in acceptable errors. AI may have the disadvantage of limited social capital and likeability compared to human co-workers, and therefore, less potential for forgiveness. Future AI development and implementation require further investigation of the fear of AI’s unknown errors to enhance the trustworthiness of perceiving AI as a co-worker. Benchmark tools, transparency, and explainability are also needed to evaluate AI algorithms in clinical implementations to ensure acceptable performance. The British Institute of Radiology. 2023-04-13 /pmc/articles/PMC10301708/ /pubmed/37389001 http://dx.doi.org/10.1259/bjro.20220053 Text en © 2023 The Authors. Published by the British Institute of Radiology https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Opinion Lenskjold, Anders Nybing, Janus Uhd Trampedach, Charlotte Galsgaard, Astrid Brejnebøl, Mathias Willadsen Raaschou, Henriette Rose, Martin Høyer Boesen, Mikael Should artificial intelligence have lower acceptable error rates than humans? |
title | Should artificial intelligence have lower acceptable error rates than humans? |
title_full | Should artificial intelligence have lower acceptable error rates than humans? |
title_fullStr | Should artificial intelligence have lower acceptable error rates than humans? |
title_full_unstemmed | Should artificial intelligence have lower acceptable error rates than humans? |
title_short | Should artificial intelligence have lower acceptable error rates than humans? |
title_sort | should artificial intelligence have lower acceptable error rates than humans? |
topic | Opinion |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10301708/ https://www.ncbi.nlm.nih.gov/pubmed/37389001 http://dx.doi.org/10.1259/bjro.20220053 |
work_keys_str_mv | AT lenskjoldanders shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT nybingjanusuhd shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT trampedachcharlotte shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT galsgaardastrid shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT brejnebølmathiaswilladsen shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT raaschouhenriette shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT rosemartinhøyer shouldartificialintelligencehaveloweracceptableerrorratesthanhumans AT boesenmikael shouldartificialintelligencehaveloweracceptableerrorratesthanhumans |