Cargando…

Human visual explanations mitigate bias in AI-based assessment of surgeon skills

Artificial intelligence (AI) systems can now reliably assess surgeon skills through videos of intraoperative surgical activity. With such systems informing future high-stakes decisions such as whether to credential surgeons and grant them the privilege to operate on patients, it is critical that the...

Descripción completa

Detalles Bibliográficos
Autores principales: Kiyasseh, Dani, Laca, Jasper, Haque, Taseen F., Otiato, Maxwell, Miles, Brian J., Wagner, Christian, Donoho, Daniel A., Trinh, Quoc-Dien, Anandkumar, Animashree, Hung, Andrew J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10063676/
https://www.ncbi.nlm.nih.gov/pubmed/36997642
http://dx.doi.org/10.1038/s41746-023-00766-2
_version_ 1785017756474671104
author Kiyasseh, Dani
Laca, Jasper
Haque, Taseen F.
Otiato, Maxwell
Miles, Brian J.
Wagner, Christian
Donoho, Daniel A.
Trinh, Quoc-Dien
Anandkumar, Animashree
Hung, Andrew J.
author_facet Kiyasseh, Dani
Laca, Jasper
Haque, Taseen F.
Otiato, Maxwell
Miles, Brian J.
Wagner, Christian
Donoho, Daniel A.
Trinh, Quoc-Dien
Anandkumar, Animashree
Hung, Andrew J.
author_sort Kiyasseh, Dani
collection PubMed
description Artificial intelligence (AI) systems can now reliably assess surgeon skills through videos of intraoperative surgical activity. With such systems informing future high-stakes decisions such as whether to credential surgeons and grant them the privilege to operate on patients, it is critical that they treat all surgeons fairly. However, it remains an open question whether surgical AI systems exhibit bias against surgeon sub-cohorts, and, if so, whether such bias can be mitigated. Here, we examine and mitigate the bias exhibited by a family of surgical AI systems—SAIS—deployed on videos of robotic surgeries from three geographically-diverse hospitals (USA and EU). We show that SAIS exhibits an underskilling bias, erroneously downgrading surgical performance, and an overskilling bias, erroneously upgrading surgical performance, at different rates across surgeon sub-cohorts. To mitigate such bias, we leverage a strategy —TWIX—which teaches an AI system to provide a visual explanation for its skill assessment that otherwise would have been provided by human experts. We show that whereas baseline strategies inconsistently mitigate algorithmic bias, TWIX can effectively mitigate the underskilling and overskilling bias while simultaneously improving the performance of these AI systems across hospitals. We discovered that these findings carry over to the training environment where we assess medical students’ skills today. Our study is a critical prerequisite to the eventual implementation of AI-augmented global surgeon credentialing programs, ensuring that all surgeons are treated fairly.
format Online
Article
Text
id pubmed-10063676
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-100636762023-04-01 Human visual explanations mitigate bias in AI-based assessment of surgeon skills Kiyasseh, Dani Laca, Jasper Haque, Taseen F. Otiato, Maxwell Miles, Brian J. Wagner, Christian Donoho, Daniel A. Trinh, Quoc-Dien Anandkumar, Animashree Hung, Andrew J. NPJ Digit Med Article Artificial intelligence (AI) systems can now reliably assess surgeon skills through videos of intraoperative surgical activity. With such systems informing future high-stakes decisions such as whether to credential surgeons and grant them the privilege to operate on patients, it is critical that they treat all surgeons fairly. However, it remains an open question whether surgical AI systems exhibit bias against surgeon sub-cohorts, and, if so, whether such bias can be mitigated. Here, we examine and mitigate the bias exhibited by a family of surgical AI systems—SAIS—deployed on videos of robotic surgeries from three geographically-diverse hospitals (USA and EU). We show that SAIS exhibits an underskilling bias, erroneously downgrading surgical performance, and an overskilling bias, erroneously upgrading surgical performance, at different rates across surgeon sub-cohorts. To mitigate such bias, we leverage a strategy —TWIX—which teaches an AI system to provide a visual explanation for its skill assessment that otherwise would have been provided by human experts. We show that whereas baseline strategies inconsistently mitigate algorithmic bias, TWIX can effectively mitigate the underskilling and overskilling bias while simultaneously improving the performance of these AI systems across hospitals. We discovered that these findings carry over to the training environment where we assess medical students’ skills today. Our study is a critical prerequisite to the eventual implementation of AI-augmented global surgeon credentialing programs, ensuring that all surgeons are treated fairly. Nature Publishing Group UK 2023-03-30 /pmc/articles/PMC10063676/ /pubmed/36997642 http://dx.doi.org/10.1038/s41746-023-00766-2 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Kiyasseh, Dani
Laca, Jasper
Haque, Taseen F.
Otiato, Maxwell
Miles, Brian J.
Wagner, Christian
Donoho, Daniel A.
Trinh, Quoc-Dien
Anandkumar, Animashree
Hung, Andrew J.
Human visual explanations mitigate bias in AI-based assessment of surgeon skills
title Human visual explanations mitigate bias in AI-based assessment of surgeon skills
title_full Human visual explanations mitigate bias in AI-based assessment of surgeon skills
title_fullStr Human visual explanations mitigate bias in AI-based assessment of surgeon skills
title_full_unstemmed Human visual explanations mitigate bias in AI-based assessment of surgeon skills
title_short Human visual explanations mitigate bias in AI-based assessment of surgeon skills
title_sort human visual explanations mitigate bias in ai-based assessment of surgeon skills
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10063676/
https://www.ncbi.nlm.nih.gov/pubmed/36997642
http://dx.doi.org/10.1038/s41746-023-00766-2
work_keys_str_mv AT kiyassehdani humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT lacajasper humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT haquetaseenf humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT otiatomaxwell humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT milesbrianj humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT wagnerchristian humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT donohodaniela humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT trinhquocdien humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT anandkumaranimashree humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills
AT hungandrewj humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills