Cargando…
Human visual explanations mitigate bias in AI-based assessment of surgeon skills
Artificial intelligence (AI) systems can now reliably assess surgeon skills through videos of intraoperative surgical activity. With such systems informing future high-stakes decisions such as whether to credential surgeons and grant them the privilege to operate on patients, it is critical that the...
Autores principales: | , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10063676/ https://www.ncbi.nlm.nih.gov/pubmed/36997642 http://dx.doi.org/10.1038/s41746-023-00766-2 |
_version_ | 1785017756474671104 |
---|---|
author | Kiyasseh, Dani Laca, Jasper Haque, Taseen F. Otiato, Maxwell Miles, Brian J. Wagner, Christian Donoho, Daniel A. Trinh, Quoc-Dien Anandkumar, Animashree Hung, Andrew J. |
author_facet | Kiyasseh, Dani Laca, Jasper Haque, Taseen F. Otiato, Maxwell Miles, Brian J. Wagner, Christian Donoho, Daniel A. Trinh, Quoc-Dien Anandkumar, Animashree Hung, Andrew J. |
author_sort | Kiyasseh, Dani |
collection | PubMed |
description | Artificial intelligence (AI) systems can now reliably assess surgeon skills through videos of intraoperative surgical activity. With such systems informing future high-stakes decisions such as whether to credential surgeons and grant them the privilege to operate on patients, it is critical that they treat all surgeons fairly. However, it remains an open question whether surgical AI systems exhibit bias against surgeon sub-cohorts, and, if so, whether such bias can be mitigated. Here, we examine and mitigate the bias exhibited by a family of surgical AI systems—SAIS—deployed on videos of robotic surgeries from three geographically-diverse hospitals (USA and EU). We show that SAIS exhibits an underskilling bias, erroneously downgrading surgical performance, and an overskilling bias, erroneously upgrading surgical performance, at different rates across surgeon sub-cohorts. To mitigate such bias, we leverage a strategy —TWIX—which teaches an AI system to provide a visual explanation for its skill assessment that otherwise would have been provided by human experts. We show that whereas baseline strategies inconsistently mitigate algorithmic bias, TWIX can effectively mitigate the underskilling and overskilling bias while simultaneously improving the performance of these AI systems across hospitals. We discovered that these findings carry over to the training environment where we assess medical students’ skills today. Our study is a critical prerequisite to the eventual implementation of AI-augmented global surgeon credentialing programs, ensuring that all surgeons are treated fairly. |
format | Online Article Text |
id | pubmed-10063676 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-100636762023-04-01 Human visual explanations mitigate bias in AI-based assessment of surgeon skills Kiyasseh, Dani Laca, Jasper Haque, Taseen F. Otiato, Maxwell Miles, Brian J. Wagner, Christian Donoho, Daniel A. Trinh, Quoc-Dien Anandkumar, Animashree Hung, Andrew J. NPJ Digit Med Article Artificial intelligence (AI) systems can now reliably assess surgeon skills through videos of intraoperative surgical activity. With such systems informing future high-stakes decisions such as whether to credential surgeons and grant them the privilege to operate on patients, it is critical that they treat all surgeons fairly. However, it remains an open question whether surgical AI systems exhibit bias against surgeon sub-cohorts, and, if so, whether such bias can be mitigated. Here, we examine and mitigate the bias exhibited by a family of surgical AI systems—SAIS—deployed on videos of robotic surgeries from three geographically-diverse hospitals (USA and EU). We show that SAIS exhibits an underskilling bias, erroneously downgrading surgical performance, and an overskilling bias, erroneously upgrading surgical performance, at different rates across surgeon sub-cohorts. To mitigate such bias, we leverage a strategy —TWIX—which teaches an AI system to provide a visual explanation for its skill assessment that otherwise would have been provided by human experts. We show that whereas baseline strategies inconsistently mitigate algorithmic bias, TWIX can effectively mitigate the underskilling and overskilling bias while simultaneously improving the performance of these AI systems across hospitals. We discovered that these findings carry over to the training environment where we assess medical students’ skills today. Our study is a critical prerequisite to the eventual implementation of AI-augmented global surgeon credentialing programs, ensuring that all surgeons are treated fairly. Nature Publishing Group UK 2023-03-30 /pmc/articles/PMC10063676/ /pubmed/36997642 http://dx.doi.org/10.1038/s41746-023-00766-2 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Kiyasseh, Dani Laca, Jasper Haque, Taseen F. Otiato, Maxwell Miles, Brian J. Wagner, Christian Donoho, Daniel A. Trinh, Quoc-Dien Anandkumar, Animashree Hung, Andrew J. Human visual explanations mitigate bias in AI-based assessment of surgeon skills |
title | Human visual explanations mitigate bias in AI-based assessment of surgeon skills |
title_full | Human visual explanations mitigate bias in AI-based assessment of surgeon skills |
title_fullStr | Human visual explanations mitigate bias in AI-based assessment of surgeon skills |
title_full_unstemmed | Human visual explanations mitigate bias in AI-based assessment of surgeon skills |
title_short | Human visual explanations mitigate bias in AI-based assessment of surgeon skills |
title_sort | human visual explanations mitigate bias in ai-based assessment of surgeon skills |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10063676/ https://www.ncbi.nlm.nih.gov/pubmed/36997642 http://dx.doi.org/10.1038/s41746-023-00766-2 |
work_keys_str_mv | AT kiyassehdani humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT lacajasper humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT haquetaseenf humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT otiatomaxwell humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT milesbrianj humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT wagnerchristian humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT donohodaniela humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT trinhquocdien humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT anandkumaranimashree humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills AT hungandrewj humanvisualexplanationsmitigatebiasinaibasedassessmentofsurgeonskills |