Cargando…

Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery

BACKGROUND: Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care–related advice. Because of a lack of regulation and oversight, the reliability of health care–related responses has become a topic of cont...

Descripción completa

Detalles Bibliográficos
Autores principales: Anastasio, Albert Thomas, Mills, Frederic Baker, Karavan, Mark P., Adams, Samuel B.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: SAGE Publications 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10666700/
https://www.ncbi.nlm.nih.gov/pubmed/38027458
http://dx.doi.org/10.1177/24730114231209919
_version_ 1785148987547844608
author Anastasio, Albert Thomas
Mills, Frederic Baker
Karavan, Mark P.
Adams, Samuel B.
author_facet Anastasio, Albert Thomas
Mills, Frederic Baker
Karavan, Mark P.
Adams, Samuel B.
author_sort Anastasio, Albert Thomas
collection PubMed
description BACKGROUND: Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care–related advice. Because of a lack of regulation and oversight, the reliability of health care–related responses has become a topic of controversy in the medical community. To date, no study has explored the quality of AI-derived information as it relates to common foot and ankle pathologies. This study aims to assess the quality and educational benefit of ChatGPT responses to common foot and ankle–related questions. METHODS: ChatGPT was asked a series of 5 questions, including “What is the optimal treatment for ankle arthritis?” “How should I decide on ankle arthroplasty versus ankle arthrodesis?” “Do I need surgery for Jones fracture?” “How can I prevent Charcot arthropathy?” and “Do I need to see a doctor for my ankle sprain?” Five responses (1 per each question) were included after applying the exclusion criteria. The content was graded using DISCERN (a well-validated informational analysis tool) and AIRM (a self-designed tool for exercise evaluation). RESULTS: Health care professionals graded the ChatGPT-generated responses as bottom tier 4.5% of the time, middle tier 27.3% of the time, and top tier 68.2% of the time. CONCLUSION: Although ChatGPT and other related AI platforms have become a popular means for medical information distribution, the educational value of the AI-generated responses related to foot and ankle pathologies was variable. With 4.5% of responses receiving a bottom-tier rating, 27.3% of responses receiving a middle-tier rating, and 68.2% of responses receiving a top-tier rating, health care professionals should be aware of the high viewership of variable-quality content easily accessible on ChatGPT. LEVEL OF EVIDENCE: Level III, cross sectional study.
format Online
Article
Text
id pubmed-10666700
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher SAGE Publications
record_format MEDLINE/PubMed
spelling pubmed-106667002023-11-22 Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery Anastasio, Albert Thomas Mills, Frederic Baker Karavan, Mark P. Adams, Samuel B. Foot Ankle Orthop Article BACKGROUND: Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care–related advice. Because of a lack of regulation and oversight, the reliability of health care–related responses has become a topic of controversy in the medical community. To date, no study has explored the quality of AI-derived information as it relates to common foot and ankle pathologies. This study aims to assess the quality and educational benefit of ChatGPT responses to common foot and ankle–related questions. METHODS: ChatGPT was asked a series of 5 questions, including “What is the optimal treatment for ankle arthritis?” “How should I decide on ankle arthroplasty versus ankle arthrodesis?” “Do I need surgery for Jones fracture?” “How can I prevent Charcot arthropathy?” and “Do I need to see a doctor for my ankle sprain?” Five responses (1 per each question) were included after applying the exclusion criteria. The content was graded using DISCERN (a well-validated informational analysis tool) and AIRM (a self-designed tool for exercise evaluation). RESULTS: Health care professionals graded the ChatGPT-generated responses as bottom tier 4.5% of the time, middle tier 27.3% of the time, and top tier 68.2% of the time. CONCLUSION: Although ChatGPT and other related AI platforms have become a popular means for medical information distribution, the educational value of the AI-generated responses related to foot and ankle pathologies was variable. With 4.5% of responses receiving a bottom-tier rating, 27.3% of responses receiving a middle-tier rating, and 68.2% of responses receiving a top-tier rating, health care professionals should be aware of the high viewership of variable-quality content easily accessible on ChatGPT. LEVEL OF EVIDENCE: Level III, cross sectional study. SAGE Publications 2023-11-22 /pmc/articles/PMC10666700/ /pubmed/38027458 http://dx.doi.org/10.1177/24730114231209919 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by-nc/4.0/This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License (https://creativecommons.org/licenses/by-nc/4.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).
spellingShingle Article
Anastasio, Albert Thomas
Mills, Frederic Baker
Karavan, Mark P.
Adams, Samuel B.
Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
title Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
title_full Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
title_fullStr Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
title_full_unstemmed Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
title_short Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
title_sort evaluating the quality and usability of artificial intelligence–generated responses to common patient questions in foot and ankle surgery
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10666700/
https://www.ncbi.nlm.nih.gov/pubmed/38027458
http://dx.doi.org/10.1177/24730114231209919
work_keys_str_mv AT anastasioalbertthomas evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery
AT millsfredericbaker evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery
AT karavanmarkp evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery
AT adamssamuelb evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery