Cargando…
Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery
BACKGROUND: Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care–related advice. Because of a lack of regulation and oversight, the reliability of health care–related responses has become a topic of cont...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
SAGE Publications
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10666700/ https://www.ncbi.nlm.nih.gov/pubmed/38027458 http://dx.doi.org/10.1177/24730114231209919 |
_version_ | 1785148987547844608 |
---|---|
author | Anastasio, Albert Thomas Mills, Frederic Baker Karavan, Mark P. Adams, Samuel B. |
author_facet | Anastasio, Albert Thomas Mills, Frederic Baker Karavan, Mark P. Adams, Samuel B. |
author_sort | Anastasio, Albert Thomas |
collection | PubMed |
description | BACKGROUND: Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care–related advice. Because of a lack of regulation and oversight, the reliability of health care–related responses has become a topic of controversy in the medical community. To date, no study has explored the quality of AI-derived information as it relates to common foot and ankle pathologies. This study aims to assess the quality and educational benefit of ChatGPT responses to common foot and ankle–related questions. METHODS: ChatGPT was asked a series of 5 questions, including “What is the optimal treatment for ankle arthritis?” “How should I decide on ankle arthroplasty versus ankle arthrodesis?” “Do I need surgery for Jones fracture?” “How can I prevent Charcot arthropathy?” and “Do I need to see a doctor for my ankle sprain?” Five responses (1 per each question) were included after applying the exclusion criteria. The content was graded using DISCERN (a well-validated informational analysis tool) and AIRM (a self-designed tool for exercise evaluation). RESULTS: Health care professionals graded the ChatGPT-generated responses as bottom tier 4.5% of the time, middle tier 27.3% of the time, and top tier 68.2% of the time. CONCLUSION: Although ChatGPT and other related AI platforms have become a popular means for medical information distribution, the educational value of the AI-generated responses related to foot and ankle pathologies was variable. With 4.5% of responses receiving a bottom-tier rating, 27.3% of responses receiving a middle-tier rating, and 68.2% of responses receiving a top-tier rating, health care professionals should be aware of the high viewership of variable-quality content easily accessible on ChatGPT. LEVEL OF EVIDENCE: Level III, cross sectional study. |
format | Online Article Text |
id | pubmed-10666700 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | SAGE Publications |
record_format | MEDLINE/PubMed |
spelling | pubmed-106667002023-11-22 Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery Anastasio, Albert Thomas Mills, Frederic Baker Karavan, Mark P. Adams, Samuel B. Foot Ankle Orthop Article BACKGROUND: Artificial intelligence (AI) platforms, such as ChatGPT, have become increasingly popular outlets for the consumption and distribution of health care–related advice. Because of a lack of regulation and oversight, the reliability of health care–related responses has become a topic of controversy in the medical community. To date, no study has explored the quality of AI-derived information as it relates to common foot and ankle pathologies. This study aims to assess the quality and educational benefit of ChatGPT responses to common foot and ankle–related questions. METHODS: ChatGPT was asked a series of 5 questions, including “What is the optimal treatment for ankle arthritis?” “How should I decide on ankle arthroplasty versus ankle arthrodesis?” “Do I need surgery for Jones fracture?” “How can I prevent Charcot arthropathy?” and “Do I need to see a doctor for my ankle sprain?” Five responses (1 per each question) were included after applying the exclusion criteria. The content was graded using DISCERN (a well-validated informational analysis tool) and AIRM (a self-designed tool for exercise evaluation). RESULTS: Health care professionals graded the ChatGPT-generated responses as bottom tier 4.5% of the time, middle tier 27.3% of the time, and top tier 68.2% of the time. CONCLUSION: Although ChatGPT and other related AI platforms have become a popular means for medical information distribution, the educational value of the AI-generated responses related to foot and ankle pathologies was variable. With 4.5% of responses receiving a bottom-tier rating, 27.3% of responses receiving a middle-tier rating, and 68.2% of responses receiving a top-tier rating, health care professionals should be aware of the high viewership of variable-quality content easily accessible on ChatGPT. LEVEL OF EVIDENCE: Level III, cross sectional study. SAGE Publications 2023-11-22 /pmc/articles/PMC10666700/ /pubmed/38027458 http://dx.doi.org/10.1177/24730114231209919 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by-nc/4.0/This article is distributed under the terms of the Creative Commons Attribution-NonCommercial 4.0 License (https://creativecommons.org/licenses/by-nc/4.0/) which permits non-commercial use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage). |
spellingShingle | Article Anastasio, Albert Thomas Mills, Frederic Baker Karavan, Mark P. Adams, Samuel B. Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery |
title | Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery |
title_full | Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery |
title_fullStr | Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery |
title_full_unstemmed | Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery |
title_short | Evaluating the Quality and Usability of Artificial Intelligence–Generated Responses to Common Patient Questions in Foot and Ankle Surgery |
title_sort | evaluating the quality and usability of artificial intelligence–generated responses to common patient questions in foot and ankle surgery |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10666700/ https://www.ncbi.nlm.nih.gov/pubmed/38027458 http://dx.doi.org/10.1177/24730114231209919 |
work_keys_str_mv | AT anastasioalbertthomas evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery AT millsfredericbaker evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery AT karavanmarkp evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery AT adamssamuelb evaluatingthequalityandusabilityofartificialintelligencegeneratedresponsestocommonpatientquestionsinfootandanklesurgery |