Cargando…

ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions

PURPOSE: With the increasing adoption of artificial intelligence (AI) in various domains, including healthcare, there is growing acceptance and interest in consulting AI models to provide medical information and advice. This study aimed to evaluate the accuracy of ChatGPT’s responses to practice qui...

Descripción completa

Detalles Bibliográficos
Autores principales: Hoch, Cosima C., Wollenberg, Barbara, Lüers, Jan-Christoffer, Knoedler, Samuel, Knoedler, Leonard, Frank, Konstantin, Cotofana, Sebastian, Alfertshofer, Michael
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer Berlin Heidelberg 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10382366/
https://www.ncbi.nlm.nih.gov/pubmed/37285018
http://dx.doi.org/10.1007/s00405-023-08051-4
_version_ 1785080665194102784
author Hoch, Cosima C.
Wollenberg, Barbara
Lüers, Jan-Christoffer
Knoedler, Samuel
Knoedler, Leonard
Frank, Konstantin
Cotofana, Sebastian
Alfertshofer, Michael
author_facet Hoch, Cosima C.
Wollenberg, Barbara
Lüers, Jan-Christoffer
Knoedler, Samuel
Knoedler, Leonard
Frank, Konstantin
Cotofana, Sebastian
Alfertshofer, Michael
author_sort Hoch, Cosima C.
collection PubMed
description PURPOSE: With the increasing adoption of artificial intelligence (AI) in various domains, including healthcare, there is growing acceptance and interest in consulting AI models to provide medical information and advice. This study aimed to evaluate the accuracy of ChatGPT’s responses to practice quiz questions designed for otolaryngology board certification and decipher potential performance disparities across different otolaryngology subspecialties. METHODS: A dataset covering 15 otolaryngology subspecialties was collected from an online learning platform funded by the German Society of Oto-Rhino-Laryngology, Head and Neck Surgery, designed for board certification examination preparation. These questions were entered into ChatGPT, with its responses being analyzed for accuracy and variance in performance. RESULTS: The dataset included 2576 questions (479 multiple-choice and 2097 single-choice), of which 57% (n = 1475) were answered correctly by ChatGPT. An in-depth analysis of question style revealed that single-choice questions were associated with a significantly higher rate (p < 0.001) of correct responses (n = 1313; 63%) compared to multiple-choice questions (n = 162; 34%). Stratified by question categories, ChatGPT yielded the highest rate of correct responses (n = 151; 72%) in the field of allergology, whereas 7 out of 10 questions (n = 65; 71%) on legal otolaryngology aspects were answered incorrectly. CONCLUSION: The study reveals ChatGPT’s potential as a supplementary tool for otolaryngology board certification preparation. However, its propensity for errors in certain otolaryngology areas calls for further refinement. Future research should address these limitations to improve ChatGPT’s educational use. An approach, with expert collaboration, is recommended for the reliable and accurate integration of such AI models.
format Online
Article
Text
id pubmed-10382366
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Springer Berlin Heidelberg
record_format MEDLINE/PubMed
spelling pubmed-103823662023-07-30 ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions Hoch, Cosima C. Wollenberg, Barbara Lüers, Jan-Christoffer Knoedler, Samuel Knoedler, Leonard Frank, Konstantin Cotofana, Sebastian Alfertshofer, Michael Eur Arch Otorhinolaryngol Miscellaneous PURPOSE: With the increasing adoption of artificial intelligence (AI) in various domains, including healthcare, there is growing acceptance and interest in consulting AI models to provide medical information and advice. This study aimed to evaluate the accuracy of ChatGPT’s responses to practice quiz questions designed for otolaryngology board certification and decipher potential performance disparities across different otolaryngology subspecialties. METHODS: A dataset covering 15 otolaryngology subspecialties was collected from an online learning platform funded by the German Society of Oto-Rhino-Laryngology, Head and Neck Surgery, designed for board certification examination preparation. These questions were entered into ChatGPT, with its responses being analyzed for accuracy and variance in performance. RESULTS: The dataset included 2576 questions (479 multiple-choice and 2097 single-choice), of which 57% (n = 1475) were answered correctly by ChatGPT. An in-depth analysis of question style revealed that single-choice questions were associated with a significantly higher rate (p < 0.001) of correct responses (n = 1313; 63%) compared to multiple-choice questions (n = 162; 34%). Stratified by question categories, ChatGPT yielded the highest rate of correct responses (n = 151; 72%) in the field of allergology, whereas 7 out of 10 questions (n = 65; 71%) on legal otolaryngology aspects were answered incorrectly. CONCLUSION: The study reveals ChatGPT’s potential as a supplementary tool for otolaryngology board certification preparation. However, its propensity for errors in certain otolaryngology areas calls for further refinement. Future research should address these limitations to improve ChatGPT’s educational use. An approach, with expert collaboration, is recommended for the reliable and accurate integration of such AI models. Springer Berlin Heidelberg 2023-06-07 2023 /pmc/articles/PMC10382366/ /pubmed/37285018 http://dx.doi.org/10.1007/s00405-023-08051-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Miscellaneous
Hoch, Cosima C.
Wollenberg, Barbara
Lüers, Jan-Christoffer
Knoedler, Samuel
Knoedler, Leonard
Frank, Konstantin
Cotofana, Sebastian
Alfertshofer, Michael
ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
title ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
title_full ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
title_fullStr ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
title_full_unstemmed ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
title_short ChatGPT’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
title_sort chatgpt’s quiz skills in different otolaryngology subspecialties: an analysis of 2576 single-choice and multiple-choice board certification preparation questions
topic Miscellaneous
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10382366/
https://www.ncbi.nlm.nih.gov/pubmed/37285018
http://dx.doi.org/10.1007/s00405-023-08051-4
work_keys_str_mv AT hochcosimac chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT wollenbergbarbara chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT luersjanchristoffer chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT knoedlersamuel chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT knoedlerleonard chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT frankkonstantin chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT cotofanasebastian chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions
AT alfertshofermichael chatgptsquizskillsindifferentotolaryngologysubspecialtiesananalysisof2576singlechoiceandmultiplechoiceboardcertificationpreparationquestions