Cargando…
ChatGPT in medical school: how successful is AI in progress testing?
BACKGROUND: As generative artificial intelligence (AI), ChatGPT provides easy access to a wide range of information, including factual knowledge in the field of medicine. Given that knowledge acquisition is a basic determinant of physicians’ performance, teaching and testing different levels of medi...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Taylor & Francis
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10262795/ https://www.ncbi.nlm.nih.gov/pubmed/37307503 http://dx.doi.org/10.1080/10872981.2023.2220920 |
_version_ | 1785058109416275968 |
---|---|
author | Friederichs, Hendrik Friederichs, Wolf Jonas März, Maren |
author_facet | Friederichs, Hendrik Friederichs, Wolf Jonas März, Maren |
author_sort | Friederichs, Hendrik |
collection | PubMed |
description | BACKGROUND: As generative artificial intelligence (AI), ChatGPT provides easy access to a wide range of information, including factual knowledge in the field of medicine. Given that knowledge acquisition is a basic determinant of physicians’ performance, teaching and testing different levels of medical knowledge is a central task of medical schools. To measure the factual knowledge level of the ChatGPT responses, we compared the performance of ChatGPT with that of medical students in a progress test. METHODS: A total of 400 multiple-choice questions (MCQs) from the progress test in German-speaking countries were entered into ChatGPT’s user interface to obtain the percentage of correctly answered questions. We calculated the correlations of the correctness of ChatGPT responses with behavior in terms of response time, word count, and difficulty of a progress test question. RESULTS: Of the 395 responses evaluated, 65.5% of the progress test questions answered by ChatGPT were correct. On average, ChatGPT required 22.8 s (SD 17.5) for a complete response, containing 36.2 (SD 28.1) words. There was no correlation between the time used and word count with the accuracy of the ChatGPT response (correlation coefficient for time rho = −0.08, 95% CI [−0.18, 0.02], t(393) = −1.55, p = 0.121; for word count rho = −0.03, 95% CI [−0.13, 0.07], t(393) = −0.54, p = 0.592). There was a significant correlation between the difficulty index of the MCQs and the accuracy of the ChatGPT response (correlation coefficient for difficulty: rho = 0.16, 95% CI [0.06, 0.25], t(393) = 3.19, p = 0.002). CONCLUSION: ChatGPT was able to correctly answer two-thirds of all MCQs at the German state licensing exam level in Progress Test Medicine and outperformed almost all medical students in years 1–3. The ChatGPT answers can be compared with the performance of medical students in the second half of their studies. |
format | Online Article Text |
id | pubmed-10262795 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Taylor & Francis |
record_format | MEDLINE/PubMed |
spelling | pubmed-102627952023-06-15 ChatGPT in medical school: how successful is AI in progress testing? Friederichs, Hendrik Friederichs, Wolf Jonas März, Maren Med Educ Online Research Article BACKGROUND: As generative artificial intelligence (AI), ChatGPT provides easy access to a wide range of information, including factual knowledge in the field of medicine. Given that knowledge acquisition is a basic determinant of physicians’ performance, teaching and testing different levels of medical knowledge is a central task of medical schools. To measure the factual knowledge level of the ChatGPT responses, we compared the performance of ChatGPT with that of medical students in a progress test. METHODS: A total of 400 multiple-choice questions (MCQs) from the progress test in German-speaking countries were entered into ChatGPT’s user interface to obtain the percentage of correctly answered questions. We calculated the correlations of the correctness of ChatGPT responses with behavior in terms of response time, word count, and difficulty of a progress test question. RESULTS: Of the 395 responses evaluated, 65.5% of the progress test questions answered by ChatGPT were correct. On average, ChatGPT required 22.8 s (SD 17.5) for a complete response, containing 36.2 (SD 28.1) words. There was no correlation between the time used and word count with the accuracy of the ChatGPT response (correlation coefficient for time rho = −0.08, 95% CI [−0.18, 0.02], t(393) = −1.55, p = 0.121; for word count rho = −0.03, 95% CI [−0.13, 0.07], t(393) = −0.54, p = 0.592). There was a significant correlation between the difficulty index of the MCQs and the accuracy of the ChatGPT response (correlation coefficient for difficulty: rho = 0.16, 95% CI [0.06, 0.25], t(393) = 3.19, p = 0.002). CONCLUSION: ChatGPT was able to correctly answer two-thirds of all MCQs at the German state licensing exam level in Progress Test Medicine and outperformed almost all medical students in years 1–3. The ChatGPT answers can be compared with the performance of medical students in the second half of their studies. Taylor & Francis 2023-06-12 /pmc/articles/PMC10262795/ /pubmed/37307503 http://dx.doi.org/10.1080/10872981.2023.2220920 Text en © 2023 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The terms on which this article has been published allow the posting of the Accepted Manuscript in a repository by the author(s) or with their consent. |
spellingShingle | Research Article Friederichs, Hendrik Friederichs, Wolf Jonas März, Maren ChatGPT in medical school: how successful is AI in progress testing? |
title | ChatGPT in medical school: how successful is AI in progress testing? |
title_full | ChatGPT in medical school: how successful is AI in progress testing? |
title_fullStr | ChatGPT in medical school: how successful is AI in progress testing? |
title_full_unstemmed | ChatGPT in medical school: how successful is AI in progress testing? |
title_short | ChatGPT in medical school: how successful is AI in progress testing? |
title_sort | chatgpt in medical school: how successful is ai in progress testing? |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10262795/ https://www.ncbi.nlm.nih.gov/pubmed/37307503 http://dx.doi.org/10.1080/10872981.2023.2220920 |
work_keys_str_mv | AT friederichshendrik chatgptinmedicalschoolhowsuccessfulisaiinprogresstesting AT friederichswolfjonas chatgptinmedicalschoolhowsuccessfulisaiinprogresstesting AT marzmaren chatgptinmedicalschoolhowsuccessfulisaiinprogresstesting |