Cargando…

Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum

Background and objective ChatGPT is an artificial intelligence (AI) language model that has been trained to process and respond to questions across a wide range of topics. It is also capable of solving problems in medical educational topics. However, the capability of ChatGPT to accurately answer fi...

Descripción completa

Detalles Bibliográficos
Autores principales: Das, Dipmala, Kumar, Nikhil, Longjam, Langamba Angom, Sinha, Ranwir, Deb Roy, Asitava, Mondal, Himel, Gupta, Pratima
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Cureus 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10086829/
https://www.ncbi.nlm.nih.gov/pubmed/37056538
http://dx.doi.org/10.7759/cureus.36034
_version_ 1785022228090322944
author Das, Dipmala
Kumar, Nikhil
Longjam, Langamba Angom
Sinha, Ranwir
Deb Roy, Asitava
Mondal, Himel
Gupta, Pratima
author_facet Das, Dipmala
Kumar, Nikhil
Longjam, Langamba Angom
Sinha, Ranwir
Deb Roy, Asitava
Mondal, Himel
Gupta, Pratima
author_sort Das, Dipmala
collection PubMed
description Background and objective ChatGPT is an artificial intelligence (AI) language model that has been trained to process and respond to questions across a wide range of topics. It is also capable of solving problems in medical educational topics. However, the capability of ChatGPT to accurately answer first- and second-order knowledge questions in the field of microbiology has not been explored so far. Hence, in this study, we aimed to analyze the capability of ChatGPT in answering first- and second-order questions on the subject of microbiology. Materials and methods Based on the competency-based medical education (CBME) curriculum of the subject of microbiology, we prepared a set of first-order and second-order questions. For the total of eight modules in the CBME curriculum for microbiology, we prepared six first-order and six second-order knowledge questions according to the National Medical Commission-recommended CBME curriculum, amounting to a total of (8 x 12) 96 questions. The questions were checked for content validity by three expert microbiologists. These questions were used to converse with ChatGPT by a single user and responses were recorded for further analysis. The answers were scored by three microbiologists on a rating scale of 0-5. The average of three scores was taken as the final score for analysis. As the data were not normally distributed, we used a non-parametric statistical test. The overall scores were tested by a one-sample median test with hypothetical values of 4 and 5. The scores of answers to first-order and second-order questions were compared by the Mann-Whitney U test. Module-wise responses were tested by the Kruskall-Wallis test followed by the post hoc test for pairwise comparisons. Results The overall score of 96 answers was 4.04 ±0.37 (median: 4.17, Q1-Q3: 3.88-4.33) with the mean score of answers to first-order knowledge questions being 4.07 ±0.32 (median: 4.17, Q1-Q3: 4-4.33) and that of answers to second-order knowledge questions being 3.99 ±0.43 (median: 4, Q1-Q3: 3.67-4.33) (Mann-Whitney p=0.4). The score was significantly below the score of 5 (one-sample median test p<0.0001) and similar to 4 (one-sample median test p=0.09). Overall, there was a variation in median scores obtained in eight categories of topics in microbiology, indicating inconsistent performance in different topics. Conclusion The results of the study indicate that ChatGPT is capable of answering both first- and second-order knowledge questions related to the subject of microbiology. The model achieved an accuracy of approximately 80% and there was no difference between the model's capability of answering first-order questions and second-order knowledge questions. The findings of this study suggest that ChatGPT has the potential to be an effective tool for automated question-answering in the field of microbiology. However, continued improvement in the training and development of language models is necessary to enhance their performance and make them suitable for academic use.
format Online
Article
Text
id pubmed-10086829
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Cureus
record_format MEDLINE/PubMed
spelling pubmed-100868292023-04-12 Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum Das, Dipmala Kumar, Nikhil Longjam, Langamba Angom Sinha, Ranwir Deb Roy, Asitava Mondal, Himel Gupta, Pratima Cureus Medical Education Background and objective ChatGPT is an artificial intelligence (AI) language model that has been trained to process and respond to questions across a wide range of topics. It is also capable of solving problems in medical educational topics. However, the capability of ChatGPT to accurately answer first- and second-order knowledge questions in the field of microbiology has not been explored so far. Hence, in this study, we aimed to analyze the capability of ChatGPT in answering first- and second-order questions on the subject of microbiology. Materials and methods Based on the competency-based medical education (CBME) curriculum of the subject of microbiology, we prepared a set of first-order and second-order questions. For the total of eight modules in the CBME curriculum for microbiology, we prepared six first-order and six second-order knowledge questions according to the National Medical Commission-recommended CBME curriculum, amounting to a total of (8 x 12) 96 questions. The questions were checked for content validity by three expert microbiologists. These questions were used to converse with ChatGPT by a single user and responses were recorded for further analysis. The answers were scored by three microbiologists on a rating scale of 0-5. The average of three scores was taken as the final score for analysis. As the data were not normally distributed, we used a non-parametric statistical test. The overall scores were tested by a one-sample median test with hypothetical values of 4 and 5. The scores of answers to first-order and second-order questions were compared by the Mann-Whitney U test. Module-wise responses were tested by the Kruskall-Wallis test followed by the post hoc test for pairwise comparisons. Results The overall score of 96 answers was 4.04 ±0.37 (median: 4.17, Q1-Q3: 3.88-4.33) with the mean score of answers to first-order knowledge questions being 4.07 ±0.32 (median: 4.17, Q1-Q3: 4-4.33) and that of answers to second-order knowledge questions being 3.99 ±0.43 (median: 4, Q1-Q3: 3.67-4.33) (Mann-Whitney p=0.4). The score was significantly below the score of 5 (one-sample median test p<0.0001) and similar to 4 (one-sample median test p=0.09). Overall, there was a variation in median scores obtained in eight categories of topics in microbiology, indicating inconsistent performance in different topics. Conclusion The results of the study indicate that ChatGPT is capable of answering both first- and second-order knowledge questions related to the subject of microbiology. The model achieved an accuracy of approximately 80% and there was no difference between the model's capability of answering first-order questions and second-order knowledge questions. The findings of this study suggest that ChatGPT has the potential to be an effective tool for automated question-answering in the field of microbiology. However, continued improvement in the training and development of language models is necessary to enhance their performance and make them suitable for academic use. Cureus 2023-03-12 /pmc/articles/PMC10086829/ /pubmed/37056538 http://dx.doi.org/10.7759/cureus.36034 Text en Copyright © 2023, Das et al. https://creativecommons.org/licenses/by/3.0/This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Medical Education
Das, Dipmala
Kumar, Nikhil
Longjam, Langamba Angom
Sinha, Ranwir
Deb Roy, Asitava
Mondal, Himel
Gupta, Pratima
Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum
title Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum
title_full Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum
title_fullStr Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum
title_full_unstemmed Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum
title_short Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum
title_sort assessing the capability of chatgpt in answering first- and second-order knowledge questions on microbiology as per competency-based medical education curriculum
topic Medical Education
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10086829/
https://www.ncbi.nlm.nih.gov/pubmed/37056538
http://dx.doi.org/10.7759/cureus.36034
work_keys_str_mv AT dasdipmala assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum
AT kumarnikhil assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum
AT longjamlangambaangom assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum
AT sinharanwir assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum
AT debroyasitava assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum
AT mondalhimel assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum
AT guptapratima assessingthecapabilityofchatgptinansweringfirstandsecondorderknowledgequestionsonmicrobiologyaspercompetencybasedmedicaleducationcurriculum