1
|
Laohawetwanit T, Apornvirat S, Kantasiripitak C. ChatGPT as a teaching tool: Preparing pathology residents for board examination with AI-generated digestive system pathology tests. Am J Clin Pathol 2024; 162:471-479. [PMID: 38795049 DOI: 10.1093/ajcp/aqae062] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Accepted: 04/24/2024] [Indexed: 05/27/2024] Open
Abstract
OBJECTIVES To evaluate the effectiveness of ChatGPT 4 in generating multiple-choice questions (MCQs) with explanations for pathology board examinations, specifically for digestive system pathology. METHODS The customized ChatGPT 4 model was developed for MCQ and explanation generation. Expert pathologists evaluated content accuracy and relevance. These MCQs were then administered to pathology residents, followed by an analysis focusing on question difficulty, accuracy, item discrimination, and internal consistency. RESULTS The customized ChatGPT 4 generated 80 MCQs covering various gastrointestinal and hepatobiliary topics. While the MCQs demonstrated moderate to high agreement in evaluation parameters such as content accuracy, clinical relevance, and overall quality, there were issues in cognitive level and distractor quality. The explanations were generally acceptable. Involving 9 residents with a median experience of 1 year, the average score was 57.4 (71.8%). Pairwise comparisons revealed a significant difference in performance between each year group (P < .01). The test analysis showed moderate difficulty, effective item discrimination (index = 0.15), and good internal consistency (Cronbach's α = 0.74). CONCLUSIONS ChatGPT 4 demonstrated significant potential as a supplementary educational tool in medical education, especially in generating MCQs with explanations similar to those seen in board examinations. While artificial intelligence-generated content was of high quality, it necessitated refinement and expert review.
Collapse
Affiliation(s)
- Thiyaphat Laohawetwanit
- Division of Pathology, Chulabhorn International College of Medicine, Thammasat University, Pathum Thani, Thailand
- Division of Pathology, Thammasat University Hospital, Pathum Thani, Thailand
| | - Sompon Apornvirat
- Division of Pathology, Chulabhorn International College of Medicine, Thammasat University, Pathum Thani, Thailand
- Division of Pathology, Thammasat University Hospital, Pathum Thani, Thailand
| | - Charinee Kantasiripitak
- Division of Pathology, Chulabhorn International College of Medicine, Thammasat University, Pathum Thani, Thailand
- Division of Pathology, Thammasat University Hospital, Pathum Thani, Thailand
| |
Collapse
|
2
|
Mistry NP, Saeed H, Rafique S, Le T, Obaid H, Adams SJ. Large Language Models as Tools to Generate Radiology Board-Style Multiple-Choice Questions. Acad Radiol 2024; 31:3872-3878. [PMID: 39013736 DOI: 10.1016/j.acra.2024.06.046] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2024] [Revised: 06/24/2024] [Accepted: 06/27/2024] [Indexed: 07/18/2024]
Abstract
RATIONALE AND OBJECTIVES To determine the potential of large language models (LLMs) to be used as tools by radiology educators to create radiology board-style multiple choice questions (MCQs), answers, and rationales. METHODS Two LLMs (Llama 2 and GPT-4) were used to develop 104 MCQs based on the American Board of Radiology exam blueprint. Two board-certified radiologists assessed each MCQ using a 10-point Likert scale across five criteria-clarity, relevance, suitability for a board exam based on level of difficulty, quality of distractors, and adequacy of rationale. For comparison, MCQs from prior American College of Radiology (ACR) Diagnostic Radiology In-Training (DXIT) exams were also assessed using these criteria, with radiologists blinded to the question source. RESULTS Mean scores (±standard deviation) for clarity, relevance, suitability, quality of distractors, and adequacy of rationale were 8.7 (±1.4), 9.2 (±1.3), 9.0 (±1.2), 8.4 (±1.9), and 7.2 (±2.2), respectively, for Llama 2; 9.9 (±0.4), 9.9 (±0.5), 9.9 (±0.4), 9.8 (±0.5), and 9.9 (±0.3), respectively, for GPT-4; and 9.9 (±0.3), 9.9 (±0.2), 9.9 (±0.2), 9.9 (±0.4), and 9.8 (±0.6), respectively, for ACR DXIT items (p < 0.001 for Llama 2 vs. ACR DXIT across all criteria; no statistically significant difference for GPT-4 vs. ACR DXIT). The accuracy of model-generated answers was 69% for Llama 2 and 100% for GPT-4. CONCLUSION A state-of-the art LLM such as GPT-4 may be used to develop radiology board-style MCQs and rationales to enhance exam preparation materials and expand exam banks, and may allow radiology educators to further use MCQs as teaching and learning tools.
Collapse
Affiliation(s)
- Neel P Mistry
- College of Medicine, University of Saskatchewan, Saskatoon, Saskatchewan, Canada (N.P.M., H.S., H.O., S.J.A.); Department of Medical Imaging, Royal University Hospital, Saskatoon, Saskatchewan, Canada (N.P.M., H.O., S.J.A.)
| | - Huzaifa Saeed
- College of Medicine, University of Saskatchewan, Saskatoon, Saskatchewan, Canada (N.P.M., H.S., H.O., S.J.A.)
| | - Sidra Rafique
- Department of Computer Science, University of Saskatchewan, Saskatoon, Saskatchewan, Canada (S.R.)
| | - Thuy Le
- Department of Community Health and Epidemiology, University of Saskatchewan, Saskatoon, Saskatchewan, Canada (T.L.)
| | - Haron Obaid
- College of Medicine, University of Saskatchewan, Saskatoon, Saskatchewan, Canada (N.P.M., H.S., H.O., S.J.A.); Department of Medical Imaging, Royal University Hospital, Saskatoon, Saskatchewan, Canada (N.P.M., H.O., S.J.A.)
| | - Scott J Adams
- College of Medicine, University of Saskatchewan, Saskatoon, Saskatchewan, Canada (N.P.M., H.S., H.O., S.J.A.); Department of Medical Imaging, Royal University Hospital, Saskatoon, Saskatchewan, Canada (N.P.M., H.O., S.J.A.).
| |
Collapse
|
3
|
Johnson M, Ribeiro AP, Drew TM, Pereira PNR. Generative AI use in dental education: Efficient exam item writing. J Dent Educ 2023; 87 Suppl 3:1865-1866. [PMID: 37354022 DOI: 10.1002/jdd.13294] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2023] [Accepted: 06/08/2023] [Indexed: 06/25/2023]
Affiliation(s)
- Margeaux Johnson
- College of Dentistry, University of Florida, Restorative Dental Sciences, Division of Operative Dentistry, Gainesville, Florida, USA
| | - Ana P Ribeiro
- College of Dentistry, University of Florida, Restorative Dental Sciences, Division of Operative Dentistry, Gainesville, Florida, USA
| | - Tiffany M Drew
- College of Dentistry, University of Florida, Restorative Dental Sciences, Division of Operative Dentistry, Gainesville, Florida, USA
| | - Patricia N R Pereira
- College of Dentistry, University of Florida, Restorative Dental Sciences, Division of Operative Dentistry, Gainesville, Florida, USA
| |
Collapse
|
5
|
Slanetz PJ, Deitte LA. Advancing the Science of Radiology Education. J Am Coll Radiol 2022; 19:685-686. [PMID: 35395210 DOI: 10.1016/j.jacr.2022.02.034] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2022] [Accepted: 02/19/2022] [Indexed: 11/18/2022]
Affiliation(s)
- Priscilla J Slanetz
- Vice Chair of Academic Affairs in the Department of Radiology and Associate Program Director of the Diagnostic Radiology Residency, Boston University Medical Center, Boston, Massachusetts, and also is from Boston University School of Medicine, Boston, Massachusetts; Director of Early Career Faculty Development and Co-Director of the Academic Writing Program for Boston University Medical Group; President of Massachusetts Radiological Society; Vice President of the Association of University Radiologists; and Subspecialty Chair of the ACR Appropriateness Criteria Breast Imaging Panels.
| | - Lori A Deitte
- Vice Chair of Education, Department of Radiology, Vanderbilt University Medical Center, Nashville, Tennessee; Chair of the Commission on Publications and Lifelong Learning; and Member of the Board of Chancellors of the American College of Radiology
| |
Collapse
|