Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC5498683PMC
http://dx.doi.org/10.2147/AMEP.S141505DOI Listing

Publication Analysis

Top Keywords

distractors multiple-choice
4
multiple-choice questions
4
questions medical
4
medical student
4
student perspective
4
distractors
1
questions
1
medical
1
student
1
perspective
1

Similar Publications

Large language models (LLMs) such as ChatGPT and Gemini are increasingly used to generate educational content in medical education, including multiple-choice questions (MCQs), but their effectiveness compared to expert-written questions remains underexplored, particularly in anatomy. We conducted a cross-sectional, mixed-methods study involving Year 2-4 medical students at Qatar University, where participants completed and evaluated three anonymized MCQ sets-authored by ChatGPT, Google-Gemini, and a clinical anatomist-across 17 quality criteria. Descriptive and chi-square analyses were performed, and optional feedback was reviewed thematically.

View Article and Find Full Text PDF

Purpose: To compare the performance of multiple large language models (LLMs) on a practice radiography certification exam.

Method: Using an exploratory, nonexperimental approach, 200 multiple-choice question stems and options (correct answers and distractors) from a practice radiography certification exam were entered into 5 LLMs: ChatGPT (OpenAI), Claude (Anthropic), Copilot (Microsoft), Gemini (Google), and Perplexity (Perplexity AI). Responses were recorded as correct or incorrect, and overall accuracy rates were calculated for each LLM.

View Article and Find Full Text PDF

: Otolaryngologists are increasingly using head and neck ultrasound (HNUS). Determining whether a practitioner of HNUS has achieved adequate theoretical knowledge remains a challenge. This study assesses the performance of two large language models (LLMs) in generating multiple-choice questions (MCQs) for head and neck ultrasound, compared with MCQs generated by physicians.

View Article and Find Full Text PDF

Objective: To examine the impact of quality parameters in the construction of multiple-choice questions (MCQs) and their associated psychometric analysis for a selected Specialty X (SpX) in the Qualifying Residency Entry Exam (QRE) at a Postgraduate Medical Institute.

Methods: A post-validation cross-sectional analytical study was conducted using a non-probability purposive judgmental sampling technique. The SpX was chosen from one clinical specialities with the lowest exam success rates among the 52 specialities in the 2020-2023 QRE cycles.

View Article and Find Full Text PDF

Aim: To compare the item difficulty and discriminative index of multiple-choice questions (MCQs) generated by ChatGPT with those created by dental educators, based on the performance of dental students in a real exam setting.

Materials And Methods: A total of 40 MCQs-20 generated by ChatGPT 4.0 and 20 by dental educators-were developed based on the Oral Diagnosis and Radiology course content.

View Article and Find Full Text PDF