ChatGPT-4 Struggles Interpreting Radiological Anatomy in FRCR Part 1 Mock Exam, Study Finds

Published On 2024-11-29 12:37 GMT   |   Update On 2024-11-29 12:37 GMT
Advertisement

India: With artificial intelligence growing in hospitals to improve patient outcomes, Indian doctors conducted a study that evaluated the efficacy of ChatGPT-4 in answering radiological anatomy and found that ChatGPT-4 has underperformed in interpreting normal radiological anatomy. The study was published in the Indian Journal of Radiology and Imaging.

Dr. Pradosh Kumar Sarangi, MD, PDF, EDiR, Department of Radiodiagnosis, AIIMS Deoghar, Jharkhand, and the study's author, explained to Medical Dialogues, "The study evaluated ChatGPT-4's performance in identifying radiological anatomy in mock First Fellowship of the Royal College of Radiologists (FRCR) Part 1 examination questions. While the AI demonstrated 100% accuracy in identifying imaging modalities, it struggled with anatomical structure identification (4-7.5% accuracy) and sidedness (~40%). This highlights its potential as a supplementary educational tool but underscores significant limitations for clinical use.”

Advertisement

Anatomical knowledge is essential in radiology, providing the foundation for accurate image interpretation and diagnosis across imaging modalities like X-rays, CT scans, MRI, and ultrasound. As AI advances, its role in medical education and exam preparation has expanded. OpenAI's ChatGPT-4, known for its capabilities in natural language processing, has shown promise across various medical fields. The FRCR Part 1 examination, a key milestone for radiology trainees, tests their knowledge of radiological anatomy and other core topics.

Dr. Sarangi and his team conducted a study to assess ChatGPT-4's ability to identify radiological anatomy in line with the FRCR Part 1 pattern. They used 100 mock radiological anatomy questions from a free website mimicking the exam. ChatGPT-4 was tested with and without context about the exam instructions. The main question was: “Identify the structure indicated by the arrow(s).” Responses were compared to correct answers, with two expert radiologists ((>5 and 30 years of experience) rating the explanations.

Four scores were assessed: correctness, sidedness, modality identification, and approximation. The latter accounts for partial correctness when the structure identified is present, but not the specific focus of the question.

The study reveals the following findings:

  • ChatGPT-4 performed poorly in both testing conditions, with correctness scores of 4% without context and 7.5% with context.
  • It identified the imaging modality with 100% accuracy and scored over 50% on the approximation metric by recognizing structures present in the image but not indicated by the arrow.
  • ChatGPT-4 had difficulty identifying the correct side of the structure, scoring around 42% without context and 40% with context. only 32% of the responses were consistent across both settings.

Dr. Sarangi explained that “ChatGPT-4’s current limitations in accuracy and reliability restrict its utility for high-stakes clinical applications, emphasizing the need for improved training on domain-specific datasets.”

Discussing the future evolution of AI in radiology, Dr. Sarangi added,” AI in radiology is likely to evolve into hybrid systems integrating visual recognition with language models, improving interpretative skills for complex imaging. Future advancements could include real-time AI support for identifying anatomical structures during imaging reviews, improved clinical decision support tools, and personalized learning platforms for radiology students.”

Reference: Sarangi, P. K., Datta, S., Panda, B. B., Panda, S., & Mondal, H. (2024). Evaluating ChatGPT-4's performance in identifying radiological anatomy in FRCR Part 1 examination questions. Indian Journal of Radiology and Imaging. https://doi.org/10.1055/s-0044-1792040.


Tags:    
Article Source : Indian Journal of Radiology and Imaging

Disclaimer: This website is primarily for healthcare professionals. The content here does not replace medical advice and should not be used as medical, diagnostic, endorsement, treatment, or prescription advice. Medical science evolves rapidly, and we strive to keep our information current. If you find any discrepancies, please contact us at corrections@medicaldialogues.in. Read our Correction Policy here. Nothing here should be used as a substitute for medical advice, diagnosis, or treatment. We do not endorse any healthcare advice that contradicts a physician's guidance. Use of this site is subject to our Terms of Use, Privacy Policy, and Advertisement Policy. For more details, read our Full Disclaimer here.

NOTE: Join us in combating medical misinformation. If you encounter a questionable health, medical, or medical education claim, email us at factcheck@medicaldialogues.in for evaluation.

Our comments section is governed by our Comments Policy . By posting comments at Medical Dialogues you automatically agree with our Comments Policy , Terms And Conditions and Privacy Policy .

Similar News