LETTER TO THE EDITOR

Response to “Exploring the Applications of ChatGPT in Family Medicine Medical Education”

Jason A. Horton, PhD | Irshad Ally, MD

PRiMER. 2023;7:28.

Published: 8/25/2023 | DOI: 10.22454/PRiMER.2023.940827

To the Editor:

“…may you live in interesting times…”

This is indeed a very interesting time to be in medical education, and in light of the rapid evolution and broad availability of artificial intelligence (AI) platforms, this curse may ring especially true without thoughtful guidelines and establishment of best practices with sufficient flexibility to evolve with the technology. In his letter to the editor,1 Dr Hanna offers several novel applications of generative AI models, such as ChatGPT, for enhancement of the learner experience as a complement to traditional pedagogical approaches. We agree that language-generative AI technology may provide unique and realistic simulation of patient encounters, and this is the most intriguing prospect suggested by Dr Hanna. However, we have great concern regarding the use of generative AI to generate differential diagnosis exercises, facilitate exam preparation, or produce patient education materials due to the frequency of false, misleading and/or “hallucinatory” fabrications of information.2 These fabrications may expose the learner to unrealistic patient presentations while false and/or biased information could miseducate the learner or patient.3 Fabricated information and biased information may also contribute to the misdiagnosis or mismanagement of patients.4 We are also concerned with the possibility that protected health information could mistakenly be divulged in the course of dialog with an AI used to assist with clinical decision support because most, if not all, end-user agreements for AI products state that the information provided by the user is not confidential and will likely be used to train the model.5

As Dr Hanna rightly notes, enthusiasm for this rapidly evolving tool must be tempered by a sober consideration of the risks that such AI technologies may present. We share their enthusiasm for all of the promise that these technologies hold, but also cannot overstate the point that their use in the academy must, at a minimum, be supervised by an experienced preceptor who can review, identify, and remediate bad information from AI-generated educational materials. Furthermore, in the absence of best practices in this area of medical education, we encourage development of thoughtful and forward-looking guidelines for the use of these technologies that can promote responsible, ethical, and productive use, but also mitigate the adverse consequences of misuse—both accidental and intentional.

References

  1. Hanna K. Exploring the applications of ChatGPT in family medicine education: five innovative ways for faculty integration. PRiMER Peer-Rev Rep Med Educ Res. 2023;7:26. doi:10.22454/PRiMER.2023.985351
  2. De Angelis L, Baglivo F, Arzilli G, et al. ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health. Front Public Health. 2023;11:1166120. doi:10.3389/fpubh.2023.1166120
  3. Abd-Alrazaq A, AlSaad R, Alhuwail D, et al. Large language models in medical education: opportunities, challenges, and future directions. JMIR Med Educ. 2023;9:e48291. doi:10.2196/48291
  4. Sallam M. ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthcare (Basel). 2023;11(6):887. doi:10.3390/healthcare11060887
  5. Dossantos J, An J, Javan R. Eyes on AI: ChatGPT’s Transformative Potential Impact on Ophthalmology. Cureus. 2023;15(6):e40765. doi:10.7759/cureus.40765

Lead Author

Jason A. Horton, PhD

Affiliations: Departments of Orthopedic Surgery, Radiation Oncology and Cell & Developmental Biology, SUNY Upstate Medical University, Syracuse NY

Co-Authors

Irshad Ally, MD - Department of Family Medicine, SUNY Upstate Medical University, Syracuse NY

Fetching other articles...

Loading the comment form...

Submitting your comment...

There are no comments for this article.

Downloads & Info

Share

Related Content

Tags

Searching for articles...