BRIEF REPORTS

Creation and Initial Validation of the Mayo Outpatient Precepting Evaluation Tool

Deirdre Paulson, PhD | Brandon Hidaka, MD, PhD | Terri Nordin, MD

Fam Med. 2023;55(8):547-552.

DOI: 10.22454/FamMed.2023.164770

Return to Issue

Abstract

Background and Objectives: Preceptors in family medicine residencies need feedback to improve. When we found no validated, behavior-based tool to assess the outpatient precepting of family medicine residents, we sought to fill this gap by developing and initially validating the Mayo Outpatient Precepting Evaluation Tool (MOPET).

Methods: To develop the MOPET, we applied the Stanford Faculty Development Program (SFDP) theoretical framework for education, more recent work on peer review of medical teaching, and expert review of items. The residency behavioral scientist and a volunteer physician independently completed the MOPET while co-observing a precepting physician during continuity clinic sessions (N=20). We assessed the tool’s validity via interrater reliability and cross-validation with the SFDP-26.

Results: The tool demonstrated high interrater reliability for the following effective teaching behaviors: (a) allowing the resident to present without interrupting, (b) encouraging the formulation of a goal, (c) checking in on the resident’s goal, (d) using multimodal teaching aids, (e) asking to discuss the differential diagnosis, (f) asking to discuss alternative management, (g) encouraging the resident to pursue literature and/or other resources, and (h) reinforcing self-directed learning. The MOPET measures strongly correlated with most items from the SFDP-26, indicating good cross-validity.

Conclusions: The MOPET is a theoretically sound, behavior-based, reliable, and initially validated tool for peer review of outpatient family medicine resident teaching. This tool can support faculty development in outpatient clinical learning environments.

INTRODUCTION

Effective teaching, requiring feedback and deliberate practice, is central to medical residencies. 1 Skeff’s Stanford Faculty Development Program (SFDP) is a fundamental medical teaching framework 2 that has demonstrable positive effects on teaching knowledge, skills, and attitudes, 3 as well as student and resident evaluations of clinical teaching. 4, 5 No preceptor assessment tool applies well to all clinical settings. 6 Many tools reflect learners’ perceptions rather than observation of precepting. 7-9 Peer review is beneficial and less biased because it sidesteps the issues of power differential. 10, 11 Peer teaching observation tools have been validated in settings that do not generalize well to outpatient family medicine (ie, inpatient internal medicine residencies). 12-14 The behavior-based precepting tools for outpatient family medicine were validated on medical student supervision 15 and used preceptor self-assessment,16 not peer assessment. We therefore sought to create and start validating an evidence-based, peer-observation instrument to assess precepting behaviors in a family medicine residency outpatient clinic.

METHODS

Setting

We developed and initially validated the Mayo Outpatient Precepting Evaluation Tool (MOPET) in the outpatient clinic at the Mayo Clinic Family Medicine Residency–Eau Claire program in Wisconsin. This 5-5-5 program had six core physician faculty members (including the program director), one core behavioral scientist faculty member, and four community preceptors. Mayo Clinic’s Education Research Committee (#21-092) and Institutional Review Board (#21-009471) deemed the study exempt from approval.

Tool Development

While using the Patient Centered Observation Form–Clinician Version (PCOF) 17 to observe and provide feedback to residents in the continuity clinic, faculty preceptors inquired about a comparable tool for them. A literature review led to Beckman and colleagues’ work. 12, 14, 18, 19 Finding no tool directly applicable, we created the MOPET.

Beckman granted permission to use and modify the various versions of the Mayo Teaching Evaluation Form (MTEF). 12, 14, 18, 19 We selected and revised MTEF items that consistently demonstrated strong validity. We also created new items derived from the SFDP, 2 PCOF, 17 Maastricht Clinical Teaching Questionnaire, 20 and personal experience. Four core faculty members’ edits of this amalgam resulted in the studied version of the MOPET (Supplemental Figure 1). All items defined an observable, effective teaching behavior and were made quantitative to increase objectivity and reliability by counting the frequency of each behavior.

Procedure

Each observation occurred over a half day of outpatient clinic, allowing for observation of multiple residents and presentations. The behavioral scientist and one volunteer faculty physician observed a precepting physician and independently counted the frequency of teaching behaviors. A precepting encounter was defined as a time a resident entered the precepting space to discuss the care of one or multiple patients. An observed behavior could be tallied only once per precepting encounter. At the end of the half day, observers used the MOPET to share feedback with the preceptor. The behavioral scientist also completed the SFDP-26 for each half day.

We estimated needing 20 observations to achieve at least 90% power to detect a correlation of 0.70 with a type-I error rate of 5%. We quantified the interrater reliability using Spearman’s rank correlation coefficient. We applied the Mann-Whitney U test to determine whether the behavioral scientist’s observations were significantly higher or lower than the physicians’. Using Spearman’s correlation, we assessed the cross-validity of the tool between items of the SFDP-26 and the sum of the relevant MOPET items (Supplemental Table 1). We identified relevant MOPET items (or lack thereof) by consensus. We chose nonparametric tests because only two of the MOPET items and none of the SFDP-26 items had evidence of a normal distribution (Shapiro-Wilk Test, P>.05). We performed all analyses in BlueSky Statistics (R version 3.6.3). 21

RESULTS

Five core faculty members applied the MOPET as an observer alongside the behavioral scientist in 20 precepting sessions of six different preceptors. The most observed behaviors were showing genuine interest, being available and ready to precept, and allowing the resident to present without interruption (Table 1 ). The rarest behaviors were encouraging the resident to ask questions, acknowledging when the resident taught them something, encouraging the resident to pursue the literature and/or other resources, and reinforcing self-directed learning.

The high interrater reliability of the MOPET items is presented in Figure 1. However, observers did not agree on when the preceptor was supporting the resident in activities the learner found difficult. We found weak concordance for prioritization of precepting over other tasks and explaining why the resident was correct or incorrect. The behavioral scientist observed more instances of preceptors supporting the resident in activities the learner found difficult and the use of multimodal learning aids, and fewer explanations for why the resident was correct or incorrect (all P<.05).

The good cross-validity between the SFDP-26 and relevant MOPET items is presented in Figure 2. We found especially strong correlations between the SFDP-26 and MOPET for items related to resident goal formation, use of multimodal learning aids, evaluation of a learner’s ability to analyze and synthesize knowledge, and delivery of feedback. We found no evidence that the MOPET measures correlated with any of the following SFDP-26 items: expressing respect for learners, discouraging external interruptions, explaining relationships in material, or promoting self-directed learning.

DISCUSSION

Our work began the validation of an objective precepting evaluation tool for outpatient family medicine residencies, filling an important gap in family medicine education. High interrater reliability indicates that the MOPET measures consistently recognizable behaviors. Strong correlations between the MOPET and most SFDP-26 items demonstrate cross-validity. Important to note is that the SFDP-26 is a peer’s gestalt of precepting, whereas the MOPET quantifies effective teaching behaviors. Because observers were not formally trained, the interrater reliability was likely lowered; yet our findings suggest that the MOPET requires little training.

Though this study was not powered to observe behavior change among individual preceptors, the behavioral scientist and observed preceptors expressed a strong sentiment of subjective improvement. Additionally, feedback from faculty members during MOPET development and unsolicited praise from observed preceptors regarding the tool’s value during the end-of-session feedback all signified face validity.

The lack of agreement between how frequently the behavioral scientist and peer observer saw the preceptor support a resident in activities the learner found difficult was surprising, but could be improved by including more concrete examples of this behavior in future versions. Another limitation of this study was the element of human error; for example, the correlation among observers for the number of residents in each precepting session was not 1.0 (possibly because additional residents not working in clinic came in to discuss a prescription request or patient case). Although measuring behavior frequency increases the reliability of measurement, it may not be an optimal approach for evaluating precepting behavior for constructive feedback; the best teaching involves using the right combination of behaviors for the specific learner at specific times. Furthermore, evidence-based precepting behaviors are tallied with this tool even when used at inopportune times, which could lead to a less effective educational experience. Lastly, which observable behaviors are most important to target is unclear.

Future research could include (a) re-evaluating after refinement via a modified Delphi method or factor analysis; (b) measuring use of the tool’s effect on preceptor behavior, learner experience, and patient outcomes; (c) assessing external validity in other family medicine residency programs, primary care clinics in other specialties, or other learners; and (d) comparing use to other preceptor improvement tools.

In conclusion, we created and initially validated an evidence-based, peer-observation instrument to support faculty development in the quintessential learning environment of family medicine residency: the continuity clinic.

Presentation

Poster presentation, STFM Annual Conference, Indianapolis, IN, May 2022.

Funding

This study was partially funded by the 2022 Endowment for Education Research Award from the Mayo Clinic College of Medicine and Science.

Acknowledgments

The authors want to thank all those who supported and participated in this study, including faculty, preceptors, and residents.

References

  1. Bowen JL, Irby DM. Assessing quality and costs of education in the ambulatory setting: a review of the literature. Acad Med. 2002;77(7):621-680. doi:10.1097/00001888-200207000-00006
  2. Skeff KM. Enhancing teaching effectiveness and vitality in the ambulatory setting. J Gen Intern Med. 1988;3(2):S26-S33. doi:10.1007/BF02600249
  3. Skeff KM, Stratos GA, Berman J, Bergen MR. Improving clinical teaching: evaluation of a national dissemination program. Arch Intern Med. 1992;152(6):1,156-1,161. doi:10.1001/archinte.1992.00400180028004
  4. Litzelman DK, Stratos GA, Marriott DJ, Skeff KM. Factorial validation of a widely disseminated educational framework for evaluating clinical teachers. Acad Med. 1998;73(6):688-695. doi:10.1097/00001888-199806000-00016
  5. Litzelman DK, Westmoreland GR, Skeff KM, Stratos GA. Factorial validation of an educational framework using residents’ evaluations of clinician-educators. Acad Med. 1999;74(10):S25-S27. doi:10.1097/00001888-199910000-00030
  6. Fluit CR, Bolhuis S, Grol R, Laan R, Wensing M. Assessing the quality of clinical teachers: a systematic review of content and quality of questionnaires for assessing clinical teachers. J Gen Intern Med. 2010;25(12):1,337-1,345. doi:10.1007/s11606-010-1458-y
  7. Castiglioni A, Shewchuk RM, Willett LL, Heudebert GR, Centor RM. A pilot study using nominal group technique to assess residents’ perceptions of successful attending rounds. J Gen Intern Med. 2008;23(7):1,060-1,065. doi:10.1007/s11606-008-0668-z
  8. Huff NG, Roy B, Estrada CA, et al. Teaching behaviors that define highest rated attending physicians: a study of the resident perspective. Med Teach. 2014;36(11):991-996. doi:10.3109/0142159X.2014.920952
  9. Smith CA, Varkey AB, Evans AT, Reilly BM. Evaluating the performance of inpatient attending physicians: a new instrument for today’s teaching hospitals. J Gen Intern Med. 2004;19(7):766-771. doi:10.1111/j.1525-1497.2004.30269.x
  10. Hinrichs L, Judd D, Hernandez M, Rapport M. Peer review of teaching to promote a culture of excellence: A scoping review. J Phys Ther Educ. 2022;36(4):293-302. doi:10.1097/JTE.0000000000000242
  11. Olvet DM, Willey JM, Bird JB, Rabin JM, Pearlman RE, Brenner J. Third year medical students impersonalize and hedge when providing negative upward feedback to clinical faculty. Med Teach. 2021;43(6):700-708. doi:10.1080/0142159X.2021.1892619
  12. Beckman TJ, Lee MC, Rohren CH, Pankratz VS. Evaluating an instrument for the peer review of inpatient teaching. Med Teach. 2003;25(2):131-135. doi:10.1080/0142159031000092508
  13. Mookherjee S, Monash B, Wentworth KL, Sharpe BA. Faculty development for hospitalists: structured peer observation of teaching. J Hosp Med. 2014;9(4):244-250. doi:10.1002/jhm.2151
  14. Beckman TJ, Lee MC, Mandrekar JN. A comparison of clinical teaching evaluations by resident and peer physicians. Med Teach. 2004;26(4):321-325. doi:10.1080/01421590410001678984
  15. Huang WY, Dains JE, Monteiro FM, Rogers JC. Observations on the teaching and learning occurring in offices of community-based family and community medicine clerkship preceptors. Fam Med. 2004;36(2):131-136.
  16. Brink D, Power D, Leppink E. Results of a preceptor improvement project. Fam Med. 2020;52(9):647-652. doi:10.22454/FamMed.2020.675133
  17. Mauksch L. Patient centered observation form–clinician version. University of Washington Department of Family Medicine; June, 2016. Accessed 2022. https://depts.washington.edu/fammed/pcof
  18. Beckman TJ, Mandrekar JN. The interpersonal, cognitive and efficiency domains of clinical teaching: construct validity of a multi-dimensional scale. Med Educ. 2005;39(12):1,221-1,229. doi:10.1111/j.1365-2929.2005.02336.x
  19. Beckman TJ, Cook DA, Mandrekar JN. Factor instability of clinical teaching assessment scores among general internists and cardiologists. Med Educ. 2006;40(12):1,209-1,216. doi:10.1111/j.1365-2929.2006.02632.x
  20. Stalmeijer RE, Dolmans DH, Wolfhagen IH, Muijtjens AM, Scherpbier AJ. The Maastricht clinical teaching questionnaire (MCTQ) as a valid and reliable instrument for the evaluation of clinical teachers. Acad Med. 2010;85(11):1,732-1,738. doi:10.1097/ACM.0b013e3181f554d6
  21. R Foundation. The R project for statistical computing. 2020. Accessed 2022. https://www.R-project.org

Lead Author

Deirdre Paulson, PhD

Affiliations: Department of Family Medicine, Mayo Clinic Health System, Eau Claire, WI

Co-Authors

Brandon Hidaka, MD, PhD - Department of Family Medicine, Mayo Clinic Health System, Eau Claire, WI

Terri Nordin, MD - Department of Family Medicine, Mayo Clinic Health System, Eau Claire, WI

Corresponding Author

Deirdre Paulson, PhD

Correspondence: Department of Family Medicine, Mayo Clinic Health System, Eau Claire, WI

Email: paulson-odonovan.deirdre@mayo.edu

Fetching other articles...

Loading the comment form...

Submitting your comment...

There are no comments for this article.

Downloads & Info

Share

Related Content

Tags

Searching for articles...