Published in:
01-05-2016 | Original Research
Rating the Quality of Entrustable Professional Activities: Content Validation and Associations with the Clinical Context
Authors:
Jason A. Post, M.D., Christopher M. Wittich, M.D., Pharm.D., Kris G. Thomas, M.D., Denise M. Dupras, M.D., Ph.D., Andrew J. Halvorsen, M.S., Jay N. Mandrekar, Ph.D., Amy S. Oxentenko, M.D., Thomas J. Beckman, M.D.
Published in:
Journal of General Internal Medicine
|
Issue 5/2016
Login to get access
ABSTRACT
BACKGROUND
Entrustable professional activities (EPAs) have been developed to assess resident physicians with respect to Accreditation Council for Graduate Medical Education (ACGME) competencies and milestones. Although the feasibility of using EPAs has been reported, we are unaware of previous validation studies on EPAs and potential associations between EPA quality scores and characteristics of educational programs.
OBJECTIVES
Our aim was to validate an instrument for assessing the quality of EPAs for assessment of internal medicine residents, and to examine associations between EPA quality scores and features of rotations.
DESIGN
This was a prospective content validation study to design an instrument to measure the quality of EPAs that were written for assessing internal medicine residents.
PARTICIPANTS
Residency leadership at Mayo Clinic, Rochester participated in this study. This included the Program Director, Associate program directors and individual rotation directors.
INTERVENTIONS
The authors reviewed salient literature. Items were developed to reflect domains of EPAs useful for assessment. The instrument underwent further testing and refinement. Each participating rotation director created EPAs that they felt would be meaningful to assess learner performance in their area. These 229 EPAs were then assessed with the QUEPA instrument to rate the quality of each EPA.
MAIN MEASURES
Performance characteristics of the QUEPA are reported. Quality ratings of EPAs were compared to the primary ACGME competency, inpatient versus outpatient setting and specialty type.
KEY RESULTS
QUEPA tool scores demonstrated excellent reliability (ICC range 0.72 to 0.94). Higher ratings were given to inpatient versus outpatient (3.88, 3.66; p = 0.03) focused EPAs. Medical knowledge EPAs scored significantly lower than EPAs assessing other competencies (3.34, 4.00; p < 0.0001).
CONCLUSIONS
The QUEPA tool is supported by good validity evidence and may help in rating the quality of EPAs developed by individual programs. Programs should take care when writing EPAs for the outpatient setting or to assess medical knowledge, as these tended to be rated lower.