ColleaguesPlease find the latest draft of my proposal for ESCape tomorrow. I am sorry for this being late but it has changed significantly over the weekend after input from Pat O on Friday. I am looking forward to presenting to you all tomorrow. There are several areas I would most appreciate feedback (although any ideas are welcome). 1- What should be the balance of resident driven assessment and faculty driven assessment for entrustment of trainees in GME. 2- How to best assess the implementation of this system (ie: process measurements- completion rates, time until completion vs. outcomes measures- satisfaction, EPAs vs. milestones for assessment) 3- Amount of time needed to ask a data management consultant to be involved to help us prepare and collect appropriate assessment data Thanks again! Jeff Creation of a learner driven GME EPA assessment network for internal medicine residency and fellowship programs. Background Assessment of graduate medical education trainees is a high priority of training programs. Training program directors must be able to certify that graduates are able to practice independently in their respective specialty/subspecialty. Various frameworks have arisen to guide how to assess and report the competence of trainees. Increasingly, professional milestones are used by program directors to ensure that trainees are ready for independent practice In the case of internal medicine, both the body accrediting programs (ACGME) and the organization certifying individuals (American Board of Internal Medicine (ABIM)) have had major roles in structuring how program directors assess trainees in medicine specialties and subspecialties. Their philosophy has evolved over time to include greater proscription of this assessment. This evolution began in 1999 when the ACGME introduced the six core competencies (patient care, medical knowledge, systems-based practice, practice-based learning, communication and professionalism). This began a transition of philosophy of accreditation from process to training outcomes. The recent introduction of the Next Accreditation System (NAS) evolved these concepts to include milestones as subsections of the core competencies. The internal medicine community developed 142 educational milestones, now termed curricular milestones, to further describe the knowledge, skills and attitudes of observable development from a novice to a proficient trainee. The developmental milestones were recognized as being too extensive to be effectively implemented or evaluated. A second effort produced the reporting milestones, a narrative based set of 22 milestones. The 6 ACGME competency domains organize these 22 milestones such that each competency now has 2 to 5 subcompetencies. (2) An alternative approach to assessment involves using entrustable professional activities (EPAs) as a framework for assessing competence. EPAs reflect the degree of mastery of professional tasks that together constitute the work of a profession. (3) Carrying EPAs requires abilities in multiple competencies. This approach recognizes that professional tasks are complex and breaking those tasks down to discreet, measurable steps, such as might occur with a milestone based approach, loses the ability to provide a holistic view of the successful acquisition of professional skills. These EPAs are best evaluated by observation of professional skills by experienced faculty members. Current UCSF Department of Medicine Assessment Strategy: As required by the ACGME and the ABIM, all Department of Medicine (DOM) ACGME training programs assess the specialty/subspecialty specific milestones; the reporting milestones are recorded in the Evalue platform system. Programs may deploy different tools (global assessments, procedure observations, 360’s, chart stimulated recall, qualitative comments) to assess their program specific milestones but most programs depend upon observational methods to record the initial milestone assessment. Clinical competency committees (CCCs) then meet to determine the summative milestone score that is reported to the ACGME and the ABIM. The CCCs are largely dependent upon the milestone scores obtained from direct faculty observations. Analysis of Current Milestone Assessment Strategy Program CCCs are reporting issues with their newly implemented milestone assessment processes that have been in effect since AY2013-14 for the internal medicine residency program and AY2014-15 for the fellowship programs. First, compliance with completing the assessments is poor; across the DOM training programs, the overall compliance of assigned evaluations is initially only 50%.. Much administrative time is spent trying to get completed evaluations to the required 70% level and often program leadership has to step in to improve faculty compliance. Reasons faculty cite for their challenges in completing assessments is “assessment fatigue” caused by multiple end of the month evaluations being due simultaneously, the need to assess as many as 22 milestones as well as overall frustration with the Evalue platform. Second, the quality of the assessments is also in question. CCCs are now often basing their judgments too few numeric scores on each milestone category bringing their validity into question. This has led CCCs to have concerns about whether they are fully able to assess and report trainee’s performances. Similarly, it creates doubt that CCCs are able to effectively identify areas of improvement for advisors to help trainees. In response to these concerns about a milestone assessment based system and wanting to improve trainee assessment, the internal medicine residency program assessment leadership team began exploring using an EPA based assessment system in 2013. In 2013-4 we utilized a Delphi method to implement a group of 8 EPAs to help make our assessments more activity specific. (6) The EPAs are housed on the UCSFEPA.org website and contain the following assessments: •Serious Illness Communication EPA •Rounds Observation EPA •Discharge summary EPA •4 habits clinic survey •Clinic Post-Discharge EPA •Acute Care EPA •Code Leadership EPA •Multidisciplinary Rounds Inpatient Feedback These assessments are recorded utilizing Qualtrics through the UCSF MyAccess website and are accessed by the faculty to generate evaluations. Several internal medicine fellowships (Palliative Care, Geriatrics and Hematology/Oncology) have begun to use EPAS and the UCSFEPA.org platform for assessment of their trainees. Each of these EPAs can be mapped to several milestones at the same time as is shown in the following chart from proposed gynecology residency EPAs. Benefits of using EPAs includes improved overall compliance with obtaining assessments due to trainee initiation of the assessment, real time assessment by the faculty and ease with utilizing the Qualtrics platform. Although not formally compared, there is a sense that the quality of the EPA assessments may be better than the milestone based assessments. Although these, and future EPAs, consist of a robust network of evaluations, we have been unable to fully implement them due to several limitations with our assessment technology and expertise. The GME office is now able to address these challenges by adding administrative leadership and data management assistance. This help creates a firm infrastructure but to create a trainee centered, interactive assessment network we will need additional help from Information Technology (IT) to create a more seamless platform. Past Assessment Efforts: Over the last four years we have made a concerted effort to bring a coordinated IT effort to enable utilizing our EPA network. Our initial efforts utilized the Mahara portfolio platform. Conceptually this system offered several advantages including resident control over their assessments on their own customizable webpage. Material could be collected and correlated for review with trusted advisors and create a framework for individual assessment that could highlight a trainee’s skills. Unfortunately, residents did not utilize this tool for several reasons. First, residents were too busy to log on to the system to record evaluations. Second, they did not feel empowered to ask for completion of evaluations. Third, since the Mahara system was completely resident driven there was no way to reliably track resident completion of educational tasks. Efforts to move Myfolio system through the Evalue platform were similarly unsuccessful due to data management problems and the inability to design an appropriate single log-in access for faculty or residents to review. Also, residents did not have ability to control their assessment information to pick their best work to present to the CCC. Last year we moved our EPAs to the Qualtrics survey platform and housed our EPAs on a website created through consultation with Eric Widera MD, from the department of geriatrics and palliative care. This system created a much easier system to collect EPA data, but currently our faculty completion rates remain low as there is are no automatic reminder systems to help busy clinicians to remember to complete these evaluations. The EPAs that are most successful have a faculty champion, such as Heather Whelan MD, a hospitalist at the VAMC. Each month she circulates a meeting list to our ward interns to schedule time to complete this EPA. This activity has created popular feedback for our interns and is quite well received by our housestaff. This individual effort is time intensive, however, and unlikely to be sustainable leaving our EPA system longevity in doubt. During this time we began the process of seeking more information from our trainees about what their goals are for their assessment process by doing focus groups with convenience samples of residents. Focus Group Themes: Eight focus groups in our internal medicine residency (and fellowship) programs over the last three years involving approximately one-hundred trainees have demonstrated several themes regarding our assessments. These themes were derived from our focus group notes and match concerns voiced by resident trainees across the country when there is inadequate supervision or feedback. (4-5) First, residency assessment that depends on summative feedback from end of rotation, milestone-based assessments, lack specificity to enable residents to improve their professional practice. Residents would prefer more timely, qualitative, in-person and practice based feedback to help them grow professionally. Second, currently used assessment tools create scores and comments that are difficult for residents to understand or access to help them follow their progress. Third, with the advent of the Clinical Competency Committee (CCC) mandated by the ACGME, residents feel there is a lack of transparency or control over their own assessments leading to mistrust of the evaluation system. Finally, the current culture in our residency and fellowship programs does not facilitate constant, clear and continuous feedback from faculty to trainees or from trainees to faculty. Residents feel this acutely and do not feel they have permission to give constructive feedback to faculty and often feel criticized when receiving anything less than superlative feedback from faculty. Current Pilot Program: With the help of Holly Nishimura BA, and Christy Boscardin PhD, fifty two of our internal medicine Prime residents are starting a two month evaluation project where they actively solicit EPA activity based feedback from faculty or other professional staff. These evaluations will be completely resident driven and will be collected in Qualtrics. Each participating resident was assigned a unique code number by Amy Forseth BA, our division administrator. She then emailed this individual code to each participating resident. The email code is unique to each participant and is easily forwarded to observing faculty and staff. The forwarded email automatically links the observing faculty member to the EPA survey menu where they can easily choose the EPA they observed and quickly complete the evaluation via email. The data are then captured for that individual resident. This “proof of concept pilot” enables us to examine the rates of completion of EPAs, willingness of residents to participate in the process and their satisfaction with the feedback they receive. Although this pilot is an exciting advance, it only fulfills our goals of facilitating timely resident feedback, it does not address our goals of creating an accessible resident dashboard where they are able to drive their own assessments and highlight their best work for the CCC. It also only takes a small incremental step towards moving our assessment and feedback culture forward to be more open and meaningful. Proposal Goals: The goal of our proposal is to address the four listed focus group themes by restructuring our assessment system over the next five years to be primarily trainee driven and to improve our information system so that all evaluations will be available to, and controlled by, the residents who receive them. Our vision is that every trainee entering the internal medicine residency or subspecialty fellowship programs would be given an assessment “dashboard” that would collect and store information that demonstrates their progress towards independent practice throughout their training. Data would be collected through a number of sources; Evalue, EPAs from Qualtrics, inservice exam scores, clinic patient dashboards and resident portfolio data and will be mapped to milestones to facilitate reporting to the ACGME. Trainees will be able to review all their data and will be able to “release” their best work to demonstrate their entrustability on the milestones. This trainee assessment dashboard will use the same Qlikview platform used by the medical school. Therefore our project research question is: Is restructure of competency assessment to a trainee-driven process associated with improved understanding of resident’s preparedness for independent practice amongst UCSF GME trainees? Process Proposal: Create a resident-driven, EPA based evaluation network that will expand following the previously published Delphi-derived EPAs for internal medicine and medicine subspecialty fellowships. The goals of the project will use the focus group themes to create a learner driven evaluation network with a goal of addressing each of the four themes. Theme #1: To create more timely opportunities for feedback the residents will be asked to solicit evaluations from their observing faculty. Trainees will send a uniquely coded email to the faculty member that requests feedback on the EPA activity they want assessed. Residents will be required to complete at least 3 EPAs per activity per year utilizing different faculty assessors, but they can choose which of the scores and assessments they release to the CCC for final review. Theme #2: Accessibility to their feedback will be ensured through creation of a resident assessment home page for each resident in the program. The homepage will contain sections that organize and display summative data from Evalue, EPA feedback evaluations, in service exam scores and a resident portfolio that allow the resident to track and record their clinical and academic progress by uploading documents that support mastery of individual milestones. Residents cannot edit this material but can add annotations if they so desire. This homepage will be based on work done in the SOM after meeting with Maxine Papadakis MD, Karen Hauer MD, PhD and Bonnie Helvig BA regarding strengths and challenges with their current dashboard program. Importantly, this system must have the flexibility to modify and grow as we fine tune assessments of trainee activities. Dashboard homepages will contain: Each trainee’s original ERAS application, Evalue and Qualtrics data for evaluations, a schedule function for viewing past/present schedules as well as vacations with an interactive timeline to enable residents to track administrative tasks, links to online modules that the residents need to access/complete (and to be able to record when they have completed them), self-assessment tools for faculty/trainee reviews, records for communication for feedback from our CCC, individual trainee quality of care dashboard data from both inpatient and clinic practice as well as instructional links and short videos instructing how decipher data as it comes in. Please see sample homepage provided below. Theme #3: Residents will control the data contained on their homepage. Residents can choose which EPA evaluations are available for the CCC enabling them to show off their best work. Similarly, CCC reports will become available to residents after each meeting so they can follow their progress on the ACGME milestones. This is based on the theoretical work presented by Patricia O’Sullivan PhD in her portfolio seminar regarding learning theory and adult learners. Theme #4: Our hope is that along with our fledgling faculty education efforts, constant collection of resident driven EPA based data will slowly help begin an overall culture change in assessment and feedback in our department. Residents will become more comfortable asking for activity specific feedback and faculty will become more comfortable giving it. This culture change will take time and effort but the resident focus group data was clear that they want timelier, activity-driven feedback so they should be motivated to ask for it. Sample Resident Landing Page: Specific Aims: 1- Create a resident and fellow assessment network that utilizes a resident driven, EPA based evaluations. 2- Create an individual resident homepage that will become a landing page that can link residents to all their assessment materials but also to UCSF based learning resources. 3- Ensure a more organized and efficient assessment data management system. 4- Change culture for assessment and feedback to become more transparent, ongoing and meaningful system for our trainees. Evaluation Plan: Process measures 1234- Percentage of responses to emails c/w Evalue time to completion of epa evaluations c/w evalue descriptive eval of epas completed amount of time of resident use of dashboard 5- actual vs reported numbers of epas completed Outcome measures 1- satisfaction (faculty/trainees) with system (focus groups) 2- comparison of entrustment on mapped epas c/w milestones 3- number of residents entrustable via epas c/w evalue milestones what about how it influences CCC Plan for Sustainability: After the initial start-up effort and costs to construct the Dashboard system this assessment system should be able to collect all the assessment data (Evalue, Qualtrics, Portfolio, In-service scores, Qualitative comments, clinic dashboards, patient follow-up reviews), create summary milestone reports for the CCC, and be updated annually for new residency and fellow trainees each year. The estimate from the Information Services Unit (ISU) group is that it will take 6 hours, $500 annually, for trouble shooting and updates to the system which will be made available from Departmental funds. Updates and administrative work will mostly involve construction of the new trainee dashboards which can be done by our program administrator, Amy Forseth. Team Members: Jeff Kohlwes MD, MPH. School of Medicine. Department of Internal Medicine. Clinical Professor of Medicine. Director, PRIME Residency Program, Associate Program Director for Assessment and Feedback Internal Medicine Residency Program. Patricia Cornett MD. School of Medicine. Department of Internal Medicine. Clinical Professor of Medicine. Associate Chair for Education, Department of Internal Medicine. Vanessa Thompson MD. School of Medicine. Department of Internal Medicine. Assistant Clinical Professor of Medicine. General Medicine Clinic Assistant Medical Director San Francisco General Hospital, Internal Medicine Residency Director of Academic Development Irina Kryzhanovskaya MD. School of Medicine. Department of Internal Medicine. Senior Resident in Internal Medicine, Chief Resident-Elect. Ray Chen MD. School of Medicine. Department of Internal Medicine. Junior Resident in Internal Medicine Total Budget 10. Consultations Completed: Nina Jameson, Deputy Director, Program Management, Information Services Unit (ISU)- Consultation on costs on IT solutions. Maxine Papadakis MD, Bonnie Helvig BA, UME Dashboard and IT expert for demonstration on UME dashboard and problem solving on creation of GME dashboard. Karen Hauer MD, PhD consultation on assessment of EPAs and milestones. Christy Boscardin PhD, Holly Nishimura BA, Department of Medical Education- Assessment expertise and Qualtrics planning for the current Pilot Study. References: 1- Green et al, J Grad Med Educ. 2009; 1:5-20. 2- http://www.acgme.org/acgmeweb/portals/0/pdfs/milestones/internalmedicinemilestones.pdf Accessed 11/13/15 3- Ten Cate et al. Academic Medicine, Vol. 82, No. 6 / June 2007 4- Farnan et al. A systematic review: the effect of clinical supervision on patient and residency education outcomes. Acad Med. 2012 Apr;87(4):428-42. 5- Foster et al. RAFT (Resident Assessment Facilitation Team): supporting resident well-being through an integrated advising and assessment process. Fam Med. 2012 Nov-Dec;44(10):731-4 6- Hauer K, et al. Identifying entrustable professional activities in internal medicine training. J Grad Med Educ. 2013 Mar;5(1):54-9.