1
|
Gin BC, Ten Cate O, O'Sullivan PS, Boscardin C. Assessing supervisor versus trainee viewpoints of entrustment through cognitive and affective lenses: an artificial intelligence investigation of bias in feedback. ADVANCES IN HEALTH SCIENCES EDUCATION : THEORY AND PRACTICE 2024; 29:1571-1592. [PMID: 38388855 PMCID: PMC11549112 DOI: 10.1007/s10459-024-10311-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Accepted: 01/21/2024] [Indexed: 02/24/2024]
Abstract
The entrustment framework redirects assessment from considering only trainees' competence to decision-making about their readiness to perform clinical tasks independently. Since trainees and supervisors both contribute to entrustment decisions, we examined the cognitive and affective factors that underly their negotiation of trust, and whether trainee demographic characteristics may bias them. Using a document analysis approach, we adapted large language models (LLMs) to examine feedback dialogs (N = 24,187, each with an associated entrustment rating) between medical student trainees and their clinical supervisors. We compared how trainees and supervisors differentially documented feedback dialogs about similar tasks by identifying qualitative themes and quantitatively assessing their correlation with entrustment ratings. Supervisors' themes predominantly reflected skills related to patient presentations, while trainees' themes were broader-including clinical performance and personal qualities. To examine affect, we trained an LLM to measure feedback sentiment. On average, trainees used more negative language (5.3% lower probability of positive sentiment, p < 0.05) compared to supervisors, while documenting higher entrustment ratings (+ 0.08 on a 1-4 scale, p < 0.05). We also found biases tied to demographic characteristics: trainees' documentation reflected more positive sentiment in the case of male trainees (+ 1.3%, p < 0.05) and of trainees underrepresented in medicine (UIM) (+ 1.3%, p < 0.05). Entrustment ratings did not appear to reflect these biases, neither when documented by trainee nor supervisor. As such, bias appeared to influence the emotive language trainees used to document entrustment more than the degree of entrustment they experienced. Mitigating these biases is nonetheless important because they may affect trainees' assimilation into their roles and formation of trusting relationships.
Collapse
Affiliation(s)
- Brian C Gin
- Department of Pediatrics, University of California San Francisco, 550 16th St Floor 4, UCSF Box 0110, San Francisco, CA, 94158, USA.
| | - Olle Ten Cate
- Utrecht Center for Research and Development of Health Professions Education, University Medical Center, Utrecht, the Netherlands
- Department of Medicine, University of California San Francisco, San Francisco, USA
| | - Patricia S O'Sullivan
- Department of Medicine, University of California San Francisco, San Francisco, USA
- Department of Surgery, University of California San Francisco, San Francisco, USA
| | - Christy Boscardin
- Department of Medicine, University of California San Francisco, San Francisco, USA
- Department of Anesthesia, University of California San Francisco, San Francisco, USA
| |
Collapse
|
2
|
Dufayet L, Piot MA, Geoffroy PA, Oulès B, Petitjean-Brichant C, Peiffer-Smadja N, Bouzid D, Tran Dinh A, Mirault T, Faye A, Lemogne C, Ruszniewski P, Peyre H, Vodovar D. CARECOS study: Medical students' empathy as assessed with the CARE measure by examiners versus standardized patients during a formative Objective and Structured Clinical Examination (OSCE) station. MEDICAL TEACHER 2024; 46:1187-1195. [PMID: 38285021 DOI: 10.1080/0142159x.2024.2306840] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Accepted: 01/15/2024] [Indexed: 01/30/2024]
Abstract
PURPOSE To assess the Consultation And Relational Empathy (CARE) measure as a tool for examiners to assess medical students' empathy during Objective and Structured Clinical Examinations (OSCEs), as the best tool for assessing empathy during OSCEs remains unknown. METHODS We first assessed the psychometric properties of the CARE measure, completed simultaneously by examiners and standardized patients (SP, either teachers - SPteacher - or civil society members - SPcivil society), for each student, at the end of an OSCE station. We then assessed the qualitative/quantitative agreement between examiners and SP. RESULTS We included 129 students, distributed in eight groups, four groups for each SP type. The CARE measure showed satisfactory psychometric properties in the context of the study but moderate, and even poor inter-rater reliability for some items. Considering paired observations, examiners scored lower than SPs (p < 0.001) regardless of the SP type. However, the difference in score was greater when the SP was a SPteacher rather than a SPcivil society (p < 0.01). CONCLUSION Despite acceptable psychometric properties, inter-rater reliability of the CARE measure between examiners and SP was unsatisfactory. The choice of examiner as well as the type of SP seems critical to ensure a fair measure of empathy during OSCEs.
Collapse
Affiliation(s)
- Laurene Dufayet
- UFR de médecine, Université Paris Cité, Paris, France
- Unité Médico-judiciaire, Hôtel-Dieu, AP-HP, Paris, France
- Centre Antipoison de Paris, Hôpital Fernand-Widal, AP-HP, Paris, France
- INSERM, UMRS-1144, Faculté de pharmacie, Paris, France
| | - Marie-Aude Piot
- UFR de médecine, Université Paris Cité, Paris, France
- Département de psychiatrie de l'enfant et de l'adolescent, Hôpital Necker, AP-HP, Paris, France
- INSERM, UMR 1018, Université Paris-Saclay, Villejuif cedex, France
| | - Pierre-Alexis Geoffroy
- UFR de médecine, Université Paris Cité, Paris, France
- Département de psychiatrie et d'addictologie, Hôpital Bichat-Claude Bernard, AP-HP, Paris, France
- Psychiatrie & Neurosciences, Hôpital Saint-Anne, GHU Paris, Paris, France
- Université de Paris, NeuroDiderot, Inserm, FHU I2-D2, Paris, France
| | - Bénédicte Oulès
- UFR de médecine, Université Paris Cité, Paris, France
- Service de dermatologie, Hôpital Saint-Louis, AP-HP, Paris, France
| | - Clara Petitjean-Brichant
- Département de psychiatrie et d'addictologie, Hôpital Bichat-Claude Bernard, AP-HP, Paris, France
| | - Nathan Peiffer-Smadja
- UFR de médecine, Université Paris Cité, Paris, France
- Service de maladies infectieuses et tropicales, Hôpital Bichat-Claude Bernard, AP-HP, Paris, France
- Université Paris Cité, INSERM UMR1137, IAME, Paris, France
| | - Donia Bouzid
- UFR de médecine, Université Paris Cité, Paris, France
- Université Paris Cité, INSERM UMR1137, IAME, Paris, France
- Service d'accueil des urgences, Hôpital Bichat-Claude Bernard, AP-HP, Paris, France
| | - Alexy Tran Dinh
- UFR de médecine, Université Paris Cité, Paris, France
- Département d'anesthésie-réanimation, Hôpital Bichat-Claude Bernard, AP-HP, Paris, France
| | - Tristan Mirault
- UFR de médecine, Université Paris Cité, Paris, France
- Service de médecine vasculaire, Hôpital Européen Georges Pompidou, Paris, France
| | - Albert Faye
- UFR de médecine, Université Paris Cité, Paris, France
- Service de Pédiatrie générale, Maladies infectieuses et Médecine interne, Hôpital Robert Debré, AP-HP, Paris, France
| | - Cédric Lemogne
- UFR de médecine, Université Paris Cité, Paris, France
- Service de Psychiatrie de l'adulte, AP-HP, Hôpital Hôtel-Dieu, Paris, France
- Center for Research in Epidemiology and StatisticS (CRESS), Université Paris Cité and Université Sorbonne Paris Nord, INSERM, INRAE, Paris, France
| | - Philippe Ruszniewski
- UFR de médecine, Université Paris Cité, Paris, France
- Service de gastro-entérologie et pancréatologie, Hôpital Beaujon AP-HP, Clichy, France
| | - Hugo Peyre
- UFR de médecine, Université Paris Cité, Paris, France
- Service de psychiatrie de l'enfant et de l'adolescent, Hôpital Robert Debré, APHP, Paris, France
- INSERM UMR 1141, Université Paris Cité, Paris, France
| | - Dominique Vodovar
- UFR de médecine, Université Paris Cité, Paris, France
- Centre Antipoison de Paris, Hôpital Fernand-Widal, AP-HP, Paris, France
- INSERM, UMRS-1144, Faculté de pharmacie, Paris, France
| |
Collapse
|
3
|
Pan TY, Piscioneri F, Owen C. Comparison Between Surgical Trainee Self-Assessment With Supervisor Scores in Technical and Nontechnical Skills to Identify Underperforming Trainees. JOURNAL OF SURGICAL EDUCATION 2023; 80:1311-1319. [PMID: 37407352 DOI: 10.1016/j.jsurg.2023.06.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/08/2023] [Revised: 05/11/2023] [Accepted: 06/07/2023] [Indexed: 07/07/2023]
Abstract
OBJECTIVE This study aims to compare trainee self-assessment with supervisor assessment to identify differences in correlations of the demographic data and evaluate whether the instrument can be utilized to identify underperforming trainees. DESIGN A novel instrument was designed based off the Royal Australasian College of Surgeons original 9 competencies utilizing the JDocs framework and covers 48-items across all surgical competencies. A multiple regression model using age, gender, postgraduate year, IMG status, and level of training as the variables was performed with backwards elimination, and pairwise comparisons made to identify the degree and direction of influence each variable contributed to trainee and supervisor ratings. SETTING Surgical trainees employed in tertiary centers within the Australian Capital Territory and South-East New South Wales health network in Australia. PARTICIPANTS A total of 36 of 59 (61%) trainees responded. Two grossly incomplete responses were excluded from further analysis resulting in 34 completed self-assessments for analysis. There was a total of 68 supervisor assessments provided by 25 different nominated supervisors. Of these assessments, 67 were fully completed and one was partially complete. RESULTS We identified that for both self-ratings and supervisor ratings, the most significant correlation is with the postgraduate year of the trainee, with correlation identified in 7 out of 9 competencies, although in different competency domains. International Medical Graduate status is associated in 2 of 9 self-ratings, and in 3 of 9 supervisor ratings. Underperforming trainees were able to be identified through supervisor assessment but not self-assessment. CONCLUSION The supervisor assessment form was able to identify underperforming trainees. Our findings resonate with existing literature in other specialty fields, and surgical units that employ assessment forms should feel more confident in the interpretation of the data and provision of feedback to trainees.
Collapse
Affiliation(s)
- Tzong-Yang Pan
- Canberra Hospital, Australian National University, Canberra, ACT, Australia.
| | - Frank Piscioneri
- Canberra Hospital, Australian National University, Canberra, ACT, Australia
| | - Cathy Owen
- Australian National University, Canberra, ACT, Australia
| |
Collapse
|
4
|
Waheed S, Maursetter L. Evaluation Evolution: Designing Optimal Evaluations to Enhance Learning in Nephrology Fellowship. Adv Chronic Kidney Dis 2022; 29:526-533. [PMID: 36371117 DOI: 10.1053/j.ackd.2022.06.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2022] [Revised: 04/26/2022] [Accepted: 06/28/2022] [Indexed: 11/10/2022]
Abstract
Evaluations serve as the backbone of any educational program and can be broadly divided into formative and summative evaluations. Formative evaluations are "just in time" evaluations focused on informing the learning process, whereas summative evaluations compare fellows to a preset standard to determine their readiness for unsupervised practice. In the nephrology fellowship programs, evaluations assess competence in the framework of ACGME Milestones 2.0. A variety of learning venues, evaluators, and tools should be incorporated into the measurement process. It is important to determine which milestones can be best assessed in each education venue to decrease the burden of assessment fatigue. Additionally, programs can diversify the evaluators to include nurses, medical students, peers, and program coordinators in addition to faculty to provide a well-rounded assessment of the fellows and share the assessment burden. Lastly, the evaluation data should be presented to fellows in a format where it can inform goal setting. The evaluation system needs to evolve along with the changes being made in curriculum design. This will help to make fellowship learning effective and efficient.
Collapse
Affiliation(s)
- Sana Waheed
- Piedmont Nephrology and Internal Medicine, Atlanta, GA
| | - Laura Maursetter
- Division of Nephrology, Department of Medicine, University of Wisconsin School of Medicine and Public Health, Madison, WI.
| |
Collapse
|
5
|
Woods R, Singh S, Thoma B, Patocka C, Cheung W, Monteiro S, Chan TM, for the QuAL Validation collaborators. Validity evidence for the Quality of Assessment for Learning score: a quality metric for supervisor comments in Competency Based Medical Education. CANADIAN MEDICAL EDUCATION JOURNAL 2022; 13:19-35. [PMID: 36440075 PMCID: PMC9684040 DOI: 10.36834/cmej.74860] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
BACKGROUND Competency based medical education (CBME) relies on supervisor narrative comments contained within entrustable professional activities (EPA) for programmatic assessment, but the quality of these supervisor comments is unassessed. There is validity evidence supporting the QuAL (Quality of Assessment for Learning) score for rating the usefulness of short narrative comments in direct observation. OBJECTIVE We sought to establish validity evidence for the QuAL score to rate the quality of supervisor narrative comments contained within an EPA by surveying the key end-users of EPA narrative comments: residents, academic advisors, and competence committee members. METHODS In 2020, the authors randomly selected 52 de-identified narrative comments from two emergency medicine EPA databases using purposeful sampling. Six collaborators (two residents, two academic advisors, and two competence committee members) were recruited from each of four EM Residency Programs (Saskatchewan, McMaster, Ottawa, and Calgary) to rate these comments with a utility score and the QuAL score. Correlation between utility and QuAL score were calculated using Pearson's correlation coefficient. Sources of variance and reliability were calculated using a generalizability study. RESULTS All collaborators (n = 24) completed the full study. The QuAL score had a high positive correlation with the utility score amongst the residents (r = 0.80) and academic advisors (r = 0.75) and a moderately high correlation amongst competence committee members (r = 0.68). The generalizability study found that the major source of variance was the comment indicating the tool performs well across raters. CONCLUSION The QuAL score may serve as an outcome measure for program evaluation of supervisors, and as a resource for faculty development.
Collapse
Affiliation(s)
- Rob Woods
- Department of Emergency Medicine, University of Saskatchewan, Saskatchewan, Canada
| | - Sim Singh
- College of Medicine, University of Saskatchewan, Saskatchewan, Canada
| | - Brent Thoma
- Department of Emergency Medicine, University of Saskatchewan, Saskatchewan, Canada
| | - Catherine Patocka
- Department of Emergency Medicine, University of Calgary, Alberta, Canada
| | - Warren Cheung
- Department of Emergency Medicine, University of Ottawa, Ontario, Canada
| | - Sandra Monteiro
- Department of Health Research Methods Evidence and Impact, McMaster University, Ontario, Canada
| | - Teresa M Chan
- Division of Emergency Medicine and Education & Innovation, Department of Medicine, McMaster University, Ontario, Canada
| | | |
Collapse
|
6
|
Concordance of Narrative Comments with Supervision Ratings Provided During Entrustable Professional Activity Assessments. J Gen Intern Med 2022; 37:2200-2207. [PMID: 35710663 PMCID: PMC9296736 DOI: 10.1007/s11606-022-07509-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/08/2021] [Accepted: 03/24/2022] [Indexed: 10/18/2022]
Abstract
BACKGROUND Use of EPA-based entrustment-supervision ratings to determine a learner's readiness to assume patient care responsibilities is expanding. OBJECTIVE In this study, we investigate the correlation between narrative comments and supervision ratings assigned during ad hoc assessments of medical students' performance of EPA tasks. DESIGN Data from assessments completed for students enrolled in the clerkship phase over 2 academic years were used to extract a stratified random sample of 100 narrative comments for review by an expert panel. PARTICIPANTS A review panel, comprised of faculty with specific expertise related to their roles within the EPA program, provided a "gold standard" supervision rating using the comments provided by the original assessor. MAIN MEASURES Interrater reliability (IRR) between members of review panel and correlation coefficients (CC) between expert ratings and supervision ratings from original assessors. KEY RESULTS IRR among members of the expert panel ranged from .536 for comments associated with focused history taking to .833 for complete physical exam. CC (Kendall's correlation coefficient W) between panel members' assignment of supervision ratings and the ratings provided by the original assessors for history taking, physical examination, and oral presentation comments were .668, .697, and .735 respectively. The supervision ratings of the expert panel had the highest degree of correlation with ratings provided during assessments done by master assessors, faculty trained to assess students across clinical contexts. Correlation between supervision ratings provided with the narrative comments at the time of observation and supervision ratings assigned by the expert panel differed by clinical discipline, perhaps reflecting the value placed on, and perhaps the comfort level with, assessment of the task in a given specialty. CONCLUSIONS To realize the full educational and catalytic effect of EPA assessments, assessors must apply established performance expectations and provide high-quality narrative comments aligned with the criteria.
Collapse
|
7
|
Swanberg M, Woodson-Smith S, Pangaro L, Torre D, Maggio L. Factors and Interactions Influencing Direct Observation: A Literature Review Guided by Activity Theory. TEACHING AND LEARNING IN MEDICINE 2022; 34:155-166. [PMID: 34238091 DOI: 10.1080/10401334.2021.1931871] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/29/2020] [Revised: 04/19/2021] [Accepted: 05/11/2021] [Indexed: 06/13/2023]
Abstract
PhenomenonEnsuring that future physicians are competent to practice medicine is necessary for high quality patient care and safety. The shift toward competency-based education has placed renewed emphasis on direct observation via workplace-based assessments in authentic patient care contexts. Despite this interest and multiple studies focused on improving direct observation, challenges regarding the objectivity of this assessment approach remain underexplored and unresolved. Approach: We conducted a literature review of direct observation in authentic patient contexts by systematically searching databases PubMed, Embase, Web of Science, and ERIC. Included studies comprised original research conducted in the patient care context with authentic patients, either as a live encounter or a video recording of an actual encounter, which focused on factors affecting the direct observation of undergraduate medical education (UME) or graduate medical education (GME) trainees. Because the patient care context adds factors that contribute to the cognitive load of the learner and of the clinician-observer we focused our question on such contexts, which are most useful in judgments about advancement to the next level of training or practice. We excluded articles or published abstracts not conducted in the patient care context (e.g., OSCEs) or those involving simulation, allied health professionals, or non-UME/GME trainees. We also excluded studies focused on end-of-rotation evaluations and in-training evaluation reports. We extracted key data from the studies and used Activity Theory as a lens to identify factors affecting these observations and the interactions between them. Activity Theory provides a framework to understand and analyze complex human activities, the systems in which people work, and the interactions or tensions between multiple associated factors. Findings: Nineteen articles were included in the analysis; 13 involved GME learners and 6 UME learners. Of the 19, six studies were set in the operating room and four in the Emergency department. Using Activity Theory, we discovered that while numerous studies focus on rater and tool influences, very few study the impact of social elements. These are the rules that govern how the activity happens, the environment and members of the community involved in the activity and how completion of the activity is divided up among the members of the community. Insights: Viewing direct observation via workplace-based assessment through the lens of Activity Theory may enable educators to implement curricular changes to improve direct observation of assessment. Activity Theory may allow researchers to design studies to focus on the identified underexplored interactions and influences in relation to direct observation.
Collapse
Affiliation(s)
- Margaret Swanberg
- Department of Neurology, Uniformed Services University, Bethesda, Maryland, USA
| | - Sarah Woodson-Smith
- Department of Neurology, Naval Medical Center Portsmouth, Portsmouth, Virginia, USA
| | - Louis Pangaro
- Department of Medicine, Uniformed Services University, Bethesda, Maryland, USA
| | - Dario Torre
- Department of Medicine, Uniformed Services University, Bethesda, Maryland, USA
- Center for Health Professions Education, Uniformed Services University, Bethesda, Maryland, USA
| | - Lauren Maggio
- Department of Medicine, Uniformed Services University, Bethesda, Maryland, USA
- Center for Health Professions Education, Uniformed Services University, Bethesda, Maryland, USA
| |
Collapse
|
8
|
Forte M, Morson N, Mirchandani N, Grundland B, Fernando O, Rubenstein W. How Teachers Adapt Their Cognitive Strategies When Using Entrustment Scales. ACADEMIC MEDICINE : JOURNAL OF THE ASSOCIATION OF AMERICAN MEDICAL COLLEGES 2021; 96:S87-S92. [PMID: 34348370 DOI: 10.1097/acm.0000000000004287] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
PURPOSE Rater cognition is a field of study focused on individual cognitive processes used by medical teachers when completing assessments. Much has been written about the role of rater cognition in the use of traditional assessment scales. Entrustment scales (ES) are becoming the tool of choice for workplace-based assessments. It is not known how moving to an entrustment framework may cause teachers to adapt their cognitive rating strategies. This study aimed to explore this gap by asking teachers to describe their thinking when making rating decisions using a validated ES. METHOD Using purposive sampling, family medicine teachers supervising obstetrical care were invited to participate in cognitive interviews. Teachers were interviewed between December 2018 and March 2019 using retrospective verbal protocol analysis. They were asked to describe their experiences of rating residents in the last 6 months using new ES. Constructivist grounded theory guided data collection and analysis. Interviews were recorded, transcribed, and analyzed iteratively. A constant comparative approach was used to code and analyze the data until consensus was reached regarding emerging themes. RESULTS There was variability in how teachers used the ES. Faculty describe several ways in which they ultimately navigated the tool to say what they wanted to say. Four key themes emerged: (1) teachers interpreted the anchors differently based on their cognitive framework, (2) teachers differed in how they were able to cognitively shift away from traditional rating scales, (3) teachers struggled to limit assessments to a report on observed behavior, and (4) teachers contextualized their ratings. CONCLUSIONS Variability in teachers' interpretation of learner performance persists in entrustment frameworks. Rater's individual cognitive strategies and how they observe, process, and integrate their thoughts into assessments form part of a rich picture of learner progress. These insights can be harnessed to contribute to decisions regarding readiness for unsupervised practice.
Collapse
Affiliation(s)
- Milena Forte
- M. Forte is assistant professor, Postgraduate lead, Office of Education Scholarship, Maternity Care lead, Department of Family and Community Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada; ORCID: https://orcid.org/0000-0001-9387-0184
| | - Natalie Morson
- N. Morson is lecturer, Department of Family and Community Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada
| | - Natasha Mirchandani
- N. Mirchandani is assistant professor, Department of Family and Community Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada
| | - Batya Grundland
- B. Grundland is assistant professor and associate postgraduate program director, Curriculum & Remediation, Department of Family and Community Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada
| | - Oshan Fernando
- O. Fernando is research associate, Office of Education Scholarship, Department of Family and Community Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada; ORCID: https://orcid.org/0000-0002-4600-9399
| | - Warren Rubenstein
- W. Rubenstein is associate professor, Department of Family and Community Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
9
|
Robinson TJG, Wagner N, Szulewski A, Dudek N, Cheung WJ, Hall AK. Exploring the use of rating scales with entrustment anchors in workplace-based assessment. MEDICAL EDUCATION 2021; 55:1047-1055. [PMID: 34060651 DOI: 10.1111/medu.14573] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/01/2020] [Revised: 04/07/2021] [Accepted: 05/26/2021] [Indexed: 06/12/2023]
Abstract
PURPOSE Competency-based medical education (CBME) has prompted widespread implementation of workplace-based assessment (WBA) tools using entrustment anchors. This study aimed to identify factors that influence faculty's rating choices immediately following assessment and explore their experiences using WBAs with entrustment anchors, specifically the Ottawa Surgical Competency Operating Room Evaluation scale. METHOD A convenience sample of 50 semi-structured interviews with Emergency Medicine (EM) physicians from a single Canadian hospital were conducted between July and August 2019. All interviews occurred within two hours of faculty completing a WBA of a trainee. Faculty were asked what they considered when rating the trainee's performance and whether they considered an alternate rating. Two team members independently analysed interview transcripts using conventional content analysis with line-by-line coding to identify themes. RESULTS Interviews captured interactions between 70% (26/37) of full-time EM faculty and 86% (19/22) of EM trainees. Faculty most commonly identified the amount of guidance the trainee required as influencing their rating. Other variables such as clinical context, trainee experience, past experiences with the trainee, perceived competence and confidence were also identified. While most faculty did not struggle to assign ratings, some had difficulty interpreting the language of entrustment anchors, being unsure whether their assessment should be retrospective or prospective in nature, and if/how the assessment should change whether they were 'in the room' or not. CONCLUSIONS By going to the frontline during WBA encounters, this study captured authentic and honest reflections from physicians immediately engaged in assessment using entrustment anchors. While many of the factors identified are consistent with previous retrospective work, we highlight how some faculty consider factors outside the prescribed approach and struggle with the language of entrustment anchors. These results further our understanding of 'in-the-moment' assessments using entrustment anchors and may facilitate effective faculty development regarding WBA in CBME.
Collapse
Affiliation(s)
| | - Natalie Wagner
- Department of Biomedical & Molecular Sciences, Queen's University, Kingston, ON, Canada
- Office of Professional Development & Educational Scholarship, Queen's University, Kingston, ON, Canada
| | - Adam Szulewski
- Department of Emergency Medicine, Queen's University, Kingston, ON, Canada
- Department of Psychology, Queen's University, Kingston, ON, Canada
| | - Nancy Dudek
- Department of Medicine and The Ottawa Hospital, University of Ottawa, Ottawa, ON, Canada
- Royal College of Physicians and Surgeons of Canada, Ottawa, ON, Canada
| | - Warren J Cheung
- Royal College of Physicians and Surgeons of Canada, Ottawa, ON, Canada
- Department of Emergency Medicine, University of Ottawa, Ottawa, ON, Canada
| | - Andrew K Hall
- Department of Emergency Medicine, Queen's University, Kingston, ON, Canada
- Royal College of Physicians and Surgeons of Canada, Ottawa, ON, Canada
| |
Collapse
|
10
|
Lucey CR, Hauer KE, Boatright D, Fernandez A. Medical Education's Wicked Problem: Achieving Equity in Assessment for Medical Learners. ACADEMIC MEDICINE : JOURNAL OF THE ASSOCIATION OF AMERICAN MEDICAL COLLEGES 2020; 95:S98-S108. [PMID: 32889943 DOI: 10.1097/acm.0000000000003717] [Citation(s) in RCA: 76] [Impact Index Per Article: 15.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Despite a lack of intent to discriminate, physicians educated in U.S. medical schools and residency programs often take actions that systematically disadvantage minority patients. The approach to assessment of learner performance in medical education can similarly disadvantage minority learners. The adoption of holistic admissions strategies to increase the diversity of medical training programs has not been accompanied by increases in diversity in honor societies, selective residency programs, medical specialties, and medical school faculty. These observations prompt justified concerns about structural and interpersonal bias in assessment. This manuscript characterizes equity in assessment as a "wicked problem" with inherent conflicts, uncertainty, dynamic tensions, and susceptibility to contextual influences. The authors review the underlying individual and structural causes of inequity in assessment. Using an organizational model, they propose strategies to achieve equity in assessment and drive institutional and systemic improvement based on clearly articulated principles. This model addresses the culture, systems, and assessment tools necessary to achieve equitable results that reflect stated principles. Three components of equity in assessment that can be measured and evaluated to confirm success include intrinsic equity (selection and design of assessment tools), contextual equity (the learning environment in which assessment occurs), and instrumental equity (uses of assessment data for learner advancement and selection and program evaluation). A research agenda to address these challenges and controversies and demonstrate reduction in bias and discrimination in medical education is presented.
Collapse
Affiliation(s)
- Catherine R Lucey
- C.R. Lucey is executive vice dean/vice dean for education and professor of medicine, University of California, San Francisco, School of Medicine, San Francisco, California
| | - Karen E Hauer
- K.E. Hauer is professor of medicine, University of California, San Francisco, School of Medicine, San Francisco, California
| | - Dowin Boatright
- D. Boatright is assistant professor of emergency medicine, Yale University School of Medicine, New Haven, Connecticut
| | - Alicia Fernandez
- A. Fernandez is professor of medicine, University of California, San Francisco, School of Medicine, San Francisco, California
| |
Collapse
|
11
|
Dudek N. Faculty and Resident Perspectives on Using Entrustment Anchors for Workplace-Based Assessment. J Grad Med Educ 2019; 11:287-294. [PMID: 31210859 PMCID: PMC6570427 DOI: 10.4300/jgme-d-18-01003.1] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/29/2018] [Revised: 03/14/2019] [Accepted: 04/09/2019] [Indexed: 01/01/2023] Open
Abstract
BACKGROUND Research suggests that workplace-based assessment (WBA) tools using entrustment anchors provide more reliable assessments than those using traditional anchors. There is a lack of evidence describing how and why entrustment anchors work. OBJECTIVE The purpose of this study is to better understand the experience of residents and faculty with respect to traditional and entrustment anchors. METHODS We used constructivist grounded theory to guide data collection and analysis (March-December 2017) and semistructured interviews to gather reflections on anchors. Phase 1 involved residents and faculty (n = 12) who had only used assessment tools with traditional anchors. Phase 2 involved participants who had used tools with entrustment anchors (n = 10). Data were analyzed iteratively. RESULTS Participants expressed that the pragmatic language of entrustment anchors made WBA (1) concrete and justifiable; (2) transparent as they explicitly link clinical assessment and learning progress; and (3) align with training outcomes, enabling better feedback. Participants with no prior experience using entrustment anchors outlined contextual concerns regarding their use. Participants with experience described how they addressed these concerns. Participants expressed that entrustment anchors leave a gap in assessment information because they do not provide normative data. CONCLUSIONS Insights from this analysis contribute to a theoretical framework of benefits and challenges related to the adoption of entrustment anchors. This richer understanding of faculty and resident perspectives of entrustment anchors may assist WBA developers in creating more acceptable tools and inform the necessary faculty development initiatives that must accompany the use of these new WBA tools. .
Collapse
|
12
|
Blake AK. From intelligence analysis to medical education: using structured tools to manage bias. MEDICAL EDUCATION 2018; 52:244-245. [PMID: 29441637 DOI: 10.1111/medu.13490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/03/2017] [Accepted: 10/12/2017] [Indexed: 06/08/2023]
|