Georgina C. Stephens, Gabrielle Brand, Sharon Yahalom
{"title":"Whose voices are heard in health professions education validity arguments?","authors":"Georgina C. Stephens, Gabrielle Brand, Sharon Yahalom","doi":"10.1111/medu.15528","DOIUrl":null,"url":null,"abstract":"<p>The construction of validity arguments for assessments in health professions education (HPE) has been likened to a lawyer preparing for and presenting a case.<span><sup>1</sup></span> Like a lawyer curates a brief of evidence with the aim of convincing a judge or jury to make a particular decision about their client, so are health professions educators required to provide validity evidence that supports a defensible decision about a student being assessed.<span><sup>1</sup></span> Kane's argument-based validity framework,<span><sup>2</sup></span> now expanded by scholars in language testing and assessment (LTA), addresses challenges of prior conceptualisations of validity by providing a staged approach to building a validation argument according to ‘inferences’. Whereas Kane's original four inference model commences with scoring,<span><sup>2</sup></span> the expanded seven inference model as described in this issue by Dai et al.<span><sup>3</sup></span> in LTA commences with a domain description.</p><p>The goal of the domain description is to ensure that the ‘selection, design and delivery of the test tasks takes the relevant target domain into account’.<span><sup>3</sup></span> Described sources of backing for this inference include interviews or surveys of domain insiders. Starting with a domain description should provide a solid foundation for subsequent inferences made about the assessment but also begs the question of who are considered domain insiders. And whether insights from diverse groups with insider perspectives can together build a more robust and nuanced validity argument. Returning to the analogy of the lawyer's decision-making processes, there may be multiple witnesses with evidence to share, but which witnesses are called upon to provide evidence in court? Or alternatively, which witnesses are not selected out of concern that their differing perspectives may threaten the lawyer's plan for the case?</p><p>Domain insiders could be considered ‘expert witnesses’, that is, those with subject matter expertise typically built through education and professional experience, such as health professions educators with clinical and/or pedagogical expertise. While subject matter expertise is important to understanding whether assessment tasks sufficiently reflect the domain being assessed, potential differences between expert and novice (i.e., student) understandings of a domain could disrupt a validity argument. Consider assessments of uncertainty tolerance (UT): Commonly used UT scales intended to measure UT in healthcare contexts engaged expertise during scale development in the form of interviews with health professionals, reviews of construct literature and consultation with medical educator peers.<span><sup>4</sup></span> One UT scale has been used by the Association of American Medical Colleges as part of routine matriculation and graduation surveys of medical students, with the intent that the results inform medical school programmatic evaluation.<span><sup>4</sup></span> However, the perspectives of medical students were not included as part of the validity argument for this scale.<span><sup>4</sup></span></p><p>Research on the UT construct more broadly highlights potential differences between expert and student conceptions of uncertainty, with students' conceptions of uncertainty focussed on individual knowledge gaps rather than uncertainty inherent within patient care more typical of clinical experts.<span><sup>5</sup></span> By not including medical student perspectives as part of the validity argument, the meaning of scale results could potentially be misinterpreted. For example, an increase in UT scores from matriculation to graduation could represent that knowledge gaps have been filled and perceived uncertainty reduced, and not that students are better at ‘tolerating’ or managing uncertainty, an essential graduate skill needed to manage the dynamic, complex and unpredictable nature of healthcare. Hence, including test takers as domain insiders or ‘witnesses’ early in constructing a validity argument may enable those developing an assessment to proactively identify early threats to their argument before evaluating subsequent inferences in the framework.</p><p>An example of an assessment where diverse ‘witnesses’ enabled inferences to be made about the domain description, and where the field of LTA and HPE intersect, is the Occupational English Test (OET).<span><sup>6</sup></span> The OET is an English-language test designed for health professionals originally developed by LTA experts and includes writing a letter of referral to assess written clinical communication.<span><sup>7</sup></span> While updating and modifying the assessment criteria for the written component of the test, applied linguists, LTA experts and HPE researchers collaborated with health professionals (domain insiders) through an interview and focus group study with clinicians and other stakeholders to identify criteria indigenous to healthcare settings.<span><sup>6, 7</sup></span></p><p>The differing expertise of members of the research team enabled them to make recommendations about modifications to the assessment criteria based on the data.<span><sup>6</sup></span> For instance, the health professionals were vital for ensuring the establishment of professionally relevant criteria, whereas the LTA experts were crucial for defining levels of performance.<span><sup>6</sup></span> When analysing data, coding schemes were initially developed by the LTA experts and then refined through collaboration with the health professionals. The researchers in this study were transparent about how collaboration posed some challenges; however, they concluded that it ultimately enabled a greater depth and understanding of the complexity of the domain being assessed, and awareness of this complexity would not have been achieved without collaboration.<span><sup>6</sup></span> Such insights identified through the process of domain description for the OET could in turn be used to enhance HPE (e.g., workshops on clinical note writing for final-year health professions students).</p><p>Collaborative approaches to understanding the perspectives of domain insiders may lead to richer and more robust validity arguments and assessments when compared with those centred on ‘expert witnesses’. Just as neglecting to call a key witness to the stand would be considered an unacceptable omission by a lawyer, the field of HPE should more broadly consider whose voices are valued and included in validity arguments so that important insights into domains are not overlooked. This can be achieved by inviting testimony from diverse stakeholders, including health professions students and patients (healthcare consumers), on what is important to and valued by them so that HPE aligns with community needs, values and expectations of healthcare. The cross-cutting edge paper by Dai et al.<span><sup>3</sup></span> provides a timely reminder of how interdisciplinary dialogue (in this case, between HPE and LTA) can expand the epistemological landscape in HPE and research to generate new ways of thinking about validation practice in HPE.<span><sup>8</sup></span> HPE in general is moving towards transformational approaches that integrate student and patient perspectives, as exemplified by the growing use of co-design methodology.<span><sup>9, 10</sup></span> Co-design is underpinned by principles of inclusive, respectful, participative, iterative and outcomes-focussed consumer engagement.<span><sup>9</sup></span> Although the use of co-design methodology in assessment is currently limited,<span><sup>10</sup></span> co-design principles may be well suited to navigating how diverse assessment stakeholder perspectives can be explored and successfully integrated within a domain description and ultimately inform the development of curricula and assessments.</p><p>Returning to the opening analogy likening validity arguments to a lawyer constructing and presenting a case, it may be helpful for health professions researchers and educators to think more broadly about who is (or is not) given the opportunity to contribute to building validity evidence. For HPE, this may involve including ‘witnesses’ with diverse lived and learned expertise (e.g., students, patients and practicing clinicians) in collaboration with researchers with diverse backgrounds (e.g., interdisciplinary collaboration between HPE and LTA researchers). By broadly considering who is a domain insider, health professions educators and researchers may facilitate a more nuanced and holistic domain description, more robust arguments for the interpretation of test scores and ultimately the creation of richer and more relevant assessments with clear benefits for health professions educators, researchers, students and the wider community.</p><p><b>Georgina C. Stephens:</b> Conceptualization; writing—original draft. <b>Gabrielle Brand:</b> Conceptualization; writing—original draft. <b>Sharon Yahalom:</b> Conceptualization; writing—original draft.</p>","PeriodicalId":18370,"journal":{"name":"Medical Education","volume":"58 12","pages":"1429-1432"},"PeriodicalIF":4.9000,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/medu.15528","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical Education","FirstCategoryId":"95","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/medu.15528","RegionNum":1,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION, SCIENTIFIC DISCIPLINES","Score":null,"Total":0}
引用次数: 0
Abstract
The construction of validity arguments for assessments in health professions education (HPE) has been likened to a lawyer preparing for and presenting a case.1 Like a lawyer curates a brief of evidence with the aim of convincing a judge or jury to make a particular decision about their client, so are health professions educators required to provide validity evidence that supports a defensible decision about a student being assessed.1 Kane's argument-based validity framework,2 now expanded by scholars in language testing and assessment (LTA), addresses challenges of prior conceptualisations of validity by providing a staged approach to building a validation argument according to ‘inferences’. Whereas Kane's original four inference model commences with scoring,2 the expanded seven inference model as described in this issue by Dai et al.3 in LTA commences with a domain description.
The goal of the domain description is to ensure that the ‘selection, design and delivery of the test tasks takes the relevant target domain into account’.3 Described sources of backing for this inference include interviews or surveys of domain insiders. Starting with a domain description should provide a solid foundation for subsequent inferences made about the assessment but also begs the question of who are considered domain insiders. And whether insights from diverse groups with insider perspectives can together build a more robust and nuanced validity argument. Returning to the analogy of the lawyer's decision-making processes, there may be multiple witnesses with evidence to share, but which witnesses are called upon to provide evidence in court? Or alternatively, which witnesses are not selected out of concern that their differing perspectives may threaten the lawyer's plan for the case?
Domain insiders could be considered ‘expert witnesses’, that is, those with subject matter expertise typically built through education and professional experience, such as health professions educators with clinical and/or pedagogical expertise. While subject matter expertise is important to understanding whether assessment tasks sufficiently reflect the domain being assessed, potential differences between expert and novice (i.e., student) understandings of a domain could disrupt a validity argument. Consider assessments of uncertainty tolerance (UT): Commonly used UT scales intended to measure UT in healthcare contexts engaged expertise during scale development in the form of interviews with health professionals, reviews of construct literature and consultation with medical educator peers.4 One UT scale has been used by the Association of American Medical Colleges as part of routine matriculation and graduation surveys of medical students, with the intent that the results inform medical school programmatic evaluation.4 However, the perspectives of medical students were not included as part of the validity argument for this scale.4
Research on the UT construct more broadly highlights potential differences between expert and student conceptions of uncertainty, with students' conceptions of uncertainty focussed on individual knowledge gaps rather than uncertainty inherent within patient care more typical of clinical experts.5 By not including medical student perspectives as part of the validity argument, the meaning of scale results could potentially be misinterpreted. For example, an increase in UT scores from matriculation to graduation could represent that knowledge gaps have been filled and perceived uncertainty reduced, and not that students are better at ‘tolerating’ or managing uncertainty, an essential graduate skill needed to manage the dynamic, complex and unpredictable nature of healthcare. Hence, including test takers as domain insiders or ‘witnesses’ early in constructing a validity argument may enable those developing an assessment to proactively identify early threats to their argument before evaluating subsequent inferences in the framework.
An example of an assessment where diverse ‘witnesses’ enabled inferences to be made about the domain description, and where the field of LTA and HPE intersect, is the Occupational English Test (OET).6 The OET is an English-language test designed for health professionals originally developed by LTA experts and includes writing a letter of referral to assess written clinical communication.7 While updating and modifying the assessment criteria for the written component of the test, applied linguists, LTA experts and HPE researchers collaborated with health professionals (domain insiders) through an interview and focus group study with clinicians and other stakeholders to identify criteria indigenous to healthcare settings.6, 7
The differing expertise of members of the research team enabled them to make recommendations about modifications to the assessment criteria based on the data.6 For instance, the health professionals were vital for ensuring the establishment of professionally relevant criteria, whereas the LTA experts were crucial for defining levels of performance.6 When analysing data, coding schemes were initially developed by the LTA experts and then refined through collaboration with the health professionals. The researchers in this study were transparent about how collaboration posed some challenges; however, they concluded that it ultimately enabled a greater depth and understanding of the complexity of the domain being assessed, and awareness of this complexity would not have been achieved without collaboration.6 Such insights identified through the process of domain description for the OET could in turn be used to enhance HPE (e.g., workshops on clinical note writing for final-year health professions students).
Collaborative approaches to understanding the perspectives of domain insiders may lead to richer and more robust validity arguments and assessments when compared with those centred on ‘expert witnesses’. Just as neglecting to call a key witness to the stand would be considered an unacceptable omission by a lawyer, the field of HPE should more broadly consider whose voices are valued and included in validity arguments so that important insights into domains are not overlooked. This can be achieved by inviting testimony from diverse stakeholders, including health professions students and patients (healthcare consumers), on what is important to and valued by them so that HPE aligns with community needs, values and expectations of healthcare. The cross-cutting edge paper by Dai et al.3 provides a timely reminder of how interdisciplinary dialogue (in this case, between HPE and LTA) can expand the epistemological landscape in HPE and research to generate new ways of thinking about validation practice in HPE.8 HPE in general is moving towards transformational approaches that integrate student and patient perspectives, as exemplified by the growing use of co-design methodology.9, 10 Co-design is underpinned by principles of inclusive, respectful, participative, iterative and outcomes-focussed consumer engagement.9 Although the use of co-design methodology in assessment is currently limited,10 co-design principles may be well suited to navigating how diverse assessment stakeholder perspectives can be explored and successfully integrated within a domain description and ultimately inform the development of curricula and assessments.
Returning to the opening analogy likening validity arguments to a lawyer constructing and presenting a case, it may be helpful for health professions researchers and educators to think more broadly about who is (or is not) given the opportunity to contribute to building validity evidence. For HPE, this may involve including ‘witnesses’ with diverse lived and learned expertise (e.g., students, patients and practicing clinicians) in collaboration with researchers with diverse backgrounds (e.g., interdisciplinary collaboration between HPE and LTA researchers). By broadly considering who is a domain insider, health professions educators and researchers may facilitate a more nuanced and holistic domain description, more robust arguments for the interpretation of test scores and ultimately the creation of richer and more relevant assessments with clear benefits for health professions educators, researchers, students and the wider community.
期刊介绍:
Medical Education seeks to be the pre-eminent journal in the field of education for health care professionals, and publishes material of the highest quality, reflecting world wide or provocative issues and perspectives.
The journal welcomes high quality papers on all aspects of health professional education including;
-undergraduate education
-postgraduate training
-continuing professional development
-interprofessional education