首页 > 最新文献

GMS Zeitschrift fur Medizinische Ausbildung最新文献

英文 中文
Electronic acquisition of OSCE performance using tablets. 使用平板电脑电子获取欧安组织性能。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000983
Achim Hochlehnert, Jobst-Hendrik Schultz, Andreas Möltner, Sevgi Tımbıl, Konstantin Brass, Jana Jünger

Background: Objective Structured Clinical Examinations (OSCEs) often involve a considerable amount of resources in terms of materials and organization since the scores are often recorded on paper. Computer-assisted administration is an alternative with which the need for material resources can be reduced. In particular, the use of tablets seems sensible because these are easy to transport and flexible to use.

Aim: User acceptance concerning the use of tablets during OSCEs has not yet been extensively investigated. The aim of this study was to evaluate tablet-based OSCEs from the perspective of the user (examiner) and the student examinee.

Method: For two OSCEs in Internal Medicine at the University of Heidelberg, user acceptance was analyzed regarding tablet-based administration (satisfaction with functionality) and the subjective amount of effort as perceived by the examiners. Standardized questionnaires and semi-standardized interviews were conducted (complete survey of all participating examiners). In addition, for one OSCE, the subjective evaluation of this mode of assessment was gathered from a random sample of participating students in semi-standardized interviews.

Results: Overall, the examiners were very satisfied with using tablets during the assessment. The subjective amount of effort to use the tablet was found on average to be "hardly difficult". The examiners identified the advantages of this mode of administration as being in particular the ease of use and low rate of error. During the interviews of the examinees, acceptance for the use of tablets during the assessment was also detected.

Discussion: Overall, it was found that the use of tablets during OSCEs was well accepted by both examiners and examinees. We expect that this mode of assessment also offers advantages regarding assessment documentation, use of resources, and rate of error in comparison with paper-based assessments; all of these aspects should be followed up on in further studies.

背景:目的结构化临床检查(OSCEs)通常在材料和组织方面涉及相当多的资源,因为分数通常记录在纸上。计算机辅助管理是一种替代方法,可以减少对物质资源的需求。特别是,使用平板电脑似乎是明智的,因为它们易于运输和灵活使用。目的:用户对欧安会议期间使用片剂的接受程度尚未广泛调查。本研究的目的是从使用者(考官)和学生考生的角度来评价基于平板电脑的osce。方法:对于海德堡大学内科的两名osce,分析了用户对基于片剂的给药(功能满意度)和审查员感知的主观工作量的接受程度。采用标准化问卷和半标准化访谈(对所有参与审查员进行完整调查)。此外,对于一个欧安组织,对这种评估模式的主观评价是从参加半标准化访谈的学生随机抽样中收集的。结果:总体而言,审查员对片剂的使用非常满意。使用平板电脑的主观工作量平均为“几乎很难”。审查员确定了这种管理模式的优点,特别是易于使用和错误率低。在与考生的面谈中,也发现在评估期间接受使用片剂。讨论:总的来说,我们发现在oses期间片剂的使用被考官和考生都很好地接受。与纸质评估相比,我们期望这种评估模式在评估文件、资源使用和错误率方面也具有优势;所有这些方面都应该在进一步的研究中跟进。
{"title":"Electronic acquisition of OSCE performance using tablets.","authors":"Achim Hochlehnert,&nbsp;Jobst-Hendrik Schultz,&nbsp;Andreas Möltner,&nbsp;Sevgi Tımbıl,&nbsp;Konstantin Brass,&nbsp;Jana Jünger","doi":"10.3205/zma000983","DOIUrl":"https://doi.org/10.3205/zma000983","url":null,"abstract":"<p><strong>Background: </strong>Objective Structured Clinical Examinations (OSCEs) often involve a considerable amount of resources in terms of materials and organization since the scores are often recorded on paper. Computer-assisted administration is an alternative with which the need for material resources can be reduced. In particular, the use of tablets seems sensible because these are easy to transport and flexible to use.</p><p><strong>Aim: </strong>User acceptance concerning the use of tablets during OSCEs has not yet been extensively investigated. The aim of this study was to evaluate tablet-based OSCEs from the perspective of the user (examiner) and the student examinee.</p><p><strong>Method: </strong>For two OSCEs in Internal Medicine at the University of Heidelberg, user acceptance was analyzed regarding tablet-based administration (satisfaction with functionality) and the subjective amount of effort as perceived by the examiners. Standardized questionnaires and semi-standardized interviews were conducted (complete survey of all participating examiners). In addition, for one OSCE, the subjective evaluation of this mode of assessment was gathered from a random sample of participating students in semi-standardized interviews.</p><p><strong>Results: </strong>Overall, the examiners were very satisfied with using tablets during the assessment. The subjective amount of effort to use the tablet was found on average to be \"hardly difficult\". The examiners identified the advantages of this mode of administration as being in particular the ease of use and low rate of error. During the interviews of the examinees, acceptance for the use of tablets during the assessment was also detected.</p><p><strong>Discussion: </strong>Overall, it was found that the use of tablets during OSCEs was well accepted by both examiners and examinees. We expect that this mode of assessment also offers advantages regarding assessment documentation, use of resources, and rate of error in comparison with paper-based assessments; all of these aspects should be followed up on in further studies.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606489/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34102163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Development of a competency-based formative progress test with student-generated MCQs: Results from a multi-centre pilot study. 利用学生生成的 MCQ 开发基于能力的形成性进展测试:多中心试点研究的结果。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000988
Stefan Wagener, Andreas Möltner, Sevgi Tımbıl, Maryna Gornostayeva, Jobst-Hendrik Schultz, Peter Brüstle, Daniela Mohr, Anna Vander Beken, Julian Better, Martin Fries, Marc Gottschalk, Janine Günther, Laura Herrmann, Christian Kreisel, Tobias Moczko, Claudius Illg, Adam Jassowicz, Andreas Müller, Moritz Niesert, Felix Strübing, Jana Jünger

Introduction: Progress tests provide students feedback on their level of proficiency over the course of their medical studies. Peer-assisted learning and competency-based education have become increasingly important in medical education. Although progress tests have been proven to be useful as a longitudinal feedback instrument, there are currently no progress tests that have been created in cooperation with students or that focus on competency in medical education. In this study, we investigated the extent to which students can be included in the development of a progress test and demonstrated that aspects of knowledge related to competency can be represented on a competency-based progress test.

Methods: A two-dimensional blueprint for 144 multiple-choice questions (MCQs) covering groups of medical subjects and groups of competency areas was generated by three expert groups for developing the competency-based progress test. A total of 31 students from seven medical schools in Germany actively participated in this exercise. After completing an intensive and comprehensive training programme, the students generated and reviewed the test questions for the competency-based progress test using a separate platform of the ItemManagementSystem (IMS). This test was administered as a formative test to 469 students in a pilot study in November 2013 at eight medical schools in Germany. The scores were analysed for the overall test and differentiated according to the subject groups and competency areas.

Results: A pool of more than 200 MCQs was compiled by the students for pilot use, of which 118 student-generated MCQs were used in the progress test. University instructors supplemented this pool with 26 MCQs, which primarily addressed the area of scientific skills. The post-review showed that student-generated MCQs were of high quality with regard to test statistic criteria and content. Overall, the progress test displayed a very high reliability. When the academic years were compared, the progress test mapped out over the course of study not only by the overall test but also in terms of the subject groups and competency areas.

Outlook: Further development in cooperation with students will be continued. Focus will be on compiling additional questions and test formats that can represent competency at a higher skill level, such as key feature questions, situational judgement test questions and OSCE. In addition, the feedback formats will be successively expanded. The intention is also to offer the formative competency-based progress test online.

介绍:进度测试为学生提供反馈,了解他们在医学学习过程中的能力水平。同伴互助学习和能力本位教育在医学教育中变得越来越重要。虽然进度测验作为一种纵向反馈工具已被证明是有用的,但目前还没有与学生合作创建的进度测验,也没有侧重于医学教育能力的进度测验。在这项研究中,我们调查了学生参与进度测验开发的程度,并证明了与能力相关的知识可以在基于能力的进度测验中得到体现:方法:为开发基于能力的进步测试,三个专家组生成了一个包含 144 道选择题(MCQ)的二维蓝图,这些选择题涵盖了医学科目组和能力领域组。共有来自德国七所医学院的 31 名学生积极参与了这项工作。在完成密集的综合培训课程后,学生们利用项目管理系统(IMS)的独立平台生成并审核了能力进步测试的试题。2013 年 11 月,德国八所医学院对 469 名学生进行了试点研究,并将该测试作为形成性测试。对整个测试的分数进行了分析,并根据学科组和能力领域进行了区分:结果:学生编制了200多道MCQ供试点使用,其中118道学生编制的MCQ用于进度测试。大学教师在此基础上补充了 26 个 MCQ,主要涉及科学技能领域。事后审查结果表明,学生自制的 MCQ 在测试统计标准和内容方面质量较高。总体而言,进度测验显示出很高的可靠性。在对各学年的情况进行比较时,进度测试不仅在整体测试方面,而且在学科组和能力领域方面都与学习过程相吻合:展望:将继续与学生合作进行进一步的开发。重点将放在编制更多的试题和测试格式上,这些试题和格式可以代表更高技能水平的能力,如关键特征试题、情境判断试题和开放性开放考试(OSCE)。此外,还将陆续扩展反馈形式。还打算在网上提供基于形成性能力的进度测试。
{"title":"Development of a competency-based formative progress test with student-generated MCQs: Results from a multi-centre pilot study.","authors":"Stefan Wagener, Andreas Möltner, Sevgi Tımbıl, Maryna Gornostayeva, Jobst-Hendrik Schultz, Peter Brüstle, Daniela Mohr, Anna Vander Beken, Julian Better, Martin Fries, Marc Gottschalk, Janine Günther, Laura Herrmann, Christian Kreisel, Tobias Moczko, Claudius Illg, Adam Jassowicz, Andreas Müller, Moritz Niesert, Felix Strübing, Jana Jünger","doi":"10.3205/zma000988","DOIUrl":"10.3205/zma000988","url":null,"abstract":"<p><strong>Introduction: </strong>Progress tests provide students feedback on their level of proficiency over the course of their medical studies. Peer-assisted learning and competency-based education have become increasingly important in medical education. Although progress tests have been proven to be useful as a longitudinal feedback instrument, there are currently no progress tests that have been created in cooperation with students or that focus on competency in medical education. In this study, we investigated the extent to which students can be included in the development of a progress test and demonstrated that aspects of knowledge related to competency can be represented on a competency-based progress test.</p><p><strong>Methods: </strong>A two-dimensional blueprint for 144 multiple-choice questions (MCQs) covering groups of medical subjects and groups of competency areas was generated by three expert groups for developing the competency-based progress test. A total of 31 students from seven medical schools in Germany actively participated in this exercise. After completing an intensive and comprehensive training programme, the students generated and reviewed the test questions for the competency-based progress test using a separate platform of the ItemManagementSystem (IMS). This test was administered as a formative test to 469 students in a pilot study in November 2013 at eight medical schools in Germany. The scores were analysed for the overall test and differentiated according to the subject groups and competency areas.</p><p><strong>Results: </strong>A pool of more than 200 MCQs was compiled by the students for pilot use, of which 118 student-generated MCQs were used in the progress test. University instructors supplemented this pool with 26 MCQs, which primarily addressed the area of scientific skills. The post-review showed that student-generated MCQs were of high quality with regard to test statistic criteria and content. Overall, the progress test displayed a very high reliability. When the academic years were compared, the progress test mapped out over the course of study not only by the overall test but also in terms of the subject groups and competency areas.</p><p><strong>Outlook: </strong>Further development in cooperation with students will be continued. Focus will be on compiling additional questions and test formats that can represent competency at a higher skill level, such as key feature questions, situational judgement test questions and OSCE. In addition, the feedback formats will be successively expanded. The intention is also to offer the formative competency-based progress test online.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606478/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34102168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Influence of a revision course and the gender of examiners on the grades of the final ENT exam--a retrospective review of 3961 exams. 复习课程和审查员性别对耳鼻喉科期末考试成绩的影响——对3961次考试的回顾性回顾。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000980
Matthäus C Grasl, Rudolf Seemann, Michael Hanisch, Gregor Heiduschka, Karl Kremser, Dietmar Thurnher

Unlabelled: Revision courses should repeat already acquired knowledge and skills and mostly provide a basis for passing the following exam.

Aim: The aim of the study is to investigate the influence of a previously attended revision course on the grades achieved in a final exam (Ear, Nose and Throat Diseases). Additionally we ask the question whether the gender of the examiners plays a role concerning the marks or not.

Methods: 3961 exams at the Department of Ear, Nose and Throat (ENT) Diseases in Vienna were investigated, 725 with revision course (experimental group) and 3236 without previous revision course (comparison group). The revision courses were performed in a standardized way concerning form and content, interactive and case based. Both groups were examined uniform in regard to topics and time duration. 16 male and 6 female examiners were involved. The grading followed a five-level scale. The examination marks were calculated in the arithmetic mean and median value for the entire sample, gender dependence was calculated according to the Wilcoxon-Mann-Whitney-Test. The inferential statistics included single- and multiple factorial analyses of variance as well as uni- and multivariate regression models.

Results: The experimental group achieved a grade average of 2.54 compared with 2.46 for the comparison group. Splitting up into male and female examiners, an average of 2.54 and 2.58 resp. for the experimental group and 2.44 and 2.61 resp. for the comparison group resulted. Female examiner marked significantly lower grades in comparison to their male colleagues (P= 0.001926).

Conclusions: The ENT revision course did not improve the grade averages of the final ENT exam. Female examiners grade stricter than male examiners. There was no difference concerning grades 4 (pass) and 5 (fail) but female examiners grade less with mark 1.

未标注:复习课程应重复已学习的知识和技能,并主要为通过后续考试提供基础。目的:本研究的目的是调查以前参加过的复习课程对期末考试(耳鼻喉科)成绩的影响。此外,我们还提出了考官的性别是否对分数起作用的问题。方法:对维也纳市耳鼻喉科3961例体检者进行调查,其中有复习课程者725例(实验组),无复习课程者3236例(对照组)。复习课程在形式和内容上都是标准化的,互动式的,基于案例的。两组在主题和时间上的检查是一致的。涉及16名男性审查员和6名女性审查员。评分分为五级。考试成绩以整个样本的算术平均值和中位数计算,性别依赖根据Wilcoxon-Mann-Whitney-Test计算。推论统计包括单因子和多因子方差分析,以及单因子和多因子回归模型。结果:实验组平均得分2.54分,对照组平均得分2.46分。分男女考官,平均分别为2.54分和2.58分。实验组分别为2.44和2.61。结果为对照组。女考官的分数明显低于男考官(P= 0.001926)。结论:耳鼻喉科复习课程没有提高耳鼻喉科期末考试的平均成绩。女考官比男考官评分更严格。在4分(及格)和5分(不及格)方面没有差异,但女性主考官给1分的分数较少。
{"title":"Influence of a revision course and the gender of examiners on the grades of the final ENT exam--a retrospective review of 3961 exams.","authors":"Matthäus C Grasl,&nbsp;Rudolf Seemann,&nbsp;Michael Hanisch,&nbsp;Gregor Heiduschka,&nbsp;Karl Kremser,&nbsp;Dietmar Thurnher","doi":"10.3205/zma000980","DOIUrl":"https://doi.org/10.3205/zma000980","url":null,"abstract":"<p><strong>Unlabelled: </strong>Revision courses should repeat already acquired knowledge and skills and mostly provide a basis for passing the following exam.</p><p><strong>Aim: </strong>The aim of the study is to investigate the influence of a previously attended revision course on the grades achieved in a final exam (Ear, Nose and Throat Diseases). Additionally we ask the question whether the gender of the examiners plays a role concerning the marks or not.</p><p><strong>Methods: </strong>3961 exams at the Department of Ear, Nose and Throat (ENT) Diseases in Vienna were investigated, 725 with revision course (experimental group) and 3236 without previous revision course (comparison group). The revision courses were performed in a standardized way concerning form and content, interactive and case based. Both groups were examined uniform in regard to topics and time duration. 16 male and 6 female examiners were involved. The grading followed a five-level scale. The examination marks were calculated in the arithmetic mean and median value for the entire sample, gender dependence was calculated according to the Wilcoxon-Mann-Whitney-Test. The inferential statistics included single- and multiple factorial analyses of variance as well as uni- and multivariate regression models.</p><p><strong>Results: </strong>The experimental group achieved a grade average of 2.54 compared with 2.46 for the comparison group. Splitting up into male and female examiners, an average of 2.54 and 2.58 resp. for the experimental group and 2.44 and 2.61 resp. for the comparison group resulted. Female examiner marked significantly lower grades in comparison to their male colleagues (P= 0.001926).</p><p><strong>Conclusions: </strong>The ENT revision course did not improve the grade averages of the final ENT exam. Female examiners grade stricter than male examiners. There was no difference concerning grades 4 (pass) and 5 (fail) but female examiners grade less with mark 1.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606481/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34100622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The new final Clinical Skills examination in human medicine in Switzerland: Essential steps of exam development, implementation and evaluation, and central insights from the perspective of the national Working Group. 瑞士人类医学新的最终临床技能考试:考试发展、实施和评估的基本步骤,以及从国家工作组的角度来看的核心见解。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000982
Christoph Berendonk, Christian Schirlo, Gianmarco Balestra, Raphael Bonvin, Sabine Feller, Philippe Huber, Ernst Jünger, Matteo Monti, Kai Schnabel, Christine Beyeler, Sissel Guttormsen, Sören Huwendiek

Objective: Since 2011, the new national final examination in human medicine has been implemented in Switzerland, with a structured clinical-practical part in the OSCE format. From the perspective of the national Working Group, the current article describes the essential steps in the development, implementation and evaluation of the Federal Licensing Examination Clinical Skills (FLE CS) as well as the applied quality assurance measures. Finally, central insights gained from the last years are presented.

Methods: Based on the principles of action research, the FLE CS is in a constant state of further development. On the foundation of systematically documented experiences from previous years, in the Working Group, unresolved questions are discussed and resulting solution approaches are substantiated (planning), implemented in the examination (implementation) and subsequently evaluated (reflection). The presented results are the product of this iterative procedure.

Results: The FLE CS is created by experts from all faculties and subject areas in a multistage process. The examination is administered in German and French on a decentralised basis and consists of twelve interdisciplinary stations per candidate. As important quality assurance measures, the national Review Board (content validation) and the meetings of the standardised patient trainers (standardisation) have proven worthwhile. The statistical analyses show good measurement reliability and support the construct validity of the examination. Among the central insights of the past years, it has been established that the consistent implementation of the principles of action research contributes to the successful further development of the examination.

Conclusion: The centrally coordinated, collaborative-iterative process, incorporating experts from all faculties, makes a fundamental contribution to the quality of the FLE CS. The processes and insights presented here can be useful for others planning a similar undertaking.

目的:自2011年起,瑞士开始实施新的国家人体医学期末考试,并采用欧安组织格式的结构化临床实践部分。从国家工作组的角度来看,本文描述了联邦许可考试临床技能(FLE CS)的发展、实施和评估的基本步骤,以及应用的质量保证措施。最后,介绍了过去几年获得的核心见解。方法:在行动研究的基础上,持续不断地进行进一步的研究。在前几年系统记录的经验的基础上,工作组讨论未解决的问题,并证实最终的解决办法(规划),在审查中实施(实施),随后评价(反思)。给出的结果是这个迭代过程的产物。结果:FLE CS是由来自所有院系和学科领域的专家在多阶段过程中创建的。考试在分散的基础上以德语和法语进行,每位考生由12个跨学科考点组成。作为重要的质量保证措施,国家审查委员会(内容验证)和标准化患者培训师会议(标准化)已被证明是值得的。统计分析显示了较好的测量信度,支持了检验的结构效度。在过去几年的核心见解中,已经确定的是,行动研究原则的一致实施有助于考试的成功进一步发展。结论:集中协调,协作迭代的过程,包括来自所有院系的专家,对FLE CS的质量做出了根本性的贡献。这里介绍的过程和见解对计划类似任务的其他人可能很有用。
{"title":"The new final Clinical Skills examination in human medicine in Switzerland: Essential steps of exam development, implementation and evaluation, and central insights from the perspective of the national Working Group.","authors":"Christoph Berendonk,&nbsp;Christian Schirlo,&nbsp;Gianmarco Balestra,&nbsp;Raphael Bonvin,&nbsp;Sabine Feller,&nbsp;Philippe Huber,&nbsp;Ernst Jünger,&nbsp;Matteo Monti,&nbsp;Kai Schnabel,&nbsp;Christine Beyeler,&nbsp;Sissel Guttormsen,&nbsp;Sören Huwendiek","doi":"10.3205/zma000982","DOIUrl":"https://doi.org/10.3205/zma000982","url":null,"abstract":"<p><strong>Objective: </strong>Since 2011, the new national final examination in human medicine has been implemented in Switzerland, with a structured clinical-practical part in the OSCE format. From the perspective of the national Working Group, the current article describes the essential steps in the development, implementation and evaluation of the Federal Licensing Examination Clinical Skills (FLE CS) as well as the applied quality assurance measures. Finally, central insights gained from the last years are presented.</p><p><strong>Methods: </strong>Based on the principles of action research, the FLE CS is in a constant state of further development. On the foundation of systematically documented experiences from previous years, in the Working Group, unresolved questions are discussed and resulting solution approaches are substantiated (planning), implemented in the examination (implementation) and subsequently evaluated (reflection). The presented results are the product of this iterative procedure.</p><p><strong>Results: </strong>The FLE CS is created by experts from all faculties and subject areas in a multistage process. The examination is administered in German and French on a decentralised basis and consists of twelve interdisciplinary stations per candidate. As important quality assurance measures, the national Review Board (content validation) and the meetings of the standardised patient trainers (standardisation) have proven worthwhile. The statistical analyses show good measurement reliability and support the construct validity of the examination. Among the central insights of the past years, it has been established that the consistent implementation of the principles of action research contributes to the successful further development of the examination.</p><p><strong>Conclusion: </strong>The centrally coordinated, collaborative-iterative process, incorporating experts from all faculties, makes a fundamental contribution to the quality of the FLE CS. The processes and insights presented here can be useful for others planning a similar undertaking.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606485/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34100624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
The reliability of the pass/fail decision for assessments comprised of multiple components. 由多个部分组成的评估的通过/未通过决定的可靠性。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000984
Andreas Möltner, Sevgi Tımbıl, Jana Jünger

Objective: The decision having the most serious consequences for a student taking an assessment is the one to pass or fail that student. For this reason, the reliability of the pass/fail decision must be determined for high quality assessments, just as the measurement reliability of the point values. Assessments in a particular subject (graded course credit) are often composed of multiple components that must be passed independently of each other. When "conjunctively" combining separate pass/fail decisions, as with other complex decision rules for passing, adequate methods of analysis are necessary for estimating the accuracy and consistency of these classifications. To date, very few papers have addressed this issue; a generally applicable procedure was published by Douglas and Mislevy in 2010. Using the example of an assessment comprised of several parts that must be passed separately, this study analyzes the reliability underlying the decision to pass or fail students and discusses the impact of an improved method for identifying those who do not fulfill the minimum requirements.

Method: The accuracy and consistency of the decision to pass or fail an examinee in the subject cluster Internal Medicine/General Medicine/Clinical Chemistry at the University of Heidelberg's Faculty of Medicine was investigated. This cluster requires students to separately pass three components (two written exams and an OSCE), whereby students may reattempt to pass each component twice. Our analysis was carried out using the method described by Douglas and Mislevy.

Results: Frequently, when complex logical connections exist between the individual pass/fail decisions in the case of low failure rates, only a very low reliability for the overall decision to grant graded course credit can be achieved, even if high reliabilities exist for the various components. For the example analyzed here, the classification accuracy and consistency when conjunctively combining the three individual parts is relatively low with κ=0.49 or κ=0.47, despite the good reliability of over 0.75 for each of the three components. The option to repeat each component twice leads to a situation in which only about half of the candidates who do not satisfy the minimum requirements would fail the overall assessment, while the other half is able to continue their studies despite having deficient knowledge and skills.

Conclusion: The method put forth by Douglas and Mislevy allows the analysis of the decision accuracy and consistency for complex combinations of scores from different components. Even in the case of highly reliable components, it is not necessarily so that a reliable pass/fail decision has been reached - for instance in the case of low failure rates. Assessments must be administered with the explicit goal of identifying examinees that do not fulfill the minimum requirements.

目标:对参加测评的学生来说,最严重的后果就是决定该学生是及格还是不及格。因此,对于高质量的评估,必须确定通过/不通过决定的可靠性,就像点值的测量可靠性一样。某一学科的评估(分级课程学分)通常由多个部分组成,这些部分必须独立通过。与其他复杂的及格判定规则一样,当 "连带 "结合单独的及格/不及格判定时,需要适当的分析方法来估计这些分类的准确性和一致性。迄今为止,很少有论文涉及这一问题;道格拉斯和米斯莱维在 2010 年发表了一个普遍适用的程序。本研究以必须分别通过的几个部分组成的评估为例,分析了决定学生合格或不合格的基本可靠性,并讨论了改进方法对识别未达到最低要求的学生的影响:方法:研究了海德堡大学医学院内科学/普通医学/临床化学科目组中决定考生及格或不及格的准确性和一致性。该组别要求学生分别通过三门考试(两门笔试和一门 OSCE),每门考试学生都可以重考两次。我们采用道格拉斯和米斯莱维描述的方法进行了分析:结果:在不及格率较低的情况下,当各个及格/不及格决定之间存在复杂的逻辑联系时,即使各部分存在较高的信度,也往往只能获得很低的整体信度,以决定是否给予分级课程学分。就本文分析的示例而言,尽管三个组成部分的信度都超过了 0.75,但将三个单独部分结合在一起时,分类的准确性和一致性相对较低,κ=0.49 或 κ=0.47。由于每个部分都可重复两次,因此只有一半左右未达到最低要求的考生无法通过总评,而另一半考生尽管在知识和技能方面存在不足,但仍能继续学习:道格拉斯和米斯利维提出的方法可以分析不同组成部分的复杂分数组合的决策准确性和一致性。即使是高度可靠的成分,也不一定就能做出可靠的及格/不及格决定--例如,在不及格率较低的情况下。在进行评估时,必须明确目标,找出不符合最低要求的考生。
{"title":"The reliability of the pass/fail decision for assessments comprised of multiple components.","authors":"Andreas Möltner, Sevgi Tımbıl, Jana Jünger","doi":"10.3205/zma000984","DOIUrl":"10.3205/zma000984","url":null,"abstract":"<p><strong>Objective: </strong>The decision having the most serious consequences for a student taking an assessment is the one to pass or fail that student. For this reason, the reliability of the pass/fail decision must be determined for high quality assessments, just as the measurement reliability of the point values. Assessments in a particular subject (graded course credit) are often composed of multiple components that must be passed independently of each other. When \"conjunctively\" combining separate pass/fail decisions, as with other complex decision rules for passing, adequate methods of analysis are necessary for estimating the accuracy and consistency of these classifications. To date, very few papers have addressed this issue; a generally applicable procedure was published by Douglas and Mislevy in 2010. Using the example of an assessment comprised of several parts that must be passed separately, this study analyzes the reliability underlying the decision to pass or fail students and discusses the impact of an improved method for identifying those who do not fulfill the minimum requirements.</p><p><strong>Method: </strong>The accuracy and consistency of the decision to pass or fail an examinee in the subject cluster Internal Medicine/General Medicine/Clinical Chemistry at the University of Heidelberg's Faculty of Medicine was investigated. This cluster requires students to separately pass three components (two written exams and an OSCE), whereby students may reattempt to pass each component twice. Our analysis was carried out using the method described by Douglas and Mislevy.</p><p><strong>Results: </strong>Frequently, when complex logical connections exist between the individual pass/fail decisions in the case of low failure rates, only a very low reliability for the overall decision to grant graded course credit can be achieved, even if high reliabilities exist for the various components. For the example analyzed here, the classification accuracy and consistency when conjunctively combining the three individual parts is relatively low with κ=0.49 or κ=0.47, despite the good reliability of over 0.75 for each of the three components. The option to repeat each component twice leads to a situation in which only about half of the candidates who do not satisfy the minimum requirements would fail the overall assessment, while the other half is able to continue their studies despite having deficient knowledge and skills.</p><p><strong>Conclusion: </strong>The method put forth by Douglas and Mislevy allows the analysis of the decision accuracy and consistency for complex combinations of scores from different components. Even in the case of highly reliable components, it is not necessarily so that a reliable pass/fail decision has been reached - for instance in the case of low failure rates. Assessments must be administered with the explicit goal of identifying examinees that do not fulfill the minimum requirements.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606479/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34102164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
8th meeting of the medical assessment consortium UCAN: "Collaborative Perspectives for Competency-based and Quality-assured Medical Assessment". 医疗评估联盟UCAN第8次会议:"基于能力和有质量保证的医疗评估的协作观点"。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000979
Ajit Johannes Thamburaj, Konstantin Brass, Manfred Herrmann, Jana Jünger
On February 9 and February 1
{"title":"8th meeting of the medical assessment consortium UCAN: \"Collaborative Perspectives for Competency-based and Quality-assured Medical Assessment\".","authors":"Ajit Johannes Thamburaj,&nbsp;Konstantin Brass,&nbsp;Manfred Herrmann,&nbsp;Jana Jünger","doi":"10.3205/zma000979","DOIUrl":"https://doi.org/10.3205/zma000979","url":null,"abstract":"On February 9 and February 1","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606488/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34100621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Overcome the 60% passing score and improve the quality of assessment. 克服60%的通过率,提高考核质量。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000985
Ara Tekian, John Norcini

It is not unusual for institutions around the world to have fixed standards (e.g., 60%) for all of their examinations. This creates problems in the creation of examinations, since all of the content has to be chosen with an eye toward this fixed standard. As a result, the validity of the decisions based on these examinations can be adversely influenced, making them less useful for their intended purposes. Over the past several decades, many institutions have addressed this problem by using standard setting methods which are defensible, acceptable, and credible [1], [2]. Many methods are available and the major reasons to use them is to ensure that test content is appropriately selected and to be as fair to the students and other test users as possible [2], [3]. One barrier to the wider use of these methods is that some institutions object to the fact that the fixed standard (e.g., 60%) has not been applied. However, it is possible to rescale the passing score so that it is equal to the fixed standard, and then apply that same rescaling calculation to all of the test scores. This ensures that the institutional guidelines are not violated and allows the application of accepted methods of standard-setting. In turn, the application of these methods allow the content of the test to be selected without regard to a fixed standard, increases the validity of the decisions being made, and ensures a fairer and more accurate test of students.

世界各地的机构对所有考试都有固定的标准(例如,60%),这并不罕见。这就给考试的创建带来了问题,因为所有的内容都必须按照这个固定的标准来选择。因此,基于这些审查的决定的有效性可能受到不利影响,使其对其预期目的的用处更小。在过去的几十年里,许多机构通过使用可辩护的、可接受的和可信的标准制定方法来解决这个问题[1],[2]。有许多方法可供选择,使用这些方法的主要原因是为了确保适当选择测试内容,并尽可能公平地对待学生和其他测试用户[2],[3]。更广泛使用这些方法的一个障碍是,一些机构反对没有适用固定标准(例如60%)这一事实。但是,可以重新调整通过分数,使其与固定标准相等,然后将相同的重新调整计算应用于所有考试分数。这确保了机构准则不会被违反,并允许采用公认的标准制定方法。反过来,这些方法的应用使考试内容的选择不受固定标准的影响,增加了所做决定的有效性,并确保对学生进行更公平、更准确的考试。
{"title":"Overcome the 60% passing score and improve the quality of assessment.","authors":"Ara Tekian,&nbsp;John Norcini","doi":"10.3205/zma000985","DOIUrl":"https://doi.org/10.3205/zma000985","url":null,"abstract":"<p><p>It is not unusual for institutions around the world to have fixed standards (e.g., 60%) for all of their examinations. This creates problems in the creation of examinations, since all of the content has to be chosen with an eye toward this fixed standard. As a result, the validity of the decisions based on these examinations can be adversely influenced, making them less useful for their intended purposes. Over the past several decades, many institutions have addressed this problem by using standard setting methods which are defensible, acceptable, and credible [1], [2]. Many methods are available and the major reasons to use them is to ensure that test content is appropriately selected and to be as fair to the students and other test users as possible [2], [3]. One barrier to the wider use of these methods is that some institutions object to the fact that the fixed standard (e.g., 60%) has not been applied. However, it is possible to rescale the passing score so that it is equal to the fixed standard, and then apply that same rescaling calculation to all of the test scores. This ensures that the institutional guidelines are not violated and allows the application of accepted methods of standard-setting. In turn, the application of these methods allow the content of the test to be selected without regard to a fixed standard, increases the validity of the decisions being made, and ensures a fairer and more accurate test of students. </p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606480/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34102165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Impact of the Medical Faculty on Study Success in Freiburg: Results from Graduate Surveys. 弗莱堡医学院对学业成功的影响:毕业生调查结果。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000986
Silke Biller, Martin Boeker, Götz Fabry, Marianne Giesler

Aim: Using the data from graduate surveys, this study aims to analyze which factors related to teaching and learning at the Freiburg Faculty of Medicine can influence study success.

Background: Study success and the factors influencing it have long been the subject of investigation, with study success being measured in terms of easily quantifiable indicators (final grades, student satisfaction, etc.). In recent years, it has also frequently been assessed in terms of graduate competency levels. Graduate surveys are considered suitable instruments for measuring these dimensions of study success.

Method: Data from three Freiburg graduate surveys conducted one and a half years after graduation were drawn upon for the analysis. Study success was operationalized using four indicators: results on the written section of the M2 exam, self-assessment of medical expertise and scientific expertise, and student satisfaction. Using multiple regression analyses, the predictive power was calculated for selected variables, also measured by the graduate surveys, for the different study success indicators.

Results: It was possible to identify models that contribute slightly or moderately to the prediction of study success. The score earned on the university entrance qualification demonstrated itself to be the strongest predictor for forecasting the M2 written exam: R(2) is between 0.08 and 0.22 for the three surveys. Different variables specific to degree program structure and teaching are helpful for predicting medical expertise (R(2)=0.04-0.32) and student satisfaction (R(2)=0.12-0.35). The two variables, structure and curricular sequencing of the degree program and combination of theory and practice, show themselves to be significant, sample-invariant predictors (β-weight(Structure)=0.21-0.58, β-weight(Combination)=0.27-0.56). For scientific expertise, no sample-independent predictors could be determined.

Conclusion: Factors describing teaching hardly provide any assistance when predicting the written M2 exam score, which makes sense to the extent that teaching goes far beyond the heavily knowledge-based content of the written M2 exam. The lack of predictability for scientific expertise is most likely explained in that these have been only rarely included in the curriculum and often inexplicitly so. The variable combination of theory and practice appears to be significant for imparting medical expertise and the development of student satisfaction. The extent to which these relationships are practically relevant needs to be explored in further studies. A specific limitation is that the measurement of expertise and skill is based solely on self-assessments.

目的:本研究旨在利用毕业生调查数据,分析哪些与弗莱堡医学院教学相关的因素会影响学习成功:长期以来,学习成功与否及其影响因素一直是研究的主题,学习成功与否可以用易于量化的指标(期末成绩、学生满意度等)来衡量。近年来,也经常用毕业生的能力水平来评估学习成功与否。毕业生调查被认为是衡量学习成功与否的合适工具:方法:我们利用弗莱堡大学在毕业生毕业一年半后进行的三次毕业生调查数据进行分析。学习成功与否用四个指标来衡量:M2考试笔试成绩、医学专业知识和科学专业知识的自我评估以及学生满意度。通过多元回归分析,计算了所选变量对不同学习成功指标的预测能力,这些变量也通过毕业生调查进行测量:结果:可以确定对预测学习成功有轻微或适度帮助的模型。大学入学资格考试的分数是预测 M2 笔试的最强预测因子:三项调查的 R(2) 在 0.08 和 0.22 之间。学位课程结构和教学的不同变量有助于预测医学专业知识(R(2)=0.04-0.32)和学生满意度(R(2)=0.12-0.35)。学位课程的结构和课程顺序以及理论与实践的结合这两个变量本身是显著的、样本不变的预测因子(β-weight(Structure)=0.21-0.58, β-weight(Combination)=0.27-0.56)。结论:描述教学的因素很难对学生的学习产生影响:结论:在预测 M2 笔试分数时,描述教学的因素几乎没有任何帮助。科学专业知识缺乏可预测性的原因很可能是,这些知识很少被纳入课程,而且往往不明确。理论与实践的可变结合似乎对传授医学专业知识和培养学生的满意度具有重要意义。这些关系在多大程度上具有实际意义,还需要进一步研究探讨。一个具体的局限是,对专业知识和技能的衡量仅以自我评估为基础。
{"title":"Impact of the Medical Faculty on Study Success in Freiburg: Results from Graduate Surveys.","authors":"Silke Biller, Martin Boeker, Götz Fabry, Marianne Giesler","doi":"10.3205/zma000986","DOIUrl":"10.3205/zma000986","url":null,"abstract":"<p><strong>Aim: </strong>Using the data from graduate surveys, this study aims to analyze which factors related to teaching and learning at the Freiburg Faculty of Medicine can influence study success.</p><p><strong>Background: </strong>Study success and the factors influencing it have long been the subject of investigation, with study success being measured in terms of easily quantifiable indicators (final grades, student satisfaction, etc.). In recent years, it has also frequently been assessed in terms of graduate competency levels. Graduate surveys are considered suitable instruments for measuring these dimensions of study success.</p><p><strong>Method: </strong>Data from three Freiburg graduate surveys conducted one and a half years after graduation were drawn upon for the analysis. Study success was operationalized using four indicators: results on the written section of the M2 exam, self-assessment of medical expertise and scientific expertise, and student satisfaction. Using multiple regression analyses, the predictive power was calculated for selected variables, also measured by the graduate surveys, for the different study success indicators.</p><p><strong>Results: </strong>It was possible to identify models that contribute slightly or moderately to the prediction of study success. The score earned on the university entrance qualification demonstrated itself to be the strongest predictor for forecasting the M2 written exam: R(2) is between 0.08 and 0.22 for the three surveys. Different variables specific to degree program structure and teaching are helpful for predicting medical expertise (R(2)=0.04-0.32) and student satisfaction (R(2)=0.12-0.35). The two variables, structure and curricular sequencing of the degree program and combination of theory and practice, show themselves to be significant, sample-invariant predictors (β-weight(Structure)=0.21-0.58, β-weight(Combination)=0.27-0.56). For scientific expertise, no sample-independent predictors could be determined.</p><p><strong>Conclusion: </strong>Factors describing teaching hardly provide any assistance when predicting the written M2 exam score, which makes sense to the extent that teaching goes far beyond the heavily knowledge-based content of the written M2 exam. The lack of predictability for scientific expertise is most likely explained in that these have been only rarely included in the curriculum and often inexplicitly so. The variable combination of theory and practice appears to be significant for imparting medical expertise and the development of student satisfaction. The extent to which these relationships are practically relevant needs to be explored in further studies. A specific limitation is that the measurement of expertise and skill is based solely on self-assessments.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606483/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34102166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Final Oral/Practical State Examination at Freiburg Medical Faculty in 2012--Analysis of grading to test quality assurance. 2012 年弗莱堡医学院期末国家口试/实践考试--分析评分以检验质量保证。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000981
Angela Schickler, Peter Brüstle, Silke Biller

Aim: The aim of this study is to analyze the grades given for the oral/practical part of the German State Examination at the Medical Faculty of Freiburg. We examined whether or not the grades given for the written and the oral/practical examinations correlated and if differences in grading between the Freiburg University Medical Center (UMC) and the other teaching hospitals could be found. In order to improve the quality of the state examination, the medical school has been offering standardized training for examiners for several years. We evaluated whether or not trained and untrained examiners differed in their grading of the exam and how these differences have changed over time.

Methods: The results of the 2012 spring and fall exams were analyzed (N=315). The relevant data set was made available to us by the Baden-Württemberg Examination Office (Landesprüfungsamt). The data were analyzed by means of descriptive and inferential statistics.

Results: We observed a correlation of ρ=0.460** between the grades for the written and the oral/practical exams. The UMC and the teaching hospitals did not differ significantly in their grade distributions. Compared to untrained examiners, trained ones assigned the grade of "very good" less often. Furthermore, they displayed a significantly higher variance in the grades given (p=0.007, phi=0.165). This effect is stronger when concentrating specifically on those examiners who took part in the training less than a year before.

Conclusion: The results of this study suggest that the standardized training for examiners at the Medical Faculty of Freiburg is effective for quality assurance. As a consequence, more examiners should be motivated to take part in the training.

目的:本研究旨在分析弗莱堡医学院德国国家考试口试/实践部分的成绩。我们研究了笔试和口试/实践考试的成绩是否相关,以及弗莱堡大学医学中心(UMC)和其他教学医院的成绩是否存在差异。为了提高国家考试的质量,医学院多年来一直在对考官进行标准化培训。我们评估了受过培训和未受过培训的考官在考试评分方面是否存在差异,以及这些差异随着时间的推移发生了怎样的变化:我们分析了 2012 年春季和秋季考试的结果(N=315)。相关数据集由巴登一符腾堡州考试办公室(Landesprüfungsamt)提供。我们通过描述性和推论性统计对数据进行了分析:我们发现笔试和口试/实践考试成绩之间的相关性为ρ=0.460**。联大医院和教学医院在成绩分布上没有明显差异。与未经培训的考官相比,受过培训的考官给出 "很好 "等级的频率较低。此外,他们给出的分数差异明显更大(p=0.007,phi=0.165)。如果特别关注那些参加培训不到一年的主考官,这种效果会更明显:本研究的结果表明,弗莱堡医学院对主考官进行的标准化培训对保证质量是有效的。因此,应鼓励更多的主考官参加培训。
{"title":"The Final Oral/Practical State Examination at Freiburg Medical Faculty in 2012--Analysis of grading to test quality assurance.","authors":"Angela Schickler, Peter Brüstle, Silke Biller","doi":"10.3205/zma000981","DOIUrl":"10.3205/zma000981","url":null,"abstract":"<p><strong>Aim: </strong>The aim of this study is to analyze the grades given for the oral/practical part of the German State Examination at the Medical Faculty of Freiburg. We examined whether or not the grades given for the written and the oral/practical examinations correlated and if differences in grading between the Freiburg University Medical Center (UMC) and the other teaching hospitals could be found. In order to improve the quality of the state examination, the medical school has been offering standardized training for examiners for several years. We evaluated whether or not trained and untrained examiners differed in their grading of the exam and how these differences have changed over time.</p><p><strong>Methods: </strong>The results of the 2012 spring and fall exams were analyzed (N=315). The relevant data set was made available to us by the Baden-Württemberg Examination Office (Landesprüfungsamt). The data were analyzed by means of descriptive and inferential statistics.</p><p><strong>Results: </strong>We observed a correlation of ρ=0.460** between the grades for the written and the oral/practical exams. The UMC and the teaching hospitals did not differ significantly in their grade distributions. Compared to untrained examiners, trained ones assigned the grade of \"very good\" less often. Furthermore, they displayed a significantly higher variance in the grades given (p=0.007, phi=0.165). This effect is stronger when concentrating specifically on those examiners who took part in the training less than a year before.</p><p><strong>Conclusion: </strong>The results of this study suggest that the standardized training for examiners at the Medical Faculty of Freiburg is effective for quality assurance. As a consequence, more examiners should be motivated to take part in the training.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4606482/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34100623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DOPS (Direct Observation of Procedural Skills) in undergraduate skills-lab: Does it work? Analysis of skills-performance and curricular side effects. DOPS(程序技能的直接观察)在本科技能实验室:它有效吗?技能表现和课程副作用分析。
Pub Date : 2015-10-15 eCollection Date: 2015-01-01 DOI: 10.3205/zma000987
Christoph Profanter, Alexander Perathoner

Objective: Sufficient teaching and assessing clinical skills in the undergraduate setting becomes more and more important. In a surgical skills-lab course at the Medical University of Innsbruck fourth year students were teached with DOPS (direct observation of procedural skills). We analyzed whether DOPS worked or not in this setting, which performance levels could be reached compared to tutor teaching (one tutor, 5 students) and which curricular side effects could be observed.

Methods: In a prospective randomized trial in summer 2013 (April - June) four competence-level-based skills were teached in small groups during one week: surgical abdominal examination, urethral catheterization (phantom), rectal-digital examination (phantom), handling of central venous catheters. Group A was teached with DOPS, group B with a classical tutor system. Both groups underwent an OSCE (objective structured clinical examination) for assessment. 193 students were included in the study. Altogether 756 OSCE´s were carried out, 209 (27,6%) in the DOPS- and 547 (72,3%) in the tutor-group.

Results: Both groups reached high performance levels. In the first month there was a statistically significant difference (p<0,05) in performance of 95% positive OSCE items in the DOPS-group versus 88% in the tutor group. In the following months the performance rates showed no difference anymore and came to 90% in both groups. In practical skills the analysis revealed a high correspondence between positive DOPS (92,4%) and OSCE (90,8%) results.

Discussion: As shown by our data DOPS furnish high performance of clinical skills and work well in the undergraduate setting. Due to the high correspondence of DOPS and OSCE results DOPS should be considered as preferred assessment tool in a students skills-lab. The approximation of performance-rates within the months after initial superiority of DOPS could be explained by an interaction between DOPS and tutor system: DOPS elements seem to have improved tutoring and performance rates as well. DOPS in students 'skills-lab afford structured feedback and assessment without increased personnel and financial resources compared to classic small group training.

Conclusion: In summary, this study shows that DOPS represent an efficient method in teaching clinical skills. Their effects on didactic culture reach beyond the positive influence of performance rates.

目的:充分的临床技能教学和评估在本科教学中变得越来越重要。因斯布鲁克医科大学的一门外科技能实验室课程向四年级学生讲授DOPS(直接观察手术技能)。我们分析了DOPS在这种情况下是否有效,与导师教学(1名导师,5名学生)相比,可以达到哪些表现水平,以及可以观察到哪些课程副作用。方法:在2013年夏季(4 - 6月)的一项前瞻性随机试验中,在一周的时间内,以小组为单位教授四项基于能力水平的技能:外科腹部检查、尿道置管(幻像)、直肠指诊(幻像)、中心静脉置管的处理。A组采用DOPS教学,B组采用经典导师制教学。两组均接受客观结构化临床检查(OSCE)进行评估。193名学生参与了这项研究。总共进行了756次欧安组织的调查,其中ops组209次(27.6%),导师组547次(72.3%)。结果:两组均达到高水平。在第一个月,有统计学上的显著差异(p讨论:根据我们的数据显示,DOPS提供了很高的临床技能表现,在本科环境中工作得很好。由于DOPS和欧安组织结果的高度对应,DOPS应被视为学生技能实验室的首选评估工具。在DOPS取得初步优势后的几个月内,绩效率的近似可以用DOPS与导师制度之间的相互作用来解释:DOPS要素似乎也提高了辅导和绩效率。与传统的小组培训相比,学生技能实验室的DOPS提供了结构化的反馈和评估,而无需增加人力和财力资源。结论:本研究表明,DOPS是一种有效的临床技能教学方法。它们对教学文化的影响超出了表现率的积极影响。
{"title":"DOPS (Direct Observation of Procedural Skills) in undergraduate skills-lab: Does it work? Analysis of skills-performance and curricular side effects.","authors":"Christoph Profanter,&nbsp;Alexander Perathoner","doi":"10.3205/zma000987","DOIUrl":"https://doi.org/10.3205/zma000987","url":null,"abstract":"<p><strong>Objective: </strong>Sufficient teaching and assessing clinical skills in the undergraduate setting becomes more and more important. In a surgical skills-lab course at the Medical University of Innsbruck fourth year students were teached with DOPS (direct observation of procedural skills). We analyzed whether DOPS worked or not in this setting, which performance levels could be reached compared to tutor teaching (one tutor, 5 students) and which curricular side effects could be observed.</p><p><strong>Methods: </strong>In a prospective randomized trial in summer 2013 (April - June) four competence-level-based skills were teached in small groups during one week: surgical abdominal examination, urethral catheterization (phantom), rectal-digital examination (phantom), handling of central venous catheters. Group A was teached with DOPS, group B with a classical tutor system. Both groups underwent an OSCE (objective structured clinical examination) for assessment. 193 students were included in the study. Altogether 756 OSCE´s were carried out, 209 (27,6%) in the DOPS- and 547 (72,3%) in the tutor-group.</p><p><strong>Results: </strong>Both groups reached high performance levels. In the first month there was a statistically significant difference (p<0,05) in performance of 95% positive OSCE items in the DOPS-group versus 88% in the tutor group. In the following months the performance rates showed no difference anymore and came to 90% in both groups. In practical skills the analysis revealed a high correspondence between positive DOPS (92,4%) and OSCE (90,8%) results.</p><p><strong>Discussion: </strong>As shown by our data DOPS furnish high performance of clinical skills and work well in the undergraduate setting. Due to the high correspondence of DOPS and OSCE results DOPS should be considered as preferred assessment tool in a students skills-lab. The approximation of performance-rates within the months after initial superiority of DOPS could be explained by an interaction between DOPS and tutor system: DOPS elements seem to have improved tutoring and performance rates as well. DOPS in students 'skills-lab afford structured feedback and assessment without increased personnel and financial resources compared to classic small group training.</p><p><strong>Conclusion: </strong>In summary, this study shows that DOPS represent an efficient method in teaching clinical skills. Their effects on didactic culture reach beyond the positive influence of performance rates.</p>","PeriodicalId":30054,"journal":{"name":"GMS Zeitschrift fur Medizinische Ausbildung","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2015-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.3205/zma000987","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"34102167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
期刊
GMS Zeitschrift fur Medizinische Ausbildung
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1