Scoring Korean Written Responses Using English - Based Automated Computer Scoring Models and Machine Translation: A Case of Natural Selection Concept Test

M. Ha
{"title":"Scoring Korean Written Responses Using English - Based Automated Computer Scoring Models and Machine Translation: A Case of Natural Selection Concept Test","authors":"M. Ha","doi":"10.14697/JKASE.2016.36.3.0389","DOIUrl":null,"url":null,"abstract":"This study aims to test the efficacy of English-based automated computer scoring models and machine translation to score Korean college students’ written responses on natural selection concept items. To this end, I collected 128 pre-service biology teachers’ written responses on four-item instrument (total 512 written responses). The machine translation software (i.e., Google Translate) translated both original responses and spell-corrected responses. The presence/absence of five scientific ideas and three naïve ideas in both translated responses were judged by the automated computer scoring models (i.e., EvoGrader). The computer-scored results (4096 predictions) were compared with expert-scored results. The results illustrated that no significant differences in both average scores and statistical results using average scores was found between the computer-scored result and experts-scored result. The Pearson correlation coefficients of composite scores for each student between computer scoring and experts scoring were 0.848 for scientific ideas and 0.776 for naïve ideas. The inter-rater reliability indices (Cohen kappa) between computer scoring and experts scoring for linguistically simple concepts (e.g., variation, competition, and limited resources) were over 0.8. These findings reveal that the English-based automated computer scoring models and machine translation can be a promising method in scoring Korean college students’ written responses on natural selection concept items.","PeriodicalId":107400,"journal":{"name":"Journal of the Korean Association for Research in Science Education","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Korean Association for Research in Science Education","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.14697/JKASE.2016.36.3.0389","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

This study aims to test the efficacy of English-based automated computer scoring models and machine translation to score Korean college students’ written responses on natural selection concept items. To this end, I collected 128 pre-service biology teachers’ written responses on four-item instrument (total 512 written responses). The machine translation software (i.e., Google Translate) translated both original responses and spell-corrected responses. The presence/absence of five scientific ideas and three naïve ideas in both translated responses were judged by the automated computer scoring models (i.e., EvoGrader). The computer-scored results (4096 predictions) were compared with expert-scored results. The results illustrated that no significant differences in both average scores and statistical results using average scores was found between the computer-scored result and experts-scored result. The Pearson correlation coefficients of composite scores for each student between computer scoring and experts scoring were 0.848 for scientific ideas and 0.776 for naïve ideas. The inter-rater reliability indices (Cohen kappa) between computer scoring and experts scoring for linguistically simple concepts (e.g., variation, competition, and limited resources) were over 0.8. These findings reveal that the English-based automated computer scoring models and machine translation can be a promising method in scoring Korean college students’ written responses on natural selection concept items.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用基于英语的自动计算机评分模型和机器翻译对韩语书面回答进行评分:以自然选择概念测试为例
本研究旨在测试以英语为基础的自动评分模型和机器翻译对韩国大学生自然选择概念题的书面回答进行评分的有效性。为此,我收集了128名职前生物教师在四项仪器上的书面回复(共计512份书面回复)。机器翻译软件(即谷歌Translate)翻译原始回复和拼写纠正后的回复。5个科学观点和3个naïve观点在翻译后的回答中是否存在由自动计算机评分模型(即EvoGrader)判断。计算机评分的结果(4096个预测)与专家评分的结果进行了比较。结果表明,计算机评分结果与专家评分结果在平均得分和使用平均得分的统计结果上均无显著差异。每个学生的计算机得分与专家得分的Pearson相关系数分别为:科学思想为0.848,naïve思想为0.776。对于语言上简单的概念(如变异、竞争和有限资源),计算机评分和专家评分之间的评分者间信度指数(Cohen kappa)大于0.8。这些研究结果表明,基于英语的计算机自动评分模型和机器翻译可以作为一种有前途的方法来评分韩国大学生自然选择概念题的书面回答。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
2015 개정 과학과 교육과정의 적용에 대한 초·중학교 교사의 인식과 요구 STEAM 기반 학생 연구 활동 분석을 통한 과학 교육에 대한 시사점 고찰 한국, 미국, 영국, 싱가포르의 과학 교육과정 비교 - 에너지 개념을 중심으로 - A Case Study on Application of the NABI Program to Realize the 'Practice Centered Mechanism of Manifesting Character' Item Response Analysis of Energy as a Cross-Cutting Concept for Grades 3 to 9
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1