更正 "机器接管:用于发散性思维任务自动评分的各种监督学习方法的比较"

IF 2.8 2区 心理学 Q2 PSYCHOLOGY, EDUCATIONAL Journal of Creative Behavior Pub Date : 2023-12-08 DOI:10.1002/jocb.627
{"title":"更正 \"机器接管:用于发散性思维任务自动评分的各种监督学习方法的比较\"","authors":"","doi":"10.1002/jocb.627","DOIUrl":null,"url":null,"abstract":"<p>Buczak, P., Huang, H., Forthmann, B., &amp; Doebler, P. (2023). The machines take over: A comparison of various supervised learning approaches for automated scoring of divergent thinking tasks. <i>The Journal of Creative Behavior</i>, <i>57</i>, 17–36. https://doi.org/10.1002/jocb.559\n </p><p>Due to an error in our code for the data preprocessing, we only utilized the word embeddings (WEs) of the last word in sentences instead of aggregating all WEs. This affected the use of WEs as model features and resulted in incomplete data usage. Our program code has been fixed and we have rerun our simulations.</p><p>We apologize for this error.</p><p>To highlight the most important differences between the findings reported in Buczak et al. (2023) and the corrected findings, Figure 1 shows the changes in terms of RMSE for the three learners on the Hofelich-Mohr et al. (2016) data set when applied in the setting of our first simulation study using two different semantic spaces. Overall, support vector regression (SVR), extreme gradient boosting (XGB), and random forest (RF) models profited greatly from the updated WEs whenever these were directly included as features. As the “Meta” feature set did not contain WEs in any form, the corresponding models and their results remained unaffected by the change. For the “Meta- + WE-based” feature set, however, the model's performance declined slightly. Regarding the learners, the updated WEs mostly led to a noticeable jump in performance for SVR when using a low-dimensional semantic space (i.e., 50 or 100 dimensions). In these cases, SVR even outperformed XGB and RF. However, SVR suffered when using high-dimensional semantic spaces. Apart from SVR, only subtle differences emerged regarding the choice of semantic space for XGB and RF.</p><p>Similar patterns of change were observed in the results for the Silvia et al. (2008) data, and when looking at correlations instead of the prediction RMSE for both data sets (figures were excluded for the sake of brevity). For the cross-sample prediction setup in our second simulation study, this mostly held up as well. However, when training on Hofelich-Mohr et al. (2016) and predicting Silvia et al. (2008), the performance of SVR increased with increasing semantic space dimensionality when studying the RMSE. For correlation scores, SVR's performance, again, suffered when a semantic space of high dimensionality was used.</p><p>As for the models using the WEs directly, the improved performance was to be expected since the aggregated WEs of all words capture more information about the answer than the WEs of a single word. However, the reasons behind the performance decrease for the models using WE-based features seem less clear-cut. Perhaps, WE-based features such as cosine similarity, WE norm, and the number of high-loading WEs degraded slightly in quality because we added word vectors from single words, including words almost unrelated to the idea described.</p><p>Although the respective feature sets were differently affected by the updated WEs, the overall order of performances achieved by the feature sets remained mostly unchanged. However, including further features alongside WEs does not necessarily enhance model performance as before. For the Silvia et al. (2008) data, models with the combined feature set achieved slightly lower performance than models with only WEs as features. Thus, one should consider the data at hand when deciding whether the WEs need to be complemented by further features.</p><p>The fact that for our initial analysis using the WEs of an answer's last word already substantially improved the performance could be explained by the nature of the DT task answers encountered in the data we used. As can be seen in Figure 2, answers were often comprised of single words or short phrases likely to contain the answer's key information in its last word. Thus, the initial models using only the WEs of an answer's last word already performed quite well in this case.</p><p>In terms of feature importance, the updated results revealed some changes regarding the importance order. However, the partial dependency plots (PDP) for the features stimulus similarity and number of words displayed the trend initially reported. In our updated results, we have provided additional PDPs for the WE norm and the number of high-loading WE component features.</p><p>Regarding the significant performance differences of SVR concerning the dimensionality of the semantic space, we posit that the root of the issue lies in the fact that the SVR fitted in this case utilizes radial basis function kernels. It has been repeatedly stressed in the support vector literature that kernel choice matters, especially in high-dimensional settings, and that dimension reduction techniques are advisable.</p><p>Even though SVR partly outperformed XGB and RF slightly for low-dimensional semantic spaces, we still stand by our initial recommendation of XGB and RF as they displayed less sensitivity to the choice of the semantic space. If, however, SVR is to be used, reducing the dimensionality of WEs, for example, via principal components analysis, remains a viable option. Furthermore, our recommendation of using additive word composition over padding is even strengthened as the new results mostly closed the predictive performance gap present in our initial analysis. As an additional result from the correction process, we observed that typical English language responses to alternate uses tasks often place a noun central for the idea conveyed at the end of a sentence, and hence, the WEs of the last word proved to be surprisingly predictive. Apart from these aspects, the original discussion remains unaffected.</p>","PeriodicalId":39915,"journal":{"name":"Journal of Creative Behavior","volume":null,"pages":null},"PeriodicalIF":2.8000,"publicationDate":"2023-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/jocb.627","citationCount":"0","resultStr":"{\"title\":\"Correction to “The Machines Take over: A Comparison of Various Supervised Learning Approaches for Automated Scoring of Divergent Thinking Tasks”\",\"authors\":\"\",\"doi\":\"10.1002/jocb.627\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Buczak, P., Huang, H., Forthmann, B., &amp; Doebler, P. (2023). The machines take over: A comparison of various supervised learning approaches for automated scoring of divergent thinking tasks. <i>The Journal of Creative Behavior</i>, <i>57</i>, 17–36. https://doi.org/10.1002/jocb.559\\n </p><p>Due to an error in our code for the data preprocessing, we only utilized the word embeddings (WEs) of the last word in sentences instead of aggregating all WEs. This affected the use of WEs as model features and resulted in incomplete data usage. Our program code has been fixed and we have rerun our simulations.</p><p>We apologize for this error.</p><p>To highlight the most important differences between the findings reported in Buczak et al. (2023) and the corrected findings, Figure 1 shows the changes in terms of RMSE for the three learners on the Hofelich-Mohr et al. (2016) data set when applied in the setting of our first simulation study using two different semantic spaces. Overall, support vector regression (SVR), extreme gradient boosting (XGB), and random forest (RF) models profited greatly from the updated WEs whenever these were directly included as features. As the “Meta” feature set did not contain WEs in any form, the corresponding models and their results remained unaffected by the change. For the “Meta- + WE-based” feature set, however, the model's performance declined slightly. Regarding the learners, the updated WEs mostly led to a noticeable jump in performance for SVR when using a low-dimensional semantic space (i.e., 50 or 100 dimensions). In these cases, SVR even outperformed XGB and RF. However, SVR suffered when using high-dimensional semantic spaces. Apart from SVR, only subtle differences emerged regarding the choice of semantic space for XGB and RF.</p><p>Similar patterns of change were observed in the results for the Silvia et al. (2008) data, and when looking at correlations instead of the prediction RMSE for both data sets (figures were excluded for the sake of brevity). For the cross-sample prediction setup in our second simulation study, this mostly held up as well. However, when training on Hofelich-Mohr et al. (2016) and predicting Silvia et al. (2008), the performance of SVR increased with increasing semantic space dimensionality when studying the RMSE. For correlation scores, SVR's performance, again, suffered when a semantic space of high dimensionality was used.</p><p>As for the models using the WEs directly, the improved performance was to be expected since the aggregated WEs of all words capture more information about the answer than the WEs of a single word. However, the reasons behind the performance decrease for the models using WE-based features seem less clear-cut. Perhaps, WE-based features such as cosine similarity, WE norm, and the number of high-loading WEs degraded slightly in quality because we added word vectors from single words, including words almost unrelated to the idea described.</p><p>Although the respective feature sets were differently affected by the updated WEs, the overall order of performances achieved by the feature sets remained mostly unchanged. However, including further features alongside WEs does not necessarily enhance model performance as before. For the Silvia et al. (2008) data, models with the combined feature set achieved slightly lower performance than models with only WEs as features. Thus, one should consider the data at hand when deciding whether the WEs need to be complemented by further features.</p><p>The fact that for our initial analysis using the WEs of an answer's last word already substantially improved the performance could be explained by the nature of the DT task answers encountered in the data we used. As can be seen in Figure 2, answers were often comprised of single words or short phrases likely to contain the answer's key information in its last word. Thus, the initial models using only the WEs of an answer's last word already performed quite well in this case.</p><p>In terms of feature importance, the updated results revealed some changes regarding the importance order. However, the partial dependency plots (PDP) for the features stimulus similarity and number of words displayed the trend initially reported. In our updated results, we have provided additional PDPs for the WE norm and the number of high-loading WE component features.</p><p>Regarding the significant performance differences of SVR concerning the dimensionality of the semantic space, we posit that the root of the issue lies in the fact that the SVR fitted in this case utilizes radial basis function kernels. It has been repeatedly stressed in the support vector literature that kernel choice matters, especially in high-dimensional settings, and that dimension reduction techniques are advisable.</p><p>Even though SVR partly outperformed XGB and RF slightly for low-dimensional semantic spaces, we still stand by our initial recommendation of XGB and RF as they displayed less sensitivity to the choice of the semantic space. If, however, SVR is to be used, reducing the dimensionality of WEs, for example, via principal components analysis, remains a viable option. Furthermore, our recommendation of using additive word composition over padding is even strengthened as the new results mostly closed the predictive performance gap present in our initial analysis. As an additional result from the correction process, we observed that typical English language responses to alternate uses tasks often place a noun central for the idea conveyed at the end of a sentence, and hence, the WEs of the last word proved to be surprisingly predictive. Apart from these aspects, the original discussion remains unaffected.</p>\",\"PeriodicalId\":39915,\"journal\":{\"name\":\"Journal of Creative Behavior\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.8000,\"publicationDate\":\"2023-12-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1002/jocb.627\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Creative Behavior\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/jocb.627\",\"RegionNum\":2,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"PSYCHOLOGY, EDUCATIONAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Creative Behavior","FirstCategoryId":"102","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/jocb.627","RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PSYCHOLOGY, EDUCATIONAL","Score":null,"Total":0}
引用次数: 0

摘要

尽管在低维语义空间中,SVR 的部分性能略优于 XGB 和 RF,但我们仍然坚持最初推荐的 XGB 和 RF,因为它们对语义空间选择的敏感度较低。不过,如果要使用 SVR,降低 WE 的维度(例如通过主成分分析)仍然是一个可行的选择。此外,由于新的结果在很大程度上缩小了我们最初分析中存在的预测性能差距,因此我们关于使用加法构词法而非填充法的建议得到了进一步加强。校正过程中的另一个结果是,我们观察到,典型的英语交替使用任务反应通常会将一个名词作为句子末尾所要表达的思想的中心,因此,最后一个词的 WEs 被证明具有令人惊讶的预测性。除此以外,原讨论内容未受影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

摘要图片

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Correction to “The Machines Take over: A Comparison of Various Supervised Learning Approaches for Automated Scoring of Divergent Thinking Tasks”

Buczak, P., Huang, H., Forthmann, B., & Doebler, P. (2023). The machines take over: A comparison of various supervised learning approaches for automated scoring of divergent thinking tasks. The Journal of Creative Behavior, 57, 17–36. https://doi.org/10.1002/jocb.559

Due to an error in our code for the data preprocessing, we only utilized the word embeddings (WEs) of the last word in sentences instead of aggregating all WEs. This affected the use of WEs as model features and resulted in incomplete data usage. Our program code has been fixed and we have rerun our simulations.

We apologize for this error.

To highlight the most important differences between the findings reported in Buczak et al. (2023) and the corrected findings, Figure 1 shows the changes in terms of RMSE for the three learners on the Hofelich-Mohr et al. (2016) data set when applied in the setting of our first simulation study using two different semantic spaces. Overall, support vector regression (SVR), extreme gradient boosting (XGB), and random forest (RF) models profited greatly from the updated WEs whenever these were directly included as features. As the “Meta” feature set did not contain WEs in any form, the corresponding models and their results remained unaffected by the change. For the “Meta- + WE-based” feature set, however, the model's performance declined slightly. Regarding the learners, the updated WEs mostly led to a noticeable jump in performance for SVR when using a low-dimensional semantic space (i.e., 50 or 100 dimensions). In these cases, SVR even outperformed XGB and RF. However, SVR suffered when using high-dimensional semantic spaces. Apart from SVR, only subtle differences emerged regarding the choice of semantic space for XGB and RF.

Similar patterns of change were observed in the results for the Silvia et al. (2008) data, and when looking at correlations instead of the prediction RMSE for both data sets (figures were excluded for the sake of brevity). For the cross-sample prediction setup in our second simulation study, this mostly held up as well. However, when training on Hofelich-Mohr et al. (2016) and predicting Silvia et al. (2008), the performance of SVR increased with increasing semantic space dimensionality when studying the RMSE. For correlation scores, SVR's performance, again, suffered when a semantic space of high dimensionality was used.

As for the models using the WEs directly, the improved performance was to be expected since the aggregated WEs of all words capture more information about the answer than the WEs of a single word. However, the reasons behind the performance decrease for the models using WE-based features seem less clear-cut. Perhaps, WE-based features such as cosine similarity, WE norm, and the number of high-loading WEs degraded slightly in quality because we added word vectors from single words, including words almost unrelated to the idea described.

Although the respective feature sets were differently affected by the updated WEs, the overall order of performances achieved by the feature sets remained mostly unchanged. However, including further features alongside WEs does not necessarily enhance model performance as before. For the Silvia et al. (2008) data, models with the combined feature set achieved slightly lower performance than models with only WEs as features. Thus, one should consider the data at hand when deciding whether the WEs need to be complemented by further features.

The fact that for our initial analysis using the WEs of an answer's last word already substantially improved the performance could be explained by the nature of the DT task answers encountered in the data we used. As can be seen in Figure 2, answers were often comprised of single words or short phrases likely to contain the answer's key information in its last word. Thus, the initial models using only the WEs of an answer's last word already performed quite well in this case.

In terms of feature importance, the updated results revealed some changes regarding the importance order. However, the partial dependency plots (PDP) for the features stimulus similarity and number of words displayed the trend initially reported. In our updated results, we have provided additional PDPs for the WE norm and the number of high-loading WE component features.

Regarding the significant performance differences of SVR concerning the dimensionality of the semantic space, we posit that the root of the issue lies in the fact that the SVR fitted in this case utilizes radial basis function kernels. It has been repeatedly stressed in the support vector literature that kernel choice matters, especially in high-dimensional settings, and that dimension reduction techniques are advisable.

Even though SVR partly outperformed XGB and RF slightly for low-dimensional semantic spaces, we still stand by our initial recommendation of XGB and RF as they displayed less sensitivity to the choice of the semantic space. If, however, SVR is to be used, reducing the dimensionality of WEs, for example, via principal components analysis, remains a viable option. Furthermore, our recommendation of using additive word composition over padding is even strengthened as the new results mostly closed the predictive performance gap present in our initial analysis. As an additional result from the correction process, we observed that typical English language responses to alternate uses tasks often place a noun central for the idea conveyed at the end of a sentence, and hence, the WEs of the last word proved to be surprisingly predictive. Apart from these aspects, the original discussion remains unaffected.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Journal of Creative Behavior
Journal of Creative Behavior Arts and Humanities-Visual Arts and Performing Arts
CiteScore
7.50
自引率
7.70%
发文量
44
期刊介绍: The Journal of Creative Behavior is our quarterly academic journal citing the most current research in creative thinking. For nearly four decades JCB has been the benchmark scientific periodical in the field. It provides up to date cutting-edge ideas about creativity in education, psychology, business, arts and more.
期刊最新文献
Issue Information Effects of Adverse Childhood Experiences on Creativity from Life History Theory Novelty Seeking Differences in Temporal Dynamics for Novelty and Appropriateness Processing of Creative Information: An ERP Investigation Collectivism–Individualism Makes the Relationships Between Digital Games Use and Creativity Different The Silver Lining of Workaholism: Its Impact on Employees' Creativity and Presenteeism Explained
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1