感知MOOC满意度:使用机器学习和微调bert的评论挖掘方法

Q1 Social Sciences Computers and Education Artificial Intelligence Pub Date : 2025-06-01 Epub Date: 2025-01-17 DOI:10.1016/j.caeai.2025.100366
Xieling Chen , Haoran Xie , Di Zou , Gary Cheng , Xiaohui Tao , Fu Lee Wang
{"title":"感知MOOC满意度:使用机器学习和微调bert的评论挖掘方法","authors":"Xieling Chen ,&nbsp;Haoran Xie ,&nbsp;Di Zou ,&nbsp;Gary Cheng ,&nbsp;Xiaohui Tao ,&nbsp;Fu Lee Wang","doi":"10.1016/j.caeai.2025.100366","DOIUrl":null,"url":null,"abstract":"<div><div>This study investigates the application of machine learning and BERT models to identify topic categories in helpful online course reviews and uncover factors that influence the overall satisfaction of learners in massive open online courses (MOOCs). The research has three main objectives: (1) to assess the effectiveness of machine learning models in classifying review helpfulness, (2) to evaluate the performance of fine-tuned BERT models in identifying review topics, and (3) to explore the factors that influence learner satisfaction across various disciplines. The study uses a MOOC corpus containing 102,184 course reviews from 401 courses across 13 disciplines. The methodology involves three approaches: (1) machine learning for automatic classification of review helpfulness, (2) BERT models for automatic classification of review topics, and (3) multiple linear regression analysis to explore the factors influencing learner satisfaction. The results show that most machine learning models achieve precision, recall, and F1 scores above 80%, 99%, and 89%, respectively, in identifying review helpfulness. The fine-tuned BERT model outperforms baseline models with precision, recall, and F1 scores of 78.4%, 74.4%, and 75.9%, respectively, in classifying review topics. Additionally, the regression analysis identifies key factors affecting learner satisfaction, such as the positive influence of “Instructor” frequency and the negative impact of “Platforms and tools” and “Process”. These insights offer valuable guidance for educators, course designers, and platform developers, contributing to the optimization of MOOC offerings to better meet the evolving needs of learners.</div></div>","PeriodicalId":34469,"journal":{"name":"Computers and Education Artificial Intelligence","volume":"8 ","pages":"Article 100366"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Perceived MOOC satisfaction: A review mining approach using machine learning and fine-tuned BERTs\",\"authors\":\"Xieling Chen ,&nbsp;Haoran Xie ,&nbsp;Di Zou ,&nbsp;Gary Cheng ,&nbsp;Xiaohui Tao ,&nbsp;Fu Lee Wang\",\"doi\":\"10.1016/j.caeai.2025.100366\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This study investigates the application of machine learning and BERT models to identify topic categories in helpful online course reviews and uncover factors that influence the overall satisfaction of learners in massive open online courses (MOOCs). The research has three main objectives: (1) to assess the effectiveness of machine learning models in classifying review helpfulness, (2) to evaluate the performance of fine-tuned BERT models in identifying review topics, and (3) to explore the factors that influence learner satisfaction across various disciplines. The study uses a MOOC corpus containing 102,184 course reviews from 401 courses across 13 disciplines. The methodology involves three approaches: (1) machine learning for automatic classification of review helpfulness, (2) BERT models for automatic classification of review topics, and (3) multiple linear regression analysis to explore the factors influencing learner satisfaction. The results show that most machine learning models achieve precision, recall, and F1 scores above 80%, 99%, and 89%, respectively, in identifying review helpfulness. The fine-tuned BERT model outperforms baseline models with precision, recall, and F1 scores of 78.4%, 74.4%, and 75.9%, respectively, in classifying review topics. Additionally, the regression analysis identifies key factors affecting learner satisfaction, such as the positive influence of “Instructor” frequency and the negative impact of “Platforms and tools” and “Process”. These insights offer valuable guidance for educators, course designers, and platform developers, contributing to the optimization of MOOC offerings to better meet the evolving needs of learners.</div></div>\",\"PeriodicalId\":34469,\"journal\":{\"name\":\"Computers and Education Artificial Intelligence\",\"volume\":\"8 \",\"pages\":\"Article 100366\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computers and Education Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2666920X25000062\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/17 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"Social Sciences\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers and Education Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666920X25000062","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/17 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 0

摘要

本研究探讨了机器学习和BERT模型的应用,以识别有用的在线课程评论中的主题类别,并揭示影响大规模开放在线课程(MOOCs)学习者整体满意度的因素。本研究有三个主要目标:(1)评估机器学习模型在分类复习有用性方面的有效性;(2)评估微调BERT模型在识别复习主题方面的表现;(3)探索影响不同学科学习者满意度的因素。该研究使用了一个MOOC语料库,其中包含13个学科401门课程的102184门课程评论。该方法包括三种方法:(1)机器学习对复习有用性进行自动分类;(2)BERT模型对复习主题进行自动分类;(3)多元线性回归分析探讨影响学习者满意度的因素。结果表明,大多数机器学习模型在识别评论有用性方面的准确率、召回率和F1得分分别超过80%、99%和89%。在分类评论主题方面,微调后的BERT模型在精度、召回率和F1得分方面分别优于基线模型,分别为78.4%、74.4%和75.9%。此外,回归分析确定了影响学习者满意度的关键因素,如“讲师”频率的积极影响,以及“平台和工具”和“过程”的消极影响。这些见解为教育工作者、课程设计师和平台开发人员提供了宝贵的指导,有助于优化MOOC产品,更好地满足学习者不断变化的需求。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Perceived MOOC satisfaction: A review mining approach using machine learning and fine-tuned BERTs
This study investigates the application of machine learning and BERT models to identify topic categories in helpful online course reviews and uncover factors that influence the overall satisfaction of learners in massive open online courses (MOOCs). The research has three main objectives: (1) to assess the effectiveness of machine learning models in classifying review helpfulness, (2) to evaluate the performance of fine-tuned BERT models in identifying review topics, and (3) to explore the factors that influence learner satisfaction across various disciplines. The study uses a MOOC corpus containing 102,184 course reviews from 401 courses across 13 disciplines. The methodology involves three approaches: (1) machine learning for automatic classification of review helpfulness, (2) BERT models for automatic classification of review topics, and (3) multiple linear regression analysis to explore the factors influencing learner satisfaction. The results show that most machine learning models achieve precision, recall, and F1 scores above 80%, 99%, and 89%, respectively, in identifying review helpfulness. The fine-tuned BERT model outperforms baseline models with precision, recall, and F1 scores of 78.4%, 74.4%, and 75.9%, respectively, in classifying review topics. Additionally, the regression analysis identifies key factors affecting learner satisfaction, such as the positive influence of “Instructor” frequency and the negative impact of “Platforms and tools” and “Process”. These insights offer valuable guidance for educators, course designers, and platform developers, contributing to the optimization of MOOC offerings to better meet the evolving needs of learners.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
16.80
自引率
0.00%
发文量
66
审稿时长
50 days
期刊最新文献
Decoding divides: The role of socioeconomic status and personality traits in AI divides and educational inequality Scaffolding critical thinking with generative AI: Design principles for integrating large language models in higher education Opening the blackbox of LLM-based automated essay scoring: Insights into feature weighting patterns and score validity Engagement in LLM chatbot-supported learning: The pivotal roles of GenAI competency and emotion Play with AI (PL-AI): A play-centered, design-based curriculum for AI literacy in pre-K and kindergarten
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1