{"title":"Research on the application of English short essay reading emotional analysis in online English teaching under IoT scenario","authors":"Xiaoli Zhan","doi":"10.1002/itl2.535","DOIUrl":null,"url":null,"abstract":"Speech‐emotion analysis plays an important role in English teaching. The existing convolutional neural networks (CNNs) can fully explore the spatial features of speech information, and cannot effectively utilize the temporal dependence of speech signals. In addition, it is difficult to build a more efficient and robust sentiment analysis system by solely utilizing speech information. With the development of the Internet of Things (IoTs), online multimodal information, including speech, video, and text, has become more convenient. To this end, this paper proposes a novel multimodal fusion emotion analysis system. Firstly, by combining convolutional networks with Transformer encoders, the spatiotemporal dependencies of speech information are effectively utilized. To improve multimodal information fusion, we introduce the exchange‐based fusion mechanism. The experimental results on the public dataset indicate that the proposed multimodal fusion model achieves the best performance. In online English teaching, teachers can effectively improve the quality of teaching by leveraging the feedback information of students' emotional states through our proposed deep model.","PeriodicalId":509592,"journal":{"name":"Internet Technology Letters","volume":"49 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Internet Technology Letters","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/itl2.535","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Speech‐emotion analysis plays an important role in English teaching. The existing convolutional neural networks (CNNs) can fully explore the spatial features of speech information, and cannot effectively utilize the temporal dependence of speech signals. In addition, it is difficult to build a more efficient and robust sentiment analysis system by solely utilizing speech information. With the development of the Internet of Things (IoTs), online multimodal information, including speech, video, and text, has become more convenient. To this end, this paper proposes a novel multimodal fusion emotion analysis system. Firstly, by combining convolutional networks with Transformer encoders, the spatiotemporal dependencies of speech information are effectively utilized. To improve multimodal information fusion, we introduce the exchange‐based fusion mechanism. The experimental results on the public dataset indicate that the proposed multimodal fusion model achieves the best performance. In online English teaching, teachers can effectively improve the quality of teaching by leveraging the feedback information of students' emotional states through our proposed deep model.