首页 > 最新文献

Data Technologies and Applications最新文献

英文 中文
ID-SF-Fusion: a cooperative model of intent detection and slot filling for natural language understanding ID-SF-Fusion:用于自然语言理解的意图检测和槽填充合作模型
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2024-01-19 DOI: 10.1108/dta-03-2023-0088
Meng Zhu, Xiaolong Xu

Purpose

Intent detection (ID) and slot filling (SF) are two important tasks in natural language understanding. ID is to identify the main intent of a paragraph of text. The goal of SF is to extract the information that is important to the intent from the input sentence. However, most of the existing methods use sentence-level intention recognition, which has the risk of error propagation, and the relationship between intention recognition and SF is not explicitly modeled. Aiming at this problem, this paper proposes a collaborative model of ID and SF for intelligent spoken language understanding called ID-SF-Fusion.

Design/methodology/approach

ID-SF-Fusion uses Bidirectional Encoder Representation from Transformers (BERT) and Bidirectional Long Short-Term Memory (BiLSTM) to extract effective word embedding and context vectors containing the whole sentence information respectively. Fusion layer is used to provide intent–slot fusion information for SF task. In this way, the relationship between ID and SF task is fully explicitly modeled. This layer takes the result of ID and slot context vectors as input to obtain the fusion information which contains both ID result and slot information. Meanwhile, to further reduce error propagation, we use word-level ID for the ID-SF-Fusion model. Finally, two tasks of ID and SF are realized by joint optimization training.

Findings

We conducted experiments on two public datasets, Airline Travel Information Systems (ATIS) and Snips. The results show that the Intent ACC score and Slot F1 score of ID-SF-Fusion on ATIS and Snips are 98.0 per cent and 95.8 per cent, respectively, and the two indicators on Snips dataset are 98.6 per cent and 96.7 per cent, respectively. These models are superior to slot-gated, SF-ID NetWork, stack-Prop and other models. In addition, ablation experiments were performed to further analyze and discuss the proposed model.

Originality/value

This paper uses word-level intent recognition and introduces intent information into the SF process, which is a significant improvement on both data sets.

目的意图检测(ID)和槽填充(SF)是自然语言理解中的两项重要任务。意图检测的目的是识别一段文本的主要意图。槽填充(SF)的目标是从输入句子中提取对意图重要的信息。然而,现有的方法大多使用句子级的意图识别,这存在错误传播的风险,而且意图识别和 SF 之间的关系没有明确的模型。针对这一问题,本文提出了一种用于智能口语理解的 ID 和 SF 协作模型,称为 ID-SF-Fusion。设计/方法/途径ID-SF-Fusion 使用双向变压器编码器表示法(BERT)和双向长短期记忆法(BiLSTM)分别提取有效的词嵌入和包含整句信息的上下文向量。融合层用于为 SF 任务提供意图-槽融合信息。通过这种方式,ID 和 SF 任务之间的关系被完全明确地建模出来。该层将 ID 和时隙上下文向量的结果作为输入,以获得包含 ID 结果和时隙信息的融合信息。同时,为了进一步减少错误传播,我们在 ID-SF-Fusion 模型中使用了词级 ID。最后,通过联合优化训练实现 ID 和 SF 两项任务。结果表明,ID-SF-Fusion 在 ATIS 和 Snips 数据集上的 Intent ACC 得分和 Slot F1 得分分别为 98.0% 和 95.8%,在 Snips 数据集上的这两项指标分别为 98.6% 和 96.7%。这些模型优于 slot-gated、SF-ID NetWork、stack-Prop 和其他模型。此外,还进行了消融实验,对提出的模型进行了进一步的分析和讨论。原创性/价值本文采用词级意图识别,将意图信息引入 SF 流程,在两个数据集上都有显著改进。
{"title":"ID-SF-Fusion: a cooperative model of intent detection and slot filling for natural language understanding","authors":"Meng Zhu, Xiaolong Xu","doi":"10.1108/dta-03-2023-0088","DOIUrl":"https://doi.org/10.1108/dta-03-2023-0088","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>Intent detection (ID) and slot filling (SF) are two important tasks in natural language understanding. ID is to identify the main intent of a paragraph of text. The goal of SF is to extract the information that is important to the intent from the input sentence. However, most of the existing methods use sentence-level intention recognition, which has the risk of error propagation, and the relationship between intention recognition and SF is not explicitly modeled. Aiming at this problem, this paper proposes a collaborative model of ID and SF for intelligent spoken language understanding called ID-SF-Fusion.</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>ID-SF-Fusion uses Bidirectional Encoder Representation from Transformers (BERT) and Bidirectional Long Short-Term Memory (BiLSTM) to extract effective word embedding and context vectors containing the whole sentence information respectively. Fusion layer is used to provide intent–slot fusion information for SF task. In this way, the relationship between ID and SF task is fully explicitly modeled. This layer takes the result of ID and slot context vectors as input to obtain the fusion information which contains both ID result and slot information. Meanwhile, to further reduce error propagation, we use word-level ID for the ID-SF-Fusion model. Finally, two tasks of ID and SF are realized by joint optimization training.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>We conducted experiments on two public datasets, Airline Travel Information Systems (ATIS) and Snips. The results show that the Intent ACC score and Slot F1 score of ID-SF-Fusion on ATIS and Snips are 98.0 per cent and 95.8 per cent, respectively, and the two indicators on Snips dataset are 98.6 per cent and 96.7 per cent, respectively. These models are superior to slot-gated, SF-ID NetWork, stack-Prop and other models. In addition, ablation experiments were performed to further analyze and discuss the proposed model.</p><!--/ Abstract__block -->\u0000<h3>Originality/value</h3>\u0000<p>This paper uses word-level intent recognition and introduces intent information into the SF process, which is a significant improvement on both data sets.</p><!--/ Abstract__block -->","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139501012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hybrid method for forecasting coal price based on ensemble learning and deep learning with data decomposition and data enhancement 基于数据分解和数据增强的集合学习和深度学习的煤炭价格预测混合方法
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2024-01-18 DOI: 10.1108/dta-07-2023-0377
Jing Tang, Yida Guo, Yilin Han

Purpose

Coal is a critical global energy source, and fluctuations in its price significantly impact related enterprises' profitability. This study aims to develop a robust model for predicting the coal price index to enhance coal purchase strategies for coal-consuming enterprises and provide crucial information for global carbon emission reduction.

Design/methodology/approach

The proposed coal price forecasting system combines data decomposition, semi-supervised feature engineering, ensemble learning and deep learning. It addresses the challenge of merging low-resolution and high-resolution data by adaptively combining both types of data and filling in missing gaps through interpolation for internal missing data and self-supervision for initiate/terminal missing data. The system employs self-supervised learning to complete the filling of complex missing data.

Findings

The ensemble model, which combines long short-term memory, XGBoost and support vector regression, demonstrated the best prediction performance among the tested models. It exhibited superior accuracy and stability across multiple indices in two datasets, namely the Bohai-Rim steam-coal price index and coal daily settlement price.

Originality/value

The proposed coal price forecasting system stands out as it integrates data decomposition, semi-supervised feature engineering, ensemble learning and deep learning. Moreover, the system pioneers the use of self-supervised learning for filling in complex missing data, contributing to its originality and effectiveness.

目的 煤炭是全球重要的能源,其价格的波动会严重影响相关企业的盈利能力。本研究旨在开发一种稳健的煤炭价格指数预测模型,以加强煤炭消费企业的煤炭采购策略,并为全球碳减排提供重要信息。 设计/方法/途径所提出的煤炭价格预测系统结合了数据分解、半监督特征工程、集合学习和深度学习。它通过自适应地合并低分辨率数据和高分辨率数据,并通过对内部缺失数据的插值和对初始/终端缺失数据的自监督来填补缺失空白,从而解决了合并低分辨率数据和高分辨率数据的难题。该系统采用自我监督学习来完成复杂缺失数据的填补。研究结果该集合模型结合了长短期记忆、XGBoost 和支持向量回归,在测试的模型中表现出最佳的预测性能。在两个数据集(即环渤海汽煤价格指数和煤炭日结算价格)中,该模型在多个指数中表现出了卓越的准确性和稳定性。此外,该系统还开创性地使用了自监督学习来填补复杂的缺失数据,从而提高了其原创性和有效性。
{"title":"A hybrid method for forecasting coal price based on ensemble learning and deep learning with data decomposition and data enhancement","authors":"Jing Tang, Yida Guo, Yilin Han","doi":"10.1108/dta-07-2023-0377","DOIUrl":"https://doi.org/10.1108/dta-07-2023-0377","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>Coal is a critical global energy source, and fluctuations in its price significantly impact related enterprises' profitability. This study aims to develop a robust model for predicting the coal price index to enhance coal purchase strategies for coal-consuming enterprises and provide crucial information for global carbon emission reduction.</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>The proposed coal price forecasting system combines data decomposition, semi-supervised feature engineering, ensemble learning and deep learning. It addresses the challenge of merging low-resolution and high-resolution data by adaptively combining both types of data and filling in missing gaps through interpolation for internal missing data and self-supervision for initiate/terminal missing data. The system employs self-supervised learning to complete the filling of complex missing data.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>The ensemble model, which combines long short-term memory, XGBoost and support vector regression, demonstrated the best prediction performance among the tested models. It exhibited superior accuracy and stability across multiple indices in two datasets, namely the Bohai-Rim steam-coal price index and coal daily settlement price.</p><!--/ Abstract__block -->\u0000<h3>Originality/value</h3>\u0000<p>The proposed coal price forecasting system stands out as it integrates data decomposition, semi-supervised feature engineering, ensemble learning and deep learning. Moreover, the system pioneers the use of self-supervised learning for filling in complex missing data, contributing to its originality and effectiveness.</p><!--/ Abstract__block -->","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2024-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139501078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sleep arousal detection for monitoring of sleep disorders using one-dimensional convolutional neural network-based U-Net and bio-signals 利用基于一维卷积神经网络的 U-Net 和生物信号检测睡眠唤醒,监测睡眠障碍
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2024-01-12 DOI: 10.1108/dta-07-2023-0302
Priya Mishra, Aleena Swetapadma

Purpose

Sleep arousal detection is an important factor to monitor the sleep disorder.

Design/methodology/approach

Thus, a unique nth layer one-dimensional (1D) convolutional neural network-based U-Net model for automatic sleep arousal identification has been proposed.

Findings

The proposed method has achieved area under the precision–recall curve performance score of 0.498 and area under the receiver operating characteristics performance score of 0.946.

Originality/value

No other researchers have suggested U-Net-based detection of sleep arousal.

Research limitations/implications

From the experimental results, it has been found that U-Net performs better accuracy as compared to the state-of-the-art methods.

Practical implications

Sleep arousal detection is an important factor to monitor the sleep disorder. Objective of the work is to detect the sleep arousal using different physiological channels of human body.

Social implications

It will help in improving mental health by monitoring a person's sleep.

目的睡眠唤醒检测是监测睡眠障碍的一个重要因素.设计/方法学/方法因此,提出了一种独特的基于第n层一维(1D)卷积神经网络的U-Net模型,用于自动识别睡眠唤醒.研究结果所提出的方法在精确度-召回曲线下的面积性能得分为0.原创性/价值其他研究人员尚未提出基于 U-Net 的睡眠唤醒检测方法。研究局限/意义从实验结果中发现,与最先进的方法相比,U-Net 的准确性更高。这项工作的目标是利用人体的不同生理通道检测睡眠唤醒。社会意义通过监测人的睡眠,有助于改善心理健康。
{"title":"Sleep arousal detection for monitoring of sleep disorders using one-dimensional convolutional neural network-based U-Net and bio-signals","authors":"Priya Mishra, Aleena Swetapadma","doi":"10.1108/dta-07-2023-0302","DOIUrl":"https://doi.org/10.1108/dta-07-2023-0302","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>Sleep arousal detection is an important factor to monitor the sleep disorder.</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>Thus, a unique <em>n</em>th layer one-dimensional (1D) convolutional neural network-based U-Net model for automatic sleep arousal identification has been proposed.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>The proposed method has achieved area under the precision–recall curve performance score of 0.498 and area under the receiver operating characteristics performance score of 0.946.</p><!--/ Abstract__block -->\u0000<h3>Originality/value</h3>\u0000<p>No other researchers have suggested U-Net-based detection of sleep arousal.</p><!--/ Abstract__block -->\u0000<h3>Research limitations/implications</h3>\u0000<p>From the experimental results, it has been found that U-Net performs better accuracy as compared to the state-of-the-art methods.</p><!--/ Abstract__block -->\u0000<h3>Practical implications</h3>\u0000<p>Sleep arousal detection is an important factor to monitor the sleep disorder. Objective of the work is to detect the sleep arousal using different physiological channels of human body.</p><!--/ Abstract__block -->\u0000<h3>Social implications</h3>\u0000<p>It will help in improving mental health by monitoring a person's sleep.</p><!--/ Abstract__block -->","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139458889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the differences between CNNs and vision transformers for COVID-19 diagnosis using CT and chest x-ray mono- and multimodality 利用 CT 和胸部 X 射线单模态和多模态诊断 COVID-19 的 CNN 与视觉变换器之间的差异
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2024-01-10 DOI: 10.1108/dta-01-2023-0005
Sara El-Ateif, Ali Idri, José Luis Fernández-Alemán

Purpose

COVID-19 continues to spread, and cause increasing deaths. Physicians diagnose COVID-19 using not only real-time polymerase chain reaction but also the computed tomography (CT) and chest x-ray (CXR) modalities, depending on the stage of infection. However, with so many patients and so few doctors, it has become difficult to keep abreast of the disease. Deep learning models have been developed in order to assist in this respect, and vision transformers are currently state-of-the-art methods, but most techniques currently focus only on one modality (CXR).

Design/methodology/approach

This work aims to leverage the benefits of both CT and CXR to improve COVID-19 diagnosis. This paper studies the differences between using convolutional MobileNetV2, ViT DeiT and Swin Transformer models when training from scratch and pretraining on the MedNIST medical dataset rather than the ImageNet dataset of natural images. The comparison is made by reporting six performance metrics, the Scott–Knott Effect Size Difference, Wilcoxon statistical test and the Borda Count method. We also use the Grad-CAM algorithm to study the model's interpretability. Finally, the model's robustness is tested by evaluating it on Gaussian noised images.

Findings

Although pretrained MobileNetV2 was the best model in terms of performance, the best model in terms of performance, interpretability, and robustness to noise is the trained from scratch Swin Transformer using the CXR (accuracy = 93.21 per cent) and CT (accuracy = 94.14 per cent) modalities.

Originality/value

Models compared are pretrained on MedNIST and leverage both the CT and CXR modalities.

目的 COVID-19 仍在继续传播,并导致越来越多的人死亡。医生在诊断 COVID-19 时,不仅要使用实时聚合酶链反应,还要根据感染阶段使用计算机断层扫描(CT)和胸部 X 光检查(CXR)。然而,由于患者多、医生少,要及时了解病情变得十分困难。为了在这方面提供帮助,人们开发了深度学习模型,视觉转换器也是目前最先进的方法,但大多数技术目前只关注一种模式(CXR)。本文研究了使用卷积 MobileNetV2、ViT DeiT 和 Swin Transformer 模型从头开始训练与在 MedNIST 医学数据集而非自然图像的 ImageNet 数据集上进行预训练之间的差异。通过报告六项性能指标、斯科特-克诺特效应大小差、Wilcoxon 统计检验和 Borda 计数法进行比较。我们还使用 Grad-CAM 算法来研究模型的可解释性。研究结果虽然经过预训练的 MobileNetV2 是性能最佳的模型,但在性能、可解释性和对噪声的鲁棒性方面,使用 CXR(准确率 = 93.21%)和 CT(准确率 = 94.14%)模式从头开始训练的 Swin Transformer 才是最佳模型。
{"title":"On the differences between CNNs and vision transformers for COVID-19 diagnosis using CT and chest x-ray mono- and multimodality","authors":"Sara El-Ateif, Ali Idri, José Luis Fernández-Alemán","doi":"10.1108/dta-01-2023-0005","DOIUrl":"https://doi.org/10.1108/dta-01-2023-0005","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>COVID-19 continues to spread, and cause increasing deaths. Physicians diagnose COVID-19 using not only real-time polymerase chain reaction but also the computed tomography (CT) and chest x-ray (CXR) modalities, depending on the stage of infection. However, with so many patients and so few doctors, it has become difficult to keep abreast of the disease. Deep learning models have been developed in order to assist in this respect, and vision transformers are currently state-of-the-art methods, but most techniques currently focus only on one modality (CXR).</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>This work aims to leverage the benefits of both CT and CXR to improve COVID-19 diagnosis. This paper studies the differences between using convolutional MobileNetV2, ViT DeiT and Swin Transformer models when training from scratch and pretraining on the MedNIST medical dataset rather than the ImageNet dataset of natural images. The comparison is made by reporting six performance metrics, the Scott–Knott Effect Size Difference, Wilcoxon statistical test and the Borda Count method. We also use the Grad-CAM algorithm to study the model's interpretability. Finally, the model's robustness is tested by evaluating it on Gaussian noised images.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>Although pretrained MobileNetV2 was the best model in terms of performance, the best model in terms of performance, interpretability, and robustness to noise is the trained from scratch Swin Transformer using the CXR (accuracy = 93.21 per cent) and CT (accuracy = 94.14 per cent) modalities.</p><!--/ Abstract__block -->\u0000<h3>Originality/value</h3>\u0000<p>Models compared are pretrained on MedNIST and leverage both the CT and CXR modalities.</p><!--/ Abstract__block -->","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139423565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding the relationship between normative records of appeals and government hotline order dispatching: a data analysis method 了解申诉的规范性记录与政府热线派单之间的关系:一种数据分析方法
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2024-01-04 DOI: 10.1108/dta-02-2023-0029
Zicheng Zhang

Purpose

Advanced big data analysis and machine learning methods are concurrently used to unleash the value of the data generated by government hotline and help devise intelligent applications including automated process management, standard construction and more accurate dispatched orders to build high-quality government service platforms as more widely data-driven methods are in the process.

Design/methodology/approach

In this study, based on the influence of the record specifications of texts related to work orders generated by the government hotline, machine learning tools are implemented and compared to optimize classify dispatching tasks by performing exploratory studies on the hotline work order text, including linguistics analysis of text feature processing, new word discovery, text clustering and text classification.

Findings

The complexity of the content of the work order is reduced by applying more standardized writing specifications based on combining text grammar numerical features. So, order dispatch success prediction accuracy rate reaches 89.6 per cent after running the LSTM model.

Originality/value

The proposed method can help improve the current dispatching processes run by the government hotline, better guide staff to standardize the writing format of work orders, improve the accuracy of order dispatching and provide innovative support to the current mechanism.

目的 随着数据驱动方法的广泛应用,同时采用先进的大数据分析和机器学习方法,释放政务热线产生的数据价值,帮助设计包括自动化流程管理、标准建设和更精准派单在内的智能应用,打造高质量的政务服务平台。设计/方法/途径在本研究中,基于政府热线生成的工单相关文本记录规范的影响,通过对热线工单文本进行探索性研究,包括文本特征处理的语言学分析、新词发现、文本聚类和文本分类等,实施机器学习工具并进行比较,以优化派单任务的分类。原创性/价值所提出的方法有助于改善目前政府热线的派单流程,更好地引导工作人员规范工单书写格式,提高派单准确率,为现行机制提供创新支持。
{"title":"Understanding the relationship between normative records of appeals and government hotline order dispatching: a data analysis method","authors":"Zicheng Zhang","doi":"10.1108/dta-02-2023-0029","DOIUrl":"https://doi.org/10.1108/dta-02-2023-0029","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>Advanced big data analysis and machine learning methods are concurrently used to unleash the value of the data generated by government hotline and help devise intelligent applications including automated process management, standard construction and more accurate dispatched orders to build high-quality government service platforms as more widely data-driven methods are in the process.</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>In this study, based on the influence of the record specifications of texts related to work orders generated by the government hotline, machine learning tools are implemented and compared to optimize classify dispatching tasks by performing exploratory studies on the hotline work order text, including linguistics analysis of text feature processing, new word discovery, text clustering and text classification.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>The complexity of the content of the work order is reduced by applying more standardized writing specifications based on combining text grammar numerical features. So, order dispatch success prediction accuracy rate reaches 89.6 per cent after running the LSTM model.</p><!--/ Abstract__block -->\u0000<h3>Originality/value</h3>\u0000<p>The proposed method can help improve the current dispatching processes run by the government hotline, better guide staff to standardize the writing format of work orders, improve the accuracy of order dispatching and provide innovative support to the current mechanism.</p><!--/ Abstract__block -->","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2024-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139103103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimized aspect and self-attention aware LSTM for target-based semantic analysis (OAS-LSTM-TSA) 用于基于目标的语义分析的优化方面和自我注意感知 LSTM(OAS-LSTM-TSA)
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2023-12-29 DOI: 10.1108/dta-10-2022-0408
B. Vasavi, P. Dileep, Ulligaddala Srrinivasarao

Purpose

Aspect-based sentiment analysis (ASA) is a task of sentiment analysis that requires predicting aspect sentiment polarity for a given sentence. Many traditional techniques use graph-based mechanisms, which reduce prediction accuracy and introduce large amounts of noise. The other problem with graph-based mechanisms is that for some context words, the feelings change depending on the aspect, and therefore it is impossible to draw conclusions on their own. ASA is challenging because a given sentence can reveal complicated feelings about multiple aspects.

Design/methodology/approach

This research proposed an optimized attention-based DL model known as optimized aspect and self-attention aware long short-term memory for target-based semantic analysis (OAS-LSTM-TSA). The proposed model goes through three phases: preprocessing, aspect extraction and classification. Aspect extraction is done using a double-layered convolutional neural network (DL-CNN). The optimized aspect and self-attention embedded LSTM (OAS-LSTM) is used to classify aspect sentiment into three classes: positive, neutral and negative.

Findings

To detect and classify sentiment polarity of the aspect using the optimized aspect and self-attention embedded LSTM (OAS-LSTM) model. The results of the proposed method revealed that it achieves a high accuracy of 95.3 per cent for the restaurant dataset and 96.7 per cent for the laptop dataset.

Originality/value

The novelty of the research work is the addition of two effective attention layers in the network model, loss function reduction and accuracy enhancement, using a recent efficient optimization algorithm. The loss function in OAS-LSTM is minimized using the adaptive pelican optimization algorithm, thus increasing the accuracy rate. The performance of the proposed method is validated on four real-time datasets, Rest14, Lap14, Rest15 and Rest16, for various performance metrics.

目的基于方面的情感分析(ASA)是情感分析的一项任务,需要预测给定句子的方面情感极性。许多传统技术使用基于图的机制,这会降低预测准确性并引入大量噪声。基于图的机制的另一个问题是,对于某些上下文词语来说,情感会随着方面的变化而变化,因此无法单独得出结论。ASA 具有挑战性,因为一个给定的句子可能揭示出多个方面的复杂感受。这项研究提出了一种基于注意力的优化 DL 模型,称为优化方面和自我注意力感知长短时记忆目标语义分析(OAS-LSTM-TSA)。该模型分为三个阶段:预处理、方面提取和分类。方面提取是通过双层卷积神经网络(DL-CNN)完成的。使用优化的方面和自我关注嵌入式 LSTM(OAS-LSTM)将方面情感分为三类:正面、中性和负面。研究工作的新颖之处在于利用最新的高效优化算法,在网络模型中增加了两个有效的关注层,减少了损失函数并提高了准确度。利用自适应鹈鹕优化算法将 OAS-LSTM 中的损失函数最小化,从而提高了准确率。在 Rest14、Lap14、Rest15 和 Rest16 四个实时数据集上,针对各种性能指标验证了所提方法的性能。
{"title":"Optimized aspect and self-attention aware LSTM for target-based semantic analysis (OAS-LSTM-TSA)","authors":"B. Vasavi, P. Dileep, Ulligaddala Srrinivasarao","doi":"10.1108/dta-10-2022-0408","DOIUrl":"https://doi.org/10.1108/dta-10-2022-0408","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>Aspect-based sentiment analysis (ASA) is a task of sentiment analysis that requires predicting aspect sentiment polarity for a given sentence. Many traditional techniques use graph-based mechanisms, which reduce prediction accuracy and introduce large amounts of noise. The other problem with graph-based mechanisms is that for some context words, the feelings change depending on the aspect, and therefore it is impossible to draw conclusions on their own. ASA is challenging because a given sentence can reveal complicated feelings about multiple aspects.</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>This research proposed an optimized attention-based DL model known as optimized aspect and self-attention aware long short-term memory for target-based semantic analysis (OAS-LSTM-TSA). The proposed model goes through three phases: preprocessing, aspect extraction and classification. Aspect extraction is done using a double-layered convolutional neural network (DL-CNN). The optimized aspect and self-attention embedded LSTM (OAS-LSTM) is used to classify aspect sentiment into three classes: positive, neutral and negative.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>To detect and classify sentiment polarity of the aspect using the optimized aspect and self-attention embedded LSTM (OAS-LSTM) model. The results of the proposed method revealed that it achieves a high accuracy of 95.3 per cent for the restaurant dataset and 96.7 per cent for the laptop dataset.</p><!--/ Abstract__block -->\u0000<h3>Originality/value</h3>\u0000<p>The novelty of the research work is the addition of two effective attention layers in the network model, loss function reduction and accuracy enhancement, using a recent efficient optimization algorithm. The loss function in OAS-LSTM is minimized using the adaptive pelican optimization algorithm, thus increasing the accuracy rate. The performance of the proposed method is validated on four real-time datasets, Rest14, Lap14, Rest15 and Rest16, for various performance metrics.</p><!--/ Abstract__block -->","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139068275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep understanding of radiology reports: leveraging dynamic convolution in chest X-ray images 对放射学报告的深刻理解:利用胸部x射线图像的动态卷积
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2023-11-29 DOI: 10.1108/dta-07-2023-0307
Tarun Jaiswal, Manju Pandey, Priyanka Tripathi

Purpose

The purpose of this study is to investigate and demonstrate the advancements achieved in the field of chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Typical convolutional neural networks (CNNs) are unable to capture both local and global contextual information effectively and apply a uniform operation to all pixels in an image. To address this, we propose an innovative approach that integrates a dynamic convolution operation at the encoder stage, improving image encoding quality and disease detection. In addition, a decoder based on the gated recurrent unit (GRU) is used for language modeling, and an attention network is incorporated to enhance consistency. This novel combination allows for improved feature extraction, mimicking the expertise of radiologists by selectively focusing on important areas and producing coherent captions with valuable clinical information.

Design/methodology/approach

In this study, we have presented a new report generation approach that utilizes dynamic convolution applied Resnet-101 (DyCNN) as an encoder (Verelst and Tuytelaars, 2019) and GRU as a decoder (Dey and Salemt, 2017; Pan et al., 2020), along with an attention network (see Figure 1). This integration innovatively extends the capabilities of image encoding and sequential caption generation, representing a shift from conventional CNN architectures. With its ability to dynamically adapt receptive fields, the DyCNN excels at capturing features of varying scales within the CXR images. This dynamic adaptability significantly enhances the granularity of feature extraction, enabling precise representation of localized abnormalities and structural intricacies. By incorporating this flexibility into the encoding process, our model can distil meaningful and contextually rich features from the radiographic data. While the attention mechanism enables the model to selectively focus on different regions of the image during caption generation. The attention mechanism enhances the report generation process by allowing the model to assign different importance weights to different regions of the image, mimicking human perception. In parallel, the GRU-based decoder adds a critical dimension to the process by ensuring a smooth, sequential generation of captions.

Findings

The findings of this study highlight the significant advancements achieved in chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Experiments conducted using the IU-Chest X-ray datasets showed that the proposed model outperformed other state-of-the-art approaches. The model achieved notable scores, including a BLEU_1 score of 0.591, a BLEU_2 score of 0.347, a BLEU_3 score of 0.277 and a BLEU_4 score of 0.155. These results highlight the efficiency and efficacy of the model in producing pre

本研究的目的是研究和展示动态卷积编码器-解码器网络(DyCNN)在胸部x线图像字幕领域取得的进展。典型的卷积神经网络(cnn)无法有效地捕获局部和全局上下文信息,并对图像中的所有像素应用统一的操作。为了解决这个问题,我们提出了一种创新的方法,在编码器阶段集成了动态卷积操作,提高了图像编码质量和疾病检测。此外,采用基于门控循环单元(GRU)的解码器进行语言建模,并引入注意网络增强一致性。这种新颖的组合允许改进特征提取,模仿放射科医生的专业知识,选择性地关注重要领域,并产生具有有价值的临床信息的连贯字幕。在本研究中,我们提出了一种新的报告生成方法,该方法利用动态卷积将Resnet-101 (DyCNN)作为编码器(Verelst和Tuytelaars, 2019)和GRU作为解码器(Dey和Salemt, 2017;Pan等人,2020),以及一个注意力网络(见图1)。这种集成创新地扩展了图像编码和顺序标题生成的能力,代表了传统CNN架构的转变。凭借其动态适应接受域的能力,DyCNN擅长捕捉CXR图像中不同尺度的特征。这种动态适应性显著提高了特征提取的粒度,能够精确地表示局部异常和结构复杂性。通过将这种灵活性纳入编码过程,我们的模型可以从射线照相数据中提取有意义和上下文丰富的特征。而注意机制使模型能够在标题生成过程中选择性地关注图像的不同区域。注意机制通过允许模型为图像的不同区域分配不同的重要权重来增强报告生成过程,模仿人类的感知。与此同时,基于gru的解码器通过确保字幕的流畅、顺序生成,为该过程增加了一个关键维度。本研究的发现强调了通过使用动态卷积编码器-解码器网络(DyCNN)在胸部x线图像字幕方面取得的重大进展。使用u -胸部x射线数据集进行的实验表明,所提出的模型优于其他最先进的方法。模型得分显著,其中BLEU_1得分为0.591,BLEU_2得分为0.347,BLEU_3得分为0.277,BLEU_4得分为0.155。这些结果突出了该模型在生成精确的放射学报告,增强图像解释和临床决策方面的效率和功效。独创性/价值这项工作是同类工作中的第一个,它使用DyCNN作为编码器从CXR图像中提取特征。此外,利用GRU作为语言建模的解码器,并将注意机制纳入模型体系结构。
{"title":"Deep understanding of radiology reports: leveraging dynamic convolution in chest X-ray images","authors":"Tarun Jaiswal, Manju Pandey, Priyanka Tripathi","doi":"10.1108/dta-07-2023-0307","DOIUrl":"https://doi.org/10.1108/dta-07-2023-0307","url":null,"abstract":"<h3>Purpose</h3>\u0000<p>The purpose of this study is to investigate and demonstrate the advancements achieved in the field of chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Typical convolutional neural networks (CNNs) are unable to capture both local and global contextual information effectively and apply a uniform operation to all pixels in an image. To address this, we propose an innovative approach that integrates a dynamic convolution operation at the encoder stage, improving image encoding quality and disease detection. In addition, a decoder based on the gated recurrent unit (GRU) is used for language modeling, and an attention network is incorporated to enhance consistency. This novel combination allows for improved feature extraction, mimicking the expertise of radiologists by selectively focusing on important areas and producing coherent captions with valuable clinical information.</p><!--/ Abstract__block -->\u0000<h3>Design/methodology/approach</h3>\u0000<p>In this study, we have presented a new report generation approach that utilizes dynamic convolution applied Resnet-101 (DyCNN) as an encoder (Verelst and Tuytelaars, 2019) and GRU as a decoder (Dey and Salemt, 2017; Pan et al., 2020), along with an attention network (see Figure 1). This integration innovatively extends the capabilities of image encoding and sequential caption generation, representing a shift from conventional CNN architectures. With its ability to dynamically adapt receptive fields, the DyCNN excels at capturing features of varying scales within the CXR images. This dynamic adaptability significantly enhances the granularity of feature extraction, enabling precise representation of localized abnormalities and structural intricacies. By incorporating this flexibility into the encoding process, our model can distil meaningful and contextually rich features from the radiographic data. While the attention mechanism enables the model to selectively focus on different regions of the image during caption generation. The attention mechanism enhances the report generation process by allowing the model to assign different importance weights to different regions of the image, mimicking human perception. In parallel, the GRU-based decoder adds a critical dimension to the process by ensuring a smooth, sequential generation of captions.</p><!--/ Abstract__block -->\u0000<h3>Findings</h3>\u0000<p>The findings of this study highlight the significant advancements achieved in chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Experiments conducted using the IU-Chest X-ray datasets showed that the proposed model outperformed other state-of-the-art approaches. The model achieved notable scores, including a BLEU_1 score of 0.591, a BLEU_2 score of 0.347, a BLEU_3 score of 0.277 and a BLEU_4 score of 0.155. These results highlight the efficiency and efficacy of the model in producing pre","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2023-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138495379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing audiovisual data for understanding user's emotion in human−computer interaction environment 在人机交互环境中,通过分析视听数据来理解用户的情感
4区 计算机科学 Q1 Social Sciences Pub Date : 2023-11-01 DOI: 10.1108/dta-08-2023-0414
Juan Yang, Zhenkun Li, Xu Du
Purpose Although numerous signal modalities are available for emotion recognition, audio and visual modalities are the most common and predominant forms for human beings to express their emotional states in daily communication. Therefore, how to achieve automatic and accurate audiovisual emotion recognition is significantly important for developing engaging and empathetic human–computer interaction environment. However, two major challenges exist in the field of audiovisual emotion recognition: (1) how to effectively capture representations of each single modality and eliminate redundant features and (2) how to efficiently integrate information from these two modalities to generate discriminative representations. Design/methodology/approach A novel key-frame extraction-based attention fusion network (KE-AFN) is proposed for audiovisual emotion recognition. KE-AFN attempts to integrate key-frame extraction with multimodal interaction and fusion to enhance audiovisual representations and reduce redundant computation, filling the research gaps of existing approaches. Specifically, the local maximum–based content analysis is designed to extract key-frames from videos for the purpose of eliminating data redundancy. Two modules, including “Multi-head Attention-based Intra-modality Interaction Module” and “Multi-head Attention-based Cross-modality Interaction Module”, are proposed to mine and capture intra- and cross-modality interactions for further reducing data redundancy and producing more powerful multimodal representations. Findings Extensive experiments on two benchmark datasets (i.e. RAVDESS and CMU-MOSEI) demonstrate the effectiveness and rationality of KE-AFN. Specifically, (1) KE-AFN is superior to state-of-the-art baselines for audiovisual emotion recognition. (2) Exploring the supplementary and complementary information of different modalities can provide more emotional clues for better emotion recognition. (3) The proposed key-frame extraction strategy can enhance the performance by more than 2.79 per cent on accuracy. (4) Both exploring intra- and cross-modality interactions and employing attention-based audiovisual fusion can lead to better prediction performance. Originality/value The proposed KE-AFN can support the development of engaging and empathetic human–computer interaction environment.
虽然情绪识别的信号方式多种多样,但听觉和视觉是人类在日常交流中表达情绪状态的最常见和最主要的形式。因此,如何实现自动准确的视听情感识别,对于开发引人入胜、共情的人机交互环境具有重要意义。然而,在视听情感识别领域存在两个主要挑战:(1)如何有效地捕获每个单一模态的表征并消除冗余特征;(2)如何有效地整合这两个模态的信息以生成判别表征。提出了一种新的基于关键帧提取的注意力融合网络(KE-AFN)用于视听情感识别。KE-AFN试图将关键帧提取与多模态交互融合结合起来,增强视听表征,减少冗余计算,填补了现有方法的研究空白。具体而言,基于局部最大值的内容分析旨在从视频中提取关键帧,以消除数据冗余。提出了“基于多头注意力的模态内交互模块”和“基于多头注意力的跨模态交互模块”两个模块来挖掘和捕获模态内交互和跨模态交互,以进一步减少数据冗余并产生更强大的多模态表示。在两个基准数据集(即RAVDESS和CMU-MOSEI)上的大量实验证明了KE-AFN的有效性和合理性。具体而言,(1)KE-AFN在视听情感识别方面优于最先进的基线。(2)探索不同模态的互补性信息可以为更好的情绪识别提供更多的情绪线索。(3)所提出的关键帧提取策略的准确率提高了2.79%以上。(4)探索模态内和模态间的相互作用以及采用基于注意力的视听融合都可以提高预测效果。提出的KE-AFN可以支持参与和共情人机交互环境的发展。
{"title":"Analyzing audiovisual data for understanding user's emotion in human−computer interaction environment","authors":"Juan Yang, Zhenkun Li, Xu Du","doi":"10.1108/dta-08-2023-0414","DOIUrl":"https://doi.org/10.1108/dta-08-2023-0414","url":null,"abstract":"Purpose Although numerous signal modalities are available for emotion recognition, audio and visual modalities are the most common and predominant forms for human beings to express their emotional states in daily communication. Therefore, how to achieve automatic and accurate audiovisual emotion recognition is significantly important for developing engaging and empathetic human–computer interaction environment. However, two major challenges exist in the field of audiovisual emotion recognition: (1) how to effectively capture representations of each single modality and eliminate redundant features and (2) how to efficiently integrate information from these two modalities to generate discriminative representations. Design/methodology/approach A novel key-frame extraction-based attention fusion network (KE-AFN) is proposed for audiovisual emotion recognition. KE-AFN attempts to integrate key-frame extraction with multimodal interaction and fusion to enhance audiovisual representations and reduce redundant computation, filling the research gaps of existing approaches. Specifically, the local maximum–based content analysis is designed to extract key-frames from videos for the purpose of eliminating data redundancy. Two modules, including “Multi-head Attention-based Intra-modality Interaction Module” and “Multi-head Attention-based Cross-modality Interaction Module”, are proposed to mine and capture intra- and cross-modality interactions for further reducing data redundancy and producing more powerful multimodal representations. Findings Extensive experiments on two benchmark datasets (i.e. RAVDESS and CMU-MOSEI) demonstrate the effectiveness and rationality of KE-AFN. Specifically, (1) KE-AFN is superior to state-of-the-art baselines for audiovisual emotion recognition. (2) Exploring the supplementary and complementary information of different modalities can provide more emotional clues for better emotion recognition. (3) The proposed key-frame extraction strategy can enhance the performance by more than 2.79 per cent on accuracy. (4) Both exploring intra- and cross-modality interactions and employing attention-based audiovisual fusion can lead to better prediction performance. Originality/value The proposed KE-AFN can support the development of engaging and empathetic human–computer interaction environment.","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135321701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AsCDPR: a novel framework for ratings and personalized preference hotel recommendation using cross-domain and aspect-based features AsCDPR:使用跨领域和基于方面的功能进行评级和个性化偏好酒店推荐的新框架
4区 计算机科学 Q1 Social Sciences Pub Date : 2023-09-20 DOI: 10.1108/dta-03-2023-0101
Hei-Chia Wang, Army Justitia, Ching-Wen Wang
Purpose The explosion of data due to the sophistication of information and communication technology makes it simple for prospective tourists to learn about previous hotel guests' experiences. They prioritize the rating score when selecting a hotel. However, rating scores are less reliable for suggesting a personalized preference for each aspect, especially when they are in a limited number. This study aims to recommend ratings and personalized preference hotels using cross-domain and aspect-based features. Design/methodology/approach We propose an aspect-based cross-domain personalized recommendation (AsCDPR), a novel framework for rating prediction and personalized customer preference recommendations. We incorporate a cross-domain personalized approach and aspect-based features of items from the review text. We extracted aspect-based feature vectors from two domains using bidirectional long short-term memory and then mapped them by a multilayer perceptron (MLP). The cross-domain recommendation module trains MLP to analyze sentiment and predict item ratings and the polarities of the aspect based on user preferences. Findings Expanded by its synonyms, aspect-based features significantly improve the performance of sentiment analysis on accuracy and the F1-score matrix. With relatively low mean absolute error and root mean square error values, AsCDPR outperforms matrix factorization, collaborative matrix factorization, EMCDPR and Personalized transfer of user preferences for cross-domain recommendation. These values are 1.3657 and 1.6682, respectively. Research limitation/implications This study assists users in recommending hotels based on their priority preferences. Users do not need to read other people's reviews to capture the key aspects of items. This model could enhance system reliability in the hospitality industry by providing personalized recommendations. Originality/value This study introduces a new approach that embeds aspect-based features of items in a cross-domain personalized recommendation. AsCDPR predicts ratings and provides recommendations based on priority aspects of each user's preferences.
由于信息和通信技术的成熟,数据的爆炸式增长使得未来的游客可以很容易地了解以前酒店客人的体验。他们在选择酒店时优先考虑评分。然而,评级分数在建议对每个方面的个性化偏好方面不太可靠,特别是当它们的数量有限时。本研究旨在使用跨领域和基于方面的功能来推荐评级和个性化偏好的酒店。我们提出了一种基于方面的跨领域个性化推荐(AsCDPR),这是一种新的评级预测和个性化客户偏好推荐框架。我们结合了跨领域的个性化方法和审查文本中项目的基于方面的特征。我们利用双向长短期记忆从两个领域中提取基于方面的特征向量,然后用多层感知器(MLP)对它们进行映射。跨领域推荐模块训练MLP分析情感,并根据用户偏好预测项目评分和方面的极性。通过同义词的扩展,基于方面的特征显著提高了情感分析在准确性和f1得分矩阵上的表现。在跨领域推荐中,AsCDPR的平均绝对误差和均方根误差值相对较低,优于矩阵分解、协同矩阵分解、EMCDPR和用户偏好的个性化转移。这些值分别为1.3657和1.6682。本研究帮助用户根据他们的优先偏好推荐酒店。用户不需要阅读其他人的评论来获取项目的关键方面。该模型可以通过提供个性化的推荐来提高酒店业的系统可靠性。本研究提出了一种新的方法,即在跨领域个性化推荐中嵌入基于方面的商品特征。AsCDPR预测评分并根据每个用户偏好的优先级提供推荐。
{"title":"AsCDPR: a novel framework for ratings and personalized preference hotel recommendation using cross-domain and aspect-based features","authors":"Hei-Chia Wang, Army Justitia, Ching-Wen Wang","doi":"10.1108/dta-03-2023-0101","DOIUrl":"https://doi.org/10.1108/dta-03-2023-0101","url":null,"abstract":"Purpose The explosion of data due to the sophistication of information and communication technology makes it simple for prospective tourists to learn about previous hotel guests' experiences. They prioritize the rating score when selecting a hotel. However, rating scores are less reliable for suggesting a personalized preference for each aspect, especially when they are in a limited number. This study aims to recommend ratings and personalized preference hotels using cross-domain and aspect-based features. Design/methodology/approach We propose an aspect-based cross-domain personalized recommendation (AsCDPR), a novel framework for rating prediction and personalized customer preference recommendations. We incorporate a cross-domain personalized approach and aspect-based features of items from the review text. We extracted aspect-based feature vectors from two domains using bidirectional long short-term memory and then mapped them by a multilayer perceptron (MLP). The cross-domain recommendation module trains MLP to analyze sentiment and predict item ratings and the polarities of the aspect based on user preferences. Findings Expanded by its synonyms, aspect-based features significantly improve the performance of sentiment analysis on accuracy and the F1-score matrix. With relatively low mean absolute error and root mean square error values, AsCDPR outperforms matrix factorization, collaborative matrix factorization, EMCDPR and Personalized transfer of user preferences for cross-domain recommendation. These values are 1.3657 and 1.6682, respectively. Research limitation/implications This study assists users in recommending hotels based on their priority preferences. Users do not need to read other people's reviews to capture the key aspects of items. This model could enhance system reliability in the hospitality industry by providing personalized recommendations. Originality/value This study introduces a new approach that embeds aspect-based features of items in a cross-domain personalized recommendation. AsCDPR predicts ratings and provides recommendations based on priority aspects of each user's preferences.","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136308711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CA-CD: context-aware clickbait detection using new Chinese clickbait dataset with transfer learning method CA-CD:基于迁移学习方法的新中文点击诱饵数据集的上下文感知点击诱饵检测
IF 1.6 4区 计算机科学 Q1 Social Sciences Pub Date : 2023-08-29 DOI: 10.1108/dta-03-2023-0072
Hei-Chia Wang, Martinus Maslim, Hung-Yu Liu
PurposeA clickbait is a deceptive headline designed to boost ad revenue without presenting closely relevant content. There are numerous negative repercussions of clickbait, such as causing viewers to feel tricked and unhappy, causing long-term confusion, and even attracting cyber criminals. Automatic detection algorithms for clickbait have been developed to address this issue. The fact that there is only one semantic representation for the same term and a limited dataset in Chinese is a need for the existing technologies for detecting clickbait. This study aims to solve the limitations of automated clickbait detection in the Chinese dataset.Design/methodology/approachThis study combines both to train the model to capture the probable relationship between clickbait news headlines and news content. In addition, part-of-speech elements are used to generate the most appropriate semantic representation for clickbait detection, improving clickbait detection performance.FindingsThis research successfully compiled a dataset containing up to 20,896 Chinese clickbait news articles. This collection contains news headlines, articles, categories and supplementary metadata. The suggested context-aware clickbait detection (CA-CD) model outperforms existing clickbait detection approaches on many criteria, demonstrating the proposed strategy's efficacy.Originality/valueThe originality of this study resides in the newly compiled Chinese clickbait dataset and contextual semantic representation-based clickbait detection approach employing transfer learning. This method can modify the semantic representation of each word based on context and assist the model in more precisely interpreting the original meaning of news articles.
目的点击诱饵是一个欺骗性的标题,旨在提高广告收入,而不提供密切相关的内容。点击诱饵有很多负面影响,比如让观众感到被骗和不开心,造成长期混乱,甚至吸引网络犯罪分子。点击诱饵的自动检测算法已经被开发出来解决这个问题。同一术语只有一个语义表示,中文数据集有限,这是现有检测点击诱饵技术的需要。本研究旨在解决中国数据集中自动点击诱饵检测的局限性。设计/方法论/方法本研究将两者结合起来训练模型,以捕捉点击诱饵新闻标题和新闻内容之间的可能关系。此外,使用词性元素生成最适合点击诱饵检测的语义表示,提高了点击诱饵检测性能。发现这项研究成功地汇编了一个数据集,其中包含多达20896篇中国点击诱饵新闻文章。此集合包含新闻标题、文章、类别和补充元数据。所提出的上下文感知点击诱饵检测(CA-CD)模型在许多标准上优于现有的点击诱饵检测方法,证明了所提出的策略的有效性。独创性/价值本研究的独创性在于新汇编的中文点击诱饵数据集和采用迁移学习的基于上下文语义表示的点击诱饵检测方法。这种方法可以根据上下文修改每个单词的语义表示,并帮助模型更准确地解释新闻文章的原意。
{"title":"CA-CD: context-aware clickbait detection using new Chinese clickbait dataset with transfer learning method","authors":"Hei-Chia Wang, Martinus Maslim, Hung-Yu Liu","doi":"10.1108/dta-03-2023-0072","DOIUrl":"https://doi.org/10.1108/dta-03-2023-0072","url":null,"abstract":"PurposeA clickbait is a deceptive headline designed to boost ad revenue without presenting closely relevant content. There are numerous negative repercussions of clickbait, such as causing viewers to feel tricked and unhappy, causing long-term confusion, and even attracting cyber criminals. Automatic detection algorithms for clickbait have been developed to address this issue. The fact that there is only one semantic representation for the same term and a limited dataset in Chinese is a need for the existing technologies for detecting clickbait. This study aims to solve the limitations of automated clickbait detection in the Chinese dataset.Design/methodology/approachThis study combines both to train the model to capture the probable relationship between clickbait news headlines and news content. In addition, part-of-speech elements are used to generate the most appropriate semantic representation for clickbait detection, improving clickbait detection performance.FindingsThis research successfully compiled a dataset containing up to 20,896 Chinese clickbait news articles. This collection contains news headlines, articles, categories and supplementary metadata. The suggested context-aware clickbait detection (CA-CD) model outperforms existing clickbait detection approaches on many criteria, demonstrating the proposed strategy's efficacy.Originality/valueThe originality of this study resides in the newly compiled Chinese clickbait dataset and contextual semantic representation-based clickbait detection approach employing transfer learning. This method can modify the semantic representation of each word based on context and assist the model in more precisely interpreting the original meaning of news articles.","PeriodicalId":56156,"journal":{"name":"Data Technologies and Applications","volume":null,"pages":null},"PeriodicalIF":1.6,"publicationDate":"2023-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49612441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Data Technologies and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1