Shumaila Khan, Iqbal Qasim, Wahab Khan, Khursheed Aurangzeb, Javed Ali Khan, Muhammad Shahid Anwar
Sarcasm detection is challenging in natural language processing (NLP) due to its implicit nature, particularly in low-resource languages. Despite limited linguistic resources, researchers have focused on detecting sarcasm on social media platforms, leading to the development of specialized algorithms and models tailored for Urdu text. Researchers have significantly improved sarcasm detection accuracy by analysing patterns and linguistic cues unique to the language, thereby advancing NLP capabilities in low-resource languages and facilitating better communication within diverse online communities. This work introduces UrduSarcasmNet, a novel architecture using cascaded group multi-head attention, which is an innovative deep-learning approach that employs cascaded group multi-head attention techniques to enhance effectiveness. By employing a series of attention heads in a cascading manner, our model captures both local and global contexts, facilitating a more comprehensive understanding of the text. Adding a group attention mechanism enables simultaneous consideration of various sub-topics within the content, thereby enriching the model's effectiveness. The proposed UrduSarcasmNet approach is validated with the Urdu-sarcastic-tweets-dataset (UST) dataset, which has been curated for this purpose. Our experimental results on the UST dataset show that the proposed UrduSarcasmNet framework outperforms the simple-attention mechanism and other state-of-the-art models. This research significantly enhances natural language processing (NLP) and provides valuable insights for improving sarcasm recognition tools in low-resource languages like Urdu.
{"title":"A novel transformer attention-based approach for sarcasm detection","authors":"Shumaila Khan, Iqbal Qasim, Wahab Khan, Khursheed Aurangzeb, Javed Ali Khan, Muhammad Shahid Anwar","doi":"10.1111/exsy.13686","DOIUrl":"10.1111/exsy.13686","url":null,"abstract":"<p>Sarcasm detection is challenging in natural language processing (NLP) due to its implicit nature, particularly in low-resource languages. Despite limited linguistic resources, researchers have focused on detecting sarcasm on social media platforms, leading to the development of specialized algorithms and models tailored for Urdu text. Researchers have significantly improved sarcasm detection accuracy by analysing patterns and linguistic cues unique to the language, thereby advancing NLP capabilities in low-resource languages and facilitating better communication within diverse online communities. This work introduces UrduSarcasmNet, a novel architecture using cascaded group multi-head attention, which is an innovative deep-learning approach that employs cascaded group multi-head attention techniques to enhance effectiveness. By employing a series of attention heads in a cascading manner, our model captures both local and global contexts, facilitating a more comprehensive understanding of the text. Adding a group attention mechanism enables simultaneous consideration of various sub-topics within the content, thereby enriching the model's effectiveness. The proposed UrduSarcasmNet approach is validated with the Urdu-sarcastic-tweets-dataset (UST) dataset, which has been curated for this purpose. Our experimental results on the UST dataset show that the proposed UrduSarcasmNet framework outperforms the simple-attention mechanism and other state-of-the-art models. This research significantly enhances natural language processing (NLP) and provides valuable insights for improving sarcasm recognition tools in low-resource languages like Urdu.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"42 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ezaz Mohammed Al‐dahasi, Rama Khaled Alsheikh, Fakhri Alam Khan, Gwanggil Jeon
The rapid advancement of the Internet and digital payments has transformed the landscape of financial transactions, leading to both technological progress and an alarming rise in cybercrime. This study addresses the critical issue of financial fraud detection in the era of digital payments, focusing on enhancing operational risk frameworks to mitigate the increasing threats. The objective is to improve the predictive performance of fraud detection systems using machine learning techniques. The methodology involves a comprehensive data preprocessing and model creation process, including one‐hot encoding, feature selection, sampling, standardization, and tokenization. Six machine learning models are employed for fraud detection, and their hyperparameters are optimized. Evaluation metrics such as accuracy, precision, recall, and F1‐score are used to assess model performance. Results reveal that XGBoost and Random Forest outperform other models, achieving a balance between false positives and false negatives. The study meets the requirements for fraud detection systems, ensuring accuracy, scalability, adaptability, and explainability. This paper provides valuable insights into the efficacy of machine learning models for financial fraud detection and emphasizes the importance of striking a balance between false positives and false negatives.
互联网和数字支付的快速发展改变了金融交易的格局,既带来了技术进步,也导致网络犯罪的惊人增长。本研究探讨了数字支付时代金融欺诈检测的关键问题,重点是加强操作风险框架,以减轻日益增长的威胁。目的是利用机器学习技术提高欺诈检测系统的预测性能。该方法涉及全面的数据预处理和模型创建过程,包括单次编码、特征选择、采样、标准化和标记化。欺诈检测采用了六个机器学习模型,并对其超参数进行了优化。准确率、精确度、召回率和 F1 分数等评价指标用于评估模型性能。结果显示,XGBoost 和随机森林的表现优于其他模型,在误报和误报之间取得了平衡。这项研究符合欺诈检测系统的要求,确保了准确性、可扩展性、适应性和可解释性。本文就机器学习模型在金融欺诈检测中的功效提供了宝贵的见解,并强调了在误报和误报之间取得平衡的重要性。
{"title":"Optimizing fraud detection in financial transactions with machine learning and imbalance mitigation","authors":"Ezaz Mohammed Al‐dahasi, Rama Khaled Alsheikh, Fakhri Alam Khan, Gwanggil Jeon","doi":"10.1111/exsy.13682","DOIUrl":"https://doi.org/10.1111/exsy.13682","url":null,"abstract":"The rapid advancement of the Internet and digital payments has transformed the landscape of financial transactions, leading to both technological progress and an alarming rise in cybercrime. This study addresses the critical issue of financial fraud detection in the era of digital payments, focusing on enhancing operational risk frameworks to mitigate the increasing threats. The objective is to improve the predictive performance of fraud detection systems using machine learning techniques. The methodology involves a comprehensive data preprocessing and model creation process, including one‐hot encoding, feature selection, sampling, standardization, and tokenization. Six machine learning models are employed for fraud detection, and their hyperparameters are optimized. Evaluation metrics such as accuracy, precision, recall, and <jats:italic>F</jats:italic>1‐score are used to assess model performance. Results reveal that XGBoost and Random Forest outperform other models, achieving a balance between false positives and false negatives. The study meets the requirements for fraud detection systems, ensuring accuracy, scalability, adaptability, and explainability. This paper provides valuable insights into the efficacy of machine learning models for financial fraud detection and emphasizes the importance of striking a balance between false positives and false negatives.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"71 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
As digital transformation accelerates, data generated in a convergence information and communication technology (ICT) environment must be secured. This data includes confidential information such as personal and financial information, so attackers spread malware in convergence ICT environments to steal this information. To protect convergence ICT environments from diverse cyber threats, deep learning models have been utilized for malware detection. However, accurately detecting rapidly generated variants and obfuscated malware is challenging. This study proposes a three‐tier malware detection (TMaD) scheme that utilizes a cloud‐fog‐edge collaborative architecture to analyse multi‐view features of executable files and detect malware. TMaD performs signature‐based malware detection at the edge device tier, then sends executables detected as unknown or benign to the fog tier. The fog tier conducts static analysis on non‐obfuscated executables and those transferred from the previous tier to detect variant malware. Subsequently, TMaD sends executables detected as benign in the fog tier to the cloud tier, where dynamic analysis is performed on obfuscated executables and those detected as benign to identify obfuscated malware. An evaluation of TMaD's detection performance resulted in an accuracy of 94.78%, a recall of 0.9794, a precision of 0.9535, and an f1‐score of 0.9663. This performance demonstrates that TMaD, by analysing executables across several tiers and minimizing false negatives, exhibits superior detection performance compared to existing malware detection models.
{"title":"TMaD: Three‐tier malware detection using multi‐view feature for secure convergence ICT environments","authors":"Jueun Jeon, Byeonghui Jeong, Seungyeon Baek, Young‐Sik Jeong","doi":"10.1111/exsy.13684","DOIUrl":"https://doi.org/10.1111/exsy.13684","url":null,"abstract":"As digital transformation accelerates, data generated in a convergence information and communication technology (ICT) environment must be secured. This data includes confidential information such as personal and financial information, so attackers spread malware in convergence ICT environments to steal this information. To protect convergence ICT environments from diverse cyber threats, deep learning models have been utilized for malware detection. However, accurately detecting rapidly generated variants and obfuscated malware is challenging. This study proposes a three‐tier malware detection (TMaD) scheme that utilizes a cloud‐fog‐edge collaborative architecture to analyse multi‐view features of executable files and detect malware. TMaD performs signature‐based malware detection at the edge device tier, then sends executables detected as unknown or benign to the fog tier. The fog tier conducts static analysis on non‐obfuscated executables and those transferred from the previous tier to detect variant malware. Subsequently, TMaD sends executables detected as benign in the fog tier to the cloud tier, where dynamic analysis is performed on obfuscated executables and those detected as benign to identify obfuscated malware. An evaluation of TMaD's detection performance resulted in an accuracy of 94.78%, a recall of 0.9794, a precision of 0.9535, and an f1‐score of 0.9663. This performance demonstrates that TMaD, by analysing executables across several tiers and minimizing false negatives, exhibits superior detection performance compared to existing malware detection models.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"12 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141742191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Claudia García-Navarro, Manuel Pulido-Martos, Cristina Pérez-Lozano
Engagement has been defined as an attitude toward work, as a positive, satisfying, work-related state of mind characterized by high levels of vigour, dedication, and absorption. Both its definition and its assessment have been controversial; however, new methods for its assessment, including artificial intelligence (AI), have been introduced in recent years. Therefore, this research aims to determine the state of the art of AI in the study of engagement. To this end, we conducted a systematic review in accordance with PRISMA to analyse the publications to date on the use of AI for the analysis of engagement. The search, carried out in six databases, was filtered, and 15 papers were finally analysed. The results show that AI has been used mainly to assess and predict engagement levels, as well as to understand the relationships between engagement and other variables. The most commonly used AI techniques are machine learning (ML) and natural language processing (NLP), and all publications use structured and unstructured data, mainly from self-report instruments, social networks, and datasets. The accuracy of the models varies from 22% to 87%, and its main benefit has been to help both managers and HR staff understand employee engagement, although it has also contributed to research. Most of the articles have been published since 2015, and the geography has been global, with publications predominantly in India and the US. In conclusion, this study highlights the state of the art in AI for the study of engagement and concludes that the number of publications is increasing, indicating that this is possibly a new field or area of research in which important advances can be made in the study of engagement through new and novel techniques.
{"title":"The study of engagement at work from the artificial intelligence perspective: A systematic review","authors":"Claudia García-Navarro, Manuel Pulido-Martos, Cristina Pérez-Lozano","doi":"10.1111/exsy.13673","DOIUrl":"10.1111/exsy.13673","url":null,"abstract":"<p>Engagement has been defined as an attitude toward work, as a positive, satisfying, work-related state of mind characterized by high levels of vigour, dedication, and absorption. Both its definition and its assessment have been controversial; however, new methods for its assessment, including artificial intelligence (AI), have been introduced in recent years. Therefore, this research aims to determine the state of the art of AI in the study of engagement. To this end, we conducted a systematic review in accordance with PRISMA to analyse the publications to date on the use of AI for the analysis of engagement. The search, carried out in six databases, was filtered, and 15 papers were finally analysed. The results show that AI has been used mainly to assess and predict engagement levels, as well as to understand the relationships between engagement and other variables. The most commonly used AI techniques are machine learning (ML) and natural language processing (NLP), and all publications use structured and unstructured data, mainly from self-report instruments, social networks, and datasets. The accuracy of the models varies from 22% to 87%, and its main benefit has been to help both managers and HR staff understand employee engagement, although it has also contributed to research. Most of the articles have been published since 2015, and the geography has been global, with publications predominantly in India and the US. In conclusion, this study highlights the state of the art in AI for the study of engagement and concludes that the number of publications is increasing, indicating that this is possibly a new field or area of research in which important advances can be made in the study of engagement through new and novel techniques.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/exsy.13673","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141742192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Federated learning (FL), celebrated for its privacy‐preserving features, has been revealed by recent studies to harbour security vulnerabilities that jeopardize client privacy, particularly through data reconstruction attacks that enable adversaries to recover original client data. This study introduces a client‐level handwriting forgery attack method for FL based on generative adversarial networks (GANs), which reveals security vulnerabilities existing in FL systems. It should be stressed that this research is purely for academic purposes, aiming to raise concerns about privacy protection and data security, and does not encourage illegal activities. Our novel methodology assumes an adversarial scenario wherein adversaries intercept a fraction of parameter updates via victim clients’ wireless communication channels, then use this information to train GAN for data recovery. Finally, the purpose of handwriting imitation is achieved. To rigorously assess and validate our methodology, experiments were conducted using a bespoke Chinese digit dataset, facilitating in‐depth analysis and robust verification of results. Our experimental findings demonstrated enhanced data recovery effectiveness, a client‐level attack and greater versatility compared to prior art. Notably, our method maintained high attack performance even with a streamlined GAN design, yielding increased precision and significantly faster execution times compared to standard methods. Specifically, our experimental numerical results revealed a substantial boost in reconstruction accuracy by 16.7%, coupled with a 51.9% decrease in computational time compared to the latest similar techniques. Furthermore, tests on a simplified version of our GAN exhibited an average 10% enhancement in accuracy, alongside a remarkable 70% reduction in time consumption. By surmounting the limitations of previous work, this study fills crucial gaps and affirms the effectiveness of our approach in achieving high‐accuracy client‐level data reconstruction within the FL context, thereby stimulating further exploration into FL security measures.
联合学习(Federated Learning,FL)因其保护隐私的特性而备受赞誉,但最近的研究却揭示了它存在着危害客户端隐私的安全漏洞,特别是通过数据重构攻击,使对手能够恢复原始客户端数据。本研究介绍了一种基于生成式对抗网络(GANs)的 FL 客户端级笔迹伪造攻击方法,揭示了 FL 系统中存在的安全漏洞。需要强调的是,本研究纯粹出于学术目的,旨在引起人们对隐私保护和数据安全的关注,并不鼓励非法活动。我们的新方法假设了一种对抗场景,即对抗者通过受害者客户端的无线通信信道截获一部分参数更新,然后利用这些信息训练 GAN 进行数据恢复。最后,笔迹模仿的目的就达到了。为了严格评估和验证我们的方法,我们使用定制的中文数字数据集进行了实验,以便对结果进行深入分析和稳健验证。我们的实验结果表明,与现有技术相比,我们的方法提高了数据恢复的有效性、客户端级别的攻击和更大的通用性。值得注意的是,即使采用精简的 GAN 设计,我们的方法仍能保持较高的攻击性能,与标准方法相比,精度更高,执行时间更短。具体来说,我们的数值实验结果表明,与最新的类似技术相比,重建精度大幅提高了 16.7%,计算时间减少了 51.9%。此外,对我们的 GAN 简化版进行的测试表明,精度平均提高了 10%,同时耗时显著减少了 70%。通过克服以往工作的局限性,本研究填补了重要空白,并肯定了我们的方法在 FL 环境下实现高精度客户端级数据重建的有效性,从而激发了对 FL 安全措施的进一步探索。
{"title":"A generative adversarial network‐based client‐level handwriting forgery attack in federated learning scenario","authors":"Lei Shi, Han Wu, Xu Ding, Hao Xu, Sinan Pan","doi":"10.1111/exsy.13676","DOIUrl":"https://doi.org/10.1111/exsy.13676","url":null,"abstract":"Federated learning (FL), celebrated for its privacy‐preserving features, has been revealed by recent studies to harbour security vulnerabilities that jeopardize client privacy, particularly through data reconstruction attacks that enable adversaries to recover original client data. This study introduces a client‐level handwriting forgery attack method for FL based on generative adversarial networks (GANs), which reveals security vulnerabilities existing in FL systems. It should be stressed that this research is purely for academic purposes, aiming to raise concerns about privacy protection and data security, and does not encourage illegal activities. Our novel methodology assumes an adversarial scenario wherein adversaries intercept a fraction of parameter updates via victim clients’ wireless communication channels, then use this information to train GAN for data recovery. Finally, the purpose of handwriting imitation is achieved. To rigorously assess and validate our methodology, experiments were conducted using a bespoke Chinese digit dataset, facilitating in‐depth analysis and robust verification of results. Our experimental findings demonstrated enhanced data recovery effectiveness, a client‐level attack and greater versatility compared to prior art. Notably, our method maintained high attack performance even with a streamlined GAN design, yielding increased precision and significantly faster execution times compared to standard methods. Specifically, our experimental numerical results revealed a substantial boost in reconstruction accuracy by 16.7%, coupled with a 51.9% decrease in computational time compared to the latest similar techniques. Furthermore, tests on a simplified version of our GAN exhibited an average 10% enhancement in accuracy, alongside a remarkable 70% reduction in time consumption. By surmounting the limitations of previous work, this study fills crucial gaps and affirms the effectiveness of our approach in achieving high‐accuracy client‐level data reconstruction within the FL context, thereby stimulating further exploration into FL security measures.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"20 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141613269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
With the progressive augmentation of parameters in multimodal models, to optimize computational efficiency, some studies have adopted the approach of fine-tuning the unimodal pre-training model to achieve multimodal fusion tasks. However, these methods tend to rely solely on simplistic or singular fusion strategies, thereby neglecting more flexible fusion approaches. Moreover, existing methods prioritize the integration of modality features containing highly semantic information, often overlooking the influence of fusing low-level features on the outcomes. Therefore, this study introduces an innovative approach named multilevel feature fusion guided by prompts (MFF-GP), a multimodal dynamic fusion framework. It guides the dynamic neural network by prompt vectors to dynamically select the suitable fusion network for each hierarchical feature of the unimodal pre-training model. This method improves the interactions between multiple modalities and promotes a more efficient fusion of features across them. Extensive experiments on the UPMC Food 101, SNLI-VE and MM-IMDB datasets demonstrate that with only a few trainable parameters, MFF-GP achieves significant accuracy improvements compared to a newly designed PMF based on fine-tuning—specifically, an accuracy improvement of 2.15% on the UPMC Food 101 dataset and 0.82% on the SNLI-VE dataset. Further study of the results reveals that increasing the diversity of interactions between distinct modalities is critical and delivers significant performance improvements. Furthermore, for certain multimodal tasks, focusing on the low-level features is beneficial for modality integration. Our implementation is available at: https://github.com/whq2024/MFF-GP.
{"title":"Multimodal dynamic fusion framework: Multilevel feature fusion guided by prompts","authors":"Lei Pan, Huan-Qing Wu","doi":"10.1111/exsy.13668","DOIUrl":"10.1111/exsy.13668","url":null,"abstract":"<p>With the progressive augmentation of parameters in multimodal models, to optimize computational efficiency, some studies have adopted the approach of fine-tuning the unimodal pre-training model to achieve multimodal fusion tasks. However, these methods tend to rely solely on simplistic or singular fusion strategies, thereby neglecting more flexible fusion approaches. Moreover, existing methods prioritize the integration of modality features containing highly semantic information, often overlooking the influence of fusing low-level features on the outcomes. Therefore, this study introduces an innovative approach named multilevel feature fusion guided by prompts (MFF-GP), a multimodal dynamic fusion framework. It guides the dynamic neural network by prompt vectors to dynamically select the suitable fusion network for each hierarchical feature of the unimodal pre-training model. This method improves the interactions between multiple modalities and promotes a more efficient fusion of features across them. Extensive experiments on the UPMC Food 101, SNLI-VE and MM-IMDB datasets demonstrate that with only a few trainable parameters, MFF-GP achieves significant accuracy improvements compared to a newly designed PMF based on fine-tuning—specifically, an accuracy improvement of 2.15% on the UPMC Food 101 dataset and 0.82% on the SNLI-VE dataset. Further study of the results reveals that increasing the diversity of interactions between distinct modalities is critical and delivers significant performance improvements. Furthermore, for certain multimodal tasks, focusing on the low-level features is beneficial for modality integration. Our implementation is available at: https://github.com/whq2024/MFF-GP.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141613581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The current prevalence of conspiracy theories on the internet is a significant issue, tackled by many computational approaches. However, these approaches fail to recognize the relevance of distinguishing between texts which contain a conspiracy theory and texts which are simply critical and oppose mainstream narratives. Furthermore, little attention is usually paid to the role of inter-group conflict in oppositional narratives. We contribute by proposing a novel topic-agnostic annotation scheme that differentiates between conspiracies and critical texts, and that defines span-level categories of inter-group conflict. We also contribute with the multilingual XAI-DisInfodemics corpus (English and Spanish), which contains a high-quality annotation of Telegram messages related to COVID-19 (5000 messages per language). We also demonstrate the feasibility of an NLP-based automatization by performing a range of experiments that yield strong baseline solutions. Finally, we perform an analysis which demonstrates that the promotion of intergroup conflict and the presence of violence and anger are key aspects to distinguish between the two types of oppositional narratives, that is, conspiracy versus critical.
{"title":"What distinguishes conspiracy from critical narratives? A computational analysis of oppositional discourse","authors":"Damir Korenčić, Berta Chulvi, Xavier Bonet Casals, Alejandro Toselli, Mariona Taulé, Paolo Rosso","doi":"10.1111/exsy.13671","DOIUrl":"10.1111/exsy.13671","url":null,"abstract":"<p>The current prevalence of conspiracy theories on the internet is a significant issue, tackled by many computational approaches. However, these approaches fail to recognize the relevance of distinguishing between texts which contain a conspiracy theory and texts which are simply critical and oppose mainstream narratives. Furthermore, little attention is usually paid to the role of inter-group conflict in oppositional narratives. We contribute by proposing a novel topic-agnostic annotation scheme that differentiates between conspiracies and critical texts, and that defines span-level categories of inter-group conflict. We also contribute with the multilingual XAI-DisInfodemics corpus (English and Spanish), which contains a high-quality annotation of Telegram messages related to COVID-19 (5000 messages per language). We also demonstrate the feasibility of an NLP-based automatization by performing a range of experiments that yield strong baseline solutions. Finally, we perform an analysis which demonstrates that the promotion of intergroup conflict and the presence of violence and anger are key aspects to distinguish between the two types of oppositional narratives, that is, conspiracy versus critical.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141586682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The forecasting of steel futures prices is important for the steel futures market, even for the steel industry. We propose a decomposition ensemble model that incorporates the Ensemble Empirical Mode Decomposition (EEMD), Long Short-Term Memory (LSTM), Support Vector Regression (SVR), and Back Propagation (BP) neural network to forecast steel futures prices. The forecasting procedures are as follows: (1) The price data are initially decomposed into several relatively independent Intrinsic Mode Functions (IMFs) and a residue using EEMD. (2) The IMFs are then reconstructed as components representing short-term, medium-term, and long-term frequencies via fine-to-coarse. (3) LSTM, SVR, and BP neural network are utilized to forecast the short-term, medium-term, and long-term reconstructed components, respectively. (4) The prediction results for each component are simply added to the final prediction results. The accuracy of the proposed model is compared with several benchmark models by experiments and evaluated by some prediction evaluation indexes. The experimental results show that our model outperforms other models in terms of forecast accuracy, confirming its strong predictive capabilities. This study provides some suggestions for investment and decision making by participants in the steel futures market. It may promote the smooth operation of the steel futures market and shed some light on the operation of the steel industry.
钢材期货价格预测对于钢材期货市场乃至钢铁行业都非常重要。我们提出了一种分解集合模型,该模型融合了集合经验模式分解(EEMD)、长短期记忆(LSTM)、支持向量回归(SVR)和反向传播(BP)神经网络,用于预测钢材期货价格。预测程序如下(1) 首先使用 EEMD 将价格数据分解为几个相对独立的本征模式函数(IMF)和一个残差。(2) 然后通过从细到粗的方法将 IMF 重构为代表短期、中期和长期频率的成分。(3) 利用 LSTM、SVR 和 BP 神经网络分别预测重建的短期、中期和长期分量。(4) 将各分量的预测结果简单相加,得出最终预测结果。通过实验将所提出模型的准确性与几个基准模型进行比较,并通过一些预测评价指标进行评估。实验结果表明,我们的模型在预测准确率方面优于其他模型,证实了其强大的预测能力。本研究为钢铁期货市场参与者的投资和决策提供了一些建议。它可以促进钢材期货市场的平稳运行,并对钢铁行业的运行起到一定的启示作用。
{"title":"An EEMD-LSTM, SVR, and BP decomposition ensemble model for steel future prices forecasting","authors":"Sen Wu, Wei Wang, Yanan Song, Shuaiqi Liu","doi":"10.1111/exsy.13672","DOIUrl":"10.1111/exsy.13672","url":null,"abstract":"<p>The forecasting of steel futures prices is important for the steel futures market, even for the steel industry. We propose a decomposition ensemble model that incorporates the Ensemble Empirical Mode Decomposition (EEMD), Long Short-Term Memory (LSTM), Support Vector Regression (SVR), and Back Propagation (BP) neural network to forecast steel futures prices. The forecasting procedures are as follows: (1) The price data are initially decomposed into several relatively independent Intrinsic Mode Functions (IMFs) and a residue using EEMD. (2) The IMFs are then reconstructed as components representing short-term, medium-term, and long-term frequencies via fine-to-coarse. (3) LSTM, SVR, and BP neural network are utilized to forecast the short-term, medium-term, and long-term reconstructed components, respectively. (4) The prediction results for each component are simply added to the final prediction results. The accuracy of the proposed model is compared with several benchmark models by experiments and evaluated by some prediction evaluation indexes. The experimental results show that our model outperforms other models in terms of forecast accuracy, confirming its strong predictive capabilities. This study provides some suggestions for investment and decision making by participants in the steel futures market. It may promote the smooth operation of the steel futures market and shed some light on the operation of the steel industry.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141570209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The people in the world rely on social media for gathering news, and it is mainly because of the development of technology. The approaches employed in natural language processing are still deficient in judgement factors, and these techniques frequently rely upon political or social circumstances. Numerous low-level communities in the area are curious after experiencing the negative effects caused by the spread of false information in different sectors. Low-resource languages are still distracted, because these techniques are extensively employed in the English language. This work aims to provide an analysis of regional language fake news and develop a referral system with advanced techniques to identify fake news in Hindi and Tamil. This proposed model includes (a) Regional Language Text Collection; (b) Text preprocessing; (c) Feature Extraction; (d) Weighted Stacked Feature Fusion; and (e) Fake News Detection. The text data is collected from the standard datasets. The collected text data is preprocessed and given into the feature extraction, which is done by using bidirectional encoder representations from transformers (BERT), transformer networks, and seq2seq network for extracting the three sets of language text features. These extracted feature sets are inserted into the weighted stacked feature fusion model, where the three sets of extracted features are integrated with the optimized weights that are acquired through the enhanced osprey optimization algorithm (EOOA). Finally, these resultant features are given to multi-scale atrous convolution-based one-dimensional convolutional neural network with dilated long short-term memory (MACNN-DLSTM) for detecting the fake news. Throughout the result analysis, the experimentation is conducted based on the standard Tamil and Hindi datasets. Moreover, the developed model shows 92% for Hindi datasets and 96% for Tamil datasets which shows effective performance regarding accuracy measures. The experimental analysis is carried out by comparing with the conventional algorithms and detection techniques to showcase the efficiency of the developed regional language-based fake news detection model.
{"title":"Adaptive weighted feature fusion for multiscale atrous convolution-based 1DCNN with dilated LSTM-aided fake news detection using regional language text information","authors":"V Rathinapriya, J. Kalaivani","doi":"10.1111/exsy.13665","DOIUrl":"10.1111/exsy.13665","url":null,"abstract":"<p>The people in the world rely on social media for gathering news, and it is mainly because of the development of technology. The approaches employed in natural language processing are still deficient in judgement factors, and these techniques frequently rely upon political or social circumstances. Numerous low-level communities in the area are curious after experiencing the negative effects caused by the spread of false information in different sectors. Low-resource languages are still distracted, because these techniques are extensively employed in the English language. This work aims to provide an analysis of regional language fake news and develop a referral system with advanced techniques to identify fake news in Hindi and Tamil. This proposed model includes (a) Regional Language Text Collection; (b) Text preprocessing; (c) Feature Extraction; (d) Weighted Stacked Feature Fusion; and (e) Fake News Detection. The text data is collected from the standard datasets. The collected text data is preprocessed and given into the feature extraction, which is done by using bidirectional encoder representations from transformers (BERT), transformer networks, and seq2seq network for extracting the three sets of language text features. These extracted feature sets are inserted into the weighted stacked feature fusion model, where the three sets of extracted features are integrated with the optimized weights that are acquired through the enhanced osprey optimization algorithm (EOOA). Finally, these resultant features are given to multi-scale atrous convolution-based one-dimensional convolutional neural network with dilated long short-term memory (MACNN-DLSTM) for detecting the fake news. Throughout the result analysis, the experimentation is conducted based on the standard Tamil and Hindi datasets. Moreover, the developed model shows 92% for Hindi datasets and 96% for Tamil datasets which shows effective performance regarding accuracy measures. The experimental analysis is carried out by comparing with the conventional algorithms and detection techniques to showcase the efficiency of the developed regional language-based fake news detection model.</p>","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"41 11","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141551442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Eduardo M. Rodrigues, Yassine Baghoussi, João Mendes‐Moreira
Deep learning models are widely used in multivariate time series forecasting, yet, they have high computational costs. One way to reduce this cost is by reducing data dimensionality, which involves removing unimportant or low importance information with the proper method. This work presents a study on an explainability feature selection framework composed of four methods (IMV‐LSTM Tensor, LIME‐LSTM, Average SHAP‐LSTM, and Instance SHAP‐LSTM) aimed at using the LSTM black‐box model complexity to its favour, with the end goal of improving the error metrics and reducing the computational cost on a forecast task. To test the framework, three datasets with a total of 101 multivariate time series were used, with the explainability methods outperforming the baseline methods in most of the data, be it in error metrics or computation time for the LSTM model training.
{"title":"KDBI special issue: Explainability feature selection framework application for LSTM multivariate time‐series forecast self optimization","authors":"Eduardo M. Rodrigues, Yassine Baghoussi, João Mendes‐Moreira","doi":"10.1111/exsy.13674","DOIUrl":"https://doi.org/10.1111/exsy.13674","url":null,"abstract":"Deep learning models are widely used in multivariate time series forecasting, yet, they have high computational costs. One way to reduce this cost is by reducing data dimensionality, which involves removing unimportant or low importance information with the proper method. This work presents a study on an explainability feature selection framework composed of four methods (IMV‐LSTM Tensor, LIME‐LSTM, Average SHAP‐LSTM, and Instance SHAP‐LSTM) aimed at using the LSTM black‐box model complexity to its favour, with the end goal of improving the error metrics and reducing the computational cost on a forecast task. To test the framework, three datasets with a total of 101 multivariate time series were used, with the explainability methods outperforming the baseline methods in most of the data, be it in error metrics or computation time for the LSTM model training.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":"15 1","pages":""},"PeriodicalIF":3.3,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141553055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}