首页 > 最新文献

Expert Systems最新文献

英文 中文
Human activity recognition: A comprehensive review 人类活动识别:全面回顾
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-27 DOI: 10.1111/exsy.13680
Harmandeep Kaur, Veenu Rani, Munish Kumar
Human Activity Recognition (HAR) is a highly promising research area meant to automatically identify and interpret human behaviour using data received from sensors in various contexts. The potential uses of HAR are many, among them health care, sports coaching or monitoring the elderly or disabled. Nonetheless, there are numerous hurdles to be circumvented for HAR's precision and usefulness to be improved. One of the challenges is that there is no uniformity in data collection and annotation making it difficult to compare findings among different studies. Furthermore, more comprehensive datasets are necessary so as to include a wider range of human activities in different contexts while complex activities, which consist of multiple sub‐activities, are still a challenge for recognition systems. Researchers have proposed new frontiers such as multi‐modal sensor data fusion and deep learning approaches for enhancing HAR accuracy while addressing these issues. Also, we are seeing more non‐traditional applications such as robotics and virtual reality/augmented world going forward with their use cases of HAR. This article offers an extensive review on the recent advances in HAR and highlights the major challenges facing this field as well as future opportunities for further researches.
人类活动识别(HAR)是一个极具发展前景的研究领域,其目的是在各种情况下利用传感器接收到的数据自动识别和解释人类行为。人类活动识别(HAR)的潜在用途很多,其中包括医疗保健、体育指导或监测老年人或残疾人。然而,要提高 HAR 的精确度和实用性,还有许多障碍需要克服。挑战之一是数据收集和注释不统一,因此很难比较不同研究的结果。此外,还需要更全面的数据集,以包括不同环境中更广泛的人类活动,而由多个子活动组成的复杂活动对识别系统来说仍是一个挑战。研究人员提出了一些新的前沿技术,如多模态传感器数据融合和深度学习方法,以便在解决这些问题的同时提高 HAR 的准确性。此外,我们还看到更多的非传统应用,如机器人和虚拟现实/增强世界,都在使用 HAR。本文对 HAR 的最新进展进行了广泛综述,并重点介绍了该领域面临的主要挑战以及未来进一步研究的机遇。
{"title":"Human activity recognition: A comprehensive review","authors":"Harmandeep Kaur, Veenu Rani, Munish Kumar","doi":"10.1111/exsy.13680","DOIUrl":"https://doi.org/10.1111/exsy.13680","url":null,"abstract":"Human Activity Recognition (HAR) is a highly promising research area meant to automatically identify and interpret human behaviour using data received from sensors in various contexts. The potential uses of HAR are many, among them health care, sports coaching or monitoring the elderly or disabled. Nonetheless, there are numerous hurdles to be circumvented for HAR's precision and usefulness to be improved. One of the challenges is that there is no uniformity in data collection and annotation making it difficult to compare findings among different studies. Furthermore, more comprehensive datasets are necessary so as to include a wider range of human activities in different contexts while complex activities, which consist of multiple sub‐activities, are still a challenge for recognition systems. Researchers have proposed new frontiers such as multi‐modal sensor data fusion and deep learning approaches for enhancing HAR accuracy while addressing these issues. Also, we are seeing more non‐traditional applications such as robotics and virtual reality/augmented world going forward with their use cases of HAR. This article offers an extensive review on the recent advances in HAR and highlights the major challenges facing this field as well as future opportunities for further researches.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783662","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sampling approaches to reduce very frequent seasonal time series 减少非常频繁的季节性时间序列的抽样方法
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-26 DOI: 10.1111/exsy.13690
Afonso Baldo, Paulo J. S. Ferreira, João Mendes‐Moreira
With technological advancements, much data is being captured by sensors, smartphones, wearable devices, and so forth. These vast datasets are stored in data centres and utilized to forge data‐driven models for the condition monitoring of infrastructures and systems through future data mining tasks. However, these datasets often surpass the processing capabilities of traditional information systems and methodologies due to their significant size. Additionally, not all samples within these datasets contribute valuable information during the model training phase, leading to inefficiencies. The processing and training of Machine Learning algorithms become time‐consuming, and storing all the data demands excessive space, contributing to the Big Data challenge. In this paper, we propose two novel techniques to reduce large time‐series datasets into more compact versions without undermining the predictive performance of the resulting models. These methods also aim to decrease the time required for training the models and the storage space needed for the condensed datasets. We evaluated our techniques on five public datasets, employing three Machine Learning algorithms: Holt‐Winters, SARIMA, and LSTM. The outcomes indicate that for most of the datasets examined, our techniques maintain, and in several instances enhance, the forecasting accuracy of the models. Moreover, we significantly reduced the time required to train the Machine Learning algorithms employed.
随着技术的进步,传感器、智能手机、可穿戴设备等正在采集大量数据。这些庞大的数据集被存储在数据中心,并通过未来的数据挖掘任务用于建立数据驱动模型,以监测基础设施和系统的状况。然而,由于这些数据集规模庞大,往往超出了传统信息系统和方法的处理能力。此外,在模型训练阶段,这些数据集中并非所有样本都能提供有价值的信息,从而导致效率低下。机器学习算法的处理和训练变得非常耗时,而存储所有数据又需要过大的空间,这就加剧了大数据的挑战。在本文中,我们提出了两种新技术,在不影响模型预测性能的前提下,将大型时间序列数据集缩减为更紧凑的版本。这些方法还旨在减少训练模型所需的时间和压缩数据集所需的存储空间。我们采用 Holt-Winters、SARIMA 和 LSTM 三种机器学习算法,在五个公共数据集上对我们的技术进行了评估。结果表明,对于大多数受检数据集,我们的技术都保持了模型的预测准确性,并在一些情况下提高了预测准确性。此外,我们还大大缩短了训练所采用的机器学习算法所需的时间。
{"title":"Sampling approaches to reduce very frequent seasonal time series","authors":"Afonso Baldo, Paulo J. S. Ferreira, João Mendes‐Moreira","doi":"10.1111/exsy.13690","DOIUrl":"https://doi.org/10.1111/exsy.13690","url":null,"abstract":"With technological advancements, much data is being captured by sensors, smartphones, wearable devices, and so forth. These vast datasets are stored in data centres and utilized to forge data‐driven models for the condition monitoring of infrastructures and systems through future data mining tasks. However, these datasets often surpass the processing capabilities of traditional information systems and methodologies due to their significant size. Additionally, not all samples within these datasets contribute valuable information during the model training phase, leading to inefficiencies. The processing and training of Machine Learning algorithms become time‐consuming, and storing all the data demands excessive space, contributing to the Big Data challenge. In this paper, we propose two novel techniques to reduce large time‐series datasets into more compact versions without undermining the predictive performance of the resulting models. These methods also aim to decrease the time required for training the models and the storage space needed for the condensed datasets. We evaluated our techniques on five public datasets, employing three Machine Learning algorithms: Holt‐Winters, SARIMA, and LSTM. The outcomes indicate that for most of the datasets examined, our techniques maintain, and in several instances enhance, the forecasting accuracy of the models. Moreover, we significantly reduced the time required to train the Machine Learning algorithms employed.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting early depression in WZT drawing image based on deep learning 基于深度学习的 WZT 图画图像早期抑郁预测
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-25 DOI: 10.1111/exsy.13675
Kyung‐yeul Kim, Young‐bo Yang, Mi‐ra Kim, Jihie Kim, Ji Su Park
When stress causes negative behaviours to emerge in our daily lives, it is important to intervene quickly and appropriately to control the negative problem behaviours. Questionnaires, a common method of information gathering, have the disadvantage that it is difficult to get the exact information needed due to defensive or insincere responses from subjects. As an alternative to these drawbacks, projective testing using pictures can provide the necessary information more accurately than questionnaires because the subject responds subconsciously and the direct experience expressed through pictures can be more accurate than questionnaires. Analysing hand‐drawn image data with the Wartegg Zeichen Test (WZT) is not easy. In this study, we used deep learning to analyse image data represented as pictures through WZT to predict early depression. We analyse the data of 54 people who were judged as early depression and 54 people without depression, and increase the number of people without depression to 100 and 500, and aim to study in unbalanced data. We use CNN and CNN‐SVM to analyse the drawing images of WZT's initial depression with deep learning and predict the outcome of depression. The results show that the initial depression is predicted with 92%–98% accuracy on the image data directly drawn by WZT. This is the first study to automatically analyse and predict early depression in WZT based on hand‐drawn image data using deep learning models. The extraction of features from WZT images by deep learning analysis is expected to create more research opportunities through the convergence of psychotherapy and Information and Communication Technology (ICT) technology, and is expected to have high growth potential.
当压力导致我们在日常生活中出现负面行为时,必须迅速采取适当的干预措施,以控制负面问题行为。问卷调查是一种常用的信息收集方法,其缺点是很难获得所需的准确信息,因为受试者会做出防卫性或不真诚的回答。与问卷调查相比,图片投射测试能更准确地提供所需的信息,因为受试者会下意识地做出反应,而且通过图片表达的直接经验比问卷调查更准确。使用 Wartegg Zeichen 测试(WZT)分析手绘图像数据并非易事。在本研究中,我们使用深度学习来分析通过 WZT 表示为图片的图像数据,从而预测早期抑郁症。我们分析了 54 名被判定为早期抑郁症的人和 54 名未患抑郁症的人的数据,并将未患抑郁症的人数增加到 100 人和 500 人,力求在非平衡数据中进行研究。我们使用 CNN 和 CNN-SVM,通过深度学习分析 WZT 初期抑郁的绘画图像,并预测抑郁的结果。结果表明,对 WZT 直接绘制的图像数据进行初始抑郁预测的准确率为 92%-98%。这是首个基于手绘图像数据,利用深度学习模型自动分析和预测 WZT 早期抑郁的研究。通过深度学习分析从WZT图像中提取特征,有望通过心理治疗与信息通信技术(ICT)技术的融合创造更多的研究机会,具有很高的发展潜力。
{"title":"Predicting early depression in WZT drawing image based on deep learning","authors":"Kyung‐yeul Kim, Young‐bo Yang, Mi‐ra Kim, Jihie Kim, Ji Su Park","doi":"10.1111/exsy.13675","DOIUrl":"https://doi.org/10.1111/exsy.13675","url":null,"abstract":"When stress causes negative behaviours to emerge in our daily lives, it is important to intervene quickly and appropriately to control the negative problem behaviours. Questionnaires, a common method of information gathering, have the disadvantage that it is difficult to get the exact information needed due to defensive or insincere responses from subjects. As an alternative to these drawbacks, projective testing using pictures can provide the necessary information more accurately than questionnaires because the subject responds subconsciously and the direct experience expressed through pictures can be more accurate than questionnaires. Analysing hand‐drawn image data with the Wartegg Zeichen Test (WZT) is not easy. In this study, we used deep learning to analyse image data represented as pictures through WZT to predict early depression. We analyse the data of 54 people who were judged as early depression and 54 people without depression, and increase the number of people without depression to 100 and 500, and aim to study in unbalanced data. We use CNN and CNN‐SVM to analyse the drawing images of WZT's initial depression with deep learning and predict the outcome of depression. The results show that the initial depression is predicted with 92%–98% accuracy on the image data directly drawn by WZT. This is the first study to automatically analyse and predict early depression in WZT based on hand‐drawn image data using deep learning models. The extraction of features from WZT images by deep learning analysis is expected to create more research opportunities through the convergence of psychotherapy and Information and Communication Technology (ICT) technology, and is expected to have high growth potential.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel transformer attention‐based approach for sarcasm detection 基于变压器注意力的讽刺检测新方法
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-24 DOI: 10.1111/exsy.13686
Shumaila Khan, Iqbal Qasim, Wahab Khan, Khursheed Aurangzeb, Javed Ali Khan, Muhammad Shahid Anwar
Sarcasm detection is challenging in natural language processing (NLP) due to its implicit nature, particularly in low‐resource languages. Despite limited linguistic resources, researchers have focused on detecting sarcasm on social media platforms, leading to the development of specialized algorithms and models tailored for Urdu text. Researchers have significantly improved sarcasm detection accuracy by analysing patterns and linguistic cues unique to the language, thereby advancing NLP capabilities in low‐resource languages and facilitating better communication within diverse online communities. This work introduces UrduSarcasmNet, a novel architecture using cascaded group multi‐head attention, which is an innovative deep‐learning approach that employs cascaded group multi‐head attention techniques to enhance effectiveness. By employing a series of attention heads in a cascading manner, our model captures both local and global contexts, facilitating a more comprehensive understanding of the text. Adding a group attention mechanism enables simultaneous consideration of various sub‐topics within the content, thereby enriching the model's effectiveness. The proposed UrduSarcasmNet approach is validated with the Urdu‐sarcastic‐tweets‐dataset (UST) dataset, which has been curated for this purpose. Our experimental results on the UST dataset show that the proposed UrduSarcasmNet framework outperforms the simple‐attention mechanism and other state‐of‐the‐art models. This research significantly enhances natural language processing (NLP) and provides valuable insights for improving sarcasm recognition tools in low‐resource languages like Urdu.
在自然语言处理(NLP)中,由于讽刺的隐含性质,尤其是在低资源语言中,讽刺检测具有挑战性。尽管语言资源有限,但研究人员一直专注于检测社交媒体平台上的讽刺,从而开发出了专门针对乌尔都语文本的算法和模型。研究人员通过分析乌尔都语特有的模式和语言线索,大大提高了讽刺语言检测的准确性,从而推动了低资源语言的 NLP 能力,促进了不同网络社区内更好的交流。这项工作介绍了使用级联组多头注意力的新型架构 UrduSarcasmNet,这是一种创新的深度学习方法,采用了级联组多头注意力技术来提高效率。通过以级联方式使用一系列注意力头,我们的模型可以捕捉局部和全局上下文,从而促进对文本更全面的理解。通过添加群体注意机制,可以同时考虑内容中的各种子主题,从而丰富了模型的有效性。建议的 UrduSarcasmNet 方法已通过为此目的策划的 Urdu-sarcastic-tweets 数据集(UST)进行了验证。我们在 UST 数据集上的实验结果表明,所提出的 UrduSarcasmNet 框架优于简单关注机制和其他最先进的模型。这项研究大大提高了自然语言处理(NLP)能力,并为改进乌尔都语等低资源语言的讽刺语言识别工具提供了宝贵的见解。
{"title":"A novel transformer attention‐based approach for sarcasm detection","authors":"Shumaila Khan, Iqbal Qasim, Wahab Khan, Khursheed Aurangzeb, Javed Ali Khan, Muhammad Shahid Anwar","doi":"10.1111/exsy.13686","DOIUrl":"https://doi.org/10.1111/exsy.13686","url":null,"abstract":"Sarcasm detection is challenging in natural language processing (NLP) due to its implicit nature, particularly in low‐resource languages. Despite limited linguistic resources, researchers have focused on detecting sarcasm on social media platforms, leading to the development of specialized algorithms and models tailored for Urdu text. Researchers have significantly improved sarcasm detection accuracy by analysing patterns and linguistic cues unique to the language, thereby advancing NLP capabilities in low‐resource languages and facilitating better communication within diverse online communities. This work introduces UrduSarcasmNet, a novel architecture using cascaded group multi‐head attention, which is an innovative deep‐learning approach that employs cascaded group multi‐head attention techniques to enhance effectiveness. By employing a series of attention heads in a cascading manner, our model captures both local and global contexts, facilitating a more comprehensive understanding of the text. Adding a group attention mechanism enables simultaneous consideration of various sub‐topics within the content, thereby enriching the model's effectiveness. The proposed UrduSarcasmNet approach is validated with the Urdu‐sarcastic‐tweets‐dataset (UST) dataset, which has been curated for this purpose. Our experimental results on the UST dataset show that the proposed UrduSarcasmNet framework outperforms the simple‐attention mechanism and other state‐of‐the‐art models. This research significantly enhances natural language processing (NLP) and provides valuable insights for improving sarcasm recognition tools in low‐resource languages like Urdu.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing fraud detection in financial transactions with machine learning and imbalance mitigation 利用机器学习和失衡缓解优化金融交易中的欺诈检测
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-23 DOI: 10.1111/exsy.13682
Ezaz Mohammed Al‐dahasi, Rama Khaled Alsheikh, Fakhri Alam Khan, Gwanggil Jeon
The rapid advancement of the Internet and digital payments has transformed the landscape of financial transactions, leading to both technological progress and an alarming rise in cybercrime. This study addresses the critical issue of financial fraud detection in the era of digital payments, focusing on enhancing operational risk frameworks to mitigate the increasing threats. The objective is to improve the predictive performance of fraud detection systems using machine learning techniques. The methodology involves a comprehensive data preprocessing and model creation process, including one‐hot encoding, feature selection, sampling, standardization, and tokenization. Six machine learning models are employed for fraud detection, and their hyperparameters are optimized. Evaluation metrics such as accuracy, precision, recall, and F1‐score are used to assess model performance. Results reveal that XGBoost and Random Forest outperform other models, achieving a balance between false positives and false negatives. The study meets the requirements for fraud detection systems, ensuring accuracy, scalability, adaptability, and explainability. This paper provides valuable insights into the efficacy of machine learning models for financial fraud detection and emphasizes the importance of striking a balance between false positives and false negatives.
互联网和数字支付的快速发展改变了金融交易的格局,既带来了技术进步,也导致网络犯罪的惊人增长。本研究探讨了数字支付时代金融欺诈检测的关键问题,重点是加强操作风险框架,以减轻日益增长的威胁。目的是利用机器学习技术提高欺诈检测系统的预测性能。该方法涉及全面的数据预处理和模型创建过程,包括单次编码、特征选择、采样、标准化和标记化。欺诈检测采用了六个机器学习模型,并对其超参数进行了优化。准确率、精确度、召回率和 F1 分数等评价指标用于评估模型性能。结果显示,XGBoost 和随机森林的表现优于其他模型,在误报和误报之间取得了平衡。这项研究符合欺诈检测系统的要求,确保了准确性、可扩展性、适应性和可解释性。本文就机器学习模型在金融欺诈检测中的功效提供了宝贵的见解,并强调了在误报和误报之间取得平衡的重要性。
{"title":"Optimizing fraud detection in financial transactions with machine learning and imbalance mitigation","authors":"Ezaz Mohammed Al‐dahasi, Rama Khaled Alsheikh, Fakhri Alam Khan, Gwanggil Jeon","doi":"10.1111/exsy.13682","DOIUrl":"https://doi.org/10.1111/exsy.13682","url":null,"abstract":"The rapid advancement of the Internet and digital payments has transformed the landscape of financial transactions, leading to both technological progress and an alarming rise in cybercrime. This study addresses the critical issue of financial fraud detection in the era of digital payments, focusing on enhancing operational risk frameworks to mitigate the increasing threats. The objective is to improve the predictive performance of fraud detection systems using machine learning techniques. The methodology involves a comprehensive data preprocessing and model creation process, including one‐hot encoding, feature selection, sampling, standardization, and tokenization. Six machine learning models are employed for fraud detection, and their hyperparameters are optimized. Evaluation metrics such as accuracy, precision, recall, and <jats:italic>F</jats:italic>1‐score are used to assess model performance. Results reveal that XGBoost and Random Forest outperform other models, achieving a balance between false positives and false negatives. The study meets the requirements for fraud detection systems, ensuring accuracy, scalability, adaptability, and explainability. This paper provides valuable insights into the efficacy of machine learning models for financial fraud detection and emphasizes the importance of striking a balance between false positives and false negatives.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141783668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TMaD: Three‐tier malware detection using multi‐view feature for secure convergence ICT environments TMaD:利用多视角特征进行三层恶意软件检测,确保融合信息和通信技术环境的安全
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-19 DOI: 10.1111/exsy.13684
Jueun Jeon, Byeonghui Jeong, Seungyeon Baek, Young‐Sik Jeong
As digital transformation accelerates, data generated in a convergence information and communication technology (ICT) environment must be secured. This data includes confidential information such as personal and financial information, so attackers spread malware in convergence ICT environments to steal this information. To protect convergence ICT environments from diverse cyber threats, deep learning models have been utilized for malware detection. However, accurately detecting rapidly generated variants and obfuscated malware is challenging. This study proposes a three‐tier malware detection (TMaD) scheme that utilizes a cloud‐fog‐edge collaborative architecture to analyse multi‐view features of executable files and detect malware. TMaD performs signature‐based malware detection at the edge device tier, then sends executables detected as unknown or benign to the fog tier. The fog tier conducts static analysis on non‐obfuscated executables and those transferred from the previous tier to detect variant malware. Subsequently, TMaD sends executables detected as benign in the fog tier to the cloud tier, where dynamic analysis is performed on obfuscated executables and those detected as benign to identify obfuscated malware. An evaluation of TMaD's detection performance resulted in an accuracy of 94.78%, a recall of 0.9794, a precision of 0.9535, and an f1‐score of 0.9663. This performance demonstrates that TMaD, by analysing executables across several tiers and minimizing false negatives, exhibits superior detection performance compared to existing malware detection models.
随着数字化转型的加速,在融合信息和通信技术(ICT)环境中生成的数据必须得到保护。这些数据包括个人和财务信息等机密信息,因此攻击者会在融合 ICT 环境中传播恶意软件,以窃取这些信息。为了保护融合 ICT 环境免受各种网络威胁,深度学习模型已被用于恶意软件检测。然而,准确检测快速生成的变种和混淆的恶意软件具有挑战性。本研究提出了一种三层恶意软件检测(TMaD)方案,利用云-雾-边协同架构分析可执行文件的多视图特征并检测恶意软件。TMaD 在边缘设备层执行基于签名的恶意软件检测,然后将检测到的未知或良性可执行文件发送到雾层。雾层对未经混淆处理的可执行文件和从上一层传输过来的可执行文件进行静态分析,以检测变种恶意软件。随后,TMaD 将在雾层中检测到的良性可执行文件发送到云层,在云层中对经过混淆处理的可执行文件和检测到的良性可执行文件进行动态分析,以识别经过混淆处理的恶意软件。对 TMaD 检测性能的评估结果是:准确率 94.78%,召回率 0.9794,精确度 0.9535,f1 分数 0.9663。这一性能表明,TMaD 通过分析多个层级的可执行文件并最大限度地减少误判,与现有的恶意软件检测模型相比,具有更出色的检测性能。
{"title":"TMaD: Three‐tier malware detection using multi‐view feature for secure convergence ICT environments","authors":"Jueun Jeon, Byeonghui Jeong, Seungyeon Baek, Young‐Sik Jeong","doi":"10.1111/exsy.13684","DOIUrl":"https://doi.org/10.1111/exsy.13684","url":null,"abstract":"As digital transformation accelerates, data generated in a convergence information and communication technology (ICT) environment must be secured. This data includes confidential information such as personal and financial information, so attackers spread malware in convergence ICT environments to steal this information. To protect convergence ICT environments from diverse cyber threats, deep learning models have been utilized for malware detection. However, accurately detecting rapidly generated variants and obfuscated malware is challenging. This study proposes a three‐tier malware detection (TMaD) scheme that utilizes a cloud‐fog‐edge collaborative architecture to analyse multi‐view features of executable files and detect malware. TMaD performs signature‐based malware detection at the edge device tier, then sends executables detected as unknown or benign to the fog tier. The fog tier conducts static analysis on non‐obfuscated executables and those transferred from the previous tier to detect variant malware. Subsequently, TMaD sends executables detected as benign in the fog tier to the cloud tier, where dynamic analysis is performed on obfuscated executables and those detected as benign to identify obfuscated malware. An evaluation of TMaD's detection performance resulted in an accuracy of 94.78%, a recall of 0.9794, a precision of 0.9535, and an f1‐score of 0.9663. This performance demonstrates that TMaD, by analysing executables across several tiers and minimizing false negatives, exhibits superior detection performance compared to existing malware detection models.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141742191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing hybrid deep learning models for drug‐target interaction prediction: A comparative analysis of evolutionary algorithms 为药物靶点相互作用预测优化混合深度学习模型:进化算法的比较分析
IF 3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-19 DOI: 10.1111/exsy.13683
Moolchand Sharma, Aryan Bhatia, Akhil, A. Dutta, Shtwai Alsubai
In the realm of Drug‐Target Interaction (DTI) prediction, this research investigates and contrasts the efficacy of diverse evolutionary algorithms in fine‐tuning a sophisticated hybrid deep learning model. Recognizing the critical role of DTI in drug discovery and repositioning, we tackle the challenges of binary classification by reframing the problem as a regression task. Our focus lies on the Convolution Self‐Attention Network with Attention‐based bidirectional Long Short‐Term Memory Network (CSAN‐BiLSTM‐Att), a hybrid model combining convolutional neural network (CNN) blocks, self‐attention mechanisms, and bidirectional LSTM layers. To optimize this complex model, we employ Differential Evolution (DE), Particle Swarm Optimization (PSO), Memetic Particle Swarm Optimization Algorithm (MPSOA), Fire Hawk Optimization (FHO), and Artificial Hummingbird Algorithm (AHA). Through thorough comparative analysis, we evaluate the performance of these evolutionary algorithms in enhancing the CSAN‐BiLSTM‐Att model's effectiveness. By examining the strengths and weaknesses of each algorithm, our study aims to provide valuable insights into DTI prediction, identifying the most effective evolutionary algorithm for hyperparameter tuning in advanced deep learning models. Notably, Fire‐hawk optimization (FHO) emerges as particularly promising, achieving the highest Concordance Index (C‐index) as 0.974 for KIBA datasets and 0.894 for DAVIS datasets and demonstrating exceptional accuracy in ranking continuous predictions across both the datasets.
在药物-目标相互作用(DTI)预测领域,本研究调查并对比了各种进化算法在微调复杂的混合深度学习模型方面的功效。认识到 DTI 在药物发现和重新定位中的关键作用,我们通过将问题重构为回归任务来应对二元分类的挑战。我们的重点是卷积自注意力网络与基于注意力的双向长短期记忆网络(CSAN-BiLSTM-Att),这是一个结合了卷积神经网络(CNN)块、自注意力机制和双向 LSTM 层的混合模型。为了优化这一复杂模型,我们采用了差分进化(DE)、粒子群优化(PSO)、记忆粒子群优化算法(MPSOA)、火鹰优化(FHO)和人工蜂鸟算法(AHA)。通过全面的比较分析,我们评估了这些进化算法在提高 CSAN-BiLSTM-Att 模型有效性方面的性能。通过研究每种算法的优缺点,我们的研究旨在为 DTI 预测提供有价值的见解,为高级深度学习模型的超参数调整确定最有效的进化算法。值得注意的是,火鹰优化(FHO)特别有前途,它在 KIBA 数据集和 DAVIS 数据集上分别获得了 0.974 和 0.894 的最高一致性指数(C-index),并在这两个数据集的连续预测排名中表现出了非凡的准确性。
{"title":"Optimizing hybrid deep learning models for drug‐target interaction prediction: A comparative analysis of evolutionary algorithms","authors":"Moolchand Sharma, Aryan Bhatia, Akhil, A. Dutta, Shtwai Alsubai","doi":"10.1111/exsy.13683","DOIUrl":"https://doi.org/10.1111/exsy.13683","url":null,"abstract":"In the realm of Drug‐Target Interaction (DTI) prediction, this research investigates and contrasts the efficacy of diverse evolutionary algorithms in fine‐tuning a sophisticated hybrid deep learning model. Recognizing the critical role of DTI in drug discovery and repositioning, we tackle the challenges of binary classification by reframing the problem as a regression task. Our focus lies on the Convolution Self‐Attention Network with Attention‐based bidirectional Long Short‐Term Memory Network (CSAN‐BiLSTM‐Att), a hybrid model combining convolutional neural network (CNN) blocks, self‐attention mechanisms, and bidirectional LSTM layers. To optimize this complex model, we employ Differential Evolution (DE), Particle Swarm Optimization (PSO), Memetic Particle Swarm Optimization Algorithm (MPSOA), Fire Hawk Optimization (FHO), and Artificial Hummingbird Algorithm (AHA). Through thorough comparative analysis, we evaluate the performance of these evolutionary algorithms in enhancing the CSAN‐BiLSTM‐Att model's effectiveness. By examining the strengths and weaknesses of each algorithm, our study aims to provide valuable insights into DTI prediction, identifying the most effective evolutionary algorithm for hyperparameter tuning in advanced deep learning models. Notably, Fire‐hawk optimization (FHO) emerges as particularly promising, achieving the highest Concordance Index (C‐index) as 0.974 for KIBA datasets and 0.894 for DAVIS datasets and demonstrating exceptional accuracy in ranking continuous predictions across both the datasets.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141820837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The study of engagement at work from the artificial intelligence perspective: A systematic review 从人工智能角度研究工作参与度:系统回顾
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-17 DOI: 10.1111/exsy.13673
Claudia García‐Navarro, Manuel Pulido‐Martos, Cristina Pérez‐Lozano
Engagement has been defined as an attitude toward work, as a positive, satisfying, work‐related state of mind characterized by high levels of vigour, dedication, and absorption. Both its definition and its assessment have been controversial; however, new methods for its assessment, including artificial intelligence (AI), have been introduced in recent years. Therefore, this research aims to determine the state of the art of AI in the study of engagement. To this end, we conducted a systematic review in accordance with PRISMA to analyse the publications to date on the use of AI for the analysis of engagement. The search, carried out in six databases, was filtered, and 15 papers were finally analysed. The results show that AI has been used mainly to assess and predict engagement levels, as well as to understand the relationships between engagement and other variables. The most commonly used AI techniques are machine learning (ML) and natural language processing (NLP), and all publications use structured and unstructured data, mainly from self‐report instruments, social networks, and datasets. The accuracy of the models varies from 22% to 87%, and its main benefit has been to help both managers and HR staff understand employee engagement, although it has also contributed to research. Most of the articles have been published since 2015, and the geography has been global, with publications predominantly in India and the US. In conclusion, this study highlights the state of the art in AI for the study of engagement and concludes that the number of publications is increasing, indicating that this is possibly a new field or area of research in which important advances can be made in the study of engagement through new and novel techniques.
敬业度被定义为一种工作态度,是一种积极的、令人满意的、与工作相关的精神状态,其特点是精力充沛、全心投入和全身心投入。对其定义和评估一直存在争议;不过,近年来,包括人工智能(AI)在内的新评估方法已经问世。因此,本研究旨在确定人工智能在参与度研究中的应用现状。为此,我们按照 PRISMA 标准进行了一次系统性回顾,分析了迄今为止有关使用人工智能分析参与度的出版物。在六个数据库中进行的搜索经过筛选,最终分析了 15 篇论文。结果显示,人工智能主要用于评估和预测参与度水平,以及了解参与度与其他变量之间的关系。最常用的人工智能技术是机器学习(ML)和自然语言处理(NLP),所有论文都使用了结构化和非结构化数据,主要来自自我报告工具、社交网络和数据集。模型的准确率从 22% 到 87% 不等,其主要益处是帮助管理人员和人力资源部门了解员工敬业度,但也有助于研究工作。大多数文章都是 2015 年以来发表的,发表地域遍及全球,主要集中在印度和美国。总之,本研究强调了人工智能在敬业度研究方面的技术水平,并得出结论认为,发表文章的数量在不断增加,这表明这可能是一个新的研究领域或领域,通过新颖的技术可以在敬业度研究方面取得重要进展。
{"title":"The study of engagement at work from the artificial intelligence perspective: A systematic review","authors":"Claudia García‐Navarro, Manuel Pulido‐Martos, Cristina Pérez‐Lozano","doi":"10.1111/exsy.13673","DOIUrl":"https://doi.org/10.1111/exsy.13673","url":null,"abstract":"Engagement has been defined as an attitude toward work, as a positive, satisfying, work‐related state of mind characterized by high levels of vigour, dedication, and absorption. Both its definition and its assessment have been controversial; however, new methods for its assessment, including artificial intelligence (AI), have been introduced in recent years. Therefore, this research aims to determine the state of the art of AI in the study of engagement. To this end, we conducted a systematic review in accordance with PRISMA to analyse the publications to date on the use of AI for the analysis of engagement. The search, carried out in six databases, was filtered, and 15 papers were finally analysed. The results show that AI has been used mainly to assess and predict engagement levels, as well as to understand the relationships between engagement and other variables. The most commonly used AI techniques are machine learning (ML) and natural language processing (NLP), and all publications use structured and unstructured data, mainly from self‐report instruments, social networks, and datasets. The accuracy of the models varies from 22% to 87%, and its main benefit has been to help both managers and HR staff understand employee engagement, although it has also contributed to research. Most of the articles have been published since 2015, and the geography has been global, with publications predominantly in India and the US. In conclusion, this study highlights the state of the art in AI for the study of engagement and concludes that the number of publications is increasing, indicating that this is possibly a new field or area of research in which important advances can be made in the study of engagement through new and novel techniques.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141742192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A generative adversarial network‐based client‐level handwriting forgery attack in federated learning scenario 联合学习场景中基于生成式对抗网络的客户端级笔迹伪造攻击
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-12 DOI: 10.1111/exsy.13676
Lei Shi, Han Wu, Xu Ding, Hao Xu, Sinan Pan
Federated learning (FL), celebrated for its privacy‐preserving features, has been revealed by recent studies to harbour security vulnerabilities that jeopardize client privacy, particularly through data reconstruction attacks that enable adversaries to recover original client data. This study introduces a client‐level handwriting forgery attack method for FL based on generative adversarial networks (GANs), which reveals security vulnerabilities existing in FL systems. It should be stressed that this research is purely for academic purposes, aiming to raise concerns about privacy protection and data security, and does not encourage illegal activities. Our novel methodology assumes an adversarial scenario wherein adversaries intercept a fraction of parameter updates via victim clients’ wireless communication channels, then use this information to train GAN for data recovery. Finally, the purpose of handwriting imitation is achieved. To rigorously assess and validate our methodology, experiments were conducted using a bespoke Chinese digit dataset, facilitating in‐depth analysis and robust verification of results. Our experimental findings demonstrated enhanced data recovery effectiveness, a client‐level attack and greater versatility compared to prior art. Notably, our method maintained high attack performance even with a streamlined GAN design, yielding increased precision and significantly faster execution times compared to standard methods. Specifically, our experimental numerical results revealed a substantial boost in reconstruction accuracy by 16.7%, coupled with a 51.9% decrease in computational time compared to the latest similar techniques. Furthermore, tests on a simplified version of our GAN exhibited an average 10% enhancement in accuracy, alongside a remarkable 70% reduction in time consumption. By surmounting the limitations of previous work, this study fills crucial gaps and affirms the effectiveness of our approach in achieving high‐accuracy client‐level data reconstruction within the FL context, thereby stimulating further exploration into FL security measures.
联合学习(Federated Learning,FL)因其保护隐私的特性而备受赞誉,但最近的研究却揭示了它存在着危害客户端隐私的安全漏洞,特别是通过数据重构攻击,使对手能够恢复原始客户端数据。本研究介绍了一种基于生成式对抗网络(GANs)的 FL 客户端级笔迹伪造攻击方法,揭示了 FL 系统中存在的安全漏洞。需要强调的是,本研究纯粹出于学术目的,旨在引起人们对隐私保护和数据安全的关注,并不鼓励非法活动。我们的新方法假设了一种对抗场景,即对抗者通过受害者客户端的无线通信信道截获一部分参数更新,然后利用这些信息训练 GAN 进行数据恢复。最后,笔迹模仿的目的就达到了。为了严格评估和验证我们的方法,我们使用定制的中文数字数据集进行了实验,以便对结果进行深入分析和稳健验证。我们的实验结果表明,与现有技术相比,我们的方法提高了数据恢复的有效性、客户端级别的攻击和更大的通用性。值得注意的是,即使采用精简的 GAN 设计,我们的方法仍能保持较高的攻击性能,与标准方法相比,精度更高,执行时间更短。具体来说,我们的数值实验结果表明,与最新的类似技术相比,重建精度大幅提高了 16.7%,计算时间减少了 51.9%。此外,对我们的 GAN 简化版进行的测试表明,精度平均提高了 10%,同时耗时显著减少了 70%。通过克服以往工作的局限性,本研究填补了重要空白,并肯定了我们的方法在 FL 环境下实现高精度客户端级数据重建的有效性,从而激发了对 FL 安全措施的进一步探索。
{"title":"A generative adversarial network‐based client‐level handwriting forgery attack in federated learning scenario","authors":"Lei Shi, Han Wu, Xu Ding, Hao Xu, Sinan Pan","doi":"10.1111/exsy.13676","DOIUrl":"https://doi.org/10.1111/exsy.13676","url":null,"abstract":"Federated learning (FL), celebrated for its privacy‐preserving features, has been revealed by recent studies to harbour security vulnerabilities that jeopardize client privacy, particularly through data reconstruction attacks that enable adversaries to recover original client data. This study introduces a client‐level handwriting forgery attack method for FL based on generative adversarial networks (GANs), which reveals security vulnerabilities existing in FL systems. It should be stressed that this research is purely for academic purposes, aiming to raise concerns about privacy protection and data security, and does not encourage illegal activities. Our novel methodology assumes an adversarial scenario wherein adversaries intercept a fraction of parameter updates via victim clients’ wireless communication channels, then use this information to train GAN for data recovery. Finally, the purpose of handwriting imitation is achieved. To rigorously assess and validate our methodology, experiments were conducted using a bespoke Chinese digit dataset, facilitating in‐depth analysis and robust verification of results. Our experimental findings demonstrated enhanced data recovery effectiveness, a client‐level attack and greater versatility compared to prior art. Notably, our method maintained high attack performance even with a streamlined GAN design, yielding increased precision and significantly faster execution times compared to standard methods. Specifically, our experimental numerical results revealed a substantial boost in reconstruction accuracy by 16.7%, coupled with a 51.9% decrease in computational time compared to the latest similar techniques. Furthermore, tests on a simplified version of our GAN exhibited an average 10% enhancement in accuracy, alongside a remarkable 70% reduction in time consumption. By surmounting the limitations of previous work, this study fills crucial gaps and affirms the effectiveness of our approach in achieving high‐accuracy client‐level data reconstruction within the FL context, thereby stimulating further exploration into FL security measures.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141613269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal dynamic fusion framework: Multilevel feature fusion guided by prompts 多模态动态融合框架:由提示引导的多级特征融合
IF 3.3 4区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-12 DOI: 10.1111/exsy.13668
Lei Pan, Huan‐Qing Wu
With the progressive augmentation of parameters in multimodal models, to optimize computational efficiency, some studies have adopted the approach of fine‐tuning the unimodal pre‐training model to achieve multimodal fusion tasks. However, these methods tend to rely solely on simplistic or singular fusion strategies, thereby neglecting more flexible fusion approaches. Moreover, existing methods prioritize the integration of modality features containing highly semantic information, often overlooking the influence of fusing low‐level features on the outcomes. Therefore, this study introduces an innovative approach named multilevel feature fusion guided by prompts (MFF‐GP), a multimodal dynamic fusion framework. It guides the dynamic neural network by prompt vectors to dynamically select the suitable fusion network for each hierarchical feature of the unimodal pre‐training model. This method improves the interactions between multiple modalities and promotes a more efficient fusion of features across them. Extensive experiments on the UPMC Food 101, SNLI‐VE and MM‐IMDB datasets demonstrate that with only a few trainable parameters, MFF‐GP achieves significant accuracy improvements compared to a newly designed PMF based on fine‐tuning—specifically, an accuracy improvement of 2.15% on the UPMC Food 101 dataset and 0.82% on the SNLI‐VE dataset. Further study of the results reveals that increasing the diversity of interactions between distinct modalities is critical and delivers significant performance improvements. Furthermore, for certain multimodal tasks, focusing on the low‐level features is beneficial for modality integration. Our implementation is available at: https://github.com/whq2024/MFF-GP.
随着多模态模型参数的逐步增加,为了优化计算效率,一些研究采用了微调单模态预训练模型的方法来实现多模态融合任务。然而,这些方法往往只依赖于简单或单一的融合策略,从而忽略了更灵活的融合方法。此外,现有方法优先考虑包含高语义信息的模态特征的融合,往往忽略了低层次特征融合对结果的影响。因此,本研究引入了一种名为 "提示引导的多级特征融合"(MFF-GP)的创新方法,这是一种多模态动态融合框架。它通过提示向量引导动态神经网络,为单模态预训练模型的每个层次特征动态选择合适的融合网络。这种方法改善了多种模态之间的交互,促进了跨模态特征的更有效融合。在 UPMC Food 101、SNLI-VE 和 MM-IMDB 数据集上进行的大量实验表明,与基于微调的新设计 PMF 相比,MFF-GP 只需几个可训练参数就能显著提高准确率,具体来说,在 UPMC Food 101 数据集上提高了 2.15%,在 SNLI-VE 数据集上提高了 0.82%。对结果的进一步研究表明,增加不同模态之间交互的多样性至关重要,并能显著提高性能。此外,对于某些多模态任务,关注低层次特征有利于模态整合。我们的实施方案可在以下网址获取:https://github.com/whq2024/MFF-GP。
{"title":"Multimodal dynamic fusion framework: Multilevel feature fusion guided by prompts","authors":"Lei Pan, Huan‐Qing Wu","doi":"10.1111/exsy.13668","DOIUrl":"https://doi.org/10.1111/exsy.13668","url":null,"abstract":"With the progressive augmentation of parameters in multimodal models, to optimize computational efficiency, some studies have adopted the approach of fine‐tuning the unimodal pre‐training model to achieve multimodal fusion tasks. However, these methods tend to rely solely on simplistic or singular fusion strategies, thereby neglecting more flexible fusion approaches. Moreover, existing methods prioritize the integration of modality features containing highly semantic information, often overlooking the influence of fusing low‐level features on the outcomes. Therefore, this study introduces an innovative approach named multilevel feature fusion guided by prompts (MFF‐GP), a multimodal dynamic fusion framework. It guides the dynamic neural network by prompt vectors to dynamically select the suitable fusion network for each hierarchical feature of the unimodal pre‐training model. This method improves the interactions between multiple modalities and promotes a more efficient fusion of features across them. Extensive experiments on the UPMC Food 101, SNLI‐VE and MM‐IMDB datasets demonstrate that with only a few trainable parameters, MFF‐GP achieves significant accuracy improvements compared to a newly designed PMF based on fine‐tuning—specifically, an accuracy improvement of 2.15% on the UPMC Food 101 dataset and 0.82% on the SNLI‐VE dataset. Further study of the results reveals that increasing the diversity of interactions between distinct modalities is critical and delivers significant performance improvements. Furthermore, for certain multimodal tasks, focusing on the low‐level features is beneficial for modality integration. Our implementation is available at: <jats:ext-link xmlns:xlink=\"http://www.w3.org/1999/xlink\" xlink:href=\"https://github.com/whq2024/MFF-GP\">https://github.com/whq2024/MFF-GP</jats:ext-link>.","PeriodicalId":51053,"journal":{"name":"Expert Systems","volume":null,"pages":null},"PeriodicalIF":3.3,"publicationDate":"2024-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141613581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Expert Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1