Pub Date : 2024-10-01DOI: 10.1007/s10489-024-05745-y
Ran Zhu, Jian Peng, Wen Huang, Yujun He, Chengyi Tang
Graph clustering is one of the most fundamental tasks in graph learning. Recently, numerous graph clustering models based on dual network (Auto-encoder+Graph Neural Network(GNN)) architectures have emerged and achieved promising results. However, we observe several limitations in the literature: 1) simple graph neural networks that fail to capture the intricate relationships between nodes are used for graph clustering tasks; 2) heterogeneous information is inadequately interacted and merged; and 3) the clustering boundaries are fuzzy in the feature space. To address the aforementioned issues, we propose a novel graph clustering model named Dynamic Graph Attention-guided Graph Clustering with Entropy Minimization self-supervision(DGAGC-EM). Specifically, we introduce DGATE, a graph auto-encoder based on dynamic graph attention, to capture the intricate relationships among graph nodes. Additionally, we perform feature enhancement from both global and local perspectives via the proposed Global-Local Feature Enhancement (GLFE) module. Finally, we propose a self-supervised strategy based on entropy minimization theory to guide network training process to achieve better performance and produce sharper clustering boundaries. Extensive experimental results obtained on four datasets demonstrate that our method is highly competitive with the SOTA methods.
The figure presents the overall framework of proposed Dynamic Graph Attention-guided Graph Clustering with Entropy Minimization selfsupervision(DGAGC-EM). Specifically, the Dynamic Graph Attetion Auto-Encoder Module is our proposed graph auto-encoder based on dynamic graph attention, to capture the intricate relationships among graph nodes. The Auto-Encoder Module is a basic autoencoder with simple MLPs to extract embeddings from node attributes. Additionally, the proposed Global-Local Feature Enhancement (GLFE) module perform feature enhancement from both global and local perspectives. Finally, the proposed Self-supervised Module guide network training process to achieve better performance and produce sharper clustering boundaries
图聚类是图学习中最基本的任务之一。最近,出现了许多基于双网络(自动编码器+图神经网络(GNN))架构的图聚类模型,并取得了可喜的成果。然而,我们在文献中发现了几个局限性:1) 简单的图神经网络无法捕捉节点之间错综复杂的关系,因此被用于图聚类任务;2) 异构信息的交互和合并不充分;3) 聚类边界在特征空间中比较模糊。针对上述问题,我们提出了一种新型图聚类模型,名为 "熵最小化自我监督的动态图注意力引导图聚类(DGAGC-EM)"。具体来说,我们引入了基于动态图注意力的图自动编码器 DGATE,以捕捉图节点之间错综复杂的关系。此外,我们还通过提议的全局-局部特征增强(GLFE)模块,从全局和局部两个角度进行特征增强。最后,我们提出了一种基于熵最小化理论的自监督策略来指导网络训练过程,以获得更好的性能和更清晰的聚类边界。在四个数据集上获得的大量实验结果表明,我们的方法与 SOTA 方法相比具有很强的竞争力。具体来说,动态图注意力自动编码器模块是我们提出的基于动态图注意力的图自动编码器,用于捕捉图节点之间错综复杂的关系。自动编码器模块是一个基本的自动编码器,使用简单的 MLP 从节点属性中提取嵌入。此外,拟议的全局-局部特征增强(GLFE)模块可从全局和局部两个角度进行特征增强。最后,拟议的自监督模块将指导网络训练过程,以获得更好的性能,并产生更清晰的聚类边界。
{"title":"Dynamic graph attention-guided graph clustering with entropy minimization self-supervision","authors":"Ran Zhu, Jian Peng, Wen Huang, Yujun He, Chengyi Tang","doi":"10.1007/s10489-024-05745-y","DOIUrl":"10.1007/s10489-024-05745-y","url":null,"abstract":"<p>Graph clustering is one of the most fundamental tasks in graph learning. Recently, numerous graph clustering models based on dual network (Auto-encoder+Graph Neural Network(GNN)) architectures have emerged and achieved promising results. However, we observe several limitations in the literature: 1) simple graph neural networks that fail to capture the intricate relationships between nodes are used for graph clustering tasks; 2) heterogeneous information is inadequately interacted and merged; and 3) the clustering boundaries are fuzzy in the feature space. To address the aforementioned issues, we propose a novel graph clustering model named <b>D</b>ynamic <b>G</b>raph <b>A</b>ttention-guided <b>G</b>raph <b>C</b>lustering with <b>E</b>ntropy <b>M</b>inimization self-supervision(DGAGC-EM). Specifically, we introduce DGATE, a graph auto-encoder based on dynamic graph attention, to capture the intricate relationships among graph nodes. Additionally, we perform feature enhancement from both global and local perspectives via the proposed Global-Local Feature Enhancement (GLFE) module. Finally, we propose a self-supervised strategy based on entropy minimization theory to guide network training process to achieve better performance and produce sharper clustering boundaries. Extensive experimental results obtained on four datasets demonstrate that our method is highly competitive with the SOTA methods.</p><p>The figure presents the overall framework of proposed Dynamic Graph Attention-guided Graph Clustering with Entropy Minimization selfsupervision(DGAGC-EM). Specifically, the Dynamic Graph Attetion Auto-Encoder Module is our proposed graph auto-encoder based on dynamic graph attention, to capture the intricate relationships among graph nodes. The Auto-Encoder Module is a basic autoencoder with simple MLPs to extract embeddings from node attributes. Additionally, the proposed Global-Local Feature Enhancement (GLFE) module perform feature enhancement from both global and local perspectives. Finally, the proposed Self-supervised Module guide network training process to achieve better performance and produce sharper clustering boundaries</p>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12819 - 12834"},"PeriodicalIF":3.4,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-30DOI: 10.1007/s10489-024-05833-z
Maoshan Liu, Vasile Palade , Zhonglong Zheng
The popular block diagonal representation subspace clustering approach shows high effectiveness in dividing a high-dimensional data space into the corresponding subspaces. However, existing subspace clustering algorithms have some weaknesses in achieving high clustering performance. This paper presents a multi-geometric block diagonal representation subspace clustering with low-rank kernel (MBDR-LRK) method that includes two major improvements. First, as visual data often exists on a Riemannian manifold not captured by Euclidean geometry, we harness the multi-order data complementarity to develop a multi-geometric block diagonal representation (MBDR) subspace clustering. Secondly, the proposed MBDR-LRK approach ensures the low-rankness in the mapped space, by adapting the kernel matrix to a pre-defined one rather than relying on a fixed kernel as in traditional methods. The paper also presents details on the monotonic decrease of the objective function and the boundedness and convergence of the affinity matrix, and the experimental results prove the convergence of the proposed method. Based on the MATLAB development environment, the proposed MBDR-LRK algorithm outperforms other related algorithms and obtained an accuracy of 88.70% on the ORL (40 classes), 89.39% on the Extended Yale B (38 classes), 50.22% on the AR (100 classes) and 75.47% on the COIL (50 classes) datasets.
{"title":"Multi-geometric block diagonal representation subspace clustering with low-rank kernel","authors":"Maoshan Liu, Vasile Palade , Zhonglong Zheng","doi":"10.1007/s10489-024-05833-z","DOIUrl":"10.1007/s10489-024-05833-z","url":null,"abstract":"<div><p>The popular block diagonal representation subspace clustering approach shows high effectiveness in dividing a high-dimensional data space into the corresponding subspaces. However, existing subspace clustering algorithms have some weaknesses in achieving high clustering performance. This paper presents a multi-geometric block diagonal representation subspace clustering with low-rank kernel (MBDR-LRK) method that includes two major improvements. First, as visual data often exists on a Riemannian manifold not captured by Euclidean geometry, we harness the multi-order data complementarity to develop a multi-geometric block diagonal representation (MBDR) subspace clustering. Secondly, the proposed MBDR-LRK approach ensures the low-rankness in the mapped space, by adapting the kernel matrix to a pre-defined one rather than relying on a fixed kernel as in traditional methods. The paper also presents details on the monotonic decrease of the objective function and the boundedness and convergence of the affinity matrix, and the experimental results prove the convergence of the proposed method. Based on the MATLAB development environment, the proposed MBDR-LRK algorithm outperforms other related algorithms and obtained an accuracy of 88.70% on the ORL (40 classes), 89.39% on the Extended Yale B (38 classes), 50.22% on the AR (100 classes) and 75.47% on the COIL (50 classes) datasets.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12764 - 12790"},"PeriodicalIF":3.4,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-28DOI: 10.1007/s10489-024-05839-7
Xueyu Guo, Shengwei Tian, Long Yu, Xiaoyu He
Multimodal sentiment analysis has received widespread attention from the research community in recent years; it aims to use information from different modalities to predict sentiment polarity. However, the model architecture of most existing methods is fixed, and data can only flow along an established path, which leads to poor generalization of the model to different types of data. Furthermore, most methods explore only intra- or intermodal interactions and do not combine the two. In this paper, we propose the SmartRouting Attention Network (SmartRAN). SmartRAN can smartly select the data flow path on the basis of the smart routing attention module, effectively avoiding the disadvantages of poor adaptability and generalizability caused by a fixed model architecture. In addition, SmartRAN includes the learning process of both intra- and intermodal information, which can enhance the semantic consistency of comprehensive information and improve the learning ability of the model for complex relationships. Extensive experiments on two benchmark datasets, CMU-MOSI and CMU-MOSEI, prove that the proposed SmartRAN has superior performance to state-of-the-art models.
{"title":"SmartRAN: Smart Routing Attention Network for multimodal sentiment analysis","authors":"Xueyu Guo, Shengwei Tian, Long Yu, Xiaoyu He","doi":"10.1007/s10489-024-05839-7","DOIUrl":"10.1007/s10489-024-05839-7","url":null,"abstract":"<div><p>Multimodal sentiment analysis has received widespread attention from the research community in recent years; it aims to use information from different modalities to predict sentiment polarity. However, the model architecture of most existing methods is fixed, and data can only flow along an established path, which leads to poor generalization of the model to different types of data. Furthermore, most methods explore only intra- or intermodal interactions and do not combine the two. In this paper, we propose the <b>Smart</b> <b>R</b>outing <b>A</b>ttention <b>N</b>etwork (SmartRAN). SmartRAN can smartly select the data flow path on the basis of the smart routing attention module, effectively avoiding the disadvantages of poor adaptability and generalizability caused by a fixed model architecture. In addition, SmartRAN includes the learning process of both intra- and intermodal information, which can enhance the semantic consistency of comprehensive information and improve the learning ability of the model for complex relationships. Extensive experiments on two benchmark datasets, CMU-MOSI and CMU-MOSEI, prove that the proposed SmartRAN has superior performance to state-of-the-art models.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12742 - 12763"},"PeriodicalIF":3.4,"publicationDate":"2024-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Trajectory mapping onto a road network is a complex yet important task. This is because, in the presence of circular sections, Y-shaped intersections, and sections with elevated overlaps with the ground, the conditions of road networks become complicated. Consequently, trajectory mapping becomes challenging owing to the complexities of road networks and the noise generated by high positioning errors. In this study, in response to the difficulty in handling redundant noisy trajectory data in complex road network environments, a complex road network map-matching method based on trajectory structure extraction is proposed. The features of the structure are extracted from the original trajectory data to reduce the effects of redundancy and noise on matching. An adaptive screening candidate method is proposed using driver behavior to estimate the road density and reduce the matching time by selecting effective candidates. A spatiotemporal analysis function is redefined using speed and distance features, and a directional analysis function is proposed for use in combination with directional features to improve the matching accuracy of complex road networks. An experimental evaluation based on real-ground trajectory data collected using in-vehicle sensing devices is conducted to verify the effectiveness of the algorithm. Moreover, extensive experiments are performed on challenging real datasets to evaluate the proposed method, and its accuracy and efficiency are compared with those of two state-of-the-art map-matching algorithms. The experimental results confirm the effectiveness of the proposed algorithm.
{"title":"CMMTSE: Complex Road Network Map Matching Based on Trajectory Structure Extraction","authors":"Xiaohan Wang, Pei Wang, Jing Wang, Yonglong Luo, Jiaqing Chen, Junze Wu","doi":"10.1007/s10489-024-05751-0","DOIUrl":"10.1007/s10489-024-05751-0","url":null,"abstract":"<div><p>Trajectory mapping onto a road network is a complex yet important task. This is because, in the presence of circular sections, Y-shaped intersections, and sections with elevated overlaps with the ground, the conditions of road networks become complicated. Consequently, trajectory mapping becomes challenging owing to the complexities of road networks and the noise generated by high positioning errors. In this study, in response to the difficulty in handling redundant noisy trajectory data in complex road network environments, a complex road network map-matching method based on trajectory structure extraction is proposed. The features of the structure are extracted from the original trajectory data to reduce the effects of redundancy and noise on matching. An adaptive screening candidate method is proposed using driver behavior to estimate the road density and reduce the matching time by selecting effective candidates. A spatiotemporal analysis function is redefined using speed and distance features, and a directional analysis function is proposed for use in combination with directional features to improve the matching accuracy of complex road networks. An experimental evaluation based on real-ground trajectory data collected using in-vehicle sensing devices is conducted to verify the effectiveness of the algorithm. Moreover, extensive experiments are performed on challenging real datasets to evaluate the proposed method, and its accuracy and efficiency are compared with those of two state-of-the-art map-matching algorithms. The experimental results confirm the effectiveness of the proposed algorithm.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12676 - 12696"},"PeriodicalIF":3.4,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-27DOI: 10.1007/s10489-024-05826-y
Duc Nguyen, Bac Le
High-utility itemset mining (HUIM) is a technique for extracting valuable insights from data. When dealing with sensitive information, HUIM can raise privacy concerns. As a result, privacy-preserving utility mining (PPUM) has become an important research direction. PPUM involves transforming quantitative transactional databases into sanitized versions that protect sensitive data while retaining useful patterns. Researchers have previously employed stochastic optimization methods to conceal sensitive patterns in databases through the addition or deletion of transactions. However, these approaches alter the database structure. To address this issue, this paper introduces a novel approach for hiding data with stochastic optimization without changing the database structure. We design a flexible objective function to let users restrict the negative effects of PPUM according to their specific requirements. We also develop a general strategy for establishing constraint matrices. In addition, we present a stochastic algorithm that applies the ant lion optimizer along with a hybrid algorithm, which combines both exact and stochastic optimization methods, to resolve the hiding problem. The results of extensive experiments are presented, demonstrating the efficiency and flexibility of the proposed algorithms.
{"title":"Novel stochastic algorithms for privacy-preserving utility mining","authors":"Duc Nguyen, Bac Le","doi":"10.1007/s10489-024-05826-y","DOIUrl":"10.1007/s10489-024-05826-y","url":null,"abstract":"<div><p>High-utility itemset mining (HUIM) is a technique for extracting valuable insights from data. When dealing with sensitive information, HUIM can raise privacy concerns. As a result, privacy-preserving utility mining (PPUM) has become an important research direction. PPUM involves transforming quantitative transactional databases into sanitized versions that protect sensitive data while retaining useful patterns. Researchers have previously employed stochastic optimization methods to conceal sensitive patterns in databases through the addition or deletion of transactions. However, these approaches alter the database structure. To address this issue, this paper introduces a novel approach for hiding data with stochastic optimization without changing the database structure. We design a flexible objective function to let users restrict the negative effects of PPUM according to their specific requirements. We also develop a general strategy for establishing constraint matrices. In addition, we present a stochastic algorithm that applies the ant lion optimizer along with a hybrid algorithm, which combines both exact and stochastic optimization methods, to resolve the hiding problem. The results of extensive experiments are presented, demonstrating the efficiency and flexibility of the proposed algorithms.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12725 - 12741"},"PeriodicalIF":3.4,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-27DOI: 10.1007/s10489-024-05732-3
Yingying Liang, Tianyu Zhang, Yan Tu, Qian Zhao
In real-world complex group decision-making problems, preference inconsistency and opinion conflict are common and crucial challenges that need to be tackled. To promote consensus reaching, a novel group consensus reaching model is constructed considering individual satisfaction and group fairness. This study focuses on managing the group consensus reaching process based on flexible and adaptable information, modelled as distributed linguistic preference relations (DLPRs). First, a building process for DLPRs is discussed by integrating cumulative distribution functions converted from single linguistic term sets, hesitant fuzzy linguistic term sets, and comparative linguistic expressions. Furthermore, a two-stage consistency improvement method is proposed, which makes a trade-off between the frequency and magnitude of adjustments. Finally, we establish an improved group consensus model to balance individual satisfaction and group fairness, where individual satisfaction is measured by the deviation between the original and revised preferences and group fairness is measured by the deviation between individual satisfactions. The emergency response plan selection is conducted to show the validity and advantages of the proposed approach.
{"title":"A group consensus reaching model balancing individual satisfaction and group fairness for distributed linguistic preference relations","authors":"Yingying Liang, Tianyu Zhang, Yan Tu, Qian Zhao","doi":"10.1007/s10489-024-05732-3","DOIUrl":"10.1007/s10489-024-05732-3","url":null,"abstract":"<div><p>In real-world complex group decision-making problems, preference inconsistency and opinion conflict are common and crucial challenges that need to be tackled. To promote consensus reaching, a novel group consensus reaching model is constructed considering individual satisfaction and group fairness. This study focuses on managing the group consensus reaching process based on flexible and adaptable information, modelled as distributed linguistic preference relations (DLPRs). First, a building process for DLPRs is discussed by integrating cumulative distribution functions converted from single linguistic term sets, hesitant fuzzy linguistic term sets, and comparative linguistic expressions. Furthermore, a two-stage consistency improvement method is proposed, which makes a trade-off between the frequency and magnitude of adjustments. Finally, we establish an improved group consensus model to balance individual satisfaction and group fairness, where individual satisfaction is measured by the deviation between the original and revised preferences and group fairness is measured by the deviation between individual satisfactions. The emergency response plan selection is conducted to show the validity and advantages of the proposed approach.</p><h3>Graphical Abstract</h3><div><figure><div><div><picture><img></picture></div></div></figure></div></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12697 - 12724"},"PeriodicalIF":3.4,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-26DOI: 10.1007/s10489-024-05785-4
Hyunsung Kim, Seonghyeon Ko, Junghyun Bum, Duc-Tai Le, Hyunseung Choo
Radiologists often inspect hundreds of two-dimensional computed-tomography (CT) images to accurately locate lesions and make diagnoses, by classifying and labeling the ribs. However, this task is repetitive and time consuming. To effectively address this problem, we propose a multi-axial rib segmentation and sequential labeling (MARSS) method. First, we slice the CT volume into sagittal, frontal, and transverse planes for segmentation. The segmentation masks generated for each plane are then reconstructed into a single 3D segmentation mask using binarization techniques. After separating the left and right rib volumes from the entire CT volume, we cluster the connected components identified as bones and sequentially assign labels to each rib. The segmentation and sequential labeling performance of this method outperformed existing methods by up to 4.2%. The proposed automatic rib sequential labeling method enhances the efficiency of radiologists. In addition, this method provides an extended opportunity for advancements not only in rib segmentation but also in bone-fracture detection and lesion-diagnosis research.
{"title":"Automatic rib segmentation and sequential labeling via multi-axial slicing and 3D reconstruction","authors":"Hyunsung Kim, Seonghyeon Ko, Junghyun Bum, Duc-Tai Le, Hyunseung Choo","doi":"10.1007/s10489-024-05785-4","DOIUrl":"10.1007/s10489-024-05785-4","url":null,"abstract":"<div><p>Radiologists often inspect hundreds of two-dimensional computed-tomography (CT) images to accurately locate lesions and make diagnoses, by classifying and labeling the ribs. However, this task is repetitive and time consuming. To effectively address this problem, we propose a multi-axial rib segmentation and sequential labeling (MARSS) method. First, we slice the CT volume into sagittal, frontal, and transverse planes for segmentation. The segmentation masks generated for each plane are then reconstructed into a single 3D segmentation mask using binarization techniques. After separating the left and right rib volumes from the entire CT volume, we cluster the connected components identified as bones and sequentially assign labels to each rib. The segmentation and sequential labeling performance of this method outperformed existing methods by up to 4.2%. The proposed automatic rib sequential labeling method enhances the efficiency of radiologists. In addition, this method provides an extended opportunity for advancements not only in rib segmentation but also in bone-fracture detection and lesion-diagnosis research.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12644 - 12660"},"PeriodicalIF":3.4,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-26DOI: 10.1007/s10489-024-05834-y
Haodong Cheng, Yingchi Mao, Xiao Jia
Physics-informed spatial-temporal discrete sequence learning networks have great potential in solving partial differential equations and time series prediction compared to traditional fully connected PINN algorithms, and can serve as the foundation for data-driven sequence prediction modeling and inverse problem analysis. However, such existing models are unable to deal with inverse problem scenarios in which the parameters of the physical process are time-varying and unknown, while usually failing to make predictions in continuous time. In this paper, we propose a continuous time series prediction algorithm constructed by the physics-informed graph neural ordinary differential equation (PGNODE). Proposed parameterized GNODE-GRU and physics-informed loss constraints are used to explicitly characterize and solve unknown time-varying hyperparameters. The GNODE solver integrates this physical parameter to predict the sequence value at any time. This paper uses epidemic prediction tasks as a case study, and experimental results demonstrate that the proposed algorithm can effectively improve the prediction accuracy of the spread of epidemics in the future continuous time.
{"title":"A framework based on physics-informed graph neural ODE: for continuous spatial-temporal pandemic prediction","authors":"Haodong Cheng, Yingchi Mao, Xiao Jia","doi":"10.1007/s10489-024-05834-y","DOIUrl":"10.1007/s10489-024-05834-y","url":null,"abstract":"<div><p>Physics-informed spatial-temporal discrete sequence learning networks have great potential in solving partial differential equations and time series prediction compared to traditional fully connected PINN algorithms, and can serve as the foundation for data-driven sequence prediction modeling and inverse problem analysis. However, such existing models are unable to deal with inverse problem scenarios in which the parameters of the physical process are time-varying and unknown, while usually failing to make predictions in continuous time. In this paper, we propose a continuous time series prediction algorithm constructed by the physics-informed graph neural ordinary differential equation (PGNODE). Proposed parameterized GNODE-GRU and physics-informed loss constraints are used to explicitly characterize and solve unknown time-varying hyperparameters. The GNODE solver integrates this physical parameter to predict the sequence value at any time. This paper uses epidemic prediction tasks as a case study, and experimental results demonstrate that the proposed algorithm can effectively improve the prediction accuracy of the spread of epidemics in the future continuous time.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12661 - 12675"},"PeriodicalIF":3.4,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-25DOI: 10.1007/s10489-024-05841-z
Yongping Du, Runfeng Xie, Bochao Zhang, Zihao Yin
Multimodal aspect-based sentiment analysis (MABSA) aims to predict the sentiment of aspect by the fusion of different modalities such as image, text and so on. However, the availability of high-quality multimodal data remains limited. Therefore, few-shot MABSA is a new challenge. Previous works are rarely able to cope with low-resource and few-shot scenarios. In order to address the above problems, we design a Few-shot Multimodal aspect-based sentiment analysis framework based on Contrastive Finetuning (FMCF). Initially, the image modality is transformed to the corresponding textual caption to achieve the entailed semantic information and a contrastive dataset is constructed based on similarity retrieval for finetuning in the following stage. Further, a sentence encoder is trained based on SBERT, which combines supervised contrastive learning and sentence-level multi-feature fusion to complete MABSA. The experiments demonstrate that our framework achieves excellent performance in the few-shot scenarios. Importantly, with only 256 training samples and limited computational resources, the proposed method outperforms fine-tuned models that use all available data on the Twitter dataset.
{"title":"FMCF: Few-shot Multimodal aspect-based sentiment analysis framework based on Contrastive Finetuning","authors":"Yongping Du, Runfeng Xie, Bochao Zhang, Zihao Yin","doi":"10.1007/s10489-024-05841-z","DOIUrl":"10.1007/s10489-024-05841-z","url":null,"abstract":"<div><p>Multimodal aspect-based sentiment analysis (MABSA) aims to predict the sentiment of aspect by the fusion of different modalities such as image, text and so on. However, the availability of high-quality multimodal data remains limited. Therefore, few-shot MABSA is a new challenge. Previous works are rarely able to cope with low-resource and few-shot scenarios. In order to address the above problems, we design a <b>F</b>ew-shot <b>M</b>ultimodal aspect-based sentiment analysis framework based on <b>C</b>ontrastive <b>F</b>inetuning (FMCF). Initially, the image modality is transformed to the corresponding textual caption to achieve the entailed semantic information and a contrastive dataset is constructed based on similarity retrieval for finetuning in the following stage. Further, a sentence encoder is trained based on SBERT, which combines supervised contrastive learning and sentence-level multi-feature fusion to complete MABSA. The experiments demonstrate that our framework achieves excellent performance in the few-shot scenarios. Importantly, with only 256 training samples and limited computational resources, the proposed method outperforms fine-tuned models that use all available data on the Twitter dataset.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12629 - 12643"},"PeriodicalIF":3.4,"publicationDate":"2024-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-24DOI: 10.1007/s10489-024-05808-0
Francisco de Arriba-Pérez, Silvia García-Méndez, Javier Otero-Mosquera, Francisco J. González-Castaño
Cognitive and neurological impairments are very common, but only a small proportion of affected individuals are diagnosed and treated, partly because of the high costs associated with frequent screening. Detecting pre-illness stages and analyzing the progression of neurological disorders through effective and efficient intelligent systems can be beneficial for timely diagnosis and early intervention. We propose using Large Language Models to extract features from free dialogues to detect cognitive decline. These features comprise high-level reasoning content-independent features (such as comprehension, decreased awareness, increased distraction, and memory problems). Our solution comprises (i) preprocessing, (ii) feature engineering via Natural Language Processing techniques and prompt engineering, (iii) feature analysis and selection to optimize performance, and (iv) classification, supported by automatic explainability. We also explore how to improve Chatgpt’s direct cognitive impairment prediction capabilities using the best features in our models. Evaluation metrics obtained endorse the effectiveness of a mixed approach combining feature extraction with Chatgpt and a specialized Machine Learning model to detect cognitive decline within free-form conversational dialogues with older adults. Ultimately, our work may facilitate the development of an inexpensive, non-invasive, and rapid means of detecting and explaining cognitive decline.
{"title":"Explainable cognitive decline detection in free dialogues with a Machine Learning approach based on pre-trained Large Language Models","authors":"Francisco de Arriba-Pérez, Silvia García-Méndez, Javier Otero-Mosquera, Francisco J. González-Castaño","doi":"10.1007/s10489-024-05808-0","DOIUrl":"10.1007/s10489-024-05808-0","url":null,"abstract":"<div><p>Cognitive and neurological impairments are very common, but only a small proportion of affected individuals are diagnosed and treated, partly because of the high costs associated with frequent screening. Detecting pre-illness stages and analyzing the progression of neurological disorders through effective and efficient intelligent systems can be beneficial for timely diagnosis and early intervention. We propose using Large Language Models to extract features from free dialogues to detect cognitive decline. These features comprise high-level reasoning content-independent features (such as comprehension, decreased awareness, increased distraction, and memory problems). Our solution comprises (<i>i</i>) preprocessing, (<i>ii</i>) feature engineering via Natural Language Processing techniques and prompt engineering, (<i>iii</i>) feature analysis and selection to optimize performance, and (<i>iv</i>) classification, supported by automatic explainability. We also explore how to improve Chat<span>gpt</span>’s direct cognitive impairment prediction capabilities using the best features in our models. Evaluation metrics obtained endorse the effectiveness of a mixed approach combining feature extraction with Chat<span>gpt</span> and a specialized Machine Learning model to detect cognitive decline within free-form conversational dialogues with older adults. Ultimately, our work may facilitate the development of an inexpensive, non-invasive, and rapid means of detecting and explaining cognitive decline.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"54 24","pages":"12613 - 12628"},"PeriodicalIF":3.4,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}