首页 > 最新文献

Applied Soft Computing最新文献

英文 中文
LesionMix data enhancement and entropy minimization for semi-supervised lesion segmentation of lung cancer 用于肺癌半监督病灶分割的 LesionMix 数据增强和熵最小化技术
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-23 DOI: 10.1016/j.asoc.2024.112244
Determining the location and contour of the lesion is a crucial prerequisite for medical diagnosis, subsequent personalized treatment plan and prognostic prediction of lung cancer. Semi-supervised learning and data augmentation methods facilitate deep learning to be used in many fields of medical imaging. In this paper, we introduce a novel data enhancement technique called LesionMix. This method involves extracting and reusing lesions from a limited amount of labeled CT data, thereby enhancing the efficiency of utilizing those labeled data. Meanwhile, we propose a two-stage semi-supervised training strategy called Entropy Minimization LesionMix (EMLM). In the first stage, features containing lesion contour information are rapidly learned through LesionMix data augmentation. Entropy minimization strategy optimizes the model parameters to alleviate overfitting as much as possible in the second stage and improves prediction confidence. Our proposed method is validated on the public dataset LIDC-IDRI and the in-house dataset GDPHLUAD. Extensive experiments demonstrate that our method achieves promising performance and outperforms seven state-of-the-art semi-supervised models; Moreover, ablation experiments validate the effectiveness of various aspects of our approach.
确定病灶的位置和轮廓是医学诊断、后续个性化治疗方案和肺癌预后预测的重要前提。半监督学习和数据增强方法促进了深度学习在医学影像诸多领域的应用。在本文中,我们介绍了一种名为 LesionMix 的新型数据增强技术。该方法涉及从有限的标注 CT 数据中提取病灶并重复使用,从而提高这些标注数据的使用效率。同时,我们提出了一种名为熵最小化病灶混合(Entropy Minimization LesionMix,EMLM)的两阶段半监督训练策略。在第一阶段,通过 LesionMix 数据扩增快速学习包含病变轮廓信息的特征。在第二阶段,熵最小化策略会优化模型参数,尽可能减少过拟合,提高预测置信度。我们提出的方法在公共数据集 LIDC-IDRI 和内部数据集 GDPHLUAD 上进行了验证。广泛的实验证明,我们的方法取得了可喜的成绩,优于七个最先进的半监督模型;此外,消融实验也验证了我们方法各方面的有效性。
{"title":"LesionMix data enhancement and entropy minimization for semi-supervised lesion segmentation of lung cancer","authors":"","doi":"10.1016/j.asoc.2024.112244","DOIUrl":"10.1016/j.asoc.2024.112244","url":null,"abstract":"<div><div>Determining the location and contour of the lesion is a crucial prerequisite for medical diagnosis, subsequent personalized treatment plan and prognostic prediction of lung cancer. Semi-supervised learning and data augmentation methods facilitate deep learning to be used in many fields of medical imaging. In this paper, we introduce a novel data enhancement technique called LesionMix. This method involves extracting and reusing lesions from a limited amount of labeled CT data, thereby enhancing the efficiency of utilizing those labeled data. Meanwhile, we propose a two-stage semi-supervised training strategy called Entropy Minimization LesionMix (EMLM). In the first stage, features containing lesion contour information are rapidly learned through LesionMix data augmentation. Entropy minimization strategy optimizes the model parameters to alleviate overfitting as much as possible in the second stage and improves prediction confidence. Our proposed method is validated on the public dataset LIDC-IDRI and the in-house dataset GDPHLUAD. Extensive experiments demonstrate that our method achieves promising performance and outperforms seven state-of-the-art semi-supervised models; Moreover, ablation experiments validate the effectiveness of various aspects of our approach.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fuzzy Self-tuning Bees Algorithm for designing optimal product lines 设计最佳生产线的模糊自调整蜜蜂算法
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-20 DOI: 10.1016/j.asoc.2024.112228
The Product Line Design (PLD) problem is an NP-hard combinatorial optimization problem in marketing that aims at determining an optimal product line through which a firm can optimize a desired objective, like its profits or market share. Since the PLD problem has been proved to have high complexity in real-life applications, high-quality solutions have been detected by researchers who develop various optimization methods and test their performance. The Bees Algorithm (BA) is a successful swarm intelligent optimization algorithm which is based on the behavior of bees. The aim of this research is to develop and assess BA in the optimal PLD problem. In this effort, a set of fuzzy rules has been developed to autonomously compute parameters for each individual solution throughout the optimization process. The performance of two BA variants is compared with those of popular previous approaches, using both real and simulated data of customer preferences. The findings reveal that BA constitutes an enhanced alternative approach for designing optimal product lines.
产品线设计(PLD)问题是市场营销中的一个 NP 难组合优化问题,其目的是确定一个最佳产品线,通过该产品线,企业可以优化其预期目标,如利润或市场份额。由于 PLD 问题在实际应用中被证明具有很高的复杂性,因此研究人员开发了各种优化方法并测试其性能,从而找到了高质量的解决方案。蜜蜂算法(BA)是一种成功的蜂群智能优化算法,它以蜜蜂的行为为基础。本研究的目的是开发和评估用于优化 PLD 问题的 BA。在此过程中,开发了一套模糊规则,可在整个优化过程中自主计算每个单独解决方案的参数。利用客户偏好的真实数据和模拟数据,将两种 BA 变体的性能与之前流行的方法进行了比较。研究结果表明,BA 是设计最佳产品线的一种增强型替代方法。
{"title":"Fuzzy Self-tuning Bees Algorithm for designing optimal product lines","authors":"","doi":"10.1016/j.asoc.2024.112228","DOIUrl":"10.1016/j.asoc.2024.112228","url":null,"abstract":"<div><div>The Product Line Design (PLD) problem is an NP-hard combinatorial optimization problem in marketing that aims at determining an optimal product line through which a firm can optimize a desired objective, like its profits or market share. Since the PLD problem has been proved to have high complexity in real-life applications, high-quality solutions have been detected by researchers who develop various optimization methods and test their performance. The Bees Algorithm (BA) is a successful swarm intelligent optimization algorithm which is based on the behavior of bees. The aim of this research is to develop and assess BA in the optimal PLD problem. In this effort, a set of fuzzy rules has been developed to autonomously compute parameters for each individual solution throughout the optimization process. The performance of two BA variants is compared with those of popular previous approaches, using both real and simulated data of customer preferences. The findings reveal that BA constitutes an enhanced alternative approach for designing optimal product lines.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Neighbors Graph Learning for Large-Scale Data Clustering using Vector Quantization and Self-Regularization 利用矢量量化和自规则化进行大规模数据聚类的自适应邻域图学习
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-19 DOI: 10.1016/j.asoc.2024.112256
In traditional adaptive neighbors graph learning (ANGL)-based clustering, the time complexity is more than O(n2), where n is the number of data points, which is not scalable for large-scale data problems in real applications. Subsequently, ANGL adds a balance regularization to its objective function to avoid the sparse over-fitting problem in the learned similarity graph matrix. Still, the regularization may leads to many weak connections between data points in different clusters. To address these problems, we propose a new fast clustering method, namely, Adaptive Neighbors Graph Learning for Large-Scale Data Clustering using Vector Quantization and Self-Regularization (ANGL-LDC), to perform vector quantization (VQ) on original data and feed the obtained VQ data as the input in the n×n similarity graph matrix learning. Hence, the n×n similarity graph matrix learning problem is simplified to weighted m×m (mn) graph learning problem, where m is the number of distinct points and weight is the duplicate times of distinct points in VQ data. Consequently, the time complexity of ANGL-LDC is much lower than that of ANGL. At the same time, we propose a new ANGL objective function with a graph connection self-regularization mechanism, where the ANGL-LDC objective function will get an infinity value if the value of one graph connection is equal to 1. Therefore, ANGL-LDC naturally avoids obtaining the sparse over-fitting problem since we need to minimize the value of ANGL-LDC’s objective function. Experimental results on synthetic and real-world datasets demonstrate the scalability and effectiveness of ANGL-LDC.
在传统的基于自适应邻接图学习(ANGL)的聚类中,时间复杂度大于 O(n2),其中 n 是数据点的数量,这对于实际应用中的大规模数据问题来说是不可扩展的。随后,ANGL 在其目标函数中加入了平衡正则化,以避免学习到的相似性图矩阵中的稀疏过拟合问题。然而,正则化可能会导致不同聚类中的数据点之间存在许多弱连接。为了解决这些问题,我们提出了一种新的快速聚类方法,即使用矢量量化和自规整的大规模数据聚类自适应邻域图学习(ANGL-LDC),对原始数据进行矢量量化(VQ),并将获得的 VQ 数据作为 n×n 相似性图矩阵学习的输入。因此,n×n 相似性图矩阵学习问题被简化为加权 m×m (m≪n) 图学习问题,其中 m 是不同点的数量,权重是 VQ 数据中不同点的重复次数。因此,ANGL-LDC 的时间复杂度远远低于 ANGL。同时,我们提出了一种具有图连接自规则化机制的新 ANGL 目标函数,即如果一个图连接的值等于 1,ANGL-LDC 目标函数的值将为无穷大。因此,由于我们需要最小化 ANGL-LDC 目标函数的值,ANGL-LDC 自然避免了获得稀疏过拟合问题。在合成数据集和真实数据集上的实验结果证明了 ANGL-LDC 的可扩展性和有效性。
{"title":"Adaptive Neighbors Graph Learning for Large-Scale Data Clustering using Vector Quantization and Self-Regularization","authors":"","doi":"10.1016/j.asoc.2024.112256","DOIUrl":"10.1016/j.asoc.2024.112256","url":null,"abstract":"<div><div>In traditional adaptive neighbors graph learning (ANGL)-based clustering, the time complexity is more than <span><math><mrow><mi>O</mi><mrow><mo>(</mo><msup><mrow><mi>n</mi></mrow><mrow><mn>2</mn></mrow></msup><mo>)</mo></mrow></mrow></math></span>, where <span><math><mi>n</mi></math></span> is the number of data points, which is not scalable for large-scale data problems in real applications. Subsequently, ANGL adds a balance regularization to its objective function to avoid the sparse over-fitting problem in the learned similarity graph matrix. Still, the regularization may leads to many weak connections between data points in different clusters. To address these problems, we propose a new fast clustering method, namely, <strong>A</strong>daptive <strong>N</strong>eighbors <strong>G</strong>raph <strong>L</strong>earning for <strong>L</strong>arge-Scale <strong>D</strong>ata <strong>C</strong>lustering using Vector Quantization and Self-Regularization (ANGL-LDC), to perform vector quantization (VQ) on original data and feed the obtained VQ data as the input in the <span><math><mrow><mi>n</mi><mo>×</mo><mi>n</mi></mrow></math></span> similarity graph matrix learning. Hence, the <span><math><mrow><mi>n</mi><mo>×</mo><mi>n</mi></mrow></math></span> similarity graph matrix learning problem is simplified to weighted <span><math><mrow><mi>m</mi><mo>×</mo><mi>m</mi></mrow></math></span> <span><math><mrow><mo>(</mo><mi>m</mi><mo>≪</mo><mi>n</mi><mo>)</mo></mrow></math></span> graph learning problem, where <span><math><mi>m</mi></math></span> is the number of distinct points and weight is the duplicate times of distinct points in VQ data. Consequently, the time complexity of ANGL-LDC is much lower than that of ANGL. At the same time, we propose a new ANGL objective function with a graph connection self-regularization mechanism, where the ANGL-LDC objective function will get an infinity value if the value of one graph connection is equal to 1. Therefore, ANGL-LDC naturally avoids obtaining the sparse over-fitting problem since we need to minimize the value of ANGL-LDC’s objective function. Experimental results on synthetic and real-world datasets demonstrate the scalability and effectiveness of ANGL-LDC.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Competitive swarm optimizer with dynamic multi-competitions and convergence accelerator for large-scale optimization problems 针对大规模优化问题的具有动态多重竞争和收敛加速器的竞争群优化器
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-19 DOI: 10.1016/j.asoc.2024.112252
Large-scale optimizations (LSOPs) with high dimensional decision variables have become one of the most challenging problems in engineering optimization. High dimensional information causes serious interference to the algorithm optimization performance. The optimization performance of the algorithms will be seriously degraded. Competitive swarm optimizer (CSO) is a robust algorithm to tackle LSOPs. However, CSO randomly selects two particles to compare, then generates the winner and the loser. Although this search mechanism can enhance the diversity of the swarm, a single comparison is difficult to guarantee the quality of winners and losers. Therefore, there exists a risk of producing unqualified solutions. In order to enhance the quality of solution, a novel CSO with dynamic multi-competitions and convergence accelerator, namely DMCACSO, is designed in this paper. In the DMCACSO, a dynamic multi-competitions based evolutionary information is designed to pick out the losers more efficiently and improve the quality of winners. In addition, a convergence accelerator with hybrid evolutionary strategy is developed to speed up the particle search when the algorithm is a state of stagnation. The experiment results in solving large-scale benchmark functions from CEC2010 and CEC2013 indicate that the DMCACSO has competitive optimization performance by comparing with some state-of-the-art algorithms. Finally, the DMCACSO is effective in terms of quality in solving an actual feature selection problem.
具有高维决策变量的大规模优化(LSOP)已成为工程优化领域最具挑战性的问题之一。高维信息对算法优化性能造成严重干扰。算法的优化性能将严重下降。竞争群优化器(CSO)是解决 LSOPs 的一种稳健算法。不过,CSO 是随机选择两个粒子进行比较,然后产生胜者和败者。虽然这种搜索机制可以增强蜂群的多样性,但单次比较很难保证胜者和败者的质量。因此,存在产生不合格解决方案的风险。为了提高解的质量,本文设计了一种具有动态多重竞争和收敛加速器的新型 CSO,即 DMCACSO。在 DMCACSO 中,设计了一种基于动态多重竞争的进化信息,以更有效地挑选出失败者,并提高获胜者的质量。此外,还开发了混合进化策略的收敛加速器,以在算法处于停滞状态时加速粒子搜索。在求解 CEC2010 和 CEC2013 大型基准函数的实验结果表明,与一些最先进的算法相比,DMCACSO 具有极具竞争力的优化性能。最后,DMCACSO 在解决实际的特征选择问题时具有良好的质量。
{"title":"Competitive swarm optimizer with dynamic multi-competitions and convergence accelerator for large-scale optimization problems","authors":"","doi":"10.1016/j.asoc.2024.112252","DOIUrl":"10.1016/j.asoc.2024.112252","url":null,"abstract":"<div><div>Large-scale optimizations (LSOPs) with high dimensional decision variables have become one of the most challenging problems in engineering optimization. High dimensional information causes serious interference to the algorithm optimization performance. The optimization performance of the algorithms will be seriously degraded. Competitive swarm optimizer (CSO) is a robust algorithm to tackle LSOPs. However, CSO randomly selects two particles to compare, then generates the winner and the loser. Although this search mechanism can enhance the diversity of the swarm, a single comparison is difficult to guarantee the quality of winners and losers. Therefore, there exists a risk of producing unqualified solutions. In order to enhance the quality of solution, a novel CSO with dynamic multi-competitions and convergence accelerator, namely DMCACSO, is designed in this paper. In the DMCACSO, a dynamic multi-competitions based evolutionary information is designed to pick out the losers more efficiently and improve the quality of winners. In addition, a convergence accelerator with hybrid evolutionary strategy is developed to speed up the particle search when the algorithm is a state of stagnation. The experiment results in solving large-scale benchmark functions from CEC2010 and CEC2013 indicate that the DMCACSO has competitive optimization performance by comparing with some state-of-the-art algorithms. Finally, the DMCACSO is effective in terms of quality in solving an actual feature selection problem.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sparse large-scale high-order fuzzy cognitive maps guided by spearman correlation coefficient 以矛曼相关系数为指导的稀疏大规模高阶模糊认知图谱
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1016/j.asoc.2024.112253

Time series prediction is one of the most important applications of Fuzzy Cognitive Maps (FCMs). In general, the state of FCMs in forecasting depends only on the state of the previous moment, but in fact it is also affected by the past state. Hence Higher-Order Fuzzy Cognitive Maps (HFCMs) are proposed based on FCMs considering historical information and have been widely used for time series forecasting. However, using HFCMs to deal with sparse and large-scale multivariate time series are still a challenge, while large-scale data makes it difficult to determine the causal relationship between nodes because of the increased number of nodes, so it is necessary to explore the relationship between nodes to guide large-scale HFCMs learning. Therefore, a sparse large-scale HFCMs learning algorithm guided by Spearman correlation coefficient, called SG-HFCM, is proposed in the paper. The SG-HFCM model is specified as follows: First, the solving of HFCMs model is transform into a regression model and an adaptive loss function is utilized to enhance the robustness of the model. Second, l1-norm is used to improve the sparsity of the weight matrix. Third, in order to more accurately characterize the correlation relationship between variables, the Spearman correlation coefficients is added as a regular term to guide the learning of weight matrices. When calculating the Spearman correlation coefficient, through splitting domain interval method, we can better understand the characteristics of the data, and get better correlation in different small intervals, and more accurately characterize the relationship between the variables in order to guide the weight matrix. In addition, the Alternating Direction Multiplication Method and quadratic programming method are used to solve the algorithms to get better solutions for the SG-HFCM, where the quadratic programming can well ensure that the range of the weights and obtaining the optimal solution. Finally, by comparing with five algorithms, the SG-HFCM model showed an average improvement of 11.93% in prediction accuracy for GRNs, indicating that our proposed model has good predictive performance.

时间序列预测是模糊认知图(FCM)最重要的应用之一。一般来说,FCMs 在预测中的状态只取决于前一时刻的状态,但事实上它也受到过去状态的影响。因此,在考虑历史信息的 FCM 基础上提出了高阶模糊认知图(HFCM),并被广泛用于时间序列预测。然而,使用 HFCMs 处理稀疏和大规模的多元时间序列仍是一个挑战,而大规模数据由于节点数量的增加,很难确定节点之间的因果关系,因此有必要探索节点之间的关系,以指导大规模 HFCMs 的学习。因此,本文提出了一种以斯皮尔曼相关系数为指导的稀疏大规模 HFCMs 学习算法,称为 SG-HFCM。SG-HFCM 模型具体如下:首先,将 HFCMs 模型的求解转化为回归模型,并利用自适应损失函数来增强模型的鲁棒性。其次,利用 l1-norm 来改善权重矩阵的稀疏性。第三,为了更准确地描述变量之间的相关关系,加入了斯皮尔曼相关系数作为正则项来指导权重矩阵的学习。在计算斯皮尔曼相关系数时,通过分割域区间的方法,可以更好地了解数据的特点,在不同的小区间内得到更好的相关性,更准确地表征变量之间的关系,从而指导权重矩阵的学习。此外,利用交替方向乘法和二次编程法对算法进行求解,得到 SG-HFCM 的较好解,其中二次编程法可以很好地保证权重的范围,获得最优解。最后,通过与五种算法的比较,SG-HFCM 模型对 GRN 的预测准确率平均提高了 11.93%,表明我们提出的模型具有良好的预测性能。
{"title":"Sparse large-scale high-order fuzzy cognitive maps guided by spearman correlation coefficient","authors":"","doi":"10.1016/j.asoc.2024.112253","DOIUrl":"10.1016/j.asoc.2024.112253","url":null,"abstract":"<div><p>Time series prediction is one of the most important applications of Fuzzy Cognitive Maps (FCMs). In general, the state of FCMs in forecasting depends only on the state of the previous moment, but in fact it is also affected by the past state. Hence Higher-Order Fuzzy Cognitive Maps (HFCMs) are proposed based on FCMs considering historical information and have been widely used for time series forecasting. However, using HFCMs to deal with sparse and large-scale multivariate time series are still a challenge, while large-scale data makes it difficult to determine the causal relationship between nodes because of the increased number of nodes, so it is necessary to explore the relationship between nodes to guide large-scale HFCMs learning. Therefore, a sparse large-scale HFCMs learning algorithm guided by Spearman correlation coefficient, called SG-HFCM, is proposed in the paper. The SG-HFCM model is specified as follows: First, the solving of HFCMs model is transform into a regression model and an adaptive loss function is utilized to enhance the robustness of the model. Second, <span><math><msub><mrow><mi>l</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span>-norm is used to improve the sparsity of the weight matrix. Third, in order to more accurately characterize the correlation relationship between variables, the Spearman correlation coefficients is added as a regular term to guide the learning of weight matrices. When calculating the Spearman correlation coefficient, through splitting domain interval method, we can better understand the characteristics of the data, and get better correlation in different small intervals, and more accurately characterize the relationship between the variables in order to guide the weight matrix. In addition, the Alternating Direction Multiplication Method and quadratic programming method are used to solve the algorithms to get better solutions for the SG-HFCM, where the quadratic programming can well ensure that the range of the weights and obtaining the optimal solution. Finally, by comparing with five algorithms, the SG-HFCM model showed an average improvement of 11.93% in prediction accuracy for GRNs, indicating that our proposed model has good predictive performance.</p></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142272703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An implicit aspect-based sentiment analysis method using supervised contrastive learning and knowledge embedding 使用监督对比学习和知识嵌入的基于隐性方面的情感分析方法
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1016/j.asoc.2024.112233

Aspect-based sentiment analysis aims to analyze and understand people’s opinions from different aspects. Some comments do not contain explicit opinion words but still convey a clear human-perceived emotional orientation, which is known as implicit sentiment. Most previous research relies on contextual information from a text for implicit aspect-based sentiment analysis. However, little work has integrated external knowledge with contextual information. This paper proposes an implicit aspect-based sentiment analysis model combining supervised contrastive learning with knowledge-enhanced fine-tuning on BERT (BERT-SCL+KEFT). In the pre-training phase, the model utilizes supervised contrastive learning (SCL) on large-scale sentiment-annotated corpora to acquire sentiment knowledge. In the fine-tuning phase, the model uses a knowledge-enhanced fine-tuning (KEFT) method to capture explicit and implicit aspect-based sentiments. Specifically, the model utilizes knowledge embedding to embed external general knowledge information into textual entities by using knowledge graphs, enriching textual information. Finally, the model combines external knowledge and contextual features to predict the implicit sentiment in a text. The experimental results demonstrate that the proposed BERT-SCL+KEFT model outperforms other baselines on the general implicit sentiment analysis and implicit aspect-based sentiment analysis tasks. In addition, ablation experimental results show that the proposed BERT-SCL+KEFT model without the knowledge embedding module or supervised contrastive learning module significantly decreases performance, indicating the importance of these modules. All experiments validate that the proposed BERT-SCL+KEFT model effectively achieves implicit aspect-based sentiment classification.

基于方面的情感分析旨在从不同方面分析和理解人们的意见。有些评论并不包含明确的意见词,但仍能传达出明显的人类感知的情感取向,这就是所谓的隐性情感。以往的研究大多依赖文本中的上下文信息来进行基于内隐方面的情感分析。然而,很少有研究将外部知识与上下文信息相结合。本文提出了一种基于内隐方面的情感分析模型,它将有监督的对比学习与 BERT(BERT-SCL+KEFT)上的知识增强微调相结合。在预训练阶段,该模型利用大规模情感注释语料库上的监督对比学习(SCL)来获取情感知识。在微调阶段,该模型使用知识增强微调(KEFT)方法来捕捉基于方面的显性和隐性情感。具体来说,该模型利用知识嵌入技术,通过知识图谱将外部常识信息嵌入文本实体,从而丰富文本信息。最后,该模型结合外部知识和上下文特征来预测文本中的内隐情感。实验结果表明,在一般内隐情感分析和基于内隐方面的情感分析任务上,所提出的 BERT-SCL+KEFT 模型优于其他基线模型。此外,消融实验结果表明,拟议的 BERT-SCL+KEFT 模型在没有知识嵌入模块或监督对比学习模块的情况下性能明显下降,这表明了这些模块的重要性。所有实验都验证了所提出的 BERT-SCL+KEFT 模型能有效实现基于隐性方面的情感分类。
{"title":"An implicit aspect-based sentiment analysis method using supervised contrastive learning and knowledge embedding","authors":"","doi":"10.1016/j.asoc.2024.112233","DOIUrl":"10.1016/j.asoc.2024.112233","url":null,"abstract":"<div><p>Aspect-based sentiment analysis aims to analyze and understand people’s opinions from different aspects. Some comments do not contain explicit opinion words but still convey a clear human-perceived emotional orientation, which is known as implicit sentiment. Most previous research relies on contextual information from a text for implicit aspect-based sentiment analysis. However, little work has integrated external knowledge with contextual information. This paper proposes an implicit aspect-based sentiment analysis model combining supervised contrastive learning with knowledge-enhanced fine-tuning on BERT (BERT-SCL+KEFT). In the pre-training phase, the model utilizes supervised contrastive learning (SCL) on large-scale sentiment-annotated corpora to acquire sentiment knowledge. In the fine-tuning phase, the model uses a knowledge-enhanced fine-tuning (KEFT) method to capture explicit and implicit aspect-based sentiments. Specifically, the model utilizes knowledge embedding to embed external general knowledge information into textual entities by using knowledge graphs, enriching textual information. Finally, the model combines external knowledge and contextual features to predict the implicit sentiment in a text. The experimental results demonstrate that the proposed BERT-SCL+KEFT model outperforms other baselines on the general implicit sentiment analysis and implicit aspect-based sentiment analysis tasks. In addition, ablation experimental results show that the proposed BERT-SCL+KEFT model without the knowledge embedding module or supervised contrastive learning module significantly decreases performance, indicating the importance of these modules. All experiments validate that the proposed BERT-SCL+KEFT model effectively achieves implicit aspect-based sentiment classification.</p></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142272702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ELSNC: A semi-supervised community detection method with integration of embedding-enhanced links and node content in attributed networks ELSNC:一种整合了属性网络中嵌入增强链接和节点内容的半监督社区检测方法
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1016/j.asoc.2024.112250
In complex network analysis, detecting communities is becoming increasingly important. However, it is difficult to fuse multiple types of information to enhance the community-detection performance in real-world applications. Besides the nodes and the edges, a network also contains the structure of communities, its networking topological structure, and the network embeddings. Note that existing works on community detection have limited usage of all these information types in combination. In this work, we designed a novel unified model called embedding-enhanced link-based semi-supervised community detection with node content (ELSNC). ELSNC integrates the structure of the topology, the priori information, the network embeddings, and the node content. First, we employ two non-negative matrix factorization (NMF)–based stochastic models to characterize the node-community membership and the content-community membership (by performing similarity detection between a topic model and the NMF). Second, we introduce the nodes’ and networking embeddings’ topological similarity into the model as topological information. To model the topological similarity, we introduce a strong constraint (i.e., the priori information) and apply matrix completion to identify the community membership with the network embeddings’ representation ability. Finally, we present a semi-supervised community-detection method based on NMF that combines the network topology, content information, and the network embeddings. Our work’s innovation can be captured in two points: 1) As a type of semi-supervised community detection method, we extend the theory of semi-supervised methods on attributed networks and propose a unified model that integrates multiple information types. 2) The community membership obtained by the unified model simultaneously contains different information, including the topological, content, priori, and embedding information, which can more robustly be explored in the community structure in real-world scenarios. Furthermore, we performed a comprehensive evaluation of our proposed approach compared with state-of-the-art methods on both synthetic and real-world networks. The results show that our proposed method significantly outperformed the baseline methods.
在复杂网络分析中,社群检测变得越来越重要。然而,在实际应用中,很难融合多种类型的信息来提高社群检测性能。除了节点和边,网络还包含社群结构、网络拓扑结构和网络嵌入。需要注意的是,现有的社群检测工作对所有这些信息类型的组合使用非常有限。在这项工作中,我们设计了一种新颖的统一模型,称为 "基于节点内容的嵌入增强链接半监督社区检测(ELSNC)"。ELSNC 整合了拓扑结构、先验信息、网络嵌入和节点内容。首先,我们采用两个基于非负矩阵因式分解(NMF)的随机模型来描述节点-社区成员身份和内容-社区成员身份(通过在主题模型和 NMF 之间进行相似性检测)。其次,我们将节点和网络嵌入的拓扑相似性作为拓扑信息引入模型。为了对拓扑相似性进行建模,我们引入了一个强约束条件(即先验信息),并应用矩阵补全来识别具有网络嵌入表示能力的社群成员。最后,我们提出了一种基于 NMF 的半监督社区检测方法,该方法结合了网络拓扑结构、内容信息和网络嵌入。我们的工作有两点创新:1) 作为一种半监督社区检测方法,我们扩展了归属网络半监督方法的理论,提出了一种整合多种信息类型的统一模型。2) 统一模型得到的社群成员信息同时包含拓扑信息、内容信息、先验信息和嵌入信息等不同信息,可以更稳健地探索真实世界场景中的社群结构。此外,我们还在合成网络和真实世界网络上对我们提出的方法与最先进的方法进行了综合评估。结果表明,我们提出的方法明显优于基线方法。
{"title":"ELSNC: A semi-supervised community detection method with integration of embedding-enhanced links and node content in attributed networks","authors":"","doi":"10.1016/j.asoc.2024.112250","DOIUrl":"10.1016/j.asoc.2024.112250","url":null,"abstract":"<div><div>In complex network analysis, detecting communities is becoming increasingly important. However, it is difficult to fuse multiple types of information to enhance the community-detection performance in real-world applications. Besides the nodes and the edges, a network also contains the structure of communities, its networking topological structure, and the network embeddings. Note that existing works on community detection have limited usage of all these information types in combination. In this work, we designed a novel unified model called embedding-enhanced link-based semi-supervised community detection with node content (ELSNC). ELSNC integrates the structure of the topology, the priori information, the network embeddings, and the node content. First, we employ two non-negative matrix factorization (NMF)–based stochastic models to characterize the node-community membership and the content-community membership (by performing similarity detection between a topic model and the NMF). Second, we introduce the nodes’ and networking embeddings’ topological similarity into the model as topological information. To model the topological similarity, we introduce a strong constraint (<em>i</em>.<em>e</em>., the priori information) and apply matrix completion to identify the community membership with the network embeddings’ representation ability. Finally, we present a semi-supervised community-detection method based on NMF that combines the network topology, content information, and the network embeddings. Our work’s innovation can be captured in two points: 1) As a type of semi-supervised community detection method, we extend the theory of semi-supervised methods on attributed networks and propose a unified model that integrates multiple information types. 2) The community membership obtained by the unified model simultaneously contains different information, including the topological, content, priori, and embedding information, which can more robustly be explored in the community structure in real-world scenarios. Furthermore, we performed a comprehensive evaluation of our proposed approach compared with state-of-the-art methods on both synthetic and real-world networks. The results show that our proposed method significantly outperformed the baseline methods.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature radiance fields (FeRF): A multi-level feature fusion method with deep neural network for image synthesis 特征辐射场(FeRF):利用深度神经网络进行图像合成的多层次特征融合方法
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1016/j.asoc.2024.112262
Neural Radiance Field (NeRF) has brought revolutionary changes to the field of image synthesis with its unique ability to generate highly realistic multi-view consistent images from a neural scene representation. However, current NeRF-based methods still largely depend on multiple, precisely posed images, especially for complex or dynamic scenes, limiting their versatility. Furthermore, some recent strategies attempt to integrate simple feature extraction networks with volume rendering techniques to reduce multi-view dependence but create blurry outputs, highlighting the need for more sophisticated feature handling to unlock NeRF's full potential. In this paper, we propose an image synthesis method named FeRF, distinguished by its capacity to perform comprehensive feature extraction on individual unposed images and facilitate feature fusion at any stage. Additionally, we present an "elaborated-feature generation network" (EGN) composed of four modules, which is configured with two advanced feature extraction modules aimed at precisely refining and processing subtle, complex visual features from a single image. Given that the core objective of FeRF is the precise capture and processing of intricate features from the input images, we innovatively incorporated precisely designed attention mechanisms into the network architecture to optimize and highlight the importance of key feature attributes, thereby effectively enhancing their contribution to subsequent volume rendering processes. Extensive experimentation validates the outstanding qualitative and quantitative performance of our proposed network structure. In comparison to current image feature-based generalized image synthesis methods, it achieves superior reconstruction quality and level of detail.
神经辐照场(NeRF)具有从神经场景表征生成高度逼真的多视角一致图像的独特能力,为图像合成领域带来了革命性的变化。然而,目前基于 NeRF 的方法在很大程度上仍依赖于多张精确摆放的图像,尤其是复杂或动态场景,从而限制了其通用性。此外,最近的一些策略试图将简单的特征提取网络与体积渲染技术相结合,以减少对多视图的依赖,但却产生了模糊的输出结果,这突出表明需要更复杂的特征处理才能释放 NeRF 的全部潜力。在本文中,我们提出了一种名为 "FeRF "的图像合成方法,其特点是能够对单个未摆放的图像进行综合特征提取,并在任何阶段促进特征融合。此外,我们还提出了一个由四个模块组成的 "精细特征生成网络"(EGN),其中配置了两个高级特征提取模块,旨在精确提炼和处理单张图像中细微、复杂的视觉特征。鉴于 FeRF 的核心目标是从输入图像中精确捕捉和处理复杂的特征,我们创新性地将精确设计的关注机制纳入网络架构,以优化和突出关键特征属性的重要性,从而有效提高它们对后续体积渲染过程的贡献。广泛的实验验证了我们提出的网络结构在质量和数量上的卓越性能。与目前基于图像特征的广义图像合成方法相比,它能获得更高的重建质量和细节水平。
{"title":"Feature radiance fields (FeRF): A multi-level feature fusion method with deep neural network for image synthesis","authors":"","doi":"10.1016/j.asoc.2024.112262","DOIUrl":"10.1016/j.asoc.2024.112262","url":null,"abstract":"<div><div>Neural Radiance Field (NeRF) has brought revolutionary changes to the field of image synthesis with its unique ability to generate highly realistic multi-view consistent images from a neural scene representation. However, current NeRF-based methods still largely depend on multiple, precisely posed images, especially for complex or dynamic scenes, limiting their versatility. Furthermore, some recent strategies attempt to integrate simple feature extraction networks with volume rendering techniques to reduce multi-view dependence but create blurry outputs, highlighting the need for more sophisticated feature handling to unlock NeRF's full potential. In this paper, we propose an image synthesis method named FeRF, distinguished by its capacity to perform comprehensive feature extraction on individual unposed images and facilitate feature fusion at any stage. Additionally, we present an \"elaborated-feature generation network\" (EGN) composed of four modules, which is configured with two advanced feature extraction modules aimed at precisely refining and processing subtle, complex visual features from a single image. Given that the core objective of FeRF is the precise capture and processing of intricate features from the input images, we innovatively incorporated precisely designed attention mechanisms into the network architecture to optimize and highlight the importance of key feature attributes, thereby effectively enhancing their contribution to subsequent volume rendering processes. Extensive experimentation validates the outstanding qualitative and quantitative performance of our proposed network structure. In comparison to current image feature-based generalized image synthesis methods, it achieves superior reconstruction quality and level of detail.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large-scale multiple criteria group decision-making with information emendation based on unsupervised opinion evolutions 基于无监督意见演化的大规模多标准群体决策与信息修正
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1016/j.asoc.2024.112227

Large-scale multiple criteria group decision-making (MCGDM) is prevalent in diverse decision-making scenarios, involving numerous decision makers (DMs), the set of alternatives and criteria, and continuous temporal cycles. Opinions from DMs dynamically evolve through iterative interaction, leading to dynamic opinion evolutions. However, traditional MCGDM methodology usually establish the opinion formation on a static time point throughout information aggregation, which will lead to information distortion. This study develops a novel large-scale MCGDM method with information emendation based on an unsupervised opinion dynamics (UOD) model, combining with the intuitionistic fuzzy set (IFS) and the technique for order preference by similarity to an ideal solution (TOPSIS). The IFS is utilized to quantify opinions since it can effectively achieve a tradeoff between information retention and convenience of evaluation. Simultaneously, in the proposed UOD model, the weight updating mechanism is further considered to improve the interaction adequacy, and the unsupervised mechanism for interaction threshold helps to decrease the influences of subjectivity from DMs. Moreover, numerical simulations validate the UOD model’s feasibility. Finally, a school site selection problem is carried out to elaborate the effectiveness of the proposed method. This study will provide a methodological reference for solving large-scale MCGDM problems, facilitating rapid convergence of opinions within large-scale groups, and enrich the research on opinion dynamics in the field of decision-making.

大规模多标准群体决策(MCGDM)普遍存在于各种决策场景中,涉及众多决策者(DM)、备选方案和标准集以及连续的时间周期。决策制定者(DM)的意见通过迭代互动动态演化,导致意见的动态演化。然而,传统的 MCGDM 方法通常将意见形成建立在整个信息聚合过程的静态时间点上,这会导致信息失真。本研究基于无监督意见动态模型(UOD),结合直觉模糊集(IFS)和理想解相似度排序偏好技术(TOPSIS),开发了一种新型的大规模 MCGDM 方法。由于直观模糊集可以有效地在信息保留和评估便利性之间实现权衡,因此利用直观模糊集对意见进行量化。同时,在所提出的 UOD 模型中,进一步考虑了权重更新机制以提高交互的充分性,而交互阈值的无监督机制则有助于减少来自 DM 的主观性影响。此外,数值模拟验证了 UOD 模型的可行性。最后,通过一个学校选址问题来阐述所提方法的有效性。本研究将为解决大规模 MCGDM 问题提供方法参考,促进大规模群体内意见的快速收敛,并丰富决策领域的意见动态研究。
{"title":"Large-scale multiple criteria group decision-making with information emendation based on unsupervised opinion evolutions","authors":"","doi":"10.1016/j.asoc.2024.112227","DOIUrl":"10.1016/j.asoc.2024.112227","url":null,"abstract":"<div><p>Large-scale multiple criteria group decision-making (MCGDM) is prevalent in diverse decision-making scenarios, involving numerous decision makers (DMs), the set of alternatives and criteria, and continuous temporal cycles. Opinions from DMs dynamically evolve through iterative interaction, leading to dynamic opinion evolutions. However, traditional MCGDM methodology usually establish the opinion formation on a static time point throughout information aggregation, which will lead to information distortion. This study develops a novel large-scale MCGDM method with information emendation based on an unsupervised opinion dynamics (UOD) model, combining with the intuitionistic fuzzy set (IFS) and the technique for order preference by similarity to an ideal solution (TOPSIS). The IFS is utilized to quantify opinions since it can effectively achieve a tradeoff between information retention and convenience of evaluation. Simultaneously, in the proposed UOD model, the weight updating mechanism is further considered to improve the interaction adequacy, and the unsupervised mechanism for interaction threshold helps to decrease the influences of subjectivity from DMs. Moreover, numerical simulations validate the UOD model’s feasibility. Finally, a school site selection problem is carried out to elaborate the effectiveness of the proposed method. This study will provide a methodological reference for solving large-scale MCGDM problems, facilitating rapid convergence of opinions within large-scale groups, and enrich the research on opinion dynamics in the field of decision-making.</p></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142272700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VTNet: A multi-domain information fusion model for long-term multi-variate time series forecasting with application in irrigation water level VTNet:用于灌溉水位长期多变量时间序列预测的多域信息融合模型
IF 7.2 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1016/j.asoc.2024.112251

Time series forecasting is intricately tied to production and life, garnering widespread attention over an extended period. Enhancing the performance of long-term multivariate time series forecasting (MTSF) poses a highly challenging task, as it requires mining complicated and obscure temporal patterns in many aspects. For this reason, this paper proposes a long-term forecasting model based on multi-domain fusion (VTNet) to adaptively capture and refine multi-scale intra- and inter-variate dependencies. In contrast to previous techniques, we devise a dual-stream learning architecture. Firstly, the fast Fourier Transform (FFT) is adopted to extract frequency domain information. The original sequences are then transformed into 2D visual features in the temporal-frequency domain, and a 2D-TBlock is designed for multi-scale dynamic learning. Secondly, a combination of convolution and recurrent networks continues to explore the local temporal features and preserve the global trend. Finally, multi-modal circulant fusion is applied to achieve a more comprehensive and enriched feature fusion representation, further promoting overall performance. Extensive experiments are conducted on 9 public benchmark datasets and the real-world irrigation water level to showcase VTNet’s promoted performance and generalization. Moreover, VTNet yields 46.93% and 25.36% relative improvements for water level forecasting, revealing its potential application value in water-saving planning and extreme event early warning.

时间序列预测与生产和生活息息相关,长期以来受到广泛关注。提高长期多变量时间序列预测(MTSF)的性能是一项极具挑战性的任务,因为它需要从多个方面挖掘复杂而模糊的时间模式。为此,本文提出了一种基于多域融合(VTNet)的长期预测模型,以适应性地捕捉和完善多尺度变量内和变量间的依赖关系。与以往的技术不同,我们设计了一种双流学习架构。首先,采用快速傅立叶变换(FFT)提取频域信息。然后,将原始序列转换为时频域的二维视觉特征,并设计了一个二维-TBlock,用于多尺度动态学习。其次,结合卷积和递归网络,继续探索局部时域特征并保持全局趋势。最后,多模态环流融合实现了更全面、更丰富的特征融合表示,进一步提高了整体性能。我们在 9 个公共基准数据集和现实世界的灌溉水位上进行了广泛的实验,以展示 VTNet 所提升的性能和泛化能力。此外,VTNet 在水位预测方面分别获得了 46.93% 和 25.36% 的相对改进,揭示了其在节水规划和极端事件预警方面的潜在应用价值。
{"title":"VTNet: A multi-domain information fusion model for long-term multi-variate time series forecasting with application in irrigation water level","authors":"","doi":"10.1016/j.asoc.2024.112251","DOIUrl":"10.1016/j.asoc.2024.112251","url":null,"abstract":"<div><p>Time series forecasting is intricately tied to production and life, garnering widespread attention over an extended period. Enhancing the performance of long-term multivariate time series forecasting (MTSF) poses a highly challenging task, as it requires mining complicated and obscure temporal patterns in many aspects. For this reason, this paper proposes a long-term forecasting model based on multi-domain fusion (VTNet) to adaptively capture and refine multi-scale intra- and inter-variate dependencies. In contrast to previous techniques, we devise a dual-stream learning architecture. Firstly, the fast Fourier Transform (FFT) is adopted to extract frequency domain information. The original sequences are then transformed into 2D visual features in the temporal-frequency domain, and a 2D-TBlock is designed for multi-scale dynamic learning. Secondly, a combination of convolution and recurrent networks continues to explore the local temporal features and preserve the global trend. Finally, multi-modal circulant fusion is applied to achieve a more comprehensive and enriched feature fusion representation, further promoting overall performance. Extensive experiments are conducted on 9 public benchmark datasets and the real-world irrigation water level to showcase VTNet’s promoted performance and generalization. Moreover, VTNet yields 46.93% and 25.36% relative improvements for water level forecasting, revealing its potential application value in water-saving planning and extreme event early warning.</p></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":null,"pages":null},"PeriodicalIF":7.2,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142272107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Soft Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1