首页 > 最新文献

Applied Intelligence最新文献

英文 中文
A convex Kullback-Leibler divergence and critical-descriptor prototypes for semi-supervised few-shot learning 半监督少镜头学习的凸Kullback-Leibler散度和临界描述子原型
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-21 DOI: 10.1007/s10489-025-06239-1
Yukun Liu, Daming Shi

Few-shot learning has achieved great success in recent years, thanks to its requirement of limited number of labeled data. However, most of the state-of-the-art techniques of few-shot learning employ transfer learning, which still requires massive labeled data to train. To simulate the human learning mechanism, a deep model of few-shot learning is proposed to learn from one, or a few examples. First of all in this paper, we analyze and note that the problem with representative semi-supervised few-shot learning methods is getting stuck in local optimization and prototype bias problems. To address these challenges, we propose a new semi-supervised few-shot learning method with Convex Kullback-Leibler and critical descriptor prototypes, hereafter referred to as CKL. Specifically, CKL optimizes joint probability density via KL divergence, subsequently deriving a strictly convex function to facilitate global optimization in semi-supervised clustering. In addition, by incorporating dictionary learning, the critical descriptor facilitates the extraction of more prototypical features, thereby capturing more distinct feature information and avoiding the problem of prototype bias caused by limited labeled samples. Intensive experiments have been conducted on three popular benchmark datasets, and the experimental results show that this method significantly improves the classification ability of few-shot learning and obtains the most advanced performance. In the future, we will explore additional methods that can be integrated with deep learning to further uncover essential features within samples.

近年来,由于对标注数据数量的要求有限,Few-shot学习方法取得了很大的成功。然而,大多数最先进的小样本学习技术采用迁移学习,这仍然需要大量的标记数据来训练。为了模拟人类的学习机制,提出了一种从一个或几个例子中学习的深度少镜头学习模型。本文首先分析并注意到具有代表性的半监督少射学习方法陷入了局部优化和原型偏差问题。为了解决这些挑战,我们提出了一种新的半监督少镜头学习方法,采用凸Kullback-Leibler和关键描述子原型,以下简称CKL。具体来说,CKL通过KL散度来优化联合概率密度,然后推导出一个严格的凸函数,以方便半监督聚类的全局优化。此外,通过结合字典学习,关键描述符有助于提取更多的原型特征,从而捕获更清晰的特征信息,避免因标记样本有限而导致的原型偏差问题。在三个流行的基准数据集上进行了大量的实验,实验结果表明,该方法显著提高了few-shot学习的分类能力,获得了最先进的性能。在未来,我们将探索可以与深度学习集成的其他方法,以进一步揭示样本中的基本特征。
{"title":"A convex Kullback-Leibler divergence and critical-descriptor prototypes for semi-supervised few-shot learning","authors":"Yukun Liu,&nbsp;Daming Shi","doi":"10.1007/s10489-025-06239-1","DOIUrl":"10.1007/s10489-025-06239-1","url":null,"abstract":"<div><p>Few-shot learning has achieved great success in recent years, thanks to its requirement of limited number of labeled data. However, most of the state-of-the-art techniques of few-shot learning employ transfer learning, which still requires massive labeled data to train. To simulate the human learning mechanism, a deep model of few-shot learning is proposed to learn from one, or a few examples. First of all in this paper, we analyze and note that the problem with representative semi-supervised few-shot learning methods is getting stuck in local optimization and prototype bias problems. To address these challenges, we propose a new semi-supervised few-shot learning method with Convex Kullback-Leibler and critical descriptor prototypes, hereafter referred to as CKL. Specifically, CKL optimizes joint probability density via KL divergence, subsequently deriving a strictly convex function to facilitate global optimization in semi-supervised clustering. In addition, by incorporating dictionary learning, the critical descriptor facilitates the extraction of more prototypical features, thereby capturing more distinct feature information and avoiding the problem of prototype bias caused by limited labeled samples. Intensive experiments have been conducted on three popular benchmark datasets, and the experimental results show that this method significantly improves the classification ability of few-shot learning and obtains the most advanced performance. In the future, we will explore additional methods that can be integrated with deep learning to further uncover essential features within samples.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995662","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep subspace clustering via latent representation learning 基于潜在表征学习的深度子空间聚类
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-21 DOI: 10.1007/s10489-025-06255-1
Shenglei Pei, Qinghao Han, Zepu Hao, Hong Zhao

Deep subspace clustering networks (DSC-Nets), which combine deep autoencoders and self-expressive modules, have garnered widespread attention due to their outstanding performance. Within these networks, the autoencoder captures the latent representations of data by reconstructing the input data, while the self-expressive layer learns an affinity matrix based on these latent representations. This matrix guides spectral clustering, ultimately completing the clustering task. However, the latent representations learned solely through self-reconstruction by the autoencoder lack discriminative power. The quality of these latent representations directly affects the performance of the affinity matrix, which inevitably limits the clustering performance. To address this issue, we propose learning dissimilar relationships between samples using a classification module, and similar relationships using the self-expressive module. We integrate the information from both modules to construct a graph based on learned similarities, which is then embedded into the autoencoder network. Furthermore, we introduce a pseudo-label supervision module to guide the learning of higher-level similarities in the latent representations, thus achieving more discriminative latent features. Additionally, to enhance the quality of the affinity matrix, we employ an entropy norm constraint to improve connectivity within the subspaces. Experimental results on four public datasets demonstrate that our method achieves superior performance compared to other popular subspace clustering approaches.

深度子空间聚类网络(DSC-Nets)将深度自编码器和自表达模块相结合,因其优异的性能而受到广泛关注。在这些网络中,自编码器通过重建输入数据来捕获数据的潜在表示,而自表达层则根据这些潜在表示学习亲和矩阵。该矩阵指导谱聚类,最终完成聚类任务。然而,仅通过自编码器自我重建学习的潜在表征缺乏判别能力。这些潜在表示的质量直接影响亲和矩阵的性能,这不可避免地限制了聚类性能。为了解决这个问题,我们建议使用分类模块学习样本之间的不相似关系,使用自表达模块学习相似关系。我们将两个模块的信息整合成一个基于学习相似度的图,然后将其嵌入到自编码器网络中。此外,我们引入了一个伪标签监督模块来指导潜在表征中更高级别相似性的学习,从而获得更具判别性的潜在特征。此外,为了提高亲和矩阵的质量,我们采用熵范数约束来提高子空间内的连通性。在四个公共数据集上的实验结果表明,与其他流行的子空间聚类方法相比,我们的方法取得了更好的性能。
{"title":"Deep subspace clustering via latent representation learning","authors":"Shenglei Pei,&nbsp;Qinghao Han,&nbsp;Zepu Hao,&nbsp;Hong Zhao","doi":"10.1007/s10489-025-06255-1","DOIUrl":"10.1007/s10489-025-06255-1","url":null,"abstract":"<div><p>Deep subspace clustering networks (DSC-Nets), which combine deep autoencoders and self-expressive modules, have garnered widespread attention due to their outstanding performance. Within these networks, the autoencoder captures the latent representations of data by reconstructing the input data, while the self-expressive layer learns an affinity matrix based on these latent representations. This matrix guides spectral clustering, ultimately completing the clustering task. However, the latent representations learned solely through self-reconstruction by the autoencoder lack discriminative power. The quality of these latent representations directly affects the performance of the affinity matrix, which inevitably limits the clustering performance. To address this issue, we propose learning dissimilar relationships between samples using a classification module, and similar relationships using the self-expressive module. We integrate the information from both modules to construct a graph based on learned similarities, which is then embedded into the autoencoder network. Furthermore, we introduce a pseudo-label supervision module to guide the learning of higher-level similarities in the latent representations, thus achieving more discriminative latent features. Additionally, to enhance the quality of the affinity matrix, we employ an entropy norm constraint to improve connectivity within the subspaces. Experimental results on four public datasets demonstrate that our method achieves superior performance compared to other popular subspace clustering approaches.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Novel dropout approach for mitigating over-smoothing in graph neural networks 一种减轻图神经网络过度平滑的新颖dropout方法
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-21 DOI: 10.1007/s10489-025-06285-9
El houssaine Hssayni, Ali Boufssasse, Nour-Eddine Joudar, Mohamed Ettaouil

Graph Neural Networks (GNNs) have emerged as a powerful tool for analyzing structured data represented as graphs. They offer significant contributions across various domains due to their ability to effectively capture and process complex relational information. However, most existing GNNs still suffer from undesirable phenomena such as non-robustness, overfitting, and over-smoothing. These challenges have raised significant interest among researchers. In this context, this work aims to address these issues by proposing a new vision of Dropout named A-DropEdge. First, it applies a message-passing layer to ensure the connection between nodes and avoid dropping in the input. Then, the information propagates through many branches with different random configurations to enhance the aggregation process. Moreover, consistency regularization is adopted to perform self-supervised learning. The experimental results on three graph data sets including Cora, Citeseer, and PubMed show the robustness and performance of the proposed approach in mitigating the over-smoothing problem.

图神经网络(gnn)已经成为分析以图表示的结构化数据的强大工具。由于它们能够有效地捕获和处理复杂的关系信息,它们在各个领域提供了重要的贡献。然而,大多数现有gnn仍然存在非鲁棒性、过拟合和过度平滑等不良现象。这些挑战引起了研究人员的极大兴趣。在此背景下,本作品旨在通过提出一个名为a - dropedge的辍学新愿景来解决这些问题。首先,它应用消息传递层来确保节点之间的连接并避免输入丢失。然后,信息通过多个具有不同随机配置的分支进行传播,以增强聚合过程。采用一致性正则化进行自监督学习。在Cora、Citeseer和PubMed三个图数据集上的实验结果表明了该方法在缓解过度平滑问题方面的鲁棒性和性能。
{"title":"Novel dropout approach for mitigating over-smoothing in graph neural networks","authors":"El houssaine Hssayni,&nbsp;Ali Boufssasse,&nbsp;Nour-Eddine Joudar,&nbsp;Mohamed Ettaouil","doi":"10.1007/s10489-025-06285-9","DOIUrl":"10.1007/s10489-025-06285-9","url":null,"abstract":"<div><p>Graph Neural Networks (GNNs) have emerged as a powerful tool for analyzing structured data represented as graphs. They offer significant contributions across various domains due to their ability to effectively capture and process complex relational information. However, most existing GNNs still suffer from undesirable phenomena such as non-robustness, overfitting, and over-smoothing. These challenges have raised significant interest among researchers. In this context, this work aims to address these issues by proposing a new vision of Dropout named A-DropEdge. First, it applies a message-passing layer to ensure the connection between nodes and avoid dropping in the input. Then, the information propagates through many branches with different random configurations to enhance the aggregation process. Moreover, consistency regularization is adopted to perform self-supervised learning. The experimental results on three graph data sets including Cora, Citeseer, and PubMed show the robustness and performance of the proposed approach in mitigating the over-smoothing problem.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contrastive prototype loss based discriminative feature network for few-shot learning 基于对比原型损失的判别特征网络的少镜头学习
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-21 DOI: 10.1007/s10489-025-06234-6
Leilei Yan, Feihong He, Xiaohan Zheng, Li Zhang, Yiqi Zhang, Jiangzhen He, Weidong Du, Yansong Wang, Fanzhang Li

Metric-based few-shot image classification methods generally perform classification by comparing the distances between the query sample features and the prototypes of each class. These methods often focus on constructing prototype representations for each class or learning a metric, while neglecting the significance of the feature space itself. In this paper, we redirect the focus to feature space construction, with the goal of constructing a discriminative feature space for few-shot image classification tasks. To this end, we designed a contrastive prototype loss that incorporates the distribution of query samples with respect to class prototypes in the feature space, emphasizing intra-class compactness and inter-class separability, thereby guiding the model to learn a more discriminative feature space. Based on this loss, we propose a contrastive prototype loss based discriminative feature network (CPL-DFNet) to address few-shot image classification tasks. CPL-DFNet enhances sample utilization by fully leveraging the distance relationships between query samples and class prototypes in the feature space, creating more favorable conditions for few-shot image classification tasks and significantly improving classification performance. We conducted extensive experiments on both general and fine-grained few-shot image classification benchmark datasets to validate the effectiveness of the proposed CPL-DFNet method. The experimental results show that CPL-DFNet can effectively perform few-shot image classification tasks and outperforms many existing methods across various task scenarios, demonstrating significant performance advantages.

基于度量的少拍图像分类方法通常通过比较查询样本特征与每个类的原型之间的距离来进行分类。这些方法通常侧重于为每个类构建原型表示或学习度量,而忽略了特征空间本身的重要性。在本文中,我们将重点转向特征空间的构建,目的是为少量图像分类任务构建一个判别特征空间。为此,我们设计了一个对比原型损失模型,该模型结合了查询样本相对于类原型在特征空间中的分布,强调了类内的紧密性和类间的可分性,从而指导模型学习更具判别性的特征空间。基于这种损失,我们提出了一种基于对比原型损失的判别特征网络(CPL-DFNet)来解决少拍图像分类任务。cpll - dfnet通过充分利用查询样本与类原型在特征空间中的距离关系,提高了样本利用率,为少拍图像分类任务创造了更有利的条件,显著提高了分类性能。我们在一般和细粒度的少量图像分类基准数据集上进行了大量实验,以验证所提出的cpll - dfnet方法的有效性。实验结果表明,cpll - dfnet可以有效地完成少量图像分类任务,并在各种任务场景下优于现有的许多方法,显示出显著的性能优势。
{"title":"Contrastive prototype loss based discriminative feature network for few-shot learning","authors":"Leilei Yan,&nbsp;Feihong He,&nbsp;Xiaohan Zheng,&nbsp;Li Zhang,&nbsp;Yiqi Zhang,&nbsp;Jiangzhen He,&nbsp;Weidong Du,&nbsp;Yansong Wang,&nbsp;Fanzhang Li","doi":"10.1007/s10489-025-06234-6","DOIUrl":"10.1007/s10489-025-06234-6","url":null,"abstract":"<div><p>Metric-based few-shot image classification methods generally perform classification by comparing the distances between the query sample features and the prototypes of each class. These methods often focus on constructing prototype representations for each class or learning a metric, while neglecting the significance of the feature space itself. In this paper, we redirect the focus to feature space construction, with the goal of constructing a discriminative feature space for few-shot image classification tasks. To this end, we designed a contrastive prototype loss that incorporates the distribution of query samples with respect to class prototypes in the feature space, emphasizing intra-class compactness and inter-class separability, thereby guiding the model to learn a more discriminative feature space. Based on this loss, we propose a contrastive prototype loss based discriminative feature network (CPL-DFNet) to address few-shot image classification tasks. CPL-DFNet enhances sample utilization by fully leveraging the distance relationships between query samples and class prototypes in the feature space, creating more favorable conditions for few-shot image classification tasks and significantly improving classification performance. We conducted extensive experiments on both general and fine-grained few-shot image classification benchmark datasets to validate the effectiveness of the proposed CPL-DFNet method. The experimental results show that CPL-DFNet can effectively perform few-shot image classification tasks and outperforms many existing methods across various task scenarios, demonstrating significant performance advantages.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BAM-SORT: border-guided activated matching for online multi-object tracking BAM-SORT:边界引导激活匹配在线多目标跟踪
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-20 DOI: 10.1007/s10489-024-06037-1
Yuan Chao, Huaiyang Zhu, Hengyu Lu

Multi-object tracking aims at estimating object bounding boxes and identity IDs in videos. Most tracking methods combine a detector and a Kalman filter using the IoU distance as a similarity metric for association matching of the previous trajectories with the current detection box. These methods usually suffer from ID switches and fragmented trajectories in response to congested and frequently occluded scenarios. To solve this problem, in this study, a simple and effective association method is proposed. First, a bottom edge cost matrix is introduced for the utilization of depth information to improve the data association and increase the robustness in the case of occlusion. Second, an asymmetric trajectory classification mechanism is proposed to distinguish the false-postive trajectories, and an activated trajectory matching strategy is introduced to reduce the interference of noise and transient objects in tracking. Finally, the trajectory deletion strategy is improved by introducing the number of trajectory state switches to delete the trajectories caused by spurious high-scoring detection boxes in real time, as a result, the number of fragmented trajectories is also reduced. These innovations achieve excellent performance on various benchmarks, including MOT17, MOT20, and especially DanceTrack where interactions and occlusions are frequent and severe. The code and models are available at https://github.com/djdodsjsjx/BAM-SORT/.

多目标跟踪的目的是估计视频中的目标边界框和身份id。大多数跟踪方法结合检测器和卡尔曼滤波器,使用IoU距离作为先前轨迹与当前检测框的关联匹配的相似性度量。这些方法通常受到ID切换和碎片化轨迹的影响,以响应拥塞和频繁闭塞的场景。为了解决这一问题,本研究提出了一种简单有效的关联方法。首先,引入底边代价矩阵,利用深度信息改善数据关联,增强遮挡情况下的鲁棒性;其次,提出了一种非对称轨迹分类机制来区分假阳性轨迹,并引入了激活轨迹匹配策略来降低跟踪过程中噪声和瞬态目标的干扰;最后,对轨迹删除策略进行改进,引入轨迹状态切换的数量,实时删除由虚假高分检测盒造成的轨迹,从而减少轨迹碎片的数量。这些创新在各种基准测试中取得了出色的性能,包括MOT17、MOT20,特别是在相互作用和闭塞频繁且严重的DanceTrack中。代码和模型可在https://github.com/djdodsjsjx/BAM-SORT/上获得。
{"title":"BAM-SORT: border-guided activated matching for online multi-object tracking","authors":"Yuan Chao,&nbsp;Huaiyang Zhu,&nbsp;Hengyu Lu","doi":"10.1007/s10489-024-06037-1","DOIUrl":"10.1007/s10489-024-06037-1","url":null,"abstract":"<div><p>Multi-object tracking aims at estimating object bounding boxes and identity IDs in videos. Most tracking methods combine a detector and a Kalman filter using the IoU distance as a similarity metric for association matching of the previous trajectories with the current detection box. These methods usually suffer from ID switches and fragmented trajectories in response to congested and frequently occluded scenarios. To solve this problem, in this study, a simple and effective association method is proposed. First, a bottom edge cost matrix is introduced for the utilization of depth information to improve the data association and increase the robustness in the case of occlusion. Second, an asymmetric trajectory classification mechanism is proposed to distinguish the false-postive trajectories, and an activated trajectory matching strategy is introduced to reduce the interference of noise and transient objects in tracking. Finally, the trajectory deletion strategy is improved by introducing the number of trajectory state switches to delete the trajectories caused by spurious high-scoring detection boxes in real time, as a result, the number of fragmented trajectories is also reduced. These innovations achieve excellent performance on various benchmarks, including MOT17, MOT20, and especially DanceTrack where interactions and occlusions are frequent and severe. The code and models are available at https://github.com/djdodsjsjx/BAM-SORT/.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep reinforcement learning portfolio model based on mixture of experts 基于专家混合的深度强化学习组合模型
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-20 DOI: 10.1007/s10489-025-06242-6
Ziqiang Wei, Deng Chen, Yanduo Zhang, Dawei Wen, Xin Nie, Liang Xie

In the field of artificial intelligence, the portfolio management problem has received widespread attention. Portfolio models based on deep reinforcement learning enable intelligent investment decision-making. However, most models only consider modeling the temporal information of stocks, neglecting the correlation between stocks and the impact of overall market risk. Moreover, their trading strategies are often singular and fail to adapt to dynamic changes in the trading market. To address these issues, this paper proposes a Deep Reinforcement Learning Portfolio Model based on Mixture of Experts (MoEDRLPM). Firstly, a spatio-temporal adaptive embedding matrix is designed, temporal and spatial self-attention mechanisms are employed to extract the temporal information and correlations of stocks. Secondly, dynamically select the current optimal expert from the mixed expert pool through router. The expert makes decisions and aggregates to derive the portfolio weights. Next, market index data is utilized to model the current market risk and determine investment capital ratios. Finally, deep reinforcement learning is employed to optimize the portfolio strategy. This approach generates diverse trading strategies according to dynamic changes in the market environment. The proposed model is tested on the SSE50 and CSI300 datasets. Results show that the total returns of this model increase by 12% and 8%, respectively, while the Sharpe Ratios improve by 64% and 51%.

在人工智能领域,项目组合管理问题受到了广泛的关注。基于深度强化学习的投资组合模型实现了智能投资决策。然而,大多数模型只考虑股票的时间信息建模,而忽略了股票与整体市场风险影响之间的相关性。此外,他们的交易策略往往单一,不能适应交易市场的动态变化。为了解决这些问题,本文提出了一种基于混合专家的深度强化学习组合模型(MoEDRLPM)。首先,设计时空自适应嵌入矩阵,利用时空自注意机制提取股票的时间信息和相关性;其次,通过路由器从混合专家池中动态选择当前最优专家;专家通过决策和汇总来得出投资组合的权重。接下来,利用市场指数数据对当前市场风险进行建模,并确定投资资本比率。最后,利用深度强化学习对投资组合策略进行优化。这种方法根据市场环境的动态变化,产生多样化的交易策略。在SSE50和CSI300数据集上对该模型进行了测试。结果表明,该模型的总收益分别提高了12%和8%,夏普比率分别提高了64%和51%。
{"title":"Deep reinforcement learning portfolio model based on mixture of experts","authors":"Ziqiang Wei,&nbsp;Deng Chen,&nbsp;Yanduo Zhang,&nbsp;Dawei Wen,&nbsp;Xin Nie,&nbsp;Liang Xie","doi":"10.1007/s10489-025-06242-6","DOIUrl":"10.1007/s10489-025-06242-6","url":null,"abstract":"<div><p>In the field of artificial intelligence, the portfolio management problem has received widespread attention. Portfolio models based on deep reinforcement learning enable intelligent investment decision-making. However, most models only consider modeling the temporal information of stocks, neglecting the correlation between stocks and the impact of overall market risk. Moreover, their trading strategies are often singular and fail to adapt to dynamic changes in the trading market. To address these issues, this paper proposes a Deep Reinforcement Learning Portfolio Model based on Mixture of Experts (MoEDRLPM). Firstly, a spatio-temporal adaptive embedding matrix is designed, temporal and spatial self-attention mechanisms are employed to extract the temporal information and correlations of stocks. Secondly, dynamically select the current optimal expert from the mixed expert pool through router. The expert makes decisions and aggregates to derive the portfolio weights. Next, market index data is utilized to model the current market risk and determine investment capital ratios. Finally, deep reinforcement learning is employed to optimize the portfolio strategy. This approach generates diverse trading strategies according to dynamic changes in the market environment. The proposed model is tested on the SSE50 and CSI300 datasets. Results show that the total returns of this model increase by 12% and 8%, respectively, while the Sharpe Ratios improve by 64% and 51%.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic-spatial guided context propagation network for camouflaged object detection 用于伪装目标检测的语义空间引导上下文传播网络
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-20 DOI: 10.1007/s10489-025-06264-0
Junchao Ren, Qiao Zhang, Bingbing Kang, Yuxi Zhong, Min He, Yanliang Ge, Hongbo Bi

Camouflaged object detection (COD) aims to detect objects that blend in with their surroundings and is a challenging task in computer vision. High-level semantic information and low-level spatial information play important roles in localizing camouflaged objects and reinforcing spatial cues. However, current COD methods directly connect high-level features with low-level features, ignoring the importance of the respective features. In this paper, we design a Semantic-spatial guided Context Propagation Network (SCPNet) to efficiently mine semantic and spatial features while enhancing their feature representations. Firstly, we design a twin positioning module (TPM) to explore semantic cues to accurately locate camouflaged objects. Afterward, we introduce a spatial awareness module (SAM) to mine spatial cues in shallow features deeply. Finally, we develop a context propagation module (CPM) to assign semantic and spatial cues to multi-level features and enhance their feature representations. Experimental results show that our SCPNet outperforms state-of-the-art methods on three challenging datasets. Codes will be made available at https://github.com/RJC0608/SCPNet.

伪装目标检测(COD)旨在检测与周围环境融为一体的物体,是计算机视觉领域的一项具有挑战性的任务。高层语义信息和低层空间信息在定位伪装对象和强化空间线索中起着重要作用。然而,目前的COD方法直接将高级特征与低级特征联系起来,忽略了各自特征的重要性。在本文中,我们设计了一个语义空间导向的上下文传播网络(SCPNet)来有效地挖掘语义和空间特征,同时增强它们的特征表示。首先,我们设计了一个双定位模块(TPM)来探索语义线索,以准确定位伪装物体。然后,我们引入了空间感知模块(SAM)来深度挖掘浅层特征中的空间线索。最后,我们开发了一个上下文传播模块(CPM)来为多层次特征分配语义和空间线索,并增强它们的特征表示。实验结果表明,我们的SCPNet在三个具有挑战性的数据集上优于最先进的方法。代码将在https://github.com/RJC0608/SCPNet上提供。
{"title":"Semantic-spatial guided context propagation network for camouflaged object detection","authors":"Junchao Ren,&nbsp;Qiao Zhang,&nbsp;Bingbing Kang,&nbsp;Yuxi Zhong,&nbsp;Min He,&nbsp;Yanliang Ge,&nbsp;Hongbo Bi","doi":"10.1007/s10489-025-06264-0","DOIUrl":"10.1007/s10489-025-06264-0","url":null,"abstract":"<div><p>Camouflaged object detection (COD) aims to detect objects that blend in with their surroundings and is a challenging task in computer vision. High-level semantic information and low-level spatial information play important roles in localizing camouflaged objects and reinforcing spatial cues. However, current COD methods directly connect high-level features with low-level features, ignoring the importance of the respective features. In this paper, we design a <i>S</i>emantic-spatial guided <i>C</i>ontext <i>P</i>ropagation <i>N</i>etwork (<i>SCPNet</i>) to efficiently mine semantic and spatial features while enhancing their feature representations. Firstly, we design a twin positioning module (TPM) to explore semantic cues to accurately locate camouflaged objects. Afterward, we introduce a spatial awareness module (SAM) to mine spatial cues in shallow features deeply. Finally, we develop a context propagation module (CPM) to assign semantic and spatial cues to multi-level features and enhance their feature representations. Experimental results show that our SCPNet outperforms state-of-the-art methods on three challenging datasets. Codes will be made available at https://github.com/RJC0608/SCPNet.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Retraction Note: A framework to evaluate the barriers for adopting the internet of medical things using the extended generalized TODIM method under the hesitant fuzzy environment 在犹豫模糊环境下,应用扩展广义TODIM方法评价医疗物联网应用障碍的框架
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-20 DOI: 10.1007/s10489-024-06187-2
Khalid Alattas, Qun Wu
{"title":"Retraction Note: A framework to evaluate the barriers for adopting the internet of medical things using the extended generalized TODIM method under the hesitant fuzzy environment","authors":"Khalid Alattas,&nbsp;Qun Wu","doi":"10.1007/s10489-024-06187-2","DOIUrl":"10.1007/s10489-024-06187-2","url":null,"abstract":"","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A masked autoencoder network for spatiotemporal predictive learning 一种用于时空预测学习的掩码自编码器网络
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-20 DOI: 10.1007/s10489-024-06214-2
Fengzhen Sun, Weidong Jin

This paper is about predictive learning, which is generating future frames given previous images. Suffering from the vanishing gradient problem, existing methods based on RNN and CNN can’t capture the long-term dependencies effectively. To overcome the above dilemma, we present MastNet a spatiotemporal framework for long-term predictive learning. In this paper, we design a Transformer-based encoder-decoder with hierarchical structure. As for the transformer block, we adopt the spatiotemporal window based self-attention to reduce computational complexity and the spatiotemporal shifted window partitioning approach. More importantly, we build a spatiotemporal autoencoder by the random clip mask strategy, which leads to better feature mining for temporal dependencies and spatial correlations. Furthermore, we insert an auxiliary prediction head, which can help our model generate higher-quality frames. Experimental results show that the proposed MastNet achieves the best results in accuracy and long-term prediction on two spatiotemporal datasets compared with the state-of-the-art models.

这篇论文是关于预测学习的,它是根据之前的图像生成未来的帧。由于存在梯度消失的问题,现有的基于RNN和CNN的方法不能有效地捕获长期依赖关系。为了克服上述困境,我们提出了一个用于长期预测学习的时空框架。本文设计了一种基于变压器的分层结构编解码器。对于变压器块,我们采用了基于时空窗的自关注来降低计算复杂度,并采用了时空移窗划分方法。更重要的是,我们通过随机剪辑掩码策略构建了一个时空自编码器,从而更好地挖掘了时间依赖性和空间相关性的特征。此外,我们还插入了一个辅助预测头,这可以帮助我们的模型生成更高质量的帧。实验结果表明,在两个时空数据集上,与现有模型相比,所提出的MastNet在精度和长期预测方面取得了最好的结果。
{"title":"A masked autoencoder network for spatiotemporal predictive learning","authors":"Fengzhen Sun,&nbsp;Weidong Jin","doi":"10.1007/s10489-024-06214-2","DOIUrl":"10.1007/s10489-024-06214-2","url":null,"abstract":"<div><p>This paper is about predictive learning, which is generating future frames given previous images. Suffering from the vanishing gradient problem, existing methods based on RNN and CNN can’t capture the long-term dependencies effectively. To overcome the above dilemma, we present MastNet a spatiotemporal framework for long-term predictive learning. In this paper, we design a Transformer-based encoder-decoder with hierarchical structure. As for the transformer block, we adopt the spatiotemporal window based self-attention to reduce computational complexity and the spatiotemporal shifted window partitioning approach. More importantly, we build a spatiotemporal autoencoder by the random clip mask strategy, which leads to better feature mining for temporal dependencies and spatial correlations. Furthermore, we insert an auxiliary prediction head, which can help our model generate higher-quality frames. Experimental results show that the proposed MastNet achieves the best results in accuracy and long-term prediction on two spatiotemporal datasets compared with the state-of-the-art models.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sine and cosine based learning rate for gradient descent method 基于正弦和余弦的学习率梯度下降法
IF 3.4 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-20 DOI: 10.1007/s10489-025-06235-5
Krutika Verma, Abyayananda Maiti

Deep learning networks have been trained using first-order-based methods. These methods often converge more quickly when combined with an adaptive step size, but they tend to settle at suboptimal points, especially when learning occurs in a large output space. When first-order-based methods are used with a constant step size, they oscillate near the zero-gradient region, which leads to slow convergence. However, these issues are exacerbated under nonconvexity, which can significantly diminish the performance of first-order methods. In this work, we propose a novel Boltzmann Probability Weighted Sine with a Cosine distance-based Adaptive Gradient (BSCAGrad) method. The step size in this method is carefully designed to mitigate the issue of slow convergence. Furthermore, it facilitates escape from suboptimal points, enabling the optimization process to progress more efficiently toward local minima. This is achieved by combining a Boltzmann probability-weighted sine function and cosine distance to calculate the step size. The Boltzmann probability-weighted sine function acts when the gradient vanishes and the cooling parameter remains moderate, a condition typically observed near suboptimal points. Moreover, using the sine function on the exponential moving average of the weight parameters leverages geometric information from the data. The cosine distance prevents zero in the step size. Together, these components accelerate convergence, improve stability, and guide the algorithm toward a better optimal solution. A theoretical analysis of the convergence rate under both convexity and nonconvexity is provided to substantiate the findings. The experimental results from language modeling, object detection, machine translation, and image classification tasks on a real-world benchmark dataset, including CIFAR10, CIFAR100, PennTreeBank, PASCALVOC and WMT2014, demonstrate that the proposed step size outperforms traditional baseline methods.

深度学习网络使用基于一阶的方法进行训练。当与自适应步长相结合时,这些方法通常收敛得更快,但它们往往停留在次优点,特别是当学习发生在大输出空间中时。当一阶方法使用恒定步长时,它们在零梯度区域附近振荡,导致收敛缓慢。然而,这些问题在非凸性下会加剧,这可能会大大降低一阶方法的性能。在这项工作中,我们提出了一种新的玻尔兹曼概率加权正弦与基于余弦距离的自适应梯度(BSCAGrad)方法。该方法的步长经过精心设计,以缓解缓慢收敛的问题。此外,它有利于逃离次优点,使优化过程更有效地向局部最小值前进。这是通过结合玻尔兹曼概率加权正弦函数和余弦距离来计算步长来实现的。玻尔兹曼概率加权正弦函数在梯度消失和冷却参数保持适中时起作用,这种情况通常在次优点附近观察到。此外,对权重参数的指数移动平均使用正弦函数可以利用数据中的几何信息。余弦距离防止步长为零。这些组件一起加速收敛,提高稳定性,并引导算法走向更好的最优解。本文给出了在凸性和非凸性条件下收敛速度的理论分析来证实这些发现。在CIFAR10、CIFAR100、PennTreeBank、PASCALVOC和WMT2014等真实基准数据集上的语言建模、目标检测、机器翻译和图像分类任务的实验结果表明,所提出的步长方法优于传统的基线方法。
{"title":"Sine and cosine based learning rate for gradient descent method","authors":"Krutika Verma,&nbsp;Abyayananda Maiti","doi":"10.1007/s10489-025-06235-5","DOIUrl":"10.1007/s10489-025-06235-5","url":null,"abstract":"<p>Deep learning networks have been trained using first-order-based methods. These methods often converge more quickly when combined with an adaptive step size, but they tend to settle at suboptimal points, especially when learning occurs in a large output space. When first-order-based methods are used with a constant step size, they oscillate near the zero-gradient region, which leads to slow convergence. However, these issues are exacerbated under nonconvexity, which can significantly diminish the performance of first-order methods. In this work, we propose a novel <b>B</b>oltzmann Probability Weighted <b>S</b>ine with a <b>C</b>osine distance-based <b>A</b>daptive <b>Grad</b>ient (<i>BSCAGrad</i>) method. The step size in this method is carefully designed to mitigate the issue of slow convergence. Furthermore, it facilitates escape from suboptimal points, enabling the optimization process to progress more efficiently toward local minima. This is achieved by combining a Boltzmann probability-weighted sine function and cosine distance to calculate the step size. The Boltzmann probability-weighted sine function acts when the gradient vanishes and the cooling parameter remains moderate, a condition typically observed near suboptimal points. Moreover, using the sine function on the exponential moving average of the weight parameters leverages geometric information from the data. The cosine distance prevents zero in the step size. Together, these components accelerate convergence, improve stability, and guide the algorithm toward a better optimal solution. A theoretical analysis of the convergence rate under both convexity and nonconvexity is provided to substantiate the findings. The experimental results from language modeling, object detection, machine translation, and image classification tasks on a real-world benchmark dataset, including <i>CIFAR</i>10, <i>CIFAR</i>100, <i>PennTreeBank</i>, <i>PASCALVOC</i> and <i>WMT</i>2014, demonstrate that the proposed step size outperforms traditional baseline methods.</p>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 5","pages":""},"PeriodicalIF":3.4,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1