首页 > 最新文献

IEEE transactions on artificial intelligence最新文献

英文 中文
Mixture-of-Experts for Open Set Domain Adaptation: A Dual-Space Detection Approach 开集域自适应的专家混合:一种双空间检测方法
Pub Date : 2025-04-14 DOI: 10.1109/TAI.2025.3560590
Zhenbang Du;Jiayu An;Yunlu Tu;Jiahao Hong;Dongrui Wu
Open set domain adaptation (OSDA) copes with the distribution and label shifts between the source and target domains simultaneously, performing accurate classification for known classes while identifying unknown class samples in the target domain. Most existing OSDA approaches, depending on the final image feature space of deep models, require manually-tuned thresholds, and may easily misclassify unknown samples as known classes. Mixture-of-experts (MoE) could be a remedy. Within an MoE, different experts handle distinct input features, producing unique expert routing patterns for various classes in a routing feature space. As a result, unknown class samples may display different expert routing patterns to known classes. This article proposes dual-space detection, which exploits the inconsistencies between the image feature space and the routing feature space to detect unknown class samples without any threshold. A graph router is further introduced to better make use of the spatial information among the image patches. Experiments on three datasets validated the effectiveness and superiority of our approach.
开放集域自适应(OSDA)同时处理源域和目标域之间的分布和标签转移,在识别目标域中未知类样本的同时,对已知类样本进行准确分类。大多数现有的OSDA方法依赖于深度模型的最终图像特征空间,需要手动调整阈值,并且很容易将未知样本误分类为已知类。专家混合(MoE)可能是一种补救措施。在MoE中,不同的专家处理不同的输入特征,为路由特征空间中的各种类生成独特的专家路由模式。因此,未知的类样本可能会显示与已知类不同的专家路由模式。本文提出双空间检测,利用图像特征空间与路由特征空间的不一致性,不设阈值检测未知类样本。为了更好地利用图像块间的空间信息,进一步引入了图形路由器。在三个数据集上的实验验证了我们方法的有效性和优越性。
{"title":"Mixture-of-Experts for Open Set Domain Adaptation: A Dual-Space Detection Approach","authors":"Zhenbang Du;Jiayu An;Yunlu Tu;Jiahao Hong;Dongrui Wu","doi":"10.1109/TAI.2025.3560590","DOIUrl":"https://doi.org/10.1109/TAI.2025.3560590","url":null,"abstract":"Open set domain adaptation (OSDA) copes with the distribution and label shifts between the source and target domains simultaneously, performing accurate classification for known classes while identifying unknown class samples in the target domain. Most existing OSDA approaches, depending on the final image feature space of deep models, require manually-tuned thresholds, and may easily misclassify unknown samples as known classes. Mixture-of-experts (MoE) could be a remedy. Within an MoE, different experts handle distinct input features, producing unique expert routing patterns for various classes in a routing feature space. As a result, unknown class samples may display different expert routing patterns to known classes. This article proposes dual-space detection, which exploits the inconsistencies between the image feature space and the routing feature space to detect unknown class samples without any threshold. A graph router is further introduced to better make use of the spatial information among the image patches. Experiments on three datasets validated the effectiveness and superiority of our approach.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 12","pages":"3207-3216"},"PeriodicalIF":0.0,"publicationDate":"2025-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incremental Semisupervised Learning With Adaptive Locality Preservation for High-Dimensional Data 基于自适应局部保存的高维数据增量半监督学习
Pub Date : 2025-04-14 DOI: 10.1109/TAI.2025.3560592
Guojie Li;Zhiwen Yu;Kaixiang Yang;Ziwei Fan;C. L. Philip Chen
Broad learning system (BLS) has been widely researched and applied in the field of semisupervised learning. However, current semisupervised BLS methods rely on predefined graph structures. High-dimensional small-sample data, characterized by abundant redundant and noisy features with complex distribution patterns, often leads to the construction of poor-quality predefined graphs, thereby constraining the model’s performance. Additionally, the random generation of feature and enhancement nodes in BLS, combined with limited data labels, results in suboptimal model performance. To address these issues, this article first proposes a broad learning system with adaptive locality preservation (BLS-ALP). This method employs adaptive locality preservation constraints in the output space to ensure that similar samples share the same label, iteratively updating the graph structure. To further enhance the performance of BLS-ALP, an incremental ensemble framework (IBLS-ALP) is proposed. This framework effectively mitigates the impact of redundant and noisy features by using multiple random subspaces instead of the original high-dimensional space. Additionally, IBLS-ALP enhances the utilization of a small number of labels by incorporating residual labels, thereby significantly improving the model’s overall performance. Extensive experiments conducted on various high-dimensional small-sample datasets demonstrate that IBLS-ALP exhibits superior performance.
广义学习系统(BLS)在半监督学习领域得到了广泛的研究和应用。然而,目前的半监督BLS方法依赖于预定义的图结构。高维小样本数据具有丰富的冗余和噪声特征,分布模式复杂,往往导致构建质量较差的预定义图,从而制约了模型的性能。此外,BLS中特征和增强节点的随机生成,加上有限的数据标签,导致模型性能不理想。为了解决这些问题,本文首先提出了一种具有自适应局部保存的广义学习系统(BLS-ALP)。该方法在输出空间中采用自适应局域保持约束,确保相似样本共享相同的标签,迭代更新图结构。为了进一步提高BLS-ALP的性能,提出了一种增量集成框架(IBLS-ALP)。该框架利用多个随机子空间代替原有的高维空间,有效地减轻了冗余和噪声特征的影响。此外,IBLS-ALP通过加入残差标签,提高了少量标签的利用率,从而显著提高了模型的整体性能。在各种高维小样本数据集上进行的大量实验表明,IBLS-ALP具有优越的性能。
{"title":"Incremental Semisupervised Learning With Adaptive Locality Preservation for High-Dimensional Data","authors":"Guojie Li;Zhiwen Yu;Kaixiang Yang;Ziwei Fan;C. L. Philip Chen","doi":"10.1109/TAI.2025.3560592","DOIUrl":"https://doi.org/10.1109/TAI.2025.3560592","url":null,"abstract":"Broad learning system (BLS) has been widely researched and applied in the field of semisupervised learning. However, current semisupervised BLS methods rely on predefined graph structures. High-dimensional small-sample data, characterized by abundant redundant and noisy features with complex distribution patterns, often leads to the construction of poor-quality predefined graphs, thereby constraining the model’s performance. Additionally, the random generation of feature and enhancement nodes in BLS, combined with limited data labels, results in suboptimal model performance. To address these issues, this article first proposes a broad learning system with adaptive locality preservation (BLS-ALP). This method employs adaptive locality preservation constraints in the output space to ensure that similar samples share the same label, iteratively updating the graph structure. To further enhance the performance of BLS-ALP, an incremental ensemble framework (IBLS-ALP) is proposed. This framework effectively mitigates the impact of redundant and noisy features by using multiple random subspaces instead of the original high-dimensional space. Additionally, IBLS-ALP enhances the utilization of a small number of labels by incorporating residual labels, thereby significantly improving the model’s overall performance. Extensive experiments conducted on various high-dimensional small-sample datasets demonstrate that IBLS-ALP exhibits superior performance.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 11","pages":"2990-3004"},"PeriodicalIF":0.0,"publicationDate":"2025-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145455763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EEG Emotion Recognition Based on an Implicit Emotion Regulatory Mechanism 基于内隐情绪调节机制的脑电情绪识别
Pub Date : 2025-04-14 DOI: 10.1109/TAI.2025.3560593
Dongdong Li;Zhishuo Jin;Yujun Shen;Zhe Wang;Suo Jiang
One of the main challenges in electroencephalography (EEG) emotion recognition is the lack of understanding of the biological properties of the brain and how they relate to emotions. To address this issue, this article proposes an implicit emotion regulatory mechanism inspired contrastive learning framework (CLIER) for EEG emotion recognition. The framework simulates the complex relationship between emotions and the underlying neurobiological processes; to achieve this, the mechanism is mainly simulated through three parts. First, to leverage the interindividual variability of emotional expression, the emotion features of the individual are captured by a dynamic connection graph in the subject-dependent setting. Subsequently, reverse regulation is simulated by contrast learning based on label information and data augmentation to capture more biologically specific emotional features. Finally, caused by the asymmetry between the left and right hemispheres of the human brain in response to emotions, brain lateralization mutual learning facilitates the fusion of the hemispheres in determining emotions. Experiments on SEED, SEED-IV, SEED-V, and EREMUS datasets show impressive results: 93.4% accuracy on SEED, 90.2% on SEED-IV, 82.46% on SEED-V, and 41.63% on EREMUS. Employing an identical experimental protocol, our model demonstrated superior performance relative to the majority of existing methods, thus showcasing its effectiveness in the realm of EEG emotion recognition.
脑电图(EEG)情绪识别的主要挑战之一是缺乏对大脑生物学特性及其与情绪的关系的理解。为了解决这一问题,本文提出了一种基于内隐情绪调节机制的脑电情绪识别对比学习框架(CLIER)。该框架模拟了情绪与潜在神经生物学过程之间的复杂关系;为此,主要通过三个部分对该机制进行仿真。首先,为了利用情绪表达的个体间可变性,个体的情绪特征被主体依赖设置中的动态连接图捕获。随后,通过基于标签信息和数据增强的对比学习模拟反向调节,以捕获更多生物特异性情绪特征。最后,由于人类大脑左右半球对情绪的反应不对称,大脑侧化相互学习促进了左右半球在决定情绪时的融合。在SEED、SEED- iv、SEED- v和EREMUS数据集上的实验显示,SEED的准确率为93.4%,SEED- iv的准确率为90.2%,SEED- v的准确率为82.46%,EREMUS的准确率为41.63%。采用相同的实验方案,我们的模型相对于大多数现有方法显示出优越的性能,从而展示了其在EEG情感识别领域的有效性。
{"title":"EEG Emotion Recognition Based on an Implicit Emotion Regulatory Mechanism","authors":"Dongdong Li;Zhishuo Jin;Yujun Shen;Zhe Wang;Suo Jiang","doi":"10.1109/TAI.2025.3560593","DOIUrl":"https://doi.org/10.1109/TAI.2025.3560593","url":null,"abstract":"One of the main challenges in electroencephalography (EEG) emotion recognition is the lack of understanding of the biological properties of the brain and how they relate to emotions. To address this issue, this article proposes an implicit emotion regulatory mechanism inspired contrastive learning framework (CLIER) for EEG emotion recognition. The framework simulates the complex relationship between emotions and the underlying neurobiological processes; to achieve this, the mechanism is mainly simulated through three parts. First, to leverage the interindividual variability of emotional expression, the emotion features of the individual are captured by a dynamic connection graph in the subject-dependent setting. Subsequently, reverse regulation is simulated by contrast learning based on label information and data augmentation to capture more biologically specific emotional features. Finally, caused by the asymmetry between the left and right hemispheres of the human brain in response to emotions, brain lateralization mutual learning facilitates the fusion of the hemispheres in determining emotions. Experiments on SEED, SEED-IV, SEED-V, and EREMUS datasets show impressive results: 93.4% accuracy on SEED, 90.2% on SEED-IV, 82.46% on SEED-V, and 41.63% on EREMUS. Employing an identical experimental protocol, our model demonstrated superior performance relative to the majority of existing methods, thus showcasing its effectiveness in the realm of EEG emotion recognition.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 11","pages":"3005-3017"},"PeriodicalIF":0.0,"publicationDate":"2025-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145455996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Residual Learning of a Probabilistic’ Partial Least Squares Model for Predictive Data Analytics 预测数据分析中概率偏最小二乘模型的深度残差学习
Pub Date : 2025-04-11 DOI: 10.1109/TAI.2025.3560248
Zhiqiang Ge;Duxin Chen;Wenwu Yu
Recently, probabilistic latent variable models have played an important role in data analytics in various industrial application scenarios, such as process monitoring, fault diagnosis, and soft sensing. Inspired by the idea of lightweight deep learning, this article proposes a new deep residual learning method for the probabilistic’ partial least squares (PLSs) model. First, layerwise probabilistic modeling is carried out to extract supervised latent variables in different hidden layers of the deep model using a well-designed expectation-maximization algorithm for parameter optimization. Through this layerwise residual learning process, more target-related latent variables can be extracted, which are supervised by the outputs of the predictive model. Next, an additional probabilistic model is constructed for information fusion and further extraction of supervised latent variables which are highly related to the modeling target. In fact, this step can be considered as an ensemble learning strategy, which has great potentials in decreasing modeling error and reducing prediction uncertainty. A soft-sensing strategy is then developed for online prediction of key variables. The performance is evaluated using two industrial examples. Compared to the shallow probabilistic model, the performance of the deep model has been improved by 10%–20%.
近年来,概率潜变量模型在过程监控、故障诊断和软测量等各种工业应用场景的数据分析中发挥了重要作用。受轻量级深度学习思想的启发,本文提出了一种新的概率偏最小二乘(pls)模型的深度残差学习方法。首先,利用设计良好的期望最大化算法进行参数优化,进行分层概率建模,提取深度模型不同隐藏层的有监督潜在变量。通过这种分层残差学习过程,可以提取更多与目标相关的潜在变量,这些潜在变量由预测模型的输出进行监督。其次,构建一个附加的概率模型,用于信息融合和进一步提取与建模目标高度相关的监督潜变量。实际上,这一步可以看作是一种集成学习策略,在减少建模误差和降低预测不确定性方面具有很大的潜力。然后开发了一种软测量策略,用于在线预测关键变量。用两个工业实例对其性能进行了评价。与浅层概率模型相比,深层模型的性能提高了10%-20%。
{"title":"Deep Residual Learning of a Probabilistic’ Partial Least Squares Model for Predictive Data Analytics","authors":"Zhiqiang Ge;Duxin Chen;Wenwu Yu","doi":"10.1109/TAI.2025.3560248","DOIUrl":"https://doi.org/10.1109/TAI.2025.3560248","url":null,"abstract":"Recently, probabilistic latent variable models have played an important role in data analytics in various industrial application scenarios, such as process monitoring, fault diagnosis, and soft sensing. Inspired by the idea of lightweight deep learning, this article proposes a new deep residual learning method for the probabilistic’ partial least squares (PLSs) model. First, layerwise probabilistic modeling is carried out to extract supervised latent variables in different hidden layers of the deep model using a well-designed expectation-maximization algorithm for parameter optimization. Through this layerwise residual learning process, more target-related latent variables can be extracted, which are supervised by the outputs of the predictive model. Next, an additional probabilistic model is constructed for information fusion and further extraction of supervised latent variables which are highly related to the modeling target. In fact, this step can be considered as an ensemble learning strategy, which has great potentials in decreasing modeling error and reducing prediction uncertainty. A soft-sensing strategy is then developed for online prediction of key variables. The performance is evaluated using two industrial examples. Compared to the shallow probabilistic model, the performance of the deep model has been improved by 10%–20%.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 11","pages":"2977-2989"},"PeriodicalIF":0.0,"publicationDate":"2025-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145455952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Parallel Inductive Shift Learning Based Recommendation System 基于并行感应移位学习的推荐系统
Pub Date : 2025-04-09 DOI: 10.1109/TAI.2025.3558183
Nilufar Zaman;Angshuman Jana
In today’s world, online services have revolutionized human activities and thus the consumers expect their service providers to make their online experiences more fruitful by recommending the relevant services to them. In this case, it becomes really challenging for the service providers to provide recommendation to a user whose information’s and preferences are unavailable. This issue is handled by cross-domain approach, which explores similar users across various domains in the same platform. However, the main concern with this cross-domain approach is that the information needs to be available in any domain of one platform. Thus, a multidomain recommendation is designed to optimize the recommendation system performance by analyzing the information obtained from multiple platforms. However, existing multidomain recommendation model has mainly two challenges. First, there are no overlapping users to understand the similarities between them. Second, the transfer learning approach in multidomain allows the transfer of information from only the source to the target domain. Therefore, our proposed approach consider the parallel inductive shift learning (PISL) model to address these two above-mentioned challenges. For the first challenge, we have focused to identify the similarities between user–user and user–item by considering various features of user and item. For the next challenge, our proposed model analyzes the source and the target domain simultaneously and thus does a parallel transfer of information from the source to the target domain and vice versa. We have tested our model for three real-life movie and book datasets i.e. for the movie dataset we have used Movielens, Amazon, and Netflix datasets. In contrast, for the book dataset, we have used the Amazon, Good Reads, and Book Crossing dataset, which proves to outperform the other state-of-the-art approaches.
在当今世界,在线服务已经彻底改变了人类的活动,因此消费者期望他们的服务提供商通过向他们推荐相关的服务来使他们的在线体验更加富有成效。在这种情况下,服务提供者向信息和偏好不可用的用户提供推荐变得非常具有挑战性。这个问题是通过跨域方法处理的,该方法在同一平台的不同域中探索类似的用户。然而,这种跨领域方法的主要问题是信息需要在一个平台的任何领域中可用。因此,我们设计了一个多领域推荐,通过分析多个平台获得的信息来优化推荐系统的性能。然而,现有的多领域推荐模型主要存在两方面的挑战。首先,没有重叠的用户来理解它们之间的相似之处。第二,多域迁移学习方法允许信息仅从源域迁移到目标域。因此,我们提出的方法考虑并行归纳移位学习(PISL)模型来解决上述两个挑战。对于第一个挑战,我们专注于通过考虑用户和物品的各种特征来识别用户-用户和用户-物品之间的相似性。对于下一个挑战,我们提出的模型同时分析源域和目标域,从而将信息从源域并行传输到目标域,反之亦然。我们已经为三个现实生活中的电影和书籍数据集测试了我们的模型,即对于电影数据集,我们使用了Movielens, Amazon和Netflix数据集。相比之下,对于图书数据集,我们使用了Amazon、Good Reads和book Crossing数据集,事实证明,这些数据集的性能优于其他最先进的方法。
{"title":"Parallel Inductive Shift Learning Based Recommendation System","authors":"Nilufar Zaman;Angshuman Jana","doi":"10.1109/TAI.2025.3558183","DOIUrl":"https://doi.org/10.1109/TAI.2025.3558183","url":null,"abstract":"In today’s world, online services have revolutionized human activities and thus the consumers expect their service providers to make their online experiences more fruitful by recommending the relevant services to them. In this case, it becomes really challenging for the service providers to provide recommendation to a user whose information’s and preferences are unavailable. This issue is handled by cross-domain approach, which explores similar users across various domains in the same platform. However, the main concern with this cross-domain approach is that the information needs to be available in any domain of one platform. Thus, a multidomain recommendation is designed to optimize the recommendation system performance by analyzing the information obtained from multiple platforms. However, existing multidomain recommendation model has mainly two challenges. First, there are no overlapping users to understand the similarities between them. Second, the transfer learning approach in multidomain allows the transfer of information from only the source to the target domain. Therefore, our proposed approach consider the parallel inductive shift learning (PISL) model to address these two above-mentioned challenges. For the first challenge, we have focused to identify the similarities between user–user and user–item by considering various features of user and item. For the next challenge, our proposed model analyzes the source and the target domain simultaneously and thus does a parallel transfer of information from the source to the target domain and vice versa. We have tested our model for three real-life movie and book datasets i.e. for the movie dataset we have used Movielens, Amazon, and Netflix datasets. In contrast, for the book dataset, we have used the Amazon, Good Reads, and Book Crossing dataset, which proves to outperform the other state-of-the-art approaches.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 11","pages":"2953-2965"},"PeriodicalIF":0.0,"publicationDate":"2025-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145428951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deterministic–Probabilistic Approach to Neural Network Pruning 神经网络剪枝的确定性-概率方法
Pub Date : 2025-04-08 DOI: 10.1109/TAI.2025.3558718
Soumyadipta Banerjee;Jiaul H. Paik
Modern deep networks are highly over-parameterized. Thus, training and testing such models in various applications are computationally intensive with excessive memory and energy requirements. Network pruning aims to find smaller subnetworks from within these dense networks that do not compromise on the test accuracy. In this article, we present a probabilistic and deterministic pruning methodology which determines the likelihood of retention of the weight parameters by modeling the layer-specific distribution of extreme values of the weights. Our method automatically finds the sparsity in each layer, unlike existing pruning techniques which require an explicit input of the sparsity information. Experiments in the present work show that deterministic–probabilistic pruning consistently achieves high sparsity levels, ranging from 65 to 95%, while maintaining comparable or improved testing accuracy across multiple datasets such as MNIST, CIFAR-10, and Tiny ImageNet, on architectures including VGG-16, ResNet-18, and ResNet-50.
现代深度网络是高度过度参数化的。因此,在各种应用程序中训练和测试这样的模型是计算密集型的,并且需要过多的内存和能量。网络修剪的目的是在这些密集的网络中找到较小的子网,这些子网不会影响测试的准确性。在本文中,我们提出了一种概率和确定性修剪方法,该方法通过对权重极值的特定层分布建模来确定权重参数保留的可能性。我们的方法可以自动找到每一层的稀疏性,而不像现有的修剪技术需要显式输入稀疏性信息。本工作中的实验表明,确定性-概率修剪始终达到高稀疏度水平,范围从65到95%,同时在多个数据集(如MNIST, CIFAR-10和Tiny ImageNet)上保持相当或改进的测试精度,架构包括VGG-16, ResNet-18和ResNet-50。
{"title":"A Deterministic–Probabilistic Approach to Neural Network Pruning","authors":"Soumyadipta Banerjee;Jiaul H. Paik","doi":"10.1109/TAI.2025.3558718","DOIUrl":"https://doi.org/10.1109/TAI.2025.3558718","url":null,"abstract":"Modern deep networks are highly over-parameterized. Thus, training and testing such models in various applications are computationally intensive with excessive memory and energy requirements. Network pruning aims to find smaller subnetworks from within these dense networks that do not compromise on the test accuracy. In this article, we present a probabilistic and deterministic pruning methodology which determines the likelihood of retention of the weight parameters by modeling the layer-specific distribution of extreme values of the weights. Our method automatically finds the sparsity in each layer, unlike existing pruning techniques which require an explicit input of the sparsity information. Experiments in the present work show that deterministic–probabilistic pruning consistently achieves high sparsity levels, ranging from 65 to 95%, while maintaining comparable or improved testing accuracy across multiple datasets such as MNIST, CIFAR-10, and Tiny ImageNet, on architectures including VGG-16, ResNet-18, and ResNet-50.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 10","pages":"2830-2839"},"PeriodicalIF":0.0,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145196043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Artificial Intelligence Publication Information IEEE人工智能学报
Pub Date : 2025-03-31 DOI: 10.1109/TAI.2025.3551528
{"title":"IEEE Transactions on Artificial Intelligence Publication Information","authors":"","doi":"10.1109/TAI.2025.3551528","DOIUrl":"https://doi.org/10.1109/TAI.2025.3551528","url":null,"abstract":"","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 4","pages":"C2-C2"},"PeriodicalIF":0.0,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10946100","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143740222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Boosting 3-D Point Cloud Registration by Orthogonal Self-Ensemble Learning 正交自集成学习促进三维点云配准
Pub Date : 2025-03-30 DOI: 10.1109/TAI.2025.3575036
Mingzhi Yuan;Ao Shen;Yingfan Ma;Jie Du;Qiao Huang;Manning Wang
Deep learning has significantly advanced the development of point cloud registration. However, in recent years, some methods have relied on additional sensor information or complex network designs to improve registration performance, which incurs considerable computational overhead. These methods often struggle to strike a reasonable balance between computational cost and performance gains. To address this, we propose a plug-and-play orthogonal self-ensemble module designed to enhance registration performance with minimal additional overhead. Specifically, we design a novel ensemble learning strategy to mine the complementary information within the extracted features of previous methods. Unlike most ensemble learning methods, our method does not set multiple complex models for performance enhancement. Instead, it only cascades a lightweight dual-branch network after the features extracted by the original model to obtain two sets of features with more diversity. To further reduce redundancy between features and prevent the degradation of the dual-branch network, we introduce an orthogonal constraint that ensures the features output by the two branches are more complementary. Finally, by concatenating the two sets of complementary features, the final enhanced features are obtained. Compared to the original features, these enhanced features thoroughly exploit the internal information and exhibit greater distinctiveness, leading to improved registration performance. To validate the effectiveness of our method, we plug it into GeoTransformer, resulting in consistent performance improvements across 3DMatch, KITTI, and ModelNet40 datasets. Moreover, our method is compatible with other performance-enhancing methods. In conjunction with the overlap prior in PEAL, GeoTransformer achieves a new state-of-the-art performance.
深度学习极大地推动了点云配准的发展。然而,近年来,一些方法依赖于额外的传感器信息或复杂的网络设计来提高配准性能,这带来了相当大的计算开销。这些方法通常难以在计算成本和性能增益之间取得合理的平衡。为了解决这个问题,我们提出了一个即插即用的正交自集成模块,旨在以最小的额外开销提高注册性能。具体而言,我们设计了一种新的集成学习策略来挖掘先前方法提取的特征中的互补信息。与大多数集成学习方法不同,我们的方法不需要为性能增强设置多个复杂的模型。它只是在原始模型提取的特征之后级联一个轻量级的双分支网络,得到两组更多样化的特征。为了进一步减少特征之间的冗余并防止双分支网络的退化,我们引入了一个正交约束,以确保两个分支输出的特征更具互补性。最后,通过将两组互补特征串接,得到最终的增强特征。与原始特征相比,这些增强特征充分利用了内部信息,具有更强的显著性,从而提高了配准性能。为了验证我们方法的有效性,我们将其插入GeoTransformer,从而在3DMatch、KITTI和ModelNet40数据集上实现一致的性能改进。此外,我们的方法与其他性能增强方法兼容。与PEAL中的重叠先验相结合,GeoTransformer实现了新的最先进的性能。
{"title":"Boosting 3-D Point Cloud Registration by Orthogonal Self-Ensemble Learning","authors":"Mingzhi Yuan;Ao Shen;Yingfan Ma;Jie Du;Qiao Huang;Manning Wang","doi":"10.1109/TAI.2025.3575036","DOIUrl":"https://doi.org/10.1109/TAI.2025.3575036","url":null,"abstract":"Deep learning has significantly advanced the development of point cloud registration. However, in recent years, some methods have relied on additional sensor information or complex network designs to improve registration performance, which incurs considerable computational overhead. These methods often struggle to strike a reasonable balance between computational cost and performance gains. To address this, we propose a plug-and-play orthogonal self-ensemble module designed to enhance registration performance with minimal additional overhead. Specifically, we design a novel ensemble learning strategy to mine the complementary information within the extracted features of previous methods. Unlike most ensemble learning methods, our method does not set multiple complex models for performance enhancement. Instead, it only cascades a lightweight dual-branch network after the features extracted by the original model to obtain two sets of features with more diversity. To further reduce redundancy between features and prevent the degradation of the dual-branch network, we introduce an orthogonal constraint that ensures the features output by the two branches are more complementary. Finally, by concatenating the two sets of complementary features, the final enhanced features are obtained. Compared to the original features, these enhanced features thoroughly exploit the internal information and exhibit greater distinctiveness, leading to improved registration performance. To validate the effectiveness of our method, we plug it into GeoTransformer, resulting in consistent performance improvements across 3DMatch, KITTI, and ModelNet40 datasets. Moreover, our method is compatible with other performance-enhancing methods. In conjunction with the overlap prior in PEAL, GeoTransformer achieves a new state-of-the-art performance.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"7 1","pages":"375-384"},"PeriodicalIF":0.0,"publicationDate":"2025-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145898230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MALADY: Multiclass Active Learning With Auction Dynamics on Graphs MALADY:多班级主动学习与拍卖动态图
Pub Date : 2025-03-30 DOI: 10.1109/TAI.2025.3575038
Gokul Bhusal;Kevin Miller;Ekaterina Merkurjev
Active learning (AL) enhances the performance of machine learning (ML) methods, particularly in low-label rate scenarios, by judiciously selecting a limited number of unlabeled data points for labeling, with the goal of improving the performance of an underlying classifier. In this work, we introduce the multiclass AL with auction dynamics on graphs (MALADY) algorithm, which leverages an auction dynamics technique on similarity graphs for efficient AL. In particular, the proposed algorithm incorporates an AL loop using as its underlying semisupervised procedure an efficient and effective similarity graph-based auction method consisting of upper and lower bound auctions that integrate class size constraints. In addition, we introduce a novel AL acquisition function that incorporates the dual variable of the auction algorithm to measure the uncertainty in the classifier to prioritize queries near the decision boundaries between different classes. Overall, the proposed method can efficiently obtain accurate results using extremely small labeled sets containing just a few elements per class; this is crucial since labeled data are scarce for many applications. Moreover, the proposed technique can incorporate class size information, which improves accuracy even further. Last, using experiments on classification tasks and various datasets, we evaluate the performance of our proposed method and show that it exceeds that of comparison algorithms.
主动学习(AL)通过明智地选择有限数量的未标记数据点进行标记,以提高底层分类器的性能,增强了机器学习(ML)方法的性能,特别是在低标记率场景下。在这项工作中,我们引入了带有图上拍卖动态的多类人工智能(MALADY)算法,该算法利用相似图上的拍卖动态技术来实现高效的人工智能。特别是,所提出的算法结合了一个人工智能循环,作为其底层半监督过程,该循环使用了一种高效且有效的基于相似图的拍卖方法,该方法由整合了类大小约束的上界和下界拍卖组成。此外,我们引入了一种新的人工智能获取函数,该函数结合了拍卖算法的双变量来衡量分类器中的不确定性,从而在不同类别之间的决策边界附近优先考虑查询。总的来说,该方法可以使用极小的标记集,每个类只包含几个元素,从而有效地获得准确的结果;这是至关重要的,因为标记数据对于许多应用程序来说是稀缺的。此外,所提出的技术可以纳入班级规模信息,这进一步提高了准确性。最后,通过对分类任务和各种数据集的实验,我们评估了我们提出的方法的性能,并表明它优于比较算法。
{"title":"MALADY: Multiclass Active Learning With Auction Dynamics on Graphs","authors":"Gokul Bhusal;Kevin Miller;Ekaterina Merkurjev","doi":"10.1109/TAI.2025.3575038","DOIUrl":"https://doi.org/10.1109/TAI.2025.3575038","url":null,"abstract":"Active learning (AL) enhances the performance of machine learning (ML) methods, particularly in low-label rate scenarios, by judiciously selecting a limited number of unlabeled data points for labeling, with the goal of improving the performance of an underlying classifier. In this work, we introduce the multiclass AL with auction dynamics on graphs (MALADY) algorithm, which leverages an auction dynamics technique on similarity graphs for efficient AL. In particular, the proposed algorithm incorporates an AL loop using as its underlying semisupervised procedure an efficient and effective similarity graph-based auction method consisting of upper and lower bound auctions that integrate class size constraints. In addition, we introduce a novel AL acquisition function that incorporates the dual variable of the auction algorithm to measure the uncertainty in the classifier to prioritize queries near the decision boundaries between different classes. Overall, the proposed method can efficiently obtain accurate results using extremely small labeled sets containing just a few elements per class; this is crucial since labeled data are scarce for many applications. Moreover, the proposed technique can incorporate class size information, which improves accuracy even further. Last, using experiments on classification tasks and various datasets, we evaluate the performance of our proposed method and show that it exceeds that of comparison algorithms.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"7 1","pages":"385-398"},"PeriodicalIF":0.0,"publicationDate":"2025-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145898218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Multiscale Dynamic Graph Convolutional Network for Traffic Data Cognition 一种新的交通数据认知多尺度动态图卷积网络
Pub Date : 2025-03-29 DOI: 10.1109/TAI.2025.3574655
Jiyao An;Zhaohui Pu;Qingqin Liu;Lei Zhang;Md Sohel Rana
This article investigates traffic data cognitive modelling problem in real traffic scene by fully utilizing multiscale spatio-temporal dependence between multiple traffic nodes, along with a novel dynamic graph convolutional network (GCN). Most recently, the deep learning network model is weighed down by some practical problems focused on as follows: 1) The existing graph convolution operations typically aggregate information from the given k-hop neighbors; and 2) How to model the similarity of traffic data patterns among these nodes given the spatio-temporal heterogeneity of traffic data. In this article, we propose a novel hierarchical traffic data cognitive modelling framework called multiscale spatio-temporal dynamic graph convolutional network architecture (MSST-DGCN). And, a multiscale graph convolution module is first constructed to expand the receptive field of convolutional operations, by developing a novel sub-GCNs cumulative concatenation mechanism. Meanwhile, two specified dynamic graphs are designed to model the spatio-temporal correlation among these nodes from both a proximity and long-term perspective through a novel Gaussian calculation strategy, which are efficiently able to represent/cognize the dynamic similarity of traffic data patterns. Through a series of qualitative evaluations, the present model has the ability to perceive the traffic data pattern states of nodes. At last, two real world traffic datasets experiments are developed to show that the proposed approach achieves state-of-the-art traffic data cognitive performance.
本文通过充分利用多个交通节点之间的多尺度时空依赖关系,结合一种新的动态图卷积网络(GCN),研究了真实交通场景下的交通数据认知建模问题。最近,深度学习网络模型被一些实际问题所困扰,主要集中在以下几个方面:1)现有的图卷积操作通常是从给定的k-hop邻居中聚集信息;2)考虑交通数据的时空异质性,如何建立节点间交通数据模式相似性模型。在本文中,我们提出了一种新的分层交通数据认知建模框架——多尺度时空动态图卷积网络架构(MSST-DGCN)。并且,通过开发一种新的子gcns累积级联机制,首先构建了一个多尺度图卷积模块来扩展卷积运算的接受域。同时,设计了两个指定的动态图,通过一种新颖的高斯计算策略,从近距离和长期角度对这些节点之间的时空相关性进行建模,能够有效地表示/识别交通数据模式的动态相似度。通过一系列定性评价,该模型具有感知节点交通数据模式状态的能力。最后,通过两个真实世界的交通数据集实验表明,该方法达到了最先进的交通数据认知性能。
{"title":"A Novel Multiscale Dynamic Graph Convolutional Network for Traffic Data Cognition","authors":"Jiyao An;Zhaohui Pu;Qingqin Liu;Lei Zhang;Md Sohel Rana","doi":"10.1109/TAI.2025.3574655","DOIUrl":"https://doi.org/10.1109/TAI.2025.3574655","url":null,"abstract":"This article investigates traffic data cognitive modelling problem in real traffic scene by fully utilizing multiscale spatio-temporal dependence between multiple traffic nodes, along with a novel dynamic graph convolutional network (GCN). Most recently, the deep learning network model is weighed down by some practical problems focused on as follows: 1) The existing graph convolution operations typically aggregate information from the given k-hop neighbors; and 2) How to model the similarity of traffic data patterns among these nodes given the spatio-temporal heterogeneity of traffic data. In this article, we propose a novel hierarchical traffic data cognitive modelling framework called multiscale spatio-temporal dynamic graph convolutional network architecture (MSST-DGCN). And, a multiscale graph convolution module is first constructed to expand the receptive field of convolutional operations, by developing a novel sub-GCNs cumulative concatenation mechanism. Meanwhile, two specified dynamic graphs are designed to model the spatio-temporal correlation among these nodes from both a proximity and long-term perspective through a novel Gaussian calculation strategy, which are efficiently able to represent/cognize the dynamic similarity of traffic data patterns. Through a series of qualitative evaluations, the present model has the ability to perceive the traffic data pattern states of nodes. At last, two real world traffic datasets experiments are developed to show that the proposed approach achieves state-of-the-art traffic data cognitive performance.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"7 1","pages":"362-374"},"PeriodicalIF":0.0,"publicationDate":"2025-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145898231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on artificial intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1