首页 > 最新文献

Neural Networks最新文献

英文 中文
BAED: A new paradigm for few-shot graph learning with explanation in the loop 一个在循环中有解释的少镜头图学习的新范例。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-02-03 DOI: 10.1016/j.neunet.2026.108673
Chao Chen , Xujia Li , Dongsheng Hong , Shanshan Lin , Xiangwen Liao , Chuanyi Liu , Lei Chen
The challenges of training and inference in few-shot environments persist in the area of graph representation learning. The quality and quantity of labels are often insufficient due to the extensive expert knowledge required to annotate graph data. In this context, Few-Shot Graph Learning (FSGL) approaches have been developed over the years. Through sophisticated neural architectures and customized training pipelines, these approaches enhance model adaptability to new label distributions. However, compromises in the model’s robustness and interpretability can result in overfitting to noise in labeled data and degraded performance. This paper introduces the first explanation-in-the-loop framework for the FSGL problem, called BAED. We novelly employ the belief propagation algorithm to facilitate label augmentation on graphs. Then, leveraging an auxiliary graph neural network and the gradient backpropagation method, our framework effectively extracts explanatory subgraphs surrounding target nodes. The final predictions are based on these informative subgraphs while mitigating the influence of redundant information from neighboring nodes. Extensive experiments on seven benchmark datasets demonstrate superior prediction accuracy, training efficiency, and explanation quality of BAED. As a pioneer, this work highlights the potential of the explanation-based research paradigm in FSGL.
在图表示学习领域,在少镜头环境下的训练和推理仍然存在挑战。由于标注图形数据需要大量的专家知识,标签的质量和数量往往不足。在这种背景下,近年来已经开发了几次图学习(FSGL)方法。通过复杂的神经结构和定制的训练管道,这些方法增强了模型对新标签分布的适应性。然而,模型的鲁棒性和可解释性的妥协可能导致对标记数据中的噪声的过度拟合和性能下降。本文介绍了FSGL问题的第一个循环解释框架,称为BAED。我们新颖地采用了信念传播算法来方便图上的标签扩充。然后,利用辅助图神经网络和梯度反向传播方法,我们的框架有效地提取目标节点周围的解释子图。最终的预测基于这些信息子图,同时减轻了来自邻近节点的冗余信息的影响。在7个基准数据集上的大量实验表明,BAED具有较高的预测精度、训练效率和解释质量。作为先驱,这项工作突出了基于解释的研究范式在FSGL中的潜力。
{"title":"BAED: A new paradigm for few-shot graph learning with explanation in the loop","authors":"Chao Chen ,&nbsp;Xujia Li ,&nbsp;Dongsheng Hong ,&nbsp;Shanshan Lin ,&nbsp;Xiangwen Liao ,&nbsp;Chuanyi Liu ,&nbsp;Lei Chen","doi":"10.1016/j.neunet.2026.108673","DOIUrl":"10.1016/j.neunet.2026.108673","url":null,"abstract":"<div><div>The challenges of training and inference in few-shot environments persist in the area of graph representation learning. The quality and quantity of labels are often insufficient due to the extensive expert knowledge required to annotate graph data. In this context, Few-Shot Graph Learning (FSGL) approaches have been developed over the years. Through sophisticated neural architectures and customized training pipelines, these approaches enhance model adaptability to new label distributions. However, compromises in the model’s robustness and interpretability can result in overfitting to noise in labeled data and degraded performance. This paper introduces the first explanation-in-the-loop framework for the FSGL problem, called BAED. We novelly employ the belief propagation algorithm to facilitate label augmentation on graphs. Then, leveraging an auxiliary graph neural network and the gradient backpropagation method, our framework effectively extracts explanatory subgraphs surrounding target nodes. The final predictions are based on these informative subgraphs while mitigating the influence of redundant information from neighboring nodes. Extensive experiments on seven benchmark datasets demonstrate superior prediction accuracy, training efficiency, and explanation quality of BAED. As a pioneer, this work highlights the potential of the explanation-based research paradigm in FSGL.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108673"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146158712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A punishment neural network-based acceleration-level joint drift-free scheme for solving constrained motion planning problem of redundant robotic manipulators 一种基于惩罚神经网络的冗余机器人关节无漂移约束运动规划方案。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-02-04 DOI: 10.1016/j.neunet.2026.108684
Zhijun Zhang, Xitong Gao, Jinjia Guo
To solve the repetitive motion problem of redundant robotic manipulators, a punishment neural network-based acceleration-level joint drift-free (PNN-ALJDF) scheme is designed. Traditional joint physical limits constraints are fixed and lack margin. Thus, a novel joint acceleration time-varying constraint is considered in the PNN-ALJDF scheme to avoid the joint state exceeding the physical limits. In addition, to ensure that redundant robotic manipulators can periodically return to the initial pose, a joint drift-free criterion is designed. Furthermore, the joint drift-free criterion, kinematics equation and joint acceleration time-varying constraint are formulated globally as an acceleration-level joint drift-free (ALJDF) scheme by a time-varying quadratic programming approach. Then, the ALJDF scheme is solved by the designed punishment neural network. Thus, the proposed PNN-ALJDF scheme is composed of the ALJDF scheme and punishment neural network. Finally, the simulations demonstrate that the PNN-ALJDF scheme avoids joints from drifting, and the states of joints are all within the acceleration time-varying constraint. In addition, the proposed PNN-ALJDF has higher solution accuracy than the linear variational inequalities-based primal-dual neural network.
为解决冗余机器人的重复运动问题,设计了一种基于惩罚神经网络的加速度级关节无漂移(PNN-ALJDF)方案。传统的关节物理极限约束是固定的,缺乏余量。因此,在PNN-ALJDF方案中考虑了一种新的关节加速度时变约束,以避免关节状态超出物理极限。此外,为了保证冗余机器人能周期性地恢复到初始姿态,设计了关节无漂移准则。在此基础上,采用时变二次规划方法,将关节无漂移准则、运动学方程和关节加速度时变约束全局化为加速度级关节无漂移格式。然后,利用设计的惩罚神经网络对ALJDF方案进行求解。因此,提出的PNN-ALJDF方案由ALJDF方案和惩罚神经网络组成。最后,仿真结果表明,PNN-ALJDF方案避免了关节漂移,且关节状态均在加速度时变约束范围内。此外,与基于线性变分不等式的原始对偶神经网络相比,所提出的PNN-ALJDF具有更高的解精度。
{"title":"A punishment neural network-based acceleration-level joint drift-free scheme for solving constrained motion planning problem of redundant robotic manipulators","authors":"Zhijun Zhang,&nbsp;Xitong Gao,&nbsp;Jinjia Guo","doi":"10.1016/j.neunet.2026.108684","DOIUrl":"10.1016/j.neunet.2026.108684","url":null,"abstract":"<div><div>To solve the repetitive motion problem of redundant robotic manipulators, a punishment neural network-based acceleration-level joint drift-free (PNN-ALJDF) scheme is designed. Traditional joint physical limits constraints are fixed and lack margin. Thus, a novel joint acceleration time-varying constraint is considered in the PNN-ALJDF scheme to avoid the joint state exceeding the physical limits. In addition, to ensure that redundant robotic manipulators can periodically return to the initial pose, a joint drift-free criterion is designed. Furthermore, the joint drift-free criterion, kinematics equation and joint acceleration time-varying constraint are formulated globally as an acceleration-level joint drift-free (ALJDF) scheme by a time-varying quadratic programming approach. Then, the ALJDF scheme is solved by the designed punishment neural network. Thus, the proposed PNN-ALJDF scheme is composed of the ALJDF scheme and punishment neural network. Finally, the simulations demonstrate that the PNN-ALJDF scheme avoids joints from drifting, and the states of joints are all within the acceleration time-varying constraint. In addition, the proposed PNN-ALJDF has higher solution accuracy than the linear variational inequalities-based primal-dual neural network.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108684"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146158685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HP-GAN: Harnessing pretrained networks for GAN improvement with FakeTwins and discriminator consistency HP-GAN:利用假双胞胎和鉴别器一致性利用预训练网络进行GAN改进。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-01-29 DOI: 10.1016/j.neunet.2026.108666
Geonhui Son , Jeong Ryong Lee , Dosik Hwang
Generative Adversarial Networks (GANs) have made significant progress in enhancing the quality of image synthesis. Recent methods frequently leverage pretrained networks to calculate perceptual losses or utilize pretrained feature spaces. In this paper, we extend the capabilities of pretrained networks by incorporating innovative self-supervised learning techniques and enforcing consistency between discriminators during GAN training. Our proposed method, named HP-GAN, effectively exploits neural network priors through two primary strategies: FakeTwins and discriminator consistency. FakeTwins leverages pretrained networks as encoders to compute a self-supervised loss and applies this through the generated images to train the generator, thereby enabling the generation of more diverse and high quality images. Additionally, we introduce a consistency mechanism between discriminators that evaluate feature maps extracted from Convolutional Neural Network (CNN) and Vision Transformer (ViT) feature networks. Discriminator consistency promotes coherent learning among discriminators and enhances training robustness by aligning their assessments of image quality. Our extensive evaluation across seventeen datasets-including scenarios with large, small, and limited data, and covering a variety of image domains-demonstrates that HP-GAN consistently outperforms current state-of-the-art methods in terms of Fréchet Inception Distance (FID), achieving significant improvements in image diversity and quality. Code is available at: https://github.com/higun2/HP-GAN.
生成对抗网络(GANs)在提高图像合成质量方面取得了重大进展。最近的方法经常利用预训练的网络来计算感知损失或利用预训练的特征空间。在本文中,我们通过结合创新的自监督学习技术和在GAN训练期间强制判别器之间的一致性来扩展预训练网络的能力。我们提出的方法,称为HP-GAN,通过两个主要策略:FakeTwins和鉴别器一致性,有效地利用神经网络先验。FakeTwins利用预训练网络作为编码器来计算自监督损失,并通过生成的图像来训练生成器,从而能够生成更多样化和高质量的图像。此外,我们引入了一种判别器之间的一致性机制,用于评估从卷积神经网络(CNN)和视觉变形(ViT)特征网络中提取的特征映射。鉴别器一致性促进鉴别器之间的连贯学习,并通过调整鉴别器对图像质量的评估来增强训练的鲁棒性。我们对17个数据集进行了广泛的评估,包括大数据、小数据和有限数据的场景,并涵盖了各种图像域,表明HP-GAN在fr起始距离(FID)方面始终优于当前最先进的方法,在图像多样性和质量方面取得了显着改善。代码可从https://github.com/higun2/HP-GAN获得。
{"title":"HP-GAN: Harnessing pretrained networks for GAN improvement with FakeTwins and discriminator consistency","authors":"Geonhui Son ,&nbsp;Jeong Ryong Lee ,&nbsp;Dosik Hwang","doi":"10.1016/j.neunet.2026.108666","DOIUrl":"10.1016/j.neunet.2026.108666","url":null,"abstract":"<div><div>Generative Adversarial Networks (GANs) have made significant progress in enhancing the quality of image synthesis. Recent methods frequently leverage pretrained networks to calculate perceptual losses or utilize pretrained feature spaces. In this paper, we extend the capabilities of pretrained networks by incorporating innovative self-supervised learning techniques and enforcing consistency between discriminators during GAN training. Our proposed method, named HP-GAN, effectively exploits neural network priors through two primary strategies: FakeTwins and discriminator consistency. FakeTwins leverages pretrained networks as encoders to compute a self-supervised loss and applies this through the generated images to train the generator, thereby enabling the generation of more diverse and high quality images. Additionally, we introduce a consistency mechanism between discriminators that evaluate feature maps extracted from Convolutional Neural Network (CNN) and Vision Transformer (ViT) feature networks. Discriminator consistency promotes coherent learning among discriminators and enhances training robustness by aligning their assessments of image quality. Our extensive evaluation across seventeen datasets-including scenarios with large, small, and limited data, and covering a variety of image domains-demonstrates that HP-GAN consistently outperforms current state-of-the-art methods in terms of Fréchet Inception Distance (FID), achieving significant improvements in image diversity and quality. Code is available at: <span><span>https://github.com/higun2/HP-GAN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108666"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146114793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing adversarial transferability via curvature-aware penalization 通过曲率感知惩罚增强对抗可转移性。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-01-29 DOI: 10.1016/j.neunet.2026.108665
Cheng Peng , Zeze Tao , Junyu Liu , Jinjia Peng
Transfer-based attack generates adversarial examples on a surrogate model and exploits the intriguing property of transferability to deceive other unknown models, making it practical for real-world scenarios. Recent research has sought to optimize the loss surface by minimizing its maximum loss, which in practice cannot be computed exactly and is instead approximated through gradient ascent. However, the loss landscape becomes increasingly non-linear during later attack stages, making the gradient ascent less effective. To address this challenge, we propose a novel attack called Curvature-Aware Penalization (CAP), which incorporates the gradient norm and the curvature-aware term as regularization terms to maintain the flatness of the loss surface. Since directly computing the Hessian matrix is computationally expensive, we utilize the finite difference method to reduce computational complexity. Specifically, we randomly sample an example from the neighborhood and interpolate gradients at three neighboring points along the example’s gradient direction to approximate the Hessian. Additionally, to reduce the variance caused by random sampling, the combined gradients are averaged over multiple stochastic samples. Comprehensive experimental results demonstrate that our CAP can not only craft adversarial examples with enhanced transferability across various network architectures but also exhibit stronger resistance to state-of-the-art adversarial defense methods. Code is available at https://github.com/PC614/CAP.
基于转移的攻击在代理模型上生成对抗性示例,并利用可转移性的有趣属性来欺骗其他未知模型,使其在现实场景中实用。最近的研究试图通过最小化其最大损失来优化损失面,而在实践中无法精确计算,而是通过梯度上升来近似计算。然而,在后期的攻击阶段,损失情况变得越来越非线性,使得梯度上升变得不那么有效。为了解决这一挑战,我们提出了一种新的攻击方法,称为曲率感知惩罚(CAP),它将梯度范数和曲率感知项作为正则化项来保持损失表面的平坦性。由于直接计算Hessian矩阵计算量大,我们采用有限差分法来降低计算复杂度。具体来说,我们从邻域中随机抽取一个样本,并沿着样本的梯度方向在三个相邻点上插值梯度来近似Hessian。此外,为了减少随机抽样引起的方差,将组合梯度在多个随机样本上平均。综合实验结果表明,我们的CAP不仅可以在各种网络架构之间制作具有增强可转移性的对抗性示例,而且还可以对最先进的对抗性防御方法表现出更强的抵抗力。代码可从https://github.com/PC614/CAP获得。
{"title":"Enhancing adversarial transferability via curvature-aware penalization","authors":"Cheng Peng ,&nbsp;Zeze Tao ,&nbsp;Junyu Liu ,&nbsp;Jinjia Peng","doi":"10.1016/j.neunet.2026.108665","DOIUrl":"10.1016/j.neunet.2026.108665","url":null,"abstract":"<div><div>Transfer-based attack generates adversarial examples on a surrogate model and exploits the intriguing property of transferability to deceive other unknown models, making it practical for real-world scenarios. Recent research has sought to optimize the loss surface by minimizing its maximum loss, which in practice cannot be computed exactly and is instead approximated through gradient ascent. However, the loss landscape becomes increasingly non-linear during later attack stages, making the gradient ascent less effective. To address this challenge, we propose a novel attack called Curvature-Aware Penalization (CAP), which incorporates the gradient norm and the curvature-aware term as regularization terms to maintain the flatness of the loss surface. Since directly computing the Hessian matrix is computationally expensive, we utilize the finite difference method to reduce computational complexity. Specifically, we randomly sample an example from the neighborhood and interpolate gradients at three neighboring points along the example’s gradient direction to approximate the Hessian. Additionally, to reduce the variance caused by random sampling, the combined gradients are averaged over multiple stochastic samples. Comprehensive experimental results demonstrate that our CAP can not only craft adversarial examples with enhanced transferability across various network architectures but also exhibit stronger resistance to state-of-the-art adversarial defense methods. Code is available at <span><span>https://github.com/PC614/CAP</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108665"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146127075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DiffMCG: A diffusion model with mask-conditioned guiding module for medical image classification DiffMCG:一种带口罩条件引导模块的医学图像分类扩散模型。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-02-05 DOI: 10.1016/j.neunet.2026.108690
Chen Guan , Haihong Ai , Weiwei Wang , Ravi P. Singh , Shiya Song
Diffusion models have application potential in medical image classification tasks due to their effectiveness in eliminating unexpected noise and perturbations from medical images. However, existing diffusion models for medical image classification utilize image features as the condition guiding diffusion model denoising, neglecting the most critical structured semantic information within medical images—namely, the mask of the lesion region. This results in suboptimal denoising performance from diffusion models, consequently impairing classification performance. To address this issue, we propose a diffusion model with the mask-conditioned guiding module called DiffMCG. Specifically, we introduce the Mask-Conditioned Guiding (MCG) module that concurrently extracts features from the medical image and its corresponding mask. Secondly, we design a U-Net denoising network based on the multi-layer perceptron (MLP) that is tailored for low-dimensional vector data and performs denoising tasks within the category label space. Furthermore, we introduce the MMD regularization constraint loss to establish a distributional relationship between the image prediction distribution, mask prediction distribution, and ground-truth label distribution within the label prediction space. This ensures the consistency of multimodal information during the diffusion process. Through analysis of comparative and ablation experiments, we validate the advantages of the MCG module in medical image classification, providing technical support for precision medical diagnostics.
扩散模型能够有效地消除医学图像中的意外噪声和扰动,在医学图像分类任务中具有很大的应用潜力。然而,现有的医学图像分类扩散模型利用图像特征作为指导扩散模型去噪的条件,忽略了医学图像中最关键的结构化语义信息,即病灶区域的掩模。这导致扩散模型的去噪性能不够理想,从而损害了分类性能。为了解决这个问题,我们提出了一个带有掩模条件引导模块的扩散模型,称为DiffMCG。具体来说,我们引入了mask - conditioned guidance (MCG)模块,该模块可以同时从医学图像及其对应的mask中提取特征。其次,我们设计了一个基于多层感知器(MLP)的U-Net去噪网络,该网络为低维向量数据量身定制,并在类别标签空间内执行去噪任务。此外,我们引入了MMD正则化约束损失,建立了图像预测分布、掩模预测分布和真地标签分布在标签预测空间内的分布关系。这保证了扩散过程中多模态信息的一致性。通过对比和消融实验分析,验证了MCG模块在医学图像分类方面的优势,为精准医学诊断提供技术支持。
{"title":"DiffMCG: A diffusion model with mask-conditioned guiding module for medical image classification","authors":"Chen Guan ,&nbsp;Haihong Ai ,&nbsp;Weiwei Wang ,&nbsp;Ravi P. Singh ,&nbsp;Shiya Song","doi":"10.1016/j.neunet.2026.108690","DOIUrl":"10.1016/j.neunet.2026.108690","url":null,"abstract":"<div><div>Diffusion models have application potential in medical image classification tasks due to their effectiveness in eliminating unexpected noise and perturbations from medical images. However, existing diffusion models for medical image classification utilize image features as the condition guiding diffusion model denoising, neglecting the most critical structured semantic information within medical images—namely, the mask of the lesion region. This results in suboptimal denoising performance from diffusion models, consequently impairing classification performance. To address this issue, we propose a diffusion model with the mask-conditioned guiding module called DiffMCG. Specifically, we introduce the Mask-Conditioned Guiding (MCG) module that concurrently extracts features from the medical image and its corresponding mask. Secondly, we design a U-Net denoising network based on the multi-layer perceptron (MLP) that is tailored for low-dimensional vector data and performs denoising tasks within the category label space. Furthermore, we introduce the MMD regularization constraint loss to establish a distributional relationship between the image prediction distribution, mask prediction distribution, and ground-truth label distribution within the label prediction space. This ensures the consistency of multimodal information during the diffusion process. Through analysis of comparative and ablation experiments, we validate the advantages of the MCG module in medical image classification, providing technical support for precision medical diagnostics.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108690"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146167526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Matrix-Transformation based Low-Rank Adaptation (MTLoRA): A brain-Inspired method for parameter-Efficient fine-Tuning 基于矩阵变换的低秩自适应(MTLoRA):一种基于大脑的参数高效微调方法
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-01-25 DOI: 10.1016/j.neunet.2026.108642
Yao Liang , Yuwei Wang , Yang Li , Yi Zeng
Parameter-efficient fine-tuning (PEFT) reduces the compute and memory demands of adapting large language models, yet standard low-rank adapters (e.g., LoRA) can lag full fine-tuning in performance and stability because they restrict updates to a fixed rank-r subspace. We propose Matrix-Transformation based Low-Rank Adaptation (MTLoRA), a brain-inspired extension that inserts a learnable r × r transformation T into the low-rank update (ΔW=BTA). By endowing the subspace with data-adapted geometry (e.g., rotations, scalings, and shears), MTLoRA reparameterizes the rank-r hypothesis class, improving its conditioning and inductive bias at negligible O(r2) overhead, and recovers LoRA when T=Ir. We instantiate four structures for T—SHIM (T=C), ICFM (T=CC), CTCM (T=CD), and DTSM (T=C+D)—providing complementary inductive biases (change of basis, PSD metric, staged mixing, dual superposition). An optimization analysis shows that T acts as a learned preconditioner within the subspace, yielding spectral-norm step-size bounds and operator-norm variance contraction that stabilize training. Empirically, MTLoRA delivers consistent gains while preserving PEFT efficiency: on GLUE (General Language Understanding Evaluation) with DeBERTaV3-base, MTLoRA improves the average over LoRA by (+2.0) points (86.9 → 88.9) and matches AdaLoRA (88.9) without any pruning schedule; on natural language generation with GPT-2 Medium, it raises BLEU on DART by (+0.95) and on WebNLG by (+0.56); and in multimodal instruction tuning with LLaVA-1.5-7B, DTSM attains the best average (69.91) with  ∼ 4.7% trainable parameters, outperforming full fine-tuning and strong PEFT baselines. These results indicate that learning geometry inside the low-rank subspace improves both effectiveness and stability, making MTLoRA a practical, plug-compatible alternative to LoRA for large-model fine-tuning.
参数有效的微调(PEFT)减少了适应大型语言模型的计算和内存需求,但是标准的低级别适配器(例如,LoRA)可能会在性能和稳定性方面滞后于完全的微调,因为它们将更新限制在固定的rank-r子空间。我们提出了基于矩阵变换的低秩自适应(MTLoRA),这是一种基于大脑的扩展,它将可学习r × r变换T插入到低秩更新(ΔW=BTA)中。通过赋予子空间数据适应几何(例如,旋转、缩放和剪切),MTLoRA重新参数化了rank-r假设类,在可忽略的O(r2)开销下改善了其条件和归纳偏差,并在T=Ir时恢复了LoRA。我们实例化了T - shim (T=C)、ICFM (T=CC)、CTCM (T=CD)和DTSM (T=C+D)的四种结构,提供了互补的归纳偏置(基的变化、PSD度量、阶段混合、对偶叠加)。优化分析表明,T作为子空间内的学习预条件,产生谱范数步长边界和算子范数方差收缩,稳定训练。从经验上看,MTLoRA在保持PEFT效率的同时提供了一致的收益:在带有debertav3碱基的GLUE(通用语言理解评估)上,MTLoRA在没有任何修剪计划的情况下,将LoRA的平均值提高了(+2.0)分(86.9 → 88.9),与AdaLoRA(88.9)相匹配;在使用GPT-2 Medium生成自然语言时,DART和WebNLG的BLEU分别提高了+0.95和+0.56;在使用llva -1.5- 7b进行多模态指令调谐时,DTSM在 ~ 4.7%可训练参数下达到最佳平均值(69.91),优于完全微调和强PEFT基线。这些结果表明,在低秩子空间内学习几何结构提高了有效性和稳定性,使MTLoRA成为一种实用的、可插入兼容的大模型微调替代LoRA。
{"title":"Matrix-Transformation based Low-Rank Adaptation (MTLoRA): A brain-Inspired method for parameter-Efficient fine-Tuning","authors":"Yao Liang ,&nbsp;Yuwei Wang ,&nbsp;Yang Li ,&nbsp;Yi Zeng","doi":"10.1016/j.neunet.2026.108642","DOIUrl":"10.1016/j.neunet.2026.108642","url":null,"abstract":"<div><div>Parameter-efficient fine-tuning (PEFT) reduces the compute and memory demands of adapting large language models, yet standard low-rank adapters (e.g., LoRA) can lag full fine-tuning in performance and stability because they restrict updates to a fixed rank-<em>r</em> subspace. We propose Matrix-Transformation based Low-Rank Adaptation (MTLoRA), a brain-inspired extension that inserts a learnable <em>r</em> × <em>r</em> transformation <em>T</em> into the low-rank update (<span><math><mrow><mstyle><mi>Δ</mi></mstyle><mi>W</mi><mo>=</mo><mi>B</mi><mi>T</mi><mi>A</mi></mrow></math></span>). By endowing the subspace with data-adapted geometry (e.g., rotations, scalings, and shears), MTLoRA reparameterizes the rank-<em>r</em> hypothesis class, improving its conditioning and inductive bias at negligible <em>O</em>(<em>r</em><sup>2</sup>) overhead, and recovers LoRA when <span><math><mrow><mi>T</mi><mo>=</mo><msub><mi>I</mi><mi>r</mi></msub></mrow></math></span>. We instantiate four structures for <em>T</em>—SHIM <span><math><mrow><mo>(</mo><mi>T</mi><mo>=</mo><mi>C</mi><mo>)</mo></mrow></math></span>, ICFM <span><math><mrow><mo>(</mo><mi>T</mi><mo>=</mo><mi>C</mi><msup><mi>C</mi><mi>⊤</mi></msup><mo>)</mo></mrow></math></span>, CTCM <span><math><mrow><mo>(</mo><mi>T</mi><mo>=</mo><mi>C</mi><mi>D</mi><mo>)</mo></mrow></math></span>, and DTSM <span><math><mrow><mo>(</mo><mi>T</mi><mo>=</mo><mi>C</mi><mo>+</mo><mi>D</mi><mo>)</mo></mrow></math></span>—providing complementary inductive biases (change of basis, PSD metric, staged mixing, dual superposition). An optimization analysis shows that <em>T</em> acts as a learned preconditioner within the subspace, yielding spectral-norm step-size bounds and operator-norm variance contraction that stabilize training. Empirically, MTLoRA delivers consistent gains while preserving PEFT efficiency: on GLUE (General Language Understanding Evaluation) with DeBERTaV3-base, MTLoRA improves the average over LoRA by (+2.0) points (86.9 → 88.9) and matches AdaLoRA (88.9) without any pruning schedule; on natural language generation with GPT-2 Medium, it raises BLEU on DART by (+0.95) and on WebNLG by (+0.56); and in multimodal instruction tuning with LLaVA-1.5-7B, DTSM attains the best average (69.91) with  ∼ 4.7% trainable parameters, outperforming full fine-tuning and strong PEFT baselines. These results indicate that learning geometry inside the low-rank subspace improves both effectiveness and stability, making MTLoRA a practical, plug-compatible alternative to LoRA for large-model fine-tuning.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108642"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-timescale representation with adaptive routing for deep tabular learning under temporal shift 基于自适应路径的深度表学习多时间尺度表示。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-01-30 DOI: 10.1016/j.neunet.2026.108670
Tianyu Wang , Maite Zhang , Mingxuan Lu , Mian Li
In real-world applications, tabular datasets often evolve over time, leading to temporal shift that degrades the long-range neural network performance. Most existing temporal encoding or adaptation solutions treat time cues as fixed auxiliary variables at a single scale. Motivated by the multi-horizon nature of temporal shifts with heterogeneous temporal dynamics, this paper presents TARS (Temporal Abstraction with Routed Scales), a novel plug-and-play method for robust tabular learning under temporal shift, applicable to various deep learning model backbones. First, an explicit temporal encoder decomposes timestamps into short-term recency, mid-term periodicity, and long-term contextual embeddings with structured memory. Next, an implicit drift encoder tracks higher-order distributional statistics at the same aligned timescales, producing drift signals that reflect ongoing temporal dynamics. These signals drive a drift-aware routing mechanism that adaptively weights the explicit temporal pathways, emphasizing the most relevant timescales under current conditions. Finally, a feature-temporal fusion layer integrates the routed temporal representation with original features, injecting context-aware bias. Extensive experiments on eight real-world datasets from the TabReD benchmark show that TARS consistently outperforms the competitive compared methods across various backbone models, achieving up to +2.38% average relative improvement on MLP, +4.08% on DCNv2, etc. Ablation studies verify the complementary contributions of all four modules. These results highlight the effectiveness of TARS for improving the temporal robustness of existing deep tabular models.
在现实世界的应用中,表格数据集经常随着时间的推移而变化,导致时间的变化,从而降低了远程神经网络的性能。大多数现有的时间编码或自适应解决方案将时间线索视为单一尺度上的固定辅助变量。摘要针对具有异构时间动态的时间转移的多视界特性,提出了一种适用于各种深度学习模型主干的时间转移鲁棒表格学习的即插即用方法TARS (temporal Abstraction with routing Scales)。首先,显式时间编码器将时间戳分解为具有结构化记忆的短期近期性、中期周期性和长期上下文嵌入。接下来,隐式漂移编码器在相同的对齐时间尺度上跟踪高阶分布统计数据,产生反映持续时间动态的漂移信号。这些信号驱动漂移感知路由机制,该机制自适应地加权显式时间路径,强调当前条件下最相关的时间尺度。最后,特征时间融合层将路由的时间表示与原始特征集成在一起,注入上下文感知偏差。在TabReD基准测试的8个真实数据集上进行的大量实验表明,TARS在各种骨干模型中始终优于竞争性比较方法,在MLP上实现了+2.38%的平均相对改进,在DCNv2等上实现了+4.08%的平均相对改进。消融研究证实了所有四个模块的互补贡献。这些结果突出了TARS在提高现有深度表格模型的时间鲁棒性方面的有效性。
{"title":"Multi-timescale representation with adaptive routing for deep tabular learning under temporal shift","authors":"Tianyu Wang ,&nbsp;Maite Zhang ,&nbsp;Mingxuan Lu ,&nbsp;Mian Li","doi":"10.1016/j.neunet.2026.108670","DOIUrl":"10.1016/j.neunet.2026.108670","url":null,"abstract":"<div><div>In real-world applications, tabular datasets often evolve over time, leading to temporal shift that degrades the long-range neural network performance. Most existing temporal encoding or adaptation solutions treat time cues as fixed auxiliary variables at a single scale. Motivated by the multi-horizon nature of temporal shifts with heterogeneous temporal dynamics, this paper presents TARS (Temporal Abstraction with Routed Scales), a novel plug-and-play method for robust tabular learning under temporal shift, applicable to various deep learning model backbones. First, an explicit temporal encoder decomposes timestamps into short-term recency, mid-term periodicity, and long-term contextual embeddings with structured memory. Next, an implicit drift encoder tracks higher-order distributional statistics at the same aligned timescales, producing drift signals that reflect ongoing temporal dynamics. These signals drive a drift-aware routing mechanism that adaptively weights the explicit temporal pathways, emphasizing the most relevant timescales under current conditions. Finally, a feature-temporal fusion layer integrates the routed temporal representation with original features, injecting context-aware bias. Extensive experiments on eight real-world datasets from the TabReD benchmark show that TARS consistently outperforms the competitive compared methods across various backbone models, achieving up to +2.38% average relative improvement on MLP, +4.08% on DCNv2, etc. Ablation studies verify the complementary contributions of all four modules. These results highlight the effectiveness of TARS for improving the temporal robustness of existing deep tabular models.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108670"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146120930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A two-level neurodynamic approach for heterogeneous networked game under event-triggered quantized mechanism 事件触发量化机制下异构网络博弈的两级神经动力学方法。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-02-04 DOI: 10.1016/j.neunet.2026.108671
Yiyao Xu , Mengxin Wang , Ruoyu Yuan , Sitian Qin
Game theory provides important theoretical and methodological support for the application expansion of deep learning in complex interaction scenarios, multi-player systems and other aspects. Networked game as a key problem has been investigated in this paper. Since it is inevitable for neighbor players to communicate with each other, key challenges in practical applications is to reduce the communication cost and improve the convergence rate. Also, as the fast development of cyber physical system, players need to obey identical intrinsic dynamics. Hence, in this paper, the communication channel is equipped with one-to-one gradient-based event trigger and logarithmic quantizer, which effectively alleviate the communication burden and reduce communication frequency. Moreover, the passivity-based strategies is used to compensate the lack of complete information, while a piecewise time-varying function is introduced to ensure prescribed time convergence. Besides, proper control input is designed for heterogeneous dynamics players to track Nash equilibrium(NE). It is proven by Lyapunov method that the two-level neurodynamic owns convergence within adjustable time. Additionally, Zeno behavior is excluded. Finally, a numerical example of connectivity control problem for autonomous mobile robots is provided to demonstrate the effectiveness of the proposed neurodynamic approach.
博弈论为深度学习在复杂交互场景、多人系统等方面的应用拓展提供了重要的理论和方法支持。本文对网络博弈这一关键问题进行了研究。由于相邻玩家之间的相互通信是不可避免的,因此降低通信成本和提高收敛速度是实际应用中的关键挑战。同时,随着网络物理系统的快速发展,玩家需要遵循相同的内在动态。因此,本文在通信信道中配置了基于一对一梯度的事件触发器和对数量化器,有效减轻了通信负担,降低了通信频率。在此基础上,采用基于无源的策略来补偿完全信息的不足,并引入分段时变函数来保证给定的时间收敛性。此外,设计了合适的控制输入,使异质动力学参与者能够跟踪纳什均衡。通过Lyapunov方法证明了两级神经动力学在可调时间内具有收敛性。此外,芝诺行为被排除在外。最后,给出了自主移动机器人连通性控制问题的数值示例,以验证所提出的神经动力学方法的有效性。
{"title":"A two-level neurodynamic approach for heterogeneous networked game under event-triggered quantized mechanism","authors":"Yiyao Xu ,&nbsp;Mengxin Wang ,&nbsp;Ruoyu Yuan ,&nbsp;Sitian Qin","doi":"10.1016/j.neunet.2026.108671","DOIUrl":"10.1016/j.neunet.2026.108671","url":null,"abstract":"<div><div>Game theory provides important theoretical and methodological support for the application expansion of deep learning in complex interaction scenarios, multi-player systems and other aspects. Networked game as a key problem has been investigated in this paper. Since it is inevitable for neighbor players to communicate with each other, key challenges in practical applications is to reduce the communication cost and improve the convergence rate. Also, as the fast development of cyber physical system, players need to obey identical intrinsic dynamics. Hence, in this paper, the communication channel is equipped with one-to-one gradient-based event trigger and logarithmic quantizer, which effectively alleviate the communication burden and reduce communication frequency. Moreover, the passivity-based strategies is used to compensate the lack of complete information, while a piecewise time-varying function is introduced to ensure prescribed time convergence. Besides, proper control input is designed for heterogeneous dynamics players to track Nash equilibrium(NE). It is proven by Lyapunov method that the two-level neurodynamic owns convergence within adjustable time. Additionally, Zeno behavior is excluded. Finally, a numerical example of connectivity control problem for autonomous mobile robots is provided to demonstrate the effectiveness of the proposed neurodynamic approach.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108671"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146158725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiple interpretation ensemble distillation for graph neural networks 图神经网络的多重解释集合蒸馏。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-02-03 DOI: 10.1016/j.neunet.2026.108674
Kang Liu , Yuqi Zhang , Shunzhi Yang , Chang-Dong Wang , Yunwen Chen , Xiaowen Ma , Zhenhua Huang
Existing graph knowledge distillation methods suffer from limited absorption of the teacher’s “dark knowledge” because they rely on simple logit alignment, which often causes overfitting or incomplete capture of underlying patterns. Additionally, relying on a single perspective severely restricts the student’s learning effectiveness and generalization ability. To address these issues, we develop a novel Multiple Interpretation Ensemble Distillation (MIED) method. It constructs a multi-interpreter composed of multiple single-layer MLPs for the student, termed the Student Interpretation (SI) component, to interpret knowledge from diversified outputs, thus avoiding representational bias from a single student output. Based on this, it introduces two effective strategies, i.e., Hybrid Sampling and Hierarchical Update. The former employs different sampling strategies for the outputs of the teacher and student (including the SI component). Specifically, the teacher’s output adopts a percentage random sampler, while the outputs of the student and SI component both leverage a positive-negative sampler. With this design, MIED can facilitate better coordination of sample selection and the learning process among the teacher, student, and SI component. The latter updates the parameters of the last layer in the student using the exponential moving average of the fused parameters of the SI component, while the parameters of other layers are updated via a regular optimizer. This enhances the robustness and generalization performance of MIED. Extensive experiments on seven real-world public datasets demonstrate that MIED outperforms existing methods in node classification tasks, resulting in an average improvement of 5.56% over GCN and 27.43% over MLP, respectively. Moreover, compared with directly using multiple students (where the number is consistent with the number of layers in the SI component), MIED achieves improvements approximately 6.00% in time, 50.00% in space, and 0.20% in accuracy. These results indicate that MIED is scalable and generalizable, and exhibits robustness on complex samples.
现有的图知识蒸馏方法对教师“暗知识”的吸收有限,因为它们依赖于简单的logit对齐,这往往会导致对底层模式的过拟合或不完全捕获。此外,依赖单一视角严重制约了学生的学习效果和泛化能力。为了解决这些问题,我们开发了一种新的多重解释集合蒸馏(MIED)方法。它为学生构建了一个由多个单层mlp组成的多解释器,称为学生解释(SI)组件,以解释来自多样化输出的知识,从而避免来自单个学生输出的代表性偏差。在此基础上,介绍了混合采样和分层更新两种有效的策略。前者对教师和学生的输出(包括SI成分)采用不同的抽样策略。具体来说,教师的输出采用百分比随机采样器,而学生和科学探究成分的输出都采用正负采样器。通过这种设计,MIED可以更好地协调教师、学生和科学探究成分之间的样本选择和学习过程。后者使用SI组件的融合参数的指数移动平均来更新学生中最后一层的参数,而其他层的参数则通过常规优化器更新。这提高了MIED的鲁棒性和泛化性能。在7个真实公共数据集上的大量实验表明,MIED在节点分类任务上优于现有方法,比GCN平均提高5.56%,比MLP平均提高27.43%。此外,与直接使用多个学生(人数与SI组件的层数一致)相比,MIED在时间上提高了约6.00%,在空间上提高了50.00%,在精度上提高了0.20%。这些结果表明,MIED具有可扩展性和可泛化性,并且对复杂样本具有鲁棒性。
{"title":"Multiple interpretation ensemble distillation for graph neural networks","authors":"Kang Liu ,&nbsp;Yuqi Zhang ,&nbsp;Shunzhi Yang ,&nbsp;Chang-Dong Wang ,&nbsp;Yunwen Chen ,&nbsp;Xiaowen Ma ,&nbsp;Zhenhua Huang","doi":"10.1016/j.neunet.2026.108674","DOIUrl":"10.1016/j.neunet.2026.108674","url":null,"abstract":"<div><div>Existing graph knowledge distillation methods suffer from limited absorption of the teacher’s “dark knowledge” because they rely on simple logit alignment, which often causes overfitting or incomplete capture of underlying patterns. Additionally, relying on a single perspective severely restricts the student’s learning effectiveness and generalization ability. To address these issues, we develop a novel Multiple Interpretation Ensemble Distillation (MIED) method. It constructs a multi-interpreter composed of multiple single-layer MLPs for the student, termed the Student Interpretation (SI) component, to interpret knowledge from diversified outputs, thus avoiding representational bias from a single student output. Based on this, it introduces two effective strategies, i.e., Hybrid Sampling and Hierarchical Update. The former employs different sampling strategies for the outputs of the teacher and student (including the SI component). Specifically, the teacher’s output adopts a percentage random sampler, while the outputs of the student and SI component both leverage a positive-negative sampler. With this design, MIED can facilitate better coordination of sample selection and the learning process among the teacher, student, and SI component. The latter updates the parameters of the last layer in the student using the exponential moving average of the fused parameters of the SI component, while the parameters of other layers are updated via a regular optimizer. This enhances the robustness and generalization performance of MIED. Extensive experiments on seven real-world public datasets demonstrate that MIED outperforms existing methods in node classification tasks, resulting in an average improvement of 5.56% over GCN and 27.43% over MLP, respectively. Moreover, compared with directly using multiple students (where the number is consistent with the number of layers in the SI component), MIED achieves improvements approximately 6.00% in time, 50.00% in space, and 0.20% in accuracy. These results indicate that MIED is scalable and generalizable, and exhibits robustness on complex samples.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108674"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146167504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An interactive axial feature selection network for medical image classification 用于医学图像分类的交互式轴向特征选择网络。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-07-01 Epub Date: 2026-01-29 DOI: 10.1016/j.neunet.2026.108661
Shuai Pang, Chunhua Hu, Juan Zhao, Haifang Yu
To address the differences and correlations between features, as well as to fully utilize the importance of salient semantics in medical image classification tasks, this paper proposes an Interactive Axial Feature Selection Network (IAFSNet), aimed at improving feature representation, effectively filtering noise during classification, thereby enhancing classification performance. The paper introduces a newly designed Feature Interaction Module (FIM), which learns spatial differences between various features and enhances the interdependence and complementarity between local spatial features and global contextual semantics. Additionally, the paper implements a novel Axial Feature Selection Module (AFSM), which filters salient feature semantics from three perspectives: horizontal, vertical, and spatial. By adjusting thresholds, salient features are emphasized while irrelevant noise is eliminated, allowing these key features to cross-aggregate layer by layer and establish interactions among them, ultimately improving classification accuracy. Experimental results on four benchmark datasets demonstrate that the proposed IAFSNet exhibits excellent classification performance and robustness, significantly outperforming many existing classification methods.
为了解决特征之间的差异和相关性,并充分利用显著语义在医学图像分类任务中的重要性,本文提出了一种交互式轴向特征选择网络(IAFSNet),旨在改进特征表示,有效过滤分类过程中的噪声,从而提高分类性能。本文介绍了一种新的特征交互模块(FIM),该模块可以学习各种特征之间的空间差异,增强局部空间特征与全局上下文语义之间的相互依存和互补。此外,本文还实现了一种新的轴向特征选择模块(AFSM),该模块从水平、垂直和空间三个角度过滤显著特征语义。通过调整阈值,突出突出的特征,消除不相关的噪声,使这些关键特征逐层交叉聚集,建立相互作用,最终提高分类精度。在4个基准数据集上的实验结果表明,所提出的IAFSNet具有优异的分类性能和鲁棒性,显著优于现有的许多分类方法。
{"title":"An interactive axial feature selection network for medical image classification","authors":"Shuai Pang,&nbsp;Chunhua Hu,&nbsp;Juan Zhao,&nbsp;Haifang Yu","doi":"10.1016/j.neunet.2026.108661","DOIUrl":"10.1016/j.neunet.2026.108661","url":null,"abstract":"<div><div>To address the differences and correlations between features, as well as to fully utilize the importance of salient semantics in medical image classification tasks, this paper proposes an Interactive Axial Feature Selection Network (IAFSNet), aimed at improving feature representation, effectively filtering noise during classification, thereby enhancing classification performance. The paper introduces a newly designed Feature Interaction Module (FIM), which learns spatial differences between various features and enhances the interdependence and complementarity between local spatial features and global contextual semantics. Additionally, the paper implements a novel Axial Feature Selection Module (AFSM), which filters salient feature semantics from three perspectives: horizontal, vertical, and spatial. By adjusting thresholds, salient features are emphasized while irrelevant noise is eliminated, allowing these key features to cross-aggregate layer by layer and establish interactions among them, ultimately improving classification accuracy. Experimental results on four benchmark datasets demonstrate that the proposed IAFSNet exhibits excellent classification performance and robustness, significantly outperforming many existing classification methods.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108661"},"PeriodicalIF":6.3,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146114790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Networks
全部 Geobiology Appl. Clay Sci. Geochim. Cosmochim. Acta J. Hydrol. Org. Geochem. Carbon Balance Manage. Contrib. Mineral. Petrol. Int. J. Biometeorol. IZV-PHYS SOLID EART+ J. Atmos. Chem. Acta Oceanolog. Sin. Acta Geophys. ACTA GEOL POL ACTA PETROL SIN ACTA GEOL SIN-ENGL AAPG Bull. Acta Geochimica Adv. Atmos. Sci. Adv. Meteorol. Am. J. Phys. Anthropol. Am. J. Sci. Am. Mineral. Annu. Rev. Earth Planet. Sci. Appl. Geochem. Aquat. Geochem. Ann. Glaciol. Archaeol. Anthropol. Sci. ARCHAEOMETRY ARCT ANTARCT ALP RES Asia-Pac. J. Atmos. Sci. ATMOSPHERE-BASEL Atmos. Res. Aust. J. Earth Sci. Atmos. Chem. Phys. Atmos. Meas. Tech. Basin Res. Big Earth Data BIOGEOSCIENCES Geostand. Geoanal. Res. GEOLOGY Geosci. J. Geochem. J. Geochem. Trans. Geosci. Front. Geol. Ore Deposits Global Biogeochem. Cycles Gondwana Res. Geochem. Int. Geol. J. Geophys. Prospect. Geosci. Model Dev. GEOL BELG GROUNDWATER Hydrogeol. J. Hydrol. Earth Syst. Sci. Hydrol. Processes Int. J. Climatol. Int. J. Earth Sci. Int. Geol. Rev. Int. J. Disaster Risk Reduct. Int. J. Geomech. Int. J. Geog. Inf. Sci. Isl. Arc J. Afr. Earth. Sci. J. Adv. Model. Earth Syst. J APPL METEOROL CLIM J. Atmos. Oceanic Technol. J. Atmos. Sol. Terr. Phys. J. Clim. J. Earth Sci. J. Earth Syst. Sci. J. Environ. Eng. Geophys. J. Geog. Sci. Mineral. Mag. Miner. Deposita Mon. Weather Rev. Nat. Hazards Earth Syst. Sci. Nat. Clim. Change Nat. Geosci. Ocean Dyn. Ocean and Coastal Research npj Clim. Atmos. Sci. Ocean Modell. Ocean Sci. Ore Geol. Rev. OCEAN SCI J Paleontol. J. PALAEOGEOGR PALAEOCL PERIOD MINERAL PETROLOGY+ Phys. Chem. Miner. Polar Sci. Prog. Oceanogr. Quat. Sci. Rev. Q. J. Eng. Geol. Hydrogeol. RADIOCARBON Pure Appl. Geophys. Resour. Geol. Rev. Geophys. Sediment. Geol.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1