This paper addresses the consensus in fractional-order networked systems with matrix-weighted coupling, where the interactions of agents are characterized by positive definite or positive semi-definite matrices. A distributed sample-based control strategy is designed, in which each agent updates its state using sampled data. Some necessary and sufficient consensus conditions are derived for both undirected and directed matrix-weighted networks, respectively. The conditions depend on the sampling period, the fractional order, the control gain strengths, as well as the eigenvalue properties of the matrix-weighted Laplacian. Notably, for undirected networks, consensus is closely related to the null space of the matrix-weighted Laplacian. For directed networks, the existence of a positive spanning tree is not necessary to reach matrix-weighted consensus. Finally, simulation examples are conducted to validate the effectiveness of the theoretical analysis.
{"title":"Matrix-weighted consensus of fractional-order networked systems via sampled-data control.","authors":"Yanyan Ye, Weiling Wang, Wenfeng Jin, Cheng Zhou, Yuanqing Wu, Zhixia Ding","doi":"10.1016/j.neunet.2026.108696","DOIUrl":"https://doi.org/10.1016/j.neunet.2026.108696","url":null,"abstract":"<p><p>This paper addresses the consensus in fractional-order networked systems with matrix-weighted coupling, where the interactions of agents are characterized by positive definite or positive semi-definite matrices. A distributed sample-based control strategy is designed, in which each agent updates its state using sampled data. Some necessary and sufficient consensus conditions are derived for both undirected and directed matrix-weighted networks, respectively. The conditions depend on the sampling period, the fractional order, the control gain strengths, as well as the eigenvalue properties of the matrix-weighted Laplacian. Notably, for undirected networks, consensus is closely related to the null space of the matrix-weighted Laplacian. For directed networks, the existence of a positive spanning tree is not necessary to reach matrix-weighted consensus. Finally, simulation examples are conducted to validate the effectiveness of the theoretical analysis.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"108696"},"PeriodicalIF":6.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146221754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-09DOI: 10.1016/j.neunet.2026.108701
Ke Wang, Yinghao Zhang, Hong-Yu Zhang, Lin Liu, Jiuyong Li, Zaiwen Feng, Debo Cheng
Pre-trained language models (PLMs) have achieved remarkable success across a wide range of natural language processing tasks, including text classification, machine translation, and question-answering systems, by leveraging vast amounts of unlabeled data to learn rich linguistic representations. However, existing models often reflect human-like biases and societal stereotypes, posing a significant challenge in their application. To address this issue, this paper proposes a novel debiasing framework called CFPLM. Unlike conventional debiasing methods, CFPLM is grounded in causal inference, aiming to identify and intervene in the factors that contribute to bias, thereby eliminating the bias in PLMs. The framework incorporates a composite loss function, which introduces a fairness penalty term to regulate the learning process of the model. Additionally, it integrates adversarial loss and entropy regularization to further optimize model performance. Experimental results demonstrate that, based on standard datasets and evaluation metrics, the proposed CFPLM method significantly reduces bias in BERT, RoBERTa, and ALBERT, while results on the GLUE benchmark indicate that enhancing model fairness does not compromise the models' language understanding capabilities.
{"title":"Learning fair representation for fine-tuning pre-trained language models.","authors":"Ke Wang, Yinghao Zhang, Hong-Yu Zhang, Lin Liu, Jiuyong Li, Zaiwen Feng, Debo Cheng","doi":"10.1016/j.neunet.2026.108701","DOIUrl":"https://doi.org/10.1016/j.neunet.2026.108701","url":null,"abstract":"<p><p>Pre-trained language models (PLMs) have achieved remarkable success across a wide range of natural language processing tasks, including text classification, machine translation, and question-answering systems, by leveraging vast amounts of unlabeled data to learn rich linguistic representations. However, existing models often reflect human-like biases and societal stereotypes, posing a significant challenge in their application. To address this issue, this paper proposes a novel debiasing framework called CFPLM. Unlike conventional debiasing methods, CFPLM is grounded in causal inference, aiming to identify and intervene in the factors that contribute to bias, thereby eliminating the bias in PLMs. The framework incorporates a composite loss function, which introduces a fairness penalty term to regulate the learning process of the model. Additionally, it integrates adversarial loss and entropy regularization to further optimize model performance. Experimental results demonstrate that, based on standard datasets and evaluation metrics, the proposed CFPLM method significantly reduces bias in BERT, RoBERTa, and ALBERT, while results on the GLUE benchmark indicate that enhancing model fairness does not compromise the models' language understanding capabilities.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"108701"},"PeriodicalIF":6.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146229387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Generative adversarial networks (GANs) have achieved remarkable success in image synthesis but faces major challenges, including mode collapse, training instability, and inefficient architecture search. Existing evolutionary GANs partially address these issues but lack semantic alignment with real data, effective weight reuse, and knowledge transfer between model generations. To overcome these limitations, we propose contrastive-guided evolutionary GANs (CGE-GAN)–a unified method introduces a novel hybrid Wasserstein-Contrastive loss function that drives generators to align semantically with real data while maintaining adversarial competitiveness. Besides, we incorporated dynamic adaptive weight sharing (DAWS) for efficient training and knowledge distillation-based crossover to preserve useful features across generations. The CGE-GAN is evaluated on CIFAR-10 and STL-10, and it achieves an Inception Score (IS) of 8.99 and 10.46, and fréchet inception distance (FID) of 9.74 and 21.86, respectively. Compared to strong baselines, CGE-GAN reduces FID by up to 1.74 points while maintaining high semantic diversity and convergence efficiency with only 0.36 GPU days. These results highlight the effectiveness of contrastive-driven evolution for generating stable and high-fidelity outputs.
{"title":"CGE-GAN: Contrastive-guided evolutionary generative adversarial networks with dynamic adaptive weight sharing","authors":"Kashif Iqbal , Xue Yu , Atifa Rafique , Muhammad Hamid , Khursheed Aurangzeb","doi":"10.1016/j.neunet.2026.108702","DOIUrl":"10.1016/j.neunet.2026.108702","url":null,"abstract":"<div><div>Generative adversarial networks (GANs) have achieved remarkable success in image synthesis but faces major challenges, including mode collapse, training instability, and inefficient architecture search. Existing evolutionary GANs partially address these issues but lack semantic alignment with real data, effective weight reuse, and knowledge transfer between model generations. To overcome these limitations, we propose contrastive-guided evolutionary GANs (CGE-GAN)–a unified method introduces a novel hybrid Wasserstein-Contrastive loss function that drives generators to align semantically with real data while maintaining adversarial competitiveness. Besides, we incorporated dynamic adaptive weight sharing (DAWS) for efficient training and knowledge distillation-based crossover to preserve useful features across generations. The CGE-GAN is evaluated on CIFAR-10 and STL-10, and it achieves an Inception Score (IS) of 8.99 and 10.46, and fréchet inception distance (FID) of 9.74 and 21.86, respectively. Compared to strong baselines, CGE-GAN reduces FID by up to 1.74 points while maintaining high semantic diversity and convergence efficiency with only 0.36 GPU days. These results highlight the effectiveness of contrastive-driven evolution for generating stable and high-fidelity outputs.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108702"},"PeriodicalIF":6.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Alzheimer’s disease (AD) is a currently incurable neurodegenerative disease, with early detection representing a high research priority. AD is characterized by progressive cognitive decline accompanied by alterations in brain functional connectivity. Based on its data structure similar to the graph, graph neural networks (GNNs) have emerged as important methods for brain function analysis and disease prediction in recent years. However, most GNN methods are limited by information loss caused by traditional functional connectivity calculation as well as common noise issues in functional magnetic resonance imaging (fMRI) data. This paper proposes a graph generation based AD classification model using resting state fMRI to address this issue. The connectome generation network with large kernels for GNN (CGLK-GNN) based AD Analysis contains a graph generation block and a GNN prediction block. The graph generation block employs decoupled convolutional networks with large kernels to extract comprehensive temporal features while preserving sequential dependencies, contrasting with previous generative GNN approaches. This module constructs the connectome graph by encoding both edge-wise correlations and node-embedded temporal features, thereby utilizing the generated graph more effectively. The subsequent GNN prediction block adopts an efficient architecture to learn these enhanced representations and perform final AD stage classification. Through independent cohort validations, CGLK-GNN outperforms state-of-the-art GNN and rsfMRI-based AD classifiers in differentiating AD status. Furthermore, CGLK-GNN demonstrates high clinical value by learning clinically relevant connectome node and connectivity features from two independent datasets.
{"title":"CGLK-GNN : A connectome generation network with large kernels for GNN based Alzheimer’s disease analysis","authors":"Wenqi Zhu , Zhong Yin , Yinghua Fu , Alzheimer's Disease Neuroimaging Initiative","doi":"10.1016/j.neunet.2026.108689","DOIUrl":"10.1016/j.neunet.2026.108689","url":null,"abstract":"<div><div>Alzheimer’s disease (AD) is a currently incurable neurodegenerative disease, with early detection representing a high research priority. AD is characterized by progressive cognitive decline accompanied by alterations in brain functional connectivity. Based on its data structure similar to the graph, graph neural networks (GNNs) have emerged as important methods for brain function analysis and disease prediction in recent years. However, most GNN methods are limited by information loss caused by traditional functional connectivity calculation as well as common noise issues in functional magnetic resonance imaging (fMRI) data. This paper proposes a graph generation based AD classification model using resting state fMRI to address this issue. The connectome generation network with large kernels for GNN (CGLK-GNN) based AD Analysis contains a graph generation block and a GNN prediction block. The graph generation block employs decoupled convolutional networks with large kernels to extract comprehensive temporal features while preserving sequential dependencies, contrasting with previous generative GNN approaches. This module constructs the connectome graph by encoding both edge-wise correlations and node-embedded temporal features, thereby utilizing the generated graph more effectively. The subsequent GNN prediction block adopts an efficient architecture to learn these enhanced representations and perform final AD stage classification. Through independent cohort validations, CGLK-GNN outperforms state-of-the-art GNN and rsfMRI-based AD classifiers in differentiating AD status. Furthermore, CGLK-GNN demonstrates high clinical value by learning clinically relevant connectome node and connectivity features from two independent datasets.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108689"},"PeriodicalIF":6.3,"publicationDate":"2026-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.neunet.2026.108691
Lei Li , Quan Zhou , Shanshan Gao , Chaoran Cui , Zhaoqiang Xia
Kinship verification aims to determine whether two individuals share a familial relationship based on facial information. Cross-gender relationships (i.e., Father-Daughter and Mother-Son) continue to face formidable challenges due to the diversity and uncertainty of genetic inheritance. Existing studies primarily focus on extracting robust features and measuring similarity, with limited attention given to the fuzziness of gender differences. To address this issue, this paper proposes a kinship verification framework based on a fuzzy neural network, which adaptively extracts gender-independent kinship features and handles relationship fuzziness to improve cross-gender verification performance. Specifically, the Swin Transformer, which has demonstrated excellent performance in facial analysis, is employed to extract initial features. A fuzzy neural network is then designed to disentangle gender and kinship features, with a gender recognition task introduced to further enhance this disentanglement and improve the gender independence of kinship features. Subsequently, a multi-metric fuzzy reasoning module is adopted to integrate kinship features, extract latent kinship cues, and leverage a contrastive loss function to effectively mine potential negative sample information, thereby significantly enhancing the model’s robustness. Experimental results on three publicly available datasets demonstrate that the proposed method achieves state-of-the-art performance.
{"title":"Gender-independent kinship verification network via fuzzy disentangling and multi-metric inference","authors":"Lei Li , Quan Zhou , Shanshan Gao , Chaoran Cui , Zhaoqiang Xia","doi":"10.1016/j.neunet.2026.108691","DOIUrl":"10.1016/j.neunet.2026.108691","url":null,"abstract":"<div><div>Kinship verification aims to determine whether two individuals share a familial relationship based on facial information. Cross-gender relationships (i.e., Father-Daughter and Mother-Son) continue to face formidable challenges due to the diversity and uncertainty of genetic inheritance. Existing studies primarily focus on extracting robust features and measuring similarity, with limited attention given to the fuzziness of gender differences. To address this issue, this paper proposes a kinship verification framework based on a fuzzy neural network, which adaptively extracts gender-independent kinship features and handles relationship fuzziness to improve cross-gender verification performance. Specifically, the Swin Transformer, which has demonstrated excellent performance in facial analysis, is employed to extract initial features. A fuzzy neural network is then designed to disentangle gender and kinship features, with a gender recognition task introduced to further enhance this disentanglement and improve the gender independence of kinship features. Subsequently, a multi-metric fuzzy reasoning module is adopted to integrate kinship features, extract latent kinship cues, and leverage a contrastive loss function to effectively mine potential negative sample information, thereby significantly enhancing the model’s robustness. Experimental results on three publicly available datasets demonstrate that the proposed method achieves state-of-the-art performance.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108691"},"PeriodicalIF":6.3,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.neunet.2026.108686
Haoquan Lu, Zhihui Lai, Heng Kong
Breast tumor images show low intra-class similarity and suffer from distribution shift, posing challenges for recognition tasks. While increasing the number of labeled training data is a common strategy to improve performance, the high cost of expert annotation is another challenge. Semi-supervised learning methods, e.g., Graph Neural Networks (GNNs), which smooth features via graph topology, have the potential to reduce the annotation costs for breast tumor datasets while achieving satisfactory classification performance. To address these challenges, we propose Graph Adiabatic Diffusion Neural Networks (GradiNet), which jointly learn discriminative graph structures for discriminative representation and simulate distribution shift environments. Specifically, we model the discriminative graph structure through a graph-learning objective function and demonstrate its effectiveness theoretically and empirically. Furthermore, we design a GNN feature propagation mechanism for the first time by incorporating the Fourier heat diffusion equation with adiabatic boundary conditions. Hence, the mechanism allows the model to adaptively simulate distribution shifts and enhance its generalization ability on both in-distribution (ID) and out-of-distribution (OOD) data. Extensive experiments on public and private breast tumor ultrasound image datasets demonstrate the superiority and effectiveness of our approach, achieving state-of-the-art performance across multiple evaluation metrics.
{"title":"Graph adiabatic diffusion neural networks for distribution-shift breast tumor image classification.","authors":"Haoquan Lu, Zhihui Lai, Heng Kong","doi":"10.1016/j.neunet.2026.108686","DOIUrl":"https://doi.org/10.1016/j.neunet.2026.108686","url":null,"abstract":"<p><p>Breast tumor images show low intra-class similarity and suffer from distribution shift, posing challenges for recognition tasks. While increasing the number of labeled training data is a common strategy to improve performance, the high cost of expert annotation is another challenge. Semi-supervised learning methods, e.g., Graph Neural Networks (GNNs), which smooth features via graph topology, have the potential to reduce the annotation costs for breast tumor datasets while achieving satisfactory classification performance. To address these challenges, we propose Graph Adiabatic Diffusion Neural Networks (GradiNet), which jointly learn discriminative graph structures for discriminative representation and simulate distribution shift environments. Specifically, we model the discriminative graph structure through a graph-learning objective function and demonstrate its effectiveness theoretically and empirically. Furthermore, we design a GNN feature propagation mechanism for the first time by incorporating the Fourier heat diffusion equation with adiabatic boundary conditions. Hence, the mechanism allows the model to adaptively simulate distribution shifts and enhance its generalization ability on both in-distribution (ID) and out-of-distribution (OOD) data. Extensive experiments on public and private breast tumor ultrasound image datasets demonstrate the superiority and effectiveness of our approach, achieving state-of-the-art performance across multiple evaluation metrics.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"108686"},"PeriodicalIF":6.3,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146221748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-05DOI: 10.1016/j.neunet.2026.108690
Chen Guan , Haihong Ai , Weiwei Wang , Ravi P. Singh , Shiya Song
Diffusion models have application potential in medical image classification tasks due to their effectiveness in eliminating unexpected noise and perturbations from medical images. However, existing diffusion models for medical image classification utilize image features as the condition guiding diffusion model denoising, neglecting the most critical structured semantic information within medical images—namely, the mask of the lesion region. This results in suboptimal denoising performance from diffusion models, consequently impairing classification performance. To address this issue, we propose a diffusion model with the mask-conditioned guiding module called DiffMCG. Specifically, we introduce the Mask-Conditioned Guiding (MCG) module that concurrently extracts features from the medical image and its corresponding mask. Secondly, we design a U-Net denoising network based on the multi-layer perceptron (MLP) that is tailored for low-dimensional vector data and performs denoising tasks within the category label space. Furthermore, we introduce the MMD regularization constraint loss to establish a distributional relationship between the image prediction distribution, mask prediction distribution, and ground-truth label distribution within the label prediction space. This ensures the consistency of multimodal information during the diffusion process. Through analysis of comparative and ablation experiments, we validate the advantages of the MCG module in medical image classification, providing technical support for precision medical diagnostics.
{"title":"DiffMCG: A diffusion model with mask-conditioned guiding module for medical image classification","authors":"Chen Guan , Haihong Ai , Weiwei Wang , Ravi P. Singh , Shiya Song","doi":"10.1016/j.neunet.2026.108690","DOIUrl":"10.1016/j.neunet.2026.108690","url":null,"abstract":"<div><div>Diffusion models have application potential in medical image classification tasks due to their effectiveness in eliminating unexpected noise and perturbations from medical images. However, existing diffusion models for medical image classification utilize image features as the condition guiding diffusion model denoising, neglecting the most critical structured semantic information within medical images—namely, the mask of the lesion region. This results in suboptimal denoising performance from diffusion models, consequently impairing classification performance. To address this issue, we propose a diffusion model with the mask-conditioned guiding module called DiffMCG. Specifically, we introduce the Mask-Conditioned Guiding (MCG) module that concurrently extracts features from the medical image and its corresponding mask. Secondly, we design a U-Net denoising network based on the multi-layer perceptron (MLP) that is tailored for low-dimensional vector data and performs denoising tasks within the category label space. Furthermore, we introduce the MMD regularization constraint loss to establish a distributional relationship between the image prediction distribution, mask prediction distribution, and ground-truth label distribution within the label prediction space. This ensures the consistency of multimodal information during the diffusion process. Through analysis of comparative and ablation experiments, we validate the advantages of the MCG module in medical image classification, providing technical support for precision medical diagnostics.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108690"},"PeriodicalIF":6.3,"publicationDate":"2026-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146167526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-05DOI: 10.1016/j.neunet.2026.108693
Fandi Gou, Haikuo Du, Yunze Cai
Safety and Restricted Communication are two critical challenges faced by practical Multi-Agent Systems (MAS). However, most Multi-Agent Reinforcement Learning (MARL) algorithms that rely solely on reward shaping are ineffective in ensuring safety, and their applicability is rather limited due to the fully connected communication. To address these issues, we propose a novel framework, Graph-based Safe MARL (GS-MARL), to enhance the safety and scalability of MARL methods. Leveraging the inherent graph structure of MAS, we design a Graph Neural Network (GNN) based on message passing to aggregate local observations and communications of varying sizes. Furthermore, we develop a constrained joint policy optimization method in the setting of local observation to improve safety. Simulation experiments demonstrate that GS-MARL achieves a better trade-off between optimality and safety compared to other methods, and in large-scale communication-limited scenarios GS-MARL achieves a success rate at least 10% higher than the leading baselines. The feasibility of our method is also verified by hardware implementation with Mecanum-wheeled vehicles. Codes and demos are available at https://github.com/finleygou/GS-MARL.
{"title":"A graph-based safe reinforcement learning method for multi-agent cooperation.","authors":"Fandi Gou, Haikuo Du, Yunze Cai","doi":"10.1016/j.neunet.2026.108693","DOIUrl":"https://doi.org/10.1016/j.neunet.2026.108693","url":null,"abstract":"<p><p>Safety and Restricted Communication are two critical challenges faced by practical Multi-Agent Systems (MAS). However, most Multi-Agent Reinforcement Learning (MARL) algorithms that rely solely on reward shaping are ineffective in ensuring safety, and their applicability is rather limited due to the fully connected communication. To address these issues, we propose a novel framework, Graph-based Safe MARL (GS-MARL), to enhance the safety and scalability of MARL methods. Leveraging the inherent graph structure of MAS, we design a Graph Neural Network (GNN) based on message passing to aggregate local observations and communications of varying sizes. Furthermore, we develop a constrained joint policy optimization method in the setting of local observation to improve safety. Simulation experiments demonstrate that GS-MARL achieves a better trade-off between optimality and safety compared to other methods, and in large-scale communication-limited scenarios GS-MARL achieves a success rate at least 10% higher than the leading baselines. The feasibility of our method is also verified by hardware implementation with Mecanum-wheeled vehicles. Codes and demos are available at https://github.com/finleygou/GS-MARL.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"108693"},"PeriodicalIF":6.3,"publicationDate":"2026-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146259854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04DOI: 10.1016/j.neunet.2026.108683
Aitor Martinez-Seras , Javier Del Ser , Aitzol Olivares-Rad , Alain Andres , Pablo Garcia-Bringas
Robustness is a fundamental aspect for developing safe and trustworthy models, particularly when they are deployed in the open world. In this work we analyze the inherent capability of one-stage object detectors to robustly operate in the presence of out-of-distribution (OoD) data. Specifically, we propose a novel detection algorithm for detecting unknown objects in image data, which leverages the features extracted by the model from each sample. Differently from other recent approaches in the literature, our proposal does not require retraining the object detector, thereby allowing for the use of pretrained models. Our proposed OoD detector exploits the application of supervised dimensionality reduction techniques to mitigate the effects of the curse of dimensionality on the features extracted by the model. Furthermore, it utilizes high-resolution feature maps to identify potential unknown objects in an unsupervised fashion. Our experiments analyze the Pareto trade-off between the performance detecting known and unknown objects resulting from different algorithmic configurations and inference confidence thresholds. We also compare the performance of our proposed algorithm to that of logits-based post-hoc OoD methods, as well as possible fusion strategies. Finally, we discuss on the competitiveness of all tested methods against state-of-the-art OoD approaches for object detection models over the recently published Unknown Object Detection benchmark. The obtained results verify that the performance of avant-garde post-hoc OoD detectors can be further improved when combined with our proposed algorithm.
{"title":"On the inherent robustness of one-stage object detection against out-of-distribution data","authors":"Aitor Martinez-Seras , Javier Del Ser , Aitzol Olivares-Rad , Alain Andres , Pablo Garcia-Bringas","doi":"10.1016/j.neunet.2026.108683","DOIUrl":"10.1016/j.neunet.2026.108683","url":null,"abstract":"<div><div>Robustness is a fundamental aspect for developing safe and trustworthy models, particularly when they are deployed in the open world. In this work we analyze the inherent capability of one-stage object detectors to robustly operate in the presence of out-of-distribution (OoD) data. Specifically, we propose a novel detection algorithm for detecting unknown objects in image data, which leverages the features extracted by the model from each sample. Differently from other recent approaches in the literature, our proposal does not require retraining the object detector, thereby allowing for the use of pretrained models. Our proposed OoD detector exploits the application of supervised dimensionality reduction techniques to mitigate the effects of the curse of dimensionality on the features extracted by the model. Furthermore, it utilizes high-resolution feature maps to identify potential unknown objects in an unsupervised fashion. Our experiments analyze the Pareto trade-off between the performance detecting known and unknown objects resulting from different algorithmic configurations and inference confidence thresholds. We also compare the performance of our proposed algorithm to that of logits-based post-hoc OoD methods, as well as possible fusion strategies. Finally, we discuss on the competitiveness of all tested methods against state-of-the-art OoD approaches for object detection models over the recently published Unknown Object Detection benchmark. The obtained results verify that the performance of avant-garde post-hoc OoD detectors can be further improved when combined with our proposed algorithm.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108683"},"PeriodicalIF":6.3,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-04DOI: 10.1016/j.neunet.2026.108684
Zhijun Zhang, Xitong Gao, Jinjia Guo
To solve the repetitive motion problem of redundant robotic manipulators, a punishment neural network-based acceleration-level joint drift-free (PNN-ALJDF) scheme is designed. Traditional joint physical limits constraints are fixed and lack margin. Thus, a novel joint acceleration time-varying constraint is considered in the PNN-ALJDF scheme to avoid the joint state exceeding the physical limits. In addition, to ensure that redundant robotic manipulators can periodically return to the initial pose, a joint drift-free criterion is designed. Furthermore, the joint drift-free criterion, kinematics equation and joint acceleration time-varying constraint are formulated globally as an acceleration-level joint drift-free (ALJDF) scheme by a time-varying quadratic programming approach. Then, the ALJDF scheme is solved by the designed punishment neural network. Thus, the proposed PNN-ALJDF scheme is composed of the ALJDF scheme and punishment neural network. Finally, the simulations demonstrate that the PNN-ALJDF scheme avoids joints from drifting, and the states of joints are all within the acceleration time-varying constraint. In addition, the proposed PNN-ALJDF has higher solution accuracy than the linear variational inequalities-based primal-dual neural network.
{"title":"A punishment neural network-based acceleration-level joint drift-free scheme for solving constrained motion planning problem of redundant robotic manipulators","authors":"Zhijun Zhang, Xitong Gao, Jinjia Guo","doi":"10.1016/j.neunet.2026.108684","DOIUrl":"10.1016/j.neunet.2026.108684","url":null,"abstract":"<div><div>To solve the repetitive motion problem of redundant robotic manipulators, a punishment neural network-based acceleration-level joint drift-free (PNN-ALJDF) scheme is designed. Traditional joint physical limits constraints are fixed and lack margin. Thus, a novel joint acceleration time-varying constraint is considered in the PNN-ALJDF scheme to avoid the joint state exceeding the physical limits. In addition, to ensure that redundant robotic manipulators can periodically return to the initial pose, a joint drift-free criterion is designed. Furthermore, the joint drift-free criterion, kinematics equation and joint acceleration time-varying constraint are formulated globally as an acceleration-level joint drift-free (ALJDF) scheme by a time-varying quadratic programming approach. Then, the ALJDF scheme is solved by the designed punishment neural network. Thus, the proposed PNN-ALJDF scheme is composed of the ALJDF scheme and punishment neural network. Finally, the simulations demonstrate that the PNN-ALJDF scheme avoids joints from drifting, and the states of joints are all within the acceleration time-varying constraint. In addition, the proposed PNN-ALJDF has higher solution accuracy than the linear variational inequalities-based primal-dual neural network.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"199 ","pages":"Article 108684"},"PeriodicalIF":6.3,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146158685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}