Diffusion models have achieved remarkable success in image generation, image super-resolution, and text-to-image synthesis. Despite their effectiveness, they face key challenges, notably long inference time and complex architectures that incur high computational costs. While various methods have been proposed to reduce inference steps and accelerate computation, the optimization of diffusion model architectures has received comparatively limited attention. To address this gap, we propose LDMOES (Lightweight Diffusion Models based on Multi-Objective Evolutionary Search), a framework that combines multi-objective evolutionary neural architecture search with knowledge distillation to design efficient UNet-based diffusion models. By adopting a modular search space, LDMOES effectively decouples architecture components for improved search efficiency. We validated our method on multiple datasets, including CIFAR-10, Tiny-ImageNet, CelebA-HQ [Formula: see text], and LSUN-church [Formula: see text]. Experiments show that LDMOES reduces multiply-accumulate operations (MACs) by approximately 40% in pixel space while outperforming the teacher model. When transferred to the larger-scale Tiny-ImageNet dataset, it still generates high-quality images with a competitive FID score of 4.16, demonstrating strong generalization ability. In latent space, MACs are reduced by about 50% with negligible performance loss. After transferring to the more complex LSUN-church dataset, the model surpasses baselines in generation quality while reducing computational cost by nearly 60%, validating the effectiveness and transferability of the multi-objective search strategy. Code and models will be available at https://github.com/GenerativeMind-arch/LDMOES.
{"title":"Lightweight Diffusion Models Based on Multi-Objective Evolutionary Neural Architecture Search.","authors":"Yu Xue, Chunxiao Jiao, Yong Zhang, Ali Wagdy Mohamed, Romany Fouad Mansour, Ferrante Neri","doi":"10.1142/S0129065725500595","DOIUrl":"10.1142/S0129065725500595","url":null,"abstract":"<p><p>Diffusion models have achieved remarkable success in image generation, image super-resolution, and text-to-image synthesis. Despite their effectiveness, they face key challenges, notably long inference time and complex architectures that incur high computational costs. While various methods have been proposed to reduce inference steps and accelerate computation, the optimization of diffusion model architectures has received comparatively limited attention. To address this gap, we propose LDMOES (<b>L</b>ightweight <b>D</b>iffusion Models based on <b>M</b>ulti-<b>O</b>bjective <b>E</b>volutionary <b>S</b>earch), a framework that combines multi-objective evolutionary neural architecture search with knowledge distillation to design efficient UNet-based diffusion models. By adopting a modular search space, LDMOES effectively decouples architecture components for improved search efficiency. We validated our method on multiple datasets, including CIFAR-10, Tiny-ImageNet, CelebA-HQ [Formula: see text], and LSUN-church [Formula: see text]. Experiments show that LDMOES reduces multiply-accumulate operations (MACs) by approximately 40% in pixel space while outperforming the teacher model. When transferred to the larger-scale Tiny-ImageNet dataset, it still generates high-quality images with a competitive FID score of 4.16, demonstrating strong generalization ability. In latent space, MACs are reduced by about 50% with negligible performance loss. After transferring to the more complex LSUN-church dataset, the model surpasses baselines in generation quality while reducing computational cost by nearly 60%, validating the effectiveness and transferability of the multi-objective search strategy. Code and models will be available at https://github.com/GenerativeMind-arch/LDMOES.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550059"},"PeriodicalIF":6.4,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144983965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-31DOI: 10.1142/S0129065726500036
Rui Guo, Beni Widarman Yus Kelana, Eman Safar Almetere, Jian Lian, Long Yang
With the increase in work-related stress, the issue of psychological pressure in occupational environments has gained increasing attention. This paper proposes an enhanced Informer stress recognition and classification method based on deep learning, which guarantees performance by integrating tailored spatial and channel attention mechanisms (SAM/CAM) with the Informer backbone. Unlike existing attention-augmented models, the proposed SAM is designed to prioritize time-sensitive physiological signal segments, while CAM dynamically weights complementary stress-related features, enabling precise capture of subtle stress-related patterns. With this dual attention mechanism, the proposed model can capture subtle changes associated with stress states accurately. To evaluate the performance of the proposed method, the experiments on one publicly available dataset were conducted. Experimental results demonstrate that the proposed method has outperformed existing approaches in terms of accuracy, recall, and F1-score for stress recognition. Additionally, we performed ablation studies to verify the contributions of spatial attention module and channel attention module to the proposed model. In conclusion, this study not only provides an effective technical means for the automatic detection of psychological stress, but also lays a foundation for the application of deep learning model in a broader range of health monitoring applications.
{"title":"Enhanced Informer Network for Stress Recognition and Classification via Spatial and Channel Attention Mechanisms.","authors":"Rui Guo, Beni Widarman Yus Kelana, Eman Safar Almetere, Jian Lian, Long Yang","doi":"10.1142/S0129065726500036","DOIUrl":"https://doi.org/10.1142/S0129065726500036","url":null,"abstract":"<p><p>With the increase in work-related stress, the issue of psychological pressure in occupational environments has gained increasing attention. This paper proposes an enhanced Informer stress recognition and classification method based on deep learning, which guarantees performance by integrating tailored spatial and channel attention mechanisms (SAM/CAM) with the Informer backbone. Unlike existing attention-augmented models, the proposed SAM is designed to prioritize time-sensitive physiological signal segments, while CAM dynamically weights complementary stress-related features, enabling precise capture of subtle stress-related patterns. With this dual attention mechanism, the proposed model can capture subtle changes associated with stress states accurately. To evaluate the performance of the proposed method, the experiments on one publicly available dataset were conducted. Experimental results demonstrate that the proposed method has outperformed existing approaches in terms of accuracy, recall, and F1-score for stress recognition. Additionally, we performed ablation studies to verify the contributions of spatial attention module and channel attention module to the proposed model. In conclusion, this study not only provides an effective technical means for the automatic detection of psychological stress, but also lays a foundation for the application of deep learning model in a broader range of health monitoring applications.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2650003"},"PeriodicalIF":6.4,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145859510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-31DOI: 10.1142/S012906572650005X
Chongfeng Wang, Brendan Z Allison, Xiao Wu, Junxian Li, Ruiyu Zhao, Weijie Chen, Xingyu Wang, Andrzej Cichocki, Jing Jin
In motor imagery (MI)-based brain-computer interfaces (BCIs), convolutional neural networks (CNNs) are widely employed to decode electroencephalogram (EEG) signals. However, due to their fixed kernel sizes and uniform attention to features, CNNs struggle to fully capture the time-frequency features of EEG signals. To address this limitation, this paper proposes the Multi-Domain Dynamic Weighted Network (MD-DWNet), which integrates multimodal complementary feature information across time, frequency, and spatial domains through a branch structure to enhance decoding performance. Specifically, MD-DWNet combines multi-band filtering, spatial convolution, and temporal variance calculation to extract spatial-spectral features, while a dual-scale CNN captures local spatiotemporal features at different time scales. A dynamic global filter is designed to optimize fused features, improving the adaptive modeling capability for dynamic changes in frequency band energy. A lightweight mixed attention mechanism selectively enhances salient channel and spatial features. The dual-branch joint loss function adaptively balances contributions through a task uncertainty mechanism, thereby enhancing optimization efficiency and generalization capability. Experimental results on the BCI Competition IV 2a, IV 2b, OpenBMI, and a self-collected laboratory dataset demonstrate that MD-DWNet achieves classification accuracies of 83.86%, 88.67%, 75.25% and 84.85%, respectively, outperforming several advanced methods and validating its superior performance in MI signal decoding.
{"title":"Multi-Domain Dynamic Weighting Network for Motor Imagery Decoding.","authors":"Chongfeng Wang, Brendan Z Allison, Xiao Wu, Junxian Li, Ruiyu Zhao, Weijie Chen, Xingyu Wang, Andrzej Cichocki, Jing Jin","doi":"10.1142/S012906572650005X","DOIUrl":"https://doi.org/10.1142/S012906572650005X","url":null,"abstract":"<p><p>In motor imagery (MI)-based brain-computer interfaces (BCIs), convolutional neural networks (CNNs) are widely employed to decode electroencephalogram (EEG) signals. However, due to their fixed kernel sizes and uniform attention to features, CNNs struggle to fully capture the time-frequency features of EEG signals. To address this limitation, this paper proposes the Multi-Domain Dynamic Weighted Network (MD-DWNet), which integrates multimodal complementary feature information across time, frequency, and spatial domains through a branch structure to enhance decoding performance. Specifically, MD-DWNet combines multi-band filtering, spatial convolution, and temporal variance calculation to extract spatial-spectral features, while a dual-scale CNN captures local spatiotemporal features at different time scales. A dynamic global filter is designed to optimize fused features, improving the adaptive modeling capability for dynamic changes in frequency band energy. A lightweight mixed attention mechanism selectively enhances salient channel and spatial features. The dual-branch joint loss function adaptively balances contributions through a task uncertainty mechanism, thereby enhancing optimization efficiency and generalization capability. Experimental results on the BCI Competition IV 2a, IV 2b, OpenBMI, and a self-collected laboratory dataset demonstrate that MD-DWNet achieves classification accuracies of 83.86%, 88.67%, 75.25% and 84.85%, respectively, outperforming several advanced methods and validating its superior performance in MI signal decoding.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2650005"},"PeriodicalIF":6.4,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145859454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-12-06DOI: 10.1142/S0129065725500662
Gema Benedicto-Rodríguez, Andrea Hongn, Carlos G Juan, Javier Garrigós-Guerrero, María Paula Bonomini, Eduardo Fernandez-Jover, Jose Manuel Ferrández-Vicente
In a world where social interaction presents challenges for children with Autism Spectrum Disorder (ASD), robots are stepping in as allies in emotional learning. This study examined how affective interactions with a humanoid robot elicited physiological responses in children with ASD, using electrodermal activity (EDA) and heart rate variability (HRV) as key indicators of emotional arousal. The objectives were to identify emotionally salient moments during human-robot interaction, assess whether certain individual characteristics - such as age or ASD severity - modulate autonomic responses, and evaluate the usefulness of wearable devices for real-time monitoring. Thirteen children participated in structured sessions involving a range of social, cognitive, and motor tasks alongside the robot Pepper. The results showed that the hugging phase (HS2) often generated greater autonomic reactivity in children, especially among younger children and those with higher levels of restlessness or a higher level of ASD. Children with level 2 ASD displayed higher sympathetic activation compared to level 1 participants, who showed more HRV stability. Age also played a role, as younger children demonstrated lower autonomic regulation. These findings highlight the relevance of physiological monitoring in detecting emotional dysregulation and tailoring robot-assisted therapy. Future developments will explore adaptive systems capable of adjusting interventions in real time to better support each child's unique needs.
{"title":"Physiological Response in Children with Autism Spectrum Disorder (ASD) During Social Robot Interaction.","authors":"Gema Benedicto-Rodríguez, Andrea Hongn, Carlos G Juan, Javier Garrigós-Guerrero, María Paula Bonomini, Eduardo Fernandez-Jover, Jose Manuel Ferrández-Vicente","doi":"10.1142/S0129065725500662","DOIUrl":"10.1142/S0129065725500662","url":null,"abstract":"<p><p>In a world where social interaction presents challenges for children with Autism Spectrum Disorder (ASD), robots are stepping in as allies in emotional learning. This study examined how affective interactions with a humanoid robot elicited physiological responses in children with ASD, using electrodermal activity (EDA) and heart rate variability (HRV) as key indicators of emotional arousal. The objectives were to identify emotionally salient moments during human-robot interaction, assess whether certain individual characteristics - such as age or ASD severity - modulate autonomic responses, and evaluate the usefulness of wearable devices for real-time monitoring. Thirteen children participated in structured sessions involving a range of social, cognitive, and motor tasks alongside the robot Pepper. The results showed that the hugging phase (HS2) often generated greater autonomic reactivity in children, especially among younger children and those with higher levels of restlessness or a higher level of ASD. Children with level 2 ASD displayed higher sympathetic activation compared to level 1 participants, who showed more HRV stability. Age also played a role, as younger children demonstrated lower autonomic regulation. These findings highlight the relevance of physiological monitoring in detecting emotional dysregulation and tailoring robot-assisted therapy. Future developments will explore adaptive systems capable of adjusting interventions in real time to better support each child's unique needs.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550066"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145688894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-10-18DOI: 10.1142/S012906572550073X
Muhammad Suffian, Cosimo Ieracitano, Francesco C Morabito, Nadia Mammone
Decoding electroencephalographic (EEG) signals is of key importance in the development of brain-computer interface (BCI) systems. However, high inter-subject variability in EEG signals requires user-specific calibration, which can be time-consuming and limit the application of deep learning approaches, due to general need of large amount of data to properly train these models. In this context, this paper proposes a multidimensional and explainable deep learning framework for fast and interpretable EEG decoding. In particular, EEG signals are projected into the spatial-spectral-temporal domain and processed using a custom three-dimensional (3D) Convolutional Neural Network, here referred to as EEGCubeNet. In this work, the method has been validated on EEGs recorded during motor BCI experiments. Namely, hand open (HO) and hand close (HC) movement planning was investigated by discriminating them from the absence of movement preparation (resting state, RE). The proposed method is based on a global- to subject-specific fine-tuning. The model is globally trained on a population of subjects and then fine-tuned on the final user, significantly reducing adaptation time. Experimental results demonstrate that EEGCubeNet achieves state-of-the-art performance (accuracy of [Formula: see text] and [Formula: see text] for HC versus RE and HO versus RE, binary classification tasks, respectively) with reduced framework complexity and with a reduced training time. In addition, to enhance transparency, a 3D occlusion sensitivity analysis-based explainability method (here named 3D xAI-OSA) that generates relevance maps revealing the most significant features to each prediction, was introduced. The data and source code are available at the following link: https://github.com/AI-Lab-UniRC/EEGCubeNet.
脑电图信号的解码是脑机接口(BCI)系统发展的关键。然而,脑电图信号的高度主体间可变性需要用户特定的校准,这可能是耗时的,并且限制了深度学习方法的应用,因为通常需要大量的数据来正确训练这些模型。在此背景下,本文提出了一个多维、可解释的深度学习框架,用于快速、可解释的脑电图解码。特别是,EEG信号被投射到空间-频谱-时间域中,并使用定制的三维(3D)卷积神经网络(这里称为EEGCubeNet)进行处理。在这项工作中,该方法已在运动脑机接口实验中记录的脑电图上得到验证。即,手张开(HO)和手闭合(HC)的运动计划通过区分它们与缺乏运动准备(RE)进行研究。所提出的方法是基于全局到特定主题的微调。该模型在一组对象上进行全局训练,然后在最终用户上进行微调,大大减少了适应时间。实验结果表明,EEGCubeNet在降低框架复杂度和减少训练时间的情况下,达到了最先进的性能(分别为HC与RE和HO与RE的二元分类任务[Formula: see text]和[Formula: see text]的准确率)。此外,为了提高透明度,引入了一种基于3D遮挡敏感性分析的可解释性方法(这里称为3D xAI-OSA),该方法生成了揭示每个预测最重要特征的相关性图。数据和源代码可从以下链接获得:https://github.com/AI-Lab-UniRC/EEGCubeNet。
{"title":"An Explainable 3D-Deep Learning Model for EEG Decoding in Brain-Computer Interface Applications.","authors":"Muhammad Suffian, Cosimo Ieracitano, Francesco C Morabito, Nadia Mammone","doi":"10.1142/S012906572550073X","DOIUrl":"10.1142/S012906572550073X","url":null,"abstract":"<p><p>Decoding electroencephalographic (EEG) signals is of key importance in the development of brain-computer interface (BCI) systems. However, high inter-subject variability in EEG signals requires user-specific calibration, which can be time-consuming and limit the application of deep learning approaches, due to general need of large amount of data to properly train these models. In this context, this paper proposes a multidimensional and explainable deep learning framework for fast and interpretable EEG decoding. In particular, EEG signals are projected into the spatial-spectral-temporal domain and processed using a custom three-dimensional (3D) Convolutional Neural Network, here referred to as <i>EEGCubeNet</i>. In this work, the method has been validated on EEGs recorded during motor BCI experiments. Namely, hand open (HO) and hand close (HC) movement planning was investigated by discriminating them from the absence of movement preparation (resting state, RE). The proposed method is based on a global- to subject-specific fine-tuning. The model is globally trained on a population of subjects and then fine-tuned on the final user, significantly reducing adaptation time. Experimental results demonstrate that <i>EEGCubeNet</i> achieves state-of-the-art performance (accuracy of [Formula: see text] and [Formula: see text] for HC versus RE and HO versus RE, binary classification tasks, respectively) with reduced framework complexity and with a reduced training time. In addition, to enhance transparency, a 3D occlusion sensitivity analysis-based explainability method (here named <i>3D xAI-OSA</i>) that generates relevance maps revealing the most significant features to each prediction, was introduced. The data and source code are available at the following link: https://github.com/AI-Lab-UniRC/EEGCubeNet.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550073"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145318871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-12-09DOI: 10.1142/S0129065725500820
Thomas Pirenne, Mansoureh Fahimi Hnazaee, Patrick Santens, Aline Moorkens, Marc M Van Hulle
Deficits in auditory perception have been widely observed in Parkinson's disease (PD) patients and the literature attributes it, in part, to impaired central auditory processing. Deep brain stimulation (DBS) of the subthalamic nucleus (STN) is a well-established therapeutic option for patients with advanced PD. Analysis of auditory evoked potentials suggested a modulatory effect of DBS on central auditory processing. To better understand the latter, we investigated whether DBS modulates auditory steady state responses (ASSR) in electroencephalography (EEG) recordings of 5 PD patients. ASSRs are neural responses along central auditory pathways phase-locked to an auditory stimulus which can serve to understand the spectral aspects of central auditory processing. In our analyses, we estimate the intensity of ASSRs with a novel method based on canonical correlation analysis (CCA) and compare them in DBS ON and OFF conditions. Our results suggest that DBS effectively reduces ASSR in patients with PD. A comparison to age-matched healthy participants suggests a pathological effect of PD on ASSRs, which is disrupted by DBS. These findings support our hypothesis that DBS suppresses central auditory processing. Further research is required to assess the symptomatic effect of this modulation, as well as which cortical and subcortical generators are most affected. A better understanding of the auditory side-effects of DBS could lead to improved treatment options.
{"title":"Subthalamic Nucleus Deep Brain Stimulation Modulates Auditory Steady State Responses in Parkinson's Disease.","authors":"Thomas Pirenne, Mansoureh Fahimi Hnazaee, Patrick Santens, Aline Moorkens, Marc M Van Hulle","doi":"10.1142/S0129065725500820","DOIUrl":"https://doi.org/10.1142/S0129065725500820","url":null,"abstract":"<p><p>Deficits in auditory perception have been widely observed in Parkinson's disease (PD) patients and the literature attributes it, in part, to impaired central auditory processing. Deep brain stimulation (DBS) of the subthalamic nucleus (STN) is a well-established therapeutic option for patients with advanced PD. Analysis of auditory evoked potentials suggested a modulatory effect of DBS on central auditory processing. To better understand the latter, we investigated whether DBS modulates auditory steady state responses (ASSR) in electroencephalography (EEG) recordings of 5 PD patients. ASSRs are neural responses along central auditory pathways phase-locked to an auditory stimulus which can serve to understand the spectral aspects of central auditory processing. In our analyses, we estimate the intensity of ASSRs with a novel method based on canonical correlation analysis (CCA) and compare them in DBS ON and OFF conditions. Our results suggest that DBS effectively reduces ASSR in patients with PD. A comparison to age-matched healthy participants suggests a pathological effect of PD on ASSRs, which is disrupted by DBS. These findings support our hypothesis that DBS suppresses central auditory processing. Further research is required to assess the symptomatic effect of this modulation, as well as which cortical and subcortical generators are most affected. A better understanding of the auditory side-effects of DBS could lead to improved treatment options.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":"35 13","pages":"2550082"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145764781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-09-18DOI: 10.1142/S0129065725500650
Ferrante Neri, Mengchen Yang, Yu Xue
In the context of neural system structure modeling and complex visual tasks, the effective integration of multi-scale features and contextual information is critical for enhancing model performance. This paper proposes a biologically inspired hybrid neural network architecture - CompEyeNet - which combines the global modeling capacity of transformers with the efficiency of lightweight convolutional structures. The backbone network, multi-attention transformer backbone network (MATBN), integrates multiple attention mechanisms to collaboratively model local details and long-range dependencies. The neck network, compound eye neck network (CENN), introduces high-resolution feature layers and efficient attention fusion modules to significantly enhance multi-scale information representation and reconstruction capability. CompEyeNet is evaluated on three authoritative medical image segmentation datasets: MICCAI-CVC-ClinicDB, ISIC2018, and MICCAI-tooth-segmentation, demonstrating its superior performance. Experimental results show that compared to models such as Deeplab, Unet, and the YOLO series, CompEyeNet achieves better performance with fewer parameters. Specifically, compared to the baseline model YOLOv11, CompEyeNet reduces the number of parameters by an average of 38.31%. On key performance metrics, the average Dice coefficient improves by 0.87%, the Jaccard index by 1.53%, Precision by 0.58%, and Recall by 1.11%. These findings verify the advantages of the proposed architecture in terms of parameter efficiency and accuracy, highlighting the broad application potential of bio-inspired attention-fusion hybrid neural networks in neural system modeling and image analysis.
{"title":"A Compound-Eye-Inspired Multi-Scale Neural Architecture with Integrated Attention Mechanisms.","authors":"Ferrante Neri, Mengchen Yang, Yu Xue","doi":"10.1142/S0129065725500650","DOIUrl":"10.1142/S0129065725500650","url":null,"abstract":"<p><p>In the context of neural system structure modeling and complex visual tasks, the effective integration of multi-scale features and contextual information is critical for enhancing model performance. This paper proposes a biologically inspired hybrid neural network architecture - CompEyeNet - which combines the global modeling capacity of transformers with the efficiency of lightweight convolutional structures. The backbone network, multi-attention transformer backbone network (MATBN), integrates multiple attention mechanisms to collaboratively model local details and long-range dependencies. The neck network, compound eye neck network (CENN), introduces high-resolution feature layers and efficient attention fusion modules to significantly enhance multi-scale information representation and reconstruction capability. CompEyeNet is evaluated on three authoritative medical image segmentation datasets: MICCAI-CVC-ClinicDB, ISIC2018, and MICCAI-tooth-segmentation, demonstrating its superior performance. Experimental results show that compared to models such as Deeplab, Unet, and the YOLO series, CompEyeNet achieves better performance with fewer parameters. Specifically, compared to the baseline model YOLOv11, CompEyeNet reduces the number of parameters by an average of 38.31%. On key performance metrics, the average Dice coefficient improves by 0.87%, the Jaccard index by 1.53%, Precision by 0.58%, and Recall by 1.11%. These findings verify the advantages of the proposed architecture in terms of parameter efficiency and accuracy, highlighting the broad application potential of bio-inspired attention-fusion hybrid neural networks in neural system modeling and image analysis.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550065"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145115668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-08-22DOI: 10.1142/S0129065725500571
José A Rodríguez-Rodríguez, Miguel A Molina-Cabello, Rafaela Benítez-Rochel, Ezequiel López-Rubio
The real-time synthesis of 3D views, facilitated by convolutional neural networks like NeX, is increasingly pivotal in various computer vision applications. These networks are trained using photographs taken from different perspectives during the training phase. However, these images may be susceptible to contamination from noise originating from the vision sensor or the surrounding environment. This research meticulously examines the impact of noise on the resulting image quality of 3D views synthesized by the NeX network. Various noise levels and scenes have been incorporated to substantiate the claim that the presence of noise significantly degrades image quality. Additionally, a new strategy is introduced to improve image quality by calculating consensus among NeX networks trained on images pre-processed with a denoising algorithm. Experimental results confirm the effectiveness of this technique, demonstrating improvements of up to 1.300 dB and 0.032 for Peak Signal Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM), respectively, under certain scenes and noise levels. Notably, the performance gains are especially significant when using synthesized images generated by NeX from noisy inputs in the consensus process.
{"title":"Consensus-Based 3D View Generation from Noisy Images.","authors":"José A Rodríguez-Rodríguez, Miguel A Molina-Cabello, Rafaela Benítez-Rochel, Ezequiel López-Rubio","doi":"10.1142/S0129065725500571","DOIUrl":"10.1142/S0129065725500571","url":null,"abstract":"<p><p>The real-time synthesis of 3D views, facilitated by convolutional neural networks like NeX, is increasingly pivotal in various computer vision applications. These networks are trained using photographs taken from different perspectives during the training phase. However, these images may be susceptible to contamination from noise originating from the vision sensor or the surrounding environment. This research meticulously examines the impact of noise on the resulting image quality of 3D views synthesized by the NeX network. Various noise levels and scenes have been incorporated to substantiate the claim that the presence of noise significantly degrades image quality. Additionally, a new strategy is introduced to improve image quality by calculating consensus among NeX networks trained on images pre-processed with a denoising algorithm. Experimental results confirm the effectiveness of this technique, demonstrating improvements of up to 1.300 dB and 0.032 for Peak Signal Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM), respectively, under certain scenes and noise levels. Notably, the performance gains are especially significant when using synthesized images generated by NeX from noisy inputs in the consensus process.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550057"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144984009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-02-03DOI: 10.1142/S0129065725020010
Zvi Kam, Giovanna Nicora
{"title":"Editorial - A Journal that Promotes Excellence Through Uncompromising Review Process: Reflection of Freedom of Speech and Scientific Publication.","authors":"Zvi Kam, Giovanna Nicora","doi":"10.1142/S0129065725020010","DOIUrl":"10.1142/S0129065725020010","url":null,"abstract":"","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2502001"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143191643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-08-27DOI: 10.1142/S0129065725500534
Gergő Bognár, Manuel Feindert, Christian Huber, Michael Lunglmayr, Mario Huemer, Péter Kovács
In this paper, we present a hybrid learning framework that integrates two model-driven AI paradigms: Deep unfolding and Variable Projections (VPs). The core idea is to unfold the iterations of VP solvers for separable nonlinear least squares (SNLLS) problems into trainable neural network layers. As a consequence, the network is capable of learning optimal nonlinear VP parameters during inference, which is a form of model-based meta-learning. Furthermore, the architecture incorporates prior knowledge of the underlying SNLLS problem, such as basis function expansions and signal structure, which enhance interpretability, reduce model size, and lower data requirements. As a case study, we adapt the proposed deep unfolded VPNet to learn ECG representations for the classification of five arrhythmias. Experimental results on the MIT-BIH Arrhythmia Database show that VPNet achieves performance comparable to state-of-the-art ECG classifiers, attaining 95% accuracy while maintaining a compact architecture. Its low computational complexity enables efficient training and inference, making it highly suitable for real-time, power-efficient edge computing applications. This is further validated through embedded implementation on STM32 microcontrollers.
{"title":"Deep Unfolded Variable Projection Networks.","authors":"Gergő Bognár, Manuel Feindert, Christian Huber, Michael Lunglmayr, Mario Huemer, Péter Kovács","doi":"10.1142/S0129065725500534","DOIUrl":"10.1142/S0129065725500534","url":null,"abstract":"<p><p>In this paper, we present a hybrid learning framework that integrates two model-driven AI paradigms: Deep unfolding and Variable Projections (VPs). The core idea is to unfold the iterations of VP solvers for separable nonlinear least squares (SNLLS) problems into trainable neural network layers. As a consequence, the network is capable of learning optimal nonlinear VP parameters during inference, which is a form of model-based meta-learning. Furthermore, the architecture incorporates prior knowledge of the underlying SNLLS problem, such as basis function expansions and signal structure, which enhance interpretability, reduce model size, and lower data requirements. As a case study, we adapt the proposed deep unfolded VPNet to learn ECG representations for the classification of five arrhythmias. Experimental results on the MIT-BIH Arrhythmia Database show that VPNet achieves performance comparable to state-of-the-art ECG classifiers, attaining 95% accuracy while maintaining a compact architecture. Its low computational complexity enables efficient training and inference, making it highly suitable for real-time, power-efficient edge computing applications. This is further validated through embedded implementation on STM32 microcontrollers.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550053"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144983999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}