Diffusion models have achieved remarkable success in image generation, image super-resolution, and text-to-image synthesis. Despite their effectiveness, they face key challenges, notably long inference time and complex architectures that incur high computational costs. While various methods have been proposed to reduce inference steps and accelerate computation, the optimization of diffusion model architectures has received comparatively limited attention. To address this gap, we propose LDMOES (Lightweight Diffusion Models based on Multi-Objective Evolutionary Search), a framework that combines multi-objective evolutionary neural architecture search with knowledge distillation to design efficient UNet-based diffusion models. By adopting a modular search space, LDMOES effectively decouples architecture components for improved search efficiency. We validated our method on multiple datasets, including CIFAR-10, Tiny-ImageNet, CelebA-HQ [Formula: see text], and LSUN-church [Formula: see text]. Experiments show that LDMOES reduces multiply-accumulate operations (MACs) by approximately 40% in pixel space while outperforming the teacher model. When transferred to the larger-scale Tiny-ImageNet dataset, it still generates high-quality images with a competitive FID score of 4.16, demonstrating strong generalization ability. In latent space, MACs are reduced by about 50% with negligible performance loss. After transferring to the more complex LSUN-church dataset, the model surpasses baselines in generation quality while reducing computational cost by nearly 60%, validating the effectiveness and transferability of the multi-objective search strategy. Code and models will be available at https://github.com/GenerativeMind-arch/LDMOES.
{"title":"Lightweight Diffusion Models Based on Multi-Objective Evolutionary Neural Architecture Search.","authors":"Yu Xue, Chunxiao Jiao, Yong Zhang, Ali Wagdy Mohamed, Romany Fouad Mansour, Ferrante Neri","doi":"10.1142/S0129065725500595","DOIUrl":"10.1142/S0129065725500595","url":null,"abstract":"<p><p>Diffusion models have achieved remarkable success in image generation, image super-resolution, and text-to-image synthesis. Despite their effectiveness, they face key challenges, notably long inference time and complex architectures that incur high computational costs. While various methods have been proposed to reduce inference steps and accelerate computation, the optimization of diffusion model architectures has received comparatively limited attention. To address this gap, we propose LDMOES (<b>L</b>ightweight <b>D</b>iffusion Models based on <b>M</b>ulti-<b>O</b>bjective <b>E</b>volutionary <b>S</b>earch), a framework that combines multi-objective evolutionary neural architecture search with knowledge distillation to design efficient UNet-based diffusion models. By adopting a modular search space, LDMOES effectively decouples architecture components for improved search efficiency. We validated our method on multiple datasets, including CIFAR-10, Tiny-ImageNet, CelebA-HQ [Formula: see text], and LSUN-church [Formula: see text]. Experiments show that LDMOES reduces multiply-accumulate operations (MACs) by approximately 40% in pixel space while outperforming the teacher model. When transferred to the larger-scale Tiny-ImageNet dataset, it still generates high-quality images with a competitive FID score of 4.16, demonstrating strong generalization ability. In latent space, MACs are reduced by about 50% with negligible performance loss. After transferring to the more complex LSUN-church dataset, the model surpasses baselines in generation quality while reducing computational cost by nearly 60%, validating the effectiveness and transferability of the multi-objective search strategy. Code and models will be available at https://github.com/GenerativeMind-arch/LDMOES.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550059"},"PeriodicalIF":6.4,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144983965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-12-06DOI: 10.1142/S0129065725500662
Gema Benedicto-Rodríguez, Andrea Hongn, Carlos G Juan, Javier Garrigós-Guerrero, María Paula Bonomini, Eduardo Fernandez-Jover, Jose Manuel Ferrández-Vicente
In a world where social interaction presents challenges for children with Autism Spectrum Disorder (ASD), robots are stepping in as allies in emotional learning. This study examined how affective interactions with a humanoid robot elicited physiological responses in children with ASD, using electrodermal activity (EDA) and heart rate variability (HRV) as key indicators of emotional arousal. The objectives were to identify emotionally salient moments during human-robot interaction, assess whether certain individual characteristics - such as age or ASD severity - modulate autonomic responses, and evaluate the usefulness of wearable devices for real-time monitoring. Thirteen children participated in structured sessions involving a range of social, cognitive, and motor tasks alongside the robot Pepper. The results showed that the hugging phase (HS2) often generated greater autonomic reactivity in children, especially among younger children and those with higher levels of restlessness or a higher level of ASD. Children with level 2 ASD displayed higher sympathetic activation compared to level 1 participants, who showed more HRV stability. Age also played a role, as younger children demonstrated lower autonomic regulation. These findings highlight the relevance of physiological monitoring in detecting emotional dysregulation and tailoring robot-assisted therapy. Future developments will explore adaptive systems capable of adjusting interventions in real time to better support each child's unique needs.
{"title":"Physiological Response in Children with Autism Spectrum Disorder (ASD) During Social Robot Interaction.","authors":"Gema Benedicto-Rodríguez, Andrea Hongn, Carlos G Juan, Javier Garrigós-Guerrero, María Paula Bonomini, Eduardo Fernandez-Jover, Jose Manuel Ferrández-Vicente","doi":"10.1142/S0129065725500662","DOIUrl":"10.1142/S0129065725500662","url":null,"abstract":"<p><p>In a world where social interaction presents challenges for children with Autism Spectrum Disorder (ASD), robots are stepping in as allies in emotional learning. This study examined how affective interactions with a humanoid robot elicited physiological responses in children with ASD, using electrodermal activity (EDA) and heart rate variability (HRV) as key indicators of emotional arousal. The objectives were to identify emotionally salient moments during human-robot interaction, assess whether certain individual characteristics - such as age or ASD severity - modulate autonomic responses, and evaluate the usefulness of wearable devices for real-time monitoring. Thirteen children participated in structured sessions involving a range of social, cognitive, and motor tasks alongside the robot Pepper. The results showed that the hugging phase (HS2) often generated greater autonomic reactivity in children, especially among younger children and those with higher levels of restlessness or a higher level of ASD. Children with level 2 ASD displayed higher sympathetic activation compared to level 1 participants, who showed more HRV stability. Age also played a role, as younger children demonstrated lower autonomic regulation. These findings highlight the relevance of physiological monitoring in detecting emotional dysregulation and tailoring robot-assisted therapy. Future developments will explore adaptive systems capable of adjusting interventions in real time to better support each child's unique needs.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550066"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145688894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-10-18DOI: 10.1142/S012906572550073X
Muhammad Suffian, Cosimo Ieracitano, Francesco C Morabito, Nadia Mammone
Decoding electroencephalographic (EEG) signals is of key importance in the development of brain-computer interface (BCI) systems. However, high inter-subject variability in EEG signals requires user-specific calibration, which can be time-consuming and limit the application of deep learning approaches, due to general need of large amount of data to properly train these models. In this context, this paper proposes a multidimensional and explainable deep learning framework for fast and interpretable EEG decoding. In particular, EEG signals are projected into the spatial-spectral-temporal domain and processed using a custom three-dimensional (3D) Convolutional Neural Network, here referred to as EEGCubeNet. In this work, the method has been validated on EEGs recorded during motor BCI experiments. Namely, hand open (HO) and hand close (HC) movement planning was investigated by discriminating them from the absence of movement preparation (resting state, RE). The proposed method is based on a global- to subject-specific fine-tuning. The model is globally trained on a population of subjects and then fine-tuned on the final user, significantly reducing adaptation time. Experimental results demonstrate that EEGCubeNet achieves state-of-the-art performance (accuracy of [Formula: see text] and [Formula: see text] for HC versus RE and HO versus RE, binary classification tasks, respectively) with reduced framework complexity and with a reduced training time. In addition, to enhance transparency, a 3D occlusion sensitivity analysis-based explainability method (here named 3D xAI-OSA) that generates relevance maps revealing the most significant features to each prediction, was introduced. The data and source code are available at the following link: https://github.com/AI-Lab-UniRC/EEGCubeNet.
脑电图信号的解码是脑机接口(BCI)系统发展的关键。然而,脑电图信号的高度主体间可变性需要用户特定的校准,这可能是耗时的,并且限制了深度学习方法的应用,因为通常需要大量的数据来正确训练这些模型。在此背景下,本文提出了一个多维、可解释的深度学习框架,用于快速、可解释的脑电图解码。特别是,EEG信号被投射到空间-频谱-时间域中,并使用定制的三维(3D)卷积神经网络(这里称为EEGCubeNet)进行处理。在这项工作中,该方法已在运动脑机接口实验中记录的脑电图上得到验证。即,手张开(HO)和手闭合(HC)的运动计划通过区分它们与缺乏运动准备(RE)进行研究。所提出的方法是基于全局到特定主题的微调。该模型在一组对象上进行全局训练,然后在最终用户上进行微调,大大减少了适应时间。实验结果表明,EEGCubeNet在降低框架复杂度和减少训练时间的情况下,达到了最先进的性能(分别为HC与RE和HO与RE的二元分类任务[Formula: see text]和[Formula: see text]的准确率)。此外,为了提高透明度,引入了一种基于3D遮挡敏感性分析的可解释性方法(这里称为3D xAI-OSA),该方法生成了揭示每个预测最重要特征的相关性图。数据和源代码可从以下链接获得:https://github.com/AI-Lab-UniRC/EEGCubeNet。
{"title":"An Explainable 3D-Deep Learning Model for EEG Decoding in Brain-Computer Interface Applications.","authors":"Muhammad Suffian, Cosimo Ieracitano, Francesco C Morabito, Nadia Mammone","doi":"10.1142/S012906572550073X","DOIUrl":"10.1142/S012906572550073X","url":null,"abstract":"<p><p>Decoding electroencephalographic (EEG) signals is of key importance in the development of brain-computer interface (BCI) systems. However, high inter-subject variability in EEG signals requires user-specific calibration, which can be time-consuming and limit the application of deep learning approaches, due to general need of large amount of data to properly train these models. In this context, this paper proposes a multidimensional and explainable deep learning framework for fast and interpretable EEG decoding. In particular, EEG signals are projected into the spatial-spectral-temporal domain and processed using a custom three-dimensional (3D) Convolutional Neural Network, here referred to as <i>EEGCubeNet</i>. In this work, the method has been validated on EEGs recorded during motor BCI experiments. Namely, hand open (HO) and hand close (HC) movement planning was investigated by discriminating them from the absence of movement preparation (resting state, RE). The proposed method is based on a global- to subject-specific fine-tuning. The model is globally trained on a population of subjects and then fine-tuned on the final user, significantly reducing adaptation time. Experimental results demonstrate that <i>EEGCubeNet</i> achieves state-of-the-art performance (accuracy of [Formula: see text] and [Formula: see text] for HC versus RE and HO versus RE, binary classification tasks, respectively) with reduced framework complexity and with a reduced training time. In addition, to enhance transparency, a 3D occlusion sensitivity analysis-based explainability method (here named <i>3D xAI-OSA</i>) that generates relevance maps revealing the most significant features to each prediction, was introduced. The data and source code are available at the following link: https://github.com/AI-Lab-UniRC/EEGCubeNet.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550073"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145318871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-12-09DOI: 10.1142/S0129065725500820
Thomas Pirenne, Mansoureh Fahimi Hnazaee, Patrick Santens, Aline Moorkens, Marc M Van Hulle
Deficits in auditory perception have been widely observed in Parkinson's disease (PD) patients and the literature attributes it, in part, to impaired central auditory processing. Deep brain stimulation (DBS) of the subthalamic nucleus (STN) is a well-established therapeutic option for patients with advanced PD. Analysis of auditory evoked potentials suggested a modulatory effect of DBS on central auditory processing. To better understand the latter, we investigated whether DBS modulates auditory steady state responses (ASSR) in electroencephalography (EEG) recordings of 5 PD patients. ASSRs are neural responses along central auditory pathways phase-locked to an auditory stimulus which can serve to understand the spectral aspects of central auditory processing. In our analyses, we estimate the intensity of ASSRs with a novel method based on canonical correlation analysis (CCA) and compare them in DBS ON and OFF conditions. Our results suggest that DBS effectively reduces ASSR in patients with PD. A comparison to age-matched healthy participants suggests a pathological effect of PD on ASSRs, which is disrupted by DBS. These findings support our hypothesis that DBS suppresses central auditory processing. Further research is required to assess the symptomatic effect of this modulation, as well as which cortical and subcortical generators are most affected. A better understanding of the auditory side-effects of DBS could lead to improved treatment options.
{"title":"Subthalamic Nucleus Deep Brain Stimulation Modulates Auditory Steady State Responses in Parkinson's Disease.","authors":"Thomas Pirenne, Mansoureh Fahimi Hnazaee, Patrick Santens, Aline Moorkens, Marc M Van Hulle","doi":"10.1142/S0129065725500820","DOIUrl":"https://doi.org/10.1142/S0129065725500820","url":null,"abstract":"<p><p>Deficits in auditory perception have been widely observed in Parkinson's disease (PD) patients and the literature attributes it, in part, to impaired central auditory processing. Deep brain stimulation (DBS) of the subthalamic nucleus (STN) is a well-established therapeutic option for patients with advanced PD. Analysis of auditory evoked potentials suggested a modulatory effect of DBS on central auditory processing. To better understand the latter, we investigated whether DBS modulates auditory steady state responses (ASSR) in electroencephalography (EEG) recordings of 5 PD patients. ASSRs are neural responses along central auditory pathways phase-locked to an auditory stimulus which can serve to understand the spectral aspects of central auditory processing. In our analyses, we estimate the intensity of ASSRs with a novel method based on canonical correlation analysis (CCA) and compare them in DBS ON and OFF conditions. Our results suggest that DBS effectively reduces ASSR in patients with PD. A comparison to age-matched healthy participants suggests a pathological effect of PD on ASSRs, which is disrupted by DBS. These findings support our hypothesis that DBS suppresses central auditory processing. Further research is required to assess the symptomatic effect of this modulation, as well as which cortical and subcortical generators are most affected. A better understanding of the auditory side-effects of DBS could lead to improved treatment options.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":"35 13","pages":"2550082"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145764781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-09-18DOI: 10.1142/S0129065725500650
Ferrante Neri, Mengchen Yang, Yu Xue
In the context of neural system structure modeling and complex visual tasks, the effective integration of multi-scale features and contextual information is critical for enhancing model performance. This paper proposes a biologically inspired hybrid neural network architecture - CompEyeNet - which combines the global modeling capacity of transformers with the efficiency of lightweight convolutional structures. The backbone network, multi-attention transformer backbone network (MATBN), integrates multiple attention mechanisms to collaboratively model local details and long-range dependencies. The neck network, compound eye neck network (CENN), introduces high-resolution feature layers and efficient attention fusion modules to significantly enhance multi-scale information representation and reconstruction capability. CompEyeNet is evaluated on three authoritative medical image segmentation datasets: MICCAI-CVC-ClinicDB, ISIC2018, and MICCAI-tooth-segmentation, demonstrating its superior performance. Experimental results show that compared to models such as Deeplab, Unet, and the YOLO series, CompEyeNet achieves better performance with fewer parameters. Specifically, compared to the baseline model YOLOv11, CompEyeNet reduces the number of parameters by an average of 38.31%. On key performance metrics, the average Dice coefficient improves by 0.87%, the Jaccard index by 1.53%, Precision by 0.58%, and Recall by 1.11%. These findings verify the advantages of the proposed architecture in terms of parameter efficiency and accuracy, highlighting the broad application potential of bio-inspired attention-fusion hybrid neural networks in neural system modeling and image analysis.
{"title":"A Compound-Eye-Inspired Multi-Scale Neural Architecture with Integrated Attention Mechanisms.","authors":"Ferrante Neri, Mengchen Yang, Yu Xue","doi":"10.1142/S0129065725500650","DOIUrl":"10.1142/S0129065725500650","url":null,"abstract":"<p><p>In the context of neural system structure modeling and complex visual tasks, the effective integration of multi-scale features and contextual information is critical for enhancing model performance. This paper proposes a biologically inspired hybrid neural network architecture - CompEyeNet - which combines the global modeling capacity of transformers with the efficiency of lightweight convolutional structures. The backbone network, multi-attention transformer backbone network (MATBN), integrates multiple attention mechanisms to collaboratively model local details and long-range dependencies. The neck network, compound eye neck network (CENN), introduces high-resolution feature layers and efficient attention fusion modules to significantly enhance multi-scale information representation and reconstruction capability. CompEyeNet is evaluated on three authoritative medical image segmentation datasets: MICCAI-CVC-ClinicDB, ISIC2018, and MICCAI-tooth-segmentation, demonstrating its superior performance. Experimental results show that compared to models such as Deeplab, Unet, and the YOLO series, CompEyeNet achieves better performance with fewer parameters. Specifically, compared to the baseline model YOLOv11, CompEyeNet reduces the number of parameters by an average of 38.31%. On key performance metrics, the average Dice coefficient improves by 0.87%, the Jaccard index by 1.53%, Precision by 0.58%, and Recall by 1.11%. These findings verify the advantages of the proposed architecture in terms of parameter efficiency and accuracy, highlighting the broad application potential of bio-inspired attention-fusion hybrid neural networks in neural system modeling and image analysis.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550065"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145115668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-08-22DOI: 10.1142/S0129065725500571
José A Rodríguez-Rodríguez, Miguel A Molina-Cabello, Rafaela Benítez-Rochel, Ezequiel López-Rubio
The real-time synthesis of 3D views, facilitated by convolutional neural networks like NeX, is increasingly pivotal in various computer vision applications. These networks are trained using photographs taken from different perspectives during the training phase. However, these images may be susceptible to contamination from noise originating from the vision sensor or the surrounding environment. This research meticulously examines the impact of noise on the resulting image quality of 3D views synthesized by the NeX network. Various noise levels and scenes have been incorporated to substantiate the claim that the presence of noise significantly degrades image quality. Additionally, a new strategy is introduced to improve image quality by calculating consensus among NeX networks trained on images pre-processed with a denoising algorithm. Experimental results confirm the effectiveness of this technique, demonstrating improvements of up to 1.300 dB and 0.032 for Peak Signal Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM), respectively, under certain scenes and noise levels. Notably, the performance gains are especially significant when using synthesized images generated by NeX from noisy inputs in the consensus process.
{"title":"Consensus-Based 3D View Generation from Noisy Images.","authors":"José A Rodríguez-Rodríguez, Miguel A Molina-Cabello, Rafaela Benítez-Rochel, Ezequiel López-Rubio","doi":"10.1142/S0129065725500571","DOIUrl":"10.1142/S0129065725500571","url":null,"abstract":"<p><p>The real-time synthesis of 3D views, facilitated by convolutional neural networks like NeX, is increasingly pivotal in various computer vision applications. These networks are trained using photographs taken from different perspectives during the training phase. However, these images may be susceptible to contamination from noise originating from the vision sensor or the surrounding environment. This research meticulously examines the impact of noise on the resulting image quality of 3D views synthesized by the NeX network. Various noise levels and scenes have been incorporated to substantiate the claim that the presence of noise significantly degrades image quality. Additionally, a new strategy is introduced to improve image quality by calculating consensus among NeX networks trained on images pre-processed with a denoising algorithm. Experimental results confirm the effectiveness of this technique, demonstrating improvements of up to 1.300 dB and 0.032 for Peak Signal Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM), respectively, under certain scenes and noise levels. Notably, the performance gains are especially significant when using synthesized images generated by NeX from noisy inputs in the consensus process.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550057"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144984009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-02-03DOI: 10.1142/S0129065725020010
Zvi Kam, Giovanna Nicora
{"title":"Editorial - A Journal that Promotes Excellence Through Uncompromising Review Process: Reflection of Freedom of Speech and Scientific Publication.","authors":"Zvi Kam, Giovanna Nicora","doi":"10.1142/S0129065725020010","DOIUrl":"10.1142/S0129065725020010","url":null,"abstract":"","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2502001"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143191643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-30Epub Date: 2025-08-27DOI: 10.1142/S0129065725500534
Gergő Bognár, Manuel Feindert, Christian Huber, Michael Lunglmayr, Mario Huemer, Péter Kovács
In this paper, we present a hybrid learning framework that integrates two model-driven AI paradigms: Deep unfolding and Variable Projections (VPs). The core idea is to unfold the iterations of VP solvers for separable nonlinear least squares (SNLLS) problems into trainable neural network layers. As a consequence, the network is capable of learning optimal nonlinear VP parameters during inference, which is a form of model-based meta-learning. Furthermore, the architecture incorporates prior knowledge of the underlying SNLLS problem, such as basis function expansions and signal structure, which enhance interpretability, reduce model size, and lower data requirements. As a case study, we adapt the proposed deep unfolded VPNet to learn ECG representations for the classification of five arrhythmias. Experimental results on the MIT-BIH Arrhythmia Database show that VPNet achieves performance comparable to state-of-the-art ECG classifiers, attaining 95% accuracy while maintaining a compact architecture. Its low computational complexity enables efficient training and inference, making it highly suitable for real-time, power-efficient edge computing applications. This is further validated through embedded implementation on STM32 microcontrollers.
{"title":"Deep Unfolded Variable Projection Networks.","authors":"Gergő Bognár, Manuel Feindert, Christian Huber, Michael Lunglmayr, Mario Huemer, Péter Kovács","doi":"10.1142/S0129065725500534","DOIUrl":"10.1142/S0129065725500534","url":null,"abstract":"<p><p>In this paper, we present a hybrid learning framework that integrates two model-driven AI paradigms: Deep unfolding and Variable Projections (VPs). The core idea is to unfold the iterations of VP solvers for separable nonlinear least squares (SNLLS) problems into trainable neural network layers. As a consequence, the network is capable of learning optimal nonlinear VP parameters during inference, which is a form of model-based meta-learning. Furthermore, the architecture incorporates prior knowledge of the underlying SNLLS problem, such as basis function expansions and signal structure, which enhance interpretability, reduce model size, and lower data requirements. As a case study, we adapt the proposed deep unfolded VPNet to learn ECG representations for the classification of five arrhythmias. Experimental results on the MIT-BIH Arrhythmia Database show that VPNet achieves performance comparable to state-of-the-art ECG classifiers, attaining 95% accuracy while maintaining a compact architecture. Its low computational complexity enables efficient training and inference, making it highly suitable for real-time, power-efficient edge computing applications. This is further validated through embedded implementation on STM32 microcontrollers.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550053"},"PeriodicalIF":6.4,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144983999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cellular traffic forecasting is a critical task that enables network operators to efficiently allocate resources and address anomalies in rapidly evolving environments. The exponential growth of data collected from base stations poses significant challenges to processing and analysis. While machine learning (ML) algorithms have emerged as powerful tools for handling these large datasets and providing accurate predictions, their environmental impact, particularly in terms of energy consumption, is often overlooked in favor of their predictive capabilities. This study investigates the potential of two bio-inspired models: Spiking Neural Networks (SNNs) and Reservoir Computing through Echo State Networks (ESNs) for cellular traffic forecasting. The evaluation focuses on both their predictive performance and energy efficiency. These models are implemented in both centralized and federated settings to analyze their effectiveness and energy consumption in decentralized systems. Additionally, we compare bio-inspired models with traditional architectures, such as Convolutional Neural Networks (CNNs) and Multi-Layer Perceptrons (MLPs), to provide a comprehensive evaluation. Using data collected from three diverse locations in Barcelona, Spain, we examine the trade-offs between predictive accuracy and energy demands across these approaches. The results indicate that bio-inspired models, such as SNNs and ESNs, can achieve significant energy savings while maintaining predictive accuracy comparable to traditional architectures. Furthermore, federated implementations were tested to evaluate their energy efficiency in decentralized settings compared to centralized systems, particularly in combination with bio-inspired models. These findings offer valuable insights into the potential of bio-inspired models for sustainable and privacy-preserving cellular traffic forecasting.
蜂窝流量预测是一项关键任务,它使网络运营商能够有效地分配资源,并在快速发展的环境中处理异常情况。从基站收集的数据呈指数级增长,对处理和分析提出了重大挑战。虽然机器学习(ML)算法已经成为处理这些大型数据集并提供准确预测的强大工具,但它们对环境的影响,特别是在能源消耗方面,往往被忽视,而倾向于它们的预测能力。本研究探讨了两种生物启发模型的潜力:脉冲神经网络(SNNs)和通过回声状态网络(ESNs)的水库计算(Reservoir Computing through Echo State Networks, ESNs)用于蜂窝流量预测。评估的重点是它们的预测性能和能源效率。这些模型在集中式和联邦设置中实现,以分析它们在分散系统中的有效性和能耗。此外,我们将生物启发模型与传统架构(如卷积神经网络(cnn)和多层感知器(mlp))进行比较,以提供全面的评估。使用从西班牙巴塞罗那三个不同地点收集的数据,我们检查了这些方法中预测准确性和能源需求之间的权衡。结果表明,生物启发模型,如snn和esn,可以在保持与传统架构相当的预测精度的同时实现显著的节能。此外,对联邦实施进行了测试,以评估其在分散环境下与集中式系统相比的能源效率,特别是与生物启发模型相结合。这些发现为生物启发模型的潜力提供了宝贵的见解,以实现可持续和隐私保护的蜂窝流量预测。
{"title":"Evaluation of Bio-Inspired Models under Different Learning Settings for Energy Efficiency in Network Traffic Prediction.","authors":"Theodoros Tsiolakis, Nikolaos Pavlidis, Vasileios Perifanis, Pavlos Efraimidis","doi":"10.1142/S0129065725500583","DOIUrl":"10.1142/S0129065725500583","url":null,"abstract":"<p><p>Cellular traffic forecasting is a critical task that enables network operators to efficiently allocate resources and address anomalies in rapidly evolving environments. The exponential growth of data collected from base stations poses significant challenges to processing and analysis. While machine learning (ML) algorithms have emerged as powerful tools for handling these large datasets and providing accurate predictions, their environmental impact, particularly in terms of energy consumption, is often overlooked in favor of their predictive capabilities. This study investigates the potential of two bio-inspired models: Spiking Neural Networks (SNNs) and Reservoir Computing through Echo State Networks (ESNs) for cellular traffic forecasting. The evaluation focuses on both their predictive performance and energy efficiency. These models are implemented in both centralized and federated settings to analyze their effectiveness and energy consumption in decentralized systems. Additionally, we compare bio-inspired models with traditional architectures, such as Convolutional Neural Networks (CNNs) and Multi-Layer Perceptrons (MLPs), to provide a comprehensive evaluation. Using data collected from three diverse locations in Barcelona, Spain, we examine the trade-offs between predictive accuracy and energy demands across these approaches. The results indicate that bio-inspired models, such as SNNs and ESNs, can achieve significant energy savings while maintaining predictive accuracy comparable to traditional architectures. Furthermore, federated implementations were tested to evaluate their energy efficiency in decentralized settings compared to centralized systems, particularly in combination with bio-inspired models. These findings offer valuable insights into the potential of bio-inspired models for sustainable and privacy-preserving cellular traffic forecasting.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550058"},"PeriodicalIF":6.4,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144983990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-15Epub Date: 2025-11-19DOI: 10.1142/S0129065725500649
Paraskevi V Tsakmaki, Sotiris Tasoulis, Spiros V Georgakopoulos, Vassilis P Plagianakos
This research looks at the use of long-short-term memory (LSTM) networks to predict psychosis, in patients within the schizophrenia spectrum, based on Heart Rate Variability (HRV) data acquired from wearable devices. Our primary objective is to test whether the personalized relapse prediction remains accurate when eliminating the artifact-removal preprocessing. We first analyzed 7 patients sleep HRV recordings (7-113 days each), and then validated the methodology on a separate 30-patient psychosis cohort from another clinical setting. In this framework, HRV characteristics are computed directly from the unprocessed time series for each patient, without artifact correction, at any stage prior to feature extraction. HRV features are then, organized into sequential inputs, where the model uses the first n-1 steps to predict the nth step. This structure allows the model to learn from temporal relationships and individual physiological trends in HRV. The sequence length used by the LSTM is optimized for each patient, allowing the model to account for individual physiological patterns. Through this, on the 7 patient cohort, the LSTM model reaches a mean F1 score of 0.9817, marking its strength across diverse patient profiles. The method provides predictions for each individual by learning from their own HRV history. Results using both traditional and state-of-the-art noise-removal techniques, like wavelet and GAN-based denoising, showed that omitting these data cleaning steps did not reduce, and in some cases even improved, prediction accuracy. These findings indicate that, for psychosis prediction based on wearable HRV data, additional data cleaning may not be necessary.
{"title":"Reducing Artifact Preprocessing in Heart Rate Variability-Based Personalized Psychosis Prediction Using Adaptive Long Short-Term Memory Models.","authors":"Paraskevi V Tsakmaki, Sotiris Tasoulis, Spiros V Georgakopoulos, Vassilis P Plagianakos","doi":"10.1142/S0129065725500649","DOIUrl":"10.1142/S0129065725500649","url":null,"abstract":"<p><p>This research looks at the use of long-short-term memory (LSTM) networks to predict psychosis, in patients within the schizophrenia spectrum, based on Heart Rate Variability (HRV) data acquired from wearable devices. Our primary objective is to test whether the personalized relapse prediction remains accurate when eliminating the artifact-removal preprocessing. We first analyzed 7 patients sleep HRV recordings (7-113 days each), and then validated the methodology on a separate 30-patient psychosis cohort from another clinical setting. In this framework, HRV characteristics are computed directly from the unprocessed time series for each patient, without artifact correction, at any stage prior to feature extraction. HRV features are then, organized into sequential inputs, where the model uses the first n-1 steps to predict the <i>n</i>th step. This structure allows the model to learn from temporal relationships and individual physiological trends in HRV. The sequence length used by the LSTM is optimized for each patient, allowing the model to account for individual physiological patterns. Through this, on the 7 patient cohort, the LSTM model reaches a mean F1 score of 0.9817, marking its strength across diverse patient profiles. The method provides predictions for each individual by learning from their own HRV history. Results using both traditional and state-of-the-art noise-removal techniques, like wavelet and GAN-based denoising, showed that omitting these data cleaning steps did not reduce, and in some cases even improved, prediction accuracy. These findings indicate that, for psychosis prediction based on wearable HRV data, additional data cleaning may not be necessary.</p>","PeriodicalId":94052,"journal":{"name":"International journal of neural systems","volume":" ","pages":"2550064"},"PeriodicalIF":6.4,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145552481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}