Pub Date : 2024-11-26eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1508297
Xinxu Lin, Mingxuan Liu, Hong Chen
Event-based cameras are suitable for human action recognition (HAR) by providing movement perception with highly dynamic range, high temporal resolution, high power efficiency and low latency. Spike Neural Networks (SNNs) are naturally suited to deal with the asynchronous and sparse data from the event cameras due to their spike-based event-driven paradigm, with less power consumption compared to artificial neural networks. In this paper, we propose two end-to-end SNNs, namely Spike-HAR and Spike-HAR++, to introduce spiking transformer into event-based HAR. Spike-HAR includes two novel blocks: a spike attention branch, which enables model to focus on regions with high spike rates, reducing the impact of noise to improve the accuracy, and a parallel spike transformer block with simplified spiking self-attention mechanism, increasing computational efficiency. To better extract crucial information from high-level features, we modify the architecture of the spike attention branch and extend it in Spike-HAR to a higher dimension, proposing Spike-HAR++ to further enhance classification performance. Comprehensive experiments were conducted on four HAR datasets: SL-Animals-DVS, N-LSA64, DVS128 Gesture and DailyAction-DVS, to demonstrate the superior performance of our proposed model. Additionally, the proposed Spike-HAR and Spike-HAR++ require only 0.03 and 0.06 mJ, respectively, to process a sequence of event frames, with model sizes of only 0.7 and 1.8 M. This efficiency positions it as a promising new SNN baseline for the HAR community. Code is available at Spike-HAR++.
{"title":"Spike-HAR++: an energy-efficient and lightweight parallel spiking transformer for event-based human action recognition.","authors":"Xinxu Lin, Mingxuan Liu, Hong Chen","doi":"10.3389/fncom.2024.1508297","DOIUrl":"10.3389/fncom.2024.1508297","url":null,"abstract":"<p><p>Event-based cameras are suitable for human action recognition (HAR) by providing movement perception with highly dynamic range, high temporal resolution, high power efficiency and low latency. Spike Neural Networks (SNNs) are naturally suited to deal with the asynchronous and sparse data from the event cameras due to their spike-based event-driven paradigm, with less power consumption compared to artificial neural networks. In this paper, we propose two end-to-end SNNs, namely Spike-HAR and Spike-HAR++, to introduce spiking transformer into event-based HAR. Spike-HAR includes two novel blocks: a spike attention branch, which enables model to focus on regions with high spike rates, reducing the impact of noise to improve the accuracy, and a parallel spike transformer block with simplified spiking self-attention mechanism, increasing computational efficiency. To better extract crucial information from high-level features, we modify the architecture of the spike attention branch and extend it in Spike-HAR to a higher dimension, proposing Spike-HAR++ to further enhance classification performance. Comprehensive experiments were conducted on four HAR datasets: SL-Animals-DVS, N-LSA64, DVS128 Gesture and DailyAction-DVS, to demonstrate the superior performance of our proposed model. Additionally, the proposed Spike-HAR and Spike-HAR++ require only 0.03 and 0.06 mJ, respectively, to process a sequence of event frames, with model sizes of only 0.7 and 1.8 M. This efficiency positions it as a promising new SNN baseline for the HAR community. Code is available at Spike-HAR++.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1508297"},"PeriodicalIF":2.1,"publicationDate":"2024-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11628275/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142806119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-20eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1417558
Pan Xia, Henry D I Abarbanel
The nucleus HVC within the avian song system produces crystalized instructions which lead to precise, learned vocalization in zebra finches (Taeniopygia guttata). This paper proposes a model of the HVC neural network based on the physiological properties of individual HVC neurons, their synaptic interactions calibrated by experimental measurements, as well as the synaptic signal into this region which triggers song production. This neural network model comprises of two major neural populations in this area: neurons projecting to the nucleus RA and interneurons. Each single neuron model of HVCRA is constructed with conductance-based ion currents of fast Na+ and K+ and a leak channel, while the interneuron model includes extra transient Ca2+ current and hyperpolarization-activated inward current. The synaptic dynamics is formed with simulated delivered neurotransmitter pulses from presynaptic cells and neurotransmitter receptor opening rates of postsynaptic neurons. We show that this network model qualitatively exhibits observed electrophysiological behaviors of neurons independent or in the network, as well as the importance of bidirectional interactions between the HVCRA neuron and the HVCI neuron. We also simulate the pulse input from A11 neuron group to HVC. This signal successfully suppresses the interneuron, which leads to sequential firing of projection neurons that matches measured burst onset, duration, and spike quantities during the zebra finch motif. The result provides a biophysically based model characterizing the dynamics and functions of the HVC neural network as a song motor, and offers a reference for synaptic coupling strength in the avian brain.
{"title":"Model of the HVC neural network as a song motor in zebra finch.","authors":"Pan Xia, Henry D I Abarbanel","doi":"10.3389/fncom.2024.1417558","DOIUrl":"10.3389/fncom.2024.1417558","url":null,"abstract":"<p><p>The nucleus HVC within the avian song system produces crystalized instructions which lead to precise, learned vocalization in zebra finches (<i>Taeniopygia guttata</i>). This paper proposes a model of the HVC neural network based on the physiological properties of individual HVC neurons, their synaptic interactions calibrated by experimental measurements, as well as the synaptic signal into this region which triggers song production. This neural network model comprises of two major neural populations in this area: neurons projecting to the nucleus RA and interneurons. Each single neuron model of HVC<sub>RA</sub> is constructed with conductance-based ion currents of fast Na<sup>+</sup> and K<sup>+</sup> and a leak channel, while the interneuron model includes extra transient Ca<sup>2+</sup> current and hyperpolarization-activated inward current. The synaptic dynamics is formed with simulated delivered neurotransmitter pulses from presynaptic cells and neurotransmitter receptor opening rates of postsynaptic neurons. We show that this network model qualitatively exhibits observed electrophysiological behaviors of neurons independent or in the network, as well as the importance of bidirectional interactions between the HVC<sub>RA</sub> neuron and the HVC<sub>I</sub> neuron. We also simulate the pulse input from A11 neuron group to HVC. This signal successfully suppresses the interneuron, which leads to sequential firing of projection neurons that matches measured burst onset, duration, and spike quantities during the zebra finch motif. The result provides a biophysically based model characterizing the dynamics and functions of the HVC neural network as a song motor, and offers a reference for synaptic coupling strength in the avian brain.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1417558"},"PeriodicalIF":2.1,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11614668/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142779799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Background: The methods used to detect epileptic seizures using electroencephalogram (EEG) signals suffer from poor accuracy in feature selection and high redundancy. This problem is addressed through the use of a novel multi-domain feature fusion and selection method (PMPSO).
Method: Discrete Wavelet Transforms (DWT) and Welch are used initially to extract features from different domains, including frequency domain, time-frequency domain, and non-linear domain. The first step in the detection process is to extract important features from different domains, such as frequency domain, time-frequency domain, and non-linear domain, using methods such as Discrete Wavelet Transform (DWT) and Welch. To extract features strongly correlated with epileptic classification detection, an improved particle swarm optimization (PSO) algorithm and Pearson correlation analysis are combined. Finally, Support Vector Machines (SVM), Artificial Neural Networks (ANN), Random Forest (RF) and XGBoost classifiers are used to construct epileptic seizure detection models based on the optimized detection features.
Result: According to experimental results, the proposed method achieves 99.32% accuracy, 99.64% specificity, 99.29% sensitivity, and 99.32% score, respectively.
Conclusion: The detection performance of the three classifiers is compared using 10-fold cross-validation. Surpassing other methods in detection accuracy. Consequently, this optimized method for epilepsy seizure detection enhances the diagnostic accuracy of epilepsy seizures.
{"title":"A novel method for optimizing epilepsy detection features through multi-domain feature fusion and selection.","authors":"Guanqing Kong, Shuang Ma, Wei Zhao, Haifeng Wang, Qingxi Fu, Jiuru Wang","doi":"10.3389/fncom.2024.1416838","DOIUrl":"10.3389/fncom.2024.1416838","url":null,"abstract":"<p><strong>Background: </strong>The methods used to detect epileptic seizures using electroencephalogram (EEG) signals suffer from poor accuracy in feature selection and high redundancy. This problem is addressed through the use of a novel multi-domain feature fusion and selection method (PMPSO).</p><p><strong>Method: </strong>Discrete Wavelet Transforms (DWT) and Welch are used initially to extract features from different domains, including frequency domain, time-frequency domain, and non-linear domain. The first step in the detection process is to extract important features from different domains, such as frequency domain, time-frequency domain, and non-linear domain, using methods such as Discrete Wavelet Transform (DWT) and Welch. To extract features strongly correlated with epileptic classification detection, an improved particle swarm optimization (PSO) algorithm and Pearson correlation analysis are combined. Finally, Support Vector Machines (SVM), Artificial Neural Networks (ANN), Random Forest (RF) and XGBoost classifiers are used to construct epileptic seizure detection models based on the optimized detection features.</p><p><strong>Result: </strong>According to experimental results, the proposed method achieves 99.32% accuracy, 99.64% specificity, 99.29% sensitivity, and 99.32% score, respectively.</p><p><strong>Conclusion: </strong>The detection performance of the three classifiers is compared using 10-fold cross-validation. Surpassing other methods in detection accuracy. Consequently, this optimized method for epilepsy seizure detection enhances the diagnostic accuracy of epilepsy seizures.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1416838"},"PeriodicalIF":2.1,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11612629/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142767773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-15eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1471229
Stein Acker, Jinqing Liang, Ninet Sinaii, Kristen Wingert, Atsuko Kurosu, Sunder Rajan, Sara Inati, William H Theodore, Nadia Biassou
Functional connectivity (FC) refers to the activation correlation between different brain regions. FC networks as typically represented as graphs with brain regions of interest (ROIs) as nodes and functional correlation as edges. Graph neural networks (GNNs) are machine learning architectures used to analyze FC graphs. However, traditional GNNs are limited in their ability to characterize FC edge attributes because they typically emphasize the importance of ROI node-based brain activation data. Line GNNs convert the edges of the original graph to nodes in the transformed graph, thereby emphasizing the FC between brain regions. We hypothesize that line GNNs will outperform traditional GNNs in FC applications. We investigated the performance of two common GNN architectures (GraphSAGE and GCN) trained on line and traditional graphs predicting task-associated FC changes across two datasets. The first dataset was from the Human Connectome Project (HCP) with 205 participants, the second was a dataset with 12 participants. The HCP dataset detailed FC changes in participants during a story-listening task, while the second dataset included the FC changes in a different auditory language task. Our findings from the HCP dataset indicated that line GNNs achieved lower mean squared error compared to traditional GNNs, with the line GraphSAGE model outperforming the traditional GraphSAGE by 18% (p < 0.0001). When applying the same models to the second dataset, both line GNNs also showed statistically significant improvements over their traditional counterparts with little to no overfitting. We believe this shows that line GNN models demonstrate promising utility in FC studies.
{"title":"Modeling functional connectivity changes during an auditory language task using line graph neural networks.","authors":"Stein Acker, Jinqing Liang, Ninet Sinaii, Kristen Wingert, Atsuko Kurosu, Sunder Rajan, Sara Inati, William H Theodore, Nadia Biassou","doi":"10.3389/fncom.2024.1471229","DOIUrl":"https://doi.org/10.3389/fncom.2024.1471229","url":null,"abstract":"<p><p>Functional connectivity (FC) refers to the activation correlation between different brain regions. FC networks as typically represented as graphs with brain regions of interest (ROIs) as nodes and functional correlation as edges. Graph neural networks (GNNs) are machine learning architectures used to analyze FC graphs. However, traditional GNNs are limited in their ability to characterize FC edge attributes because they typically emphasize the importance of ROI node-based brain activation data. Line GNNs convert the edges of the original graph to nodes in the transformed graph, thereby emphasizing the FC between brain regions. We hypothesize that line GNNs will outperform traditional GNNs in FC applications. We investigated the performance of two common GNN architectures (GraphSAGE and GCN) trained on line and traditional graphs predicting task-associated FC changes across two datasets. The first dataset was from the Human Connectome Project (HCP) with 205 participants, the second was a dataset with 12 participants. The HCP dataset detailed FC changes in participants during a story-listening task, while the second dataset included the FC changes in a different auditory language task. Our findings from the HCP dataset indicated that line GNNs achieved lower mean squared error compared to traditional GNNs, with the line GraphSAGE model outperforming the traditional GraphSAGE by 18% (<i>p</i> < 0.0001). When applying the same models to the second dataset, both line GNNs also showed statistically significant improvements over their traditional counterparts with little to no overfitting. We believe this shows that line GNN models demonstrate promising utility in FC studies.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1471229"},"PeriodicalIF":2.1,"publicationDate":"2024-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11604466/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142767777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-13eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1452457
Mahsa Dibaji, Johanna Ospel, Roberto Souza, Mariana Bento
This study leverages deep learning to analyze sex differences in brain MRI data, aiming to further advance fairness in medical imaging. We employed 3D T1-weighted Magnetic Resonance images from four diverse datasets: Calgary-Campinas-359, OASIS-3, Alzheimer's Disease Neuroimaging Initiative, and Cambridge Center for Aging and Neuroscience, ensuring a balanced representation of sexes and a broad demographic scope. Our methodology focused on minimal preprocessing to preserve the integrity of brain structures, utilizing a Convolutional Neural Network model for sex classification. The model achieved an accuracy of 87% on the test set without employing total intracranial volume (TIV) adjustment techniques. We observed that while the model exhibited biases at extreme brain sizes, it performed with less bias when the TIV distributions overlapped more. Saliency maps were used to identify brain regions significant in sex differentiation, revealing that certain supratentorial and infratentorial regions were important for predictions. Furthermore, our interdisciplinary team, comprising machine learning specialists and a radiologist, ensured diverse perspectives in validating the results. The detailed investigation of sex differences in brain MRI in this study, highlighted by the sex differences map, offers valuable insights into sex-specific aspects of medical imaging and could aid in developing sex-based bias mitigation strategies, contributing to the future development of fair AI algorithms. Awareness of the brain's differences between sexes enables more equitable AI predictions, promoting fairness in healthcare outcomes. Our code and saliency maps are available at https://github.com/mahsadibaji/sex-differences-brain-dl.
{"title":"Sex differences in brain MRI using deep learning toward fairer healthcare outcomes.","authors":"Mahsa Dibaji, Johanna Ospel, Roberto Souza, Mariana Bento","doi":"10.3389/fncom.2024.1452457","DOIUrl":"10.3389/fncom.2024.1452457","url":null,"abstract":"<p><p>This study leverages deep learning to analyze sex differences in brain MRI data, aiming to further advance fairness in medical imaging. We employed 3D T1-weighted Magnetic Resonance images from four diverse datasets: Calgary-Campinas-359, OASIS-3, Alzheimer's Disease Neuroimaging Initiative, and Cambridge Center for Aging and Neuroscience, ensuring a balanced representation of sexes and a broad demographic scope. Our methodology focused on minimal preprocessing to preserve the integrity of brain structures, utilizing a Convolutional Neural Network model for sex classification. The model achieved an accuracy of 87% on the test set without employing total intracranial volume (TIV) adjustment techniques. We observed that while the model exhibited biases at extreme brain sizes, it performed with less bias when the TIV distributions overlapped more. Saliency maps were used to identify brain regions significant in sex differentiation, revealing that certain supratentorial and infratentorial regions were important for predictions. Furthermore, our interdisciplinary team, comprising machine learning specialists and a radiologist, ensured diverse perspectives in validating the results. The detailed investigation of sex differences in brain MRI in this study, highlighted by the sex differences map, offers valuable insights into sex-specific aspects of medical imaging and could aid in developing sex-based bias mitigation strategies, contributing to the future development of fair AI algorithms. Awareness of the brain's differences between sexes enables more equitable AI predictions, promoting fairness in healthcare outcomes. Our code and saliency maps are available at https://github.com/mahsadibaji/sex-differences-brain-dl.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1452457"},"PeriodicalIF":2.1,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11598355/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142738838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-08eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1514220
Noemi Montobbio, Roberto Maffulli, Anees Abrol, Pablo Martínez-Cañada
{"title":"Editorial: Computational modeling and machine learning methods in neurodevelopment and neurodegeneration: from basic research to clinical applications.","authors":"Noemi Montobbio, Roberto Maffulli, Anees Abrol, Pablo Martínez-Cañada","doi":"10.3389/fncom.2024.1514220","DOIUrl":"10.3389/fncom.2024.1514220","url":null,"abstract":"","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1514220"},"PeriodicalIF":2.1,"publicationDate":"2024-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11581899/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142709341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-06eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1466364
Eric Chalmers, Santina Duarte, Xena Al-Hejji, Daniel Devoe, Aaron Gruber, Robert J McDonald
Deep Reinforcement Learning is a branch of artificial intelligence that uses artificial neural networks to model reward-based learning as it occurs in biological agents. Here we modify a Deep Reinforcement Learning approach by imposing a suppressive effect on the connections between neurons in the artificial network-simulating the effect of dendritic spine loss as observed in major depressive disorder (MDD). Surprisingly, this simulated spine loss is sufficient to induce a variety of MDD-like behaviors in the artificially intelligent agent, including anhedonia, increased temporal discounting, avoidance, and an altered exploration/exploitation balance. Furthermore, simulating alternative and longstanding reward-processing-centric conceptions of MDD (dysfunction of the dopamine system, altered reward discounting, context-dependent learning rates, increased exploration) does not produce the same range of MDD-like behaviors. These results support a conceptual model of MDD as a reduction of brain connectivity (and thus information-processing capacity) rather than an imbalance in monoamines-though the computational model suggests a possible explanation for the dysfunction of dopamine systems in MDD. Reversing the spine-loss effect in our computational MDD model can lead to rescue of rewarding behavior under some conditions. This supports the search for treatments that increase plasticity and synaptogenesis, and the model suggests some implications for their effective administration.
{"title":"Simulated synapse loss induces depression-like behaviors in deep reinforcement learning.","authors":"Eric Chalmers, Santina Duarte, Xena Al-Hejji, Daniel Devoe, Aaron Gruber, Robert J McDonald","doi":"10.3389/fncom.2024.1466364","DOIUrl":"10.3389/fncom.2024.1466364","url":null,"abstract":"<p><p>Deep Reinforcement Learning is a branch of artificial intelligence that uses artificial neural networks to model reward-based learning as it occurs in biological agents. Here we modify a Deep Reinforcement Learning approach by imposing a suppressive effect on the connections between neurons in the artificial network-simulating the effect of dendritic spine loss as observed in major depressive disorder (MDD). Surprisingly, this simulated spine loss is sufficient to induce a variety of MDD-like behaviors in the artificially intelligent agent, including anhedonia, increased temporal discounting, avoidance, and an altered exploration/exploitation balance. Furthermore, simulating alternative and longstanding reward-processing-centric conceptions of MDD (dysfunction of the dopamine system, altered reward discounting, context-dependent learning rates, increased exploration) does not produce the same range of MDD-like behaviors. These results support a conceptual model of MDD as a reduction of brain connectivity (and thus information-processing capacity) rather than an imbalance in monoamines-though the computational model suggests a possible explanation for the dysfunction of dopamine systems in MDD. Reversing the spine-loss effect in our computational MDD model can lead to rescue of rewarding behavior under some conditions. This supports the search for treatments that increase plasticity and synaptogenesis, and the model suggests some implications for their effective administration.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1466364"},"PeriodicalIF":2.1,"publicationDate":"2024-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11576168/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142681336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-31eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1475530
Mansoor S Raza, Mohsin Murtaza, Chi-Tsun Cheng, Muhana M A Muslam, Bader M Albahlal
The intricate interplay between driver cognitive dysfunction, mental workload (MWL), and heart rate variability (HRV) provides a captivating avenue for investigation within the domain of transportation safety studies. This article provides a systematic review and examines cognitive hindrance stemming from mental workload and heart rate variability. It scrutinizes the mental workload experienced by drivers by leveraging data gleaned from prior studies that employed heart rate monitoring systems and eye tracking technology, thereby illuminating the correlation between cognitive impairment, mental workload, and physiological indicators such as heart rate and ocular movements. The investigation is grounded in the premise that the mental workload of drivers can be assessed through physiological cues, such as heart rate and eye movements. The study discovered that HRV and infrared (IR) measurements played a crucial role in evaluating fatigue and workload for skilled drivers. However, the study overlooked potential factors contributing to cognitive impairment in drivers and could benefit from incorporating alternative indicators of cognitive workload for deeper insights. Furthermore, investigated driving simulators demonstrated that an eco-safe driving Human-Machine Interface (HMI) significantly promoted safe driving behaviors without imposing excessive mental and visual workload on drivers. Recommendations were made for future studies to consider additional indicators of cognitive workload, such as subjective assessments or task performance metrics, for a more comprehensive understanding.
{"title":"Systematic review of cognitive impairment in drivers through mental workload using physiological measures of heart rate variability.","authors":"Mansoor S Raza, Mohsin Murtaza, Chi-Tsun Cheng, Muhana M A Muslam, Bader M Albahlal","doi":"10.3389/fncom.2024.1475530","DOIUrl":"10.3389/fncom.2024.1475530","url":null,"abstract":"<p><p>The intricate interplay between driver cognitive dysfunction, mental workload (MWL), and heart rate variability (HRV) provides a captivating avenue for investigation within the domain of transportation safety studies. This article provides a systematic review and examines cognitive hindrance stemming from mental workload and heart rate variability. It scrutinizes the mental workload experienced by drivers by leveraging data gleaned from prior studies that employed heart rate monitoring systems and eye tracking technology, thereby illuminating the correlation between cognitive impairment, mental workload, and physiological indicators such as heart rate and ocular movements. The investigation is grounded in the premise that the mental workload of drivers can be assessed through physiological cues, such as heart rate and eye movements. The study discovered that HRV and infrared (IR) measurements played a crucial role in evaluating fatigue and workload for skilled drivers. However, the study overlooked potential factors contributing to cognitive impairment in drivers and could benefit from incorporating alternative indicators of cognitive workload for deeper insights. Furthermore, investigated driving simulators demonstrated that an eco-safe driving Human-Machine Interface (HMI) significantly promoted safe driving behaviors without imposing excessive mental and visual workload on drivers. Recommendations were made for future studies to consider additional indicators of cognitive workload, such as subjective assessments or task performance metrics, for a more comprehensive understanding.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1475530"},"PeriodicalIF":2.1,"publicationDate":"2024-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11560442/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142617217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-30eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1435956
Shtwai Alsubai, Abdullah Alqahtani, Abed Alanazi, Mohemmed Sha, Abdu Gumaei
Introduction: Facial expressions have become a common way for interaction among humans. People cannot comprehend and predict the emotions or expressions of individuals through simple vision. Thus, in psychology, detecting facial expressions or emotion analysis demands an assessment and evaluation of decisions for identifying the emotions of a person or any group during communication. With the recent evolution of technology, AI (Artificial Intelligence) has gained significant usage, wherein DL (Deep Learning) based algorithms are employed for detecting facial expressions.
Methods: The study proposes a system design that detects facial expressions by extracting relevant features using a Modified ResNet model. The proposed system stacks building-blocks with residual connections and employs an advanced extraction method with quantum computing, which significantly reduces computation time compared to conventional methods. The backbone stem utilizes a quantum convolutional layer comprised of several parameterized quantum-filters. Additionally, the research integrates residual connections in the ResNet-18 model with the Modified up Sampled Bottle Neck Process (MuS-BNP), retaining computational efficacy while benefiting from residual connections.
Results: The proposed model demonstrates superior performance by overcoming the issue of maximum similarity within varied facial expressions. The system's ability to accurately detect and differentiate between expressions is measured using performance metrics such as accuracy, F1-score, recall, and precision.
Discussion: This performance analysis confirms the efficacy of the proposed system, highlighting the advantages of quantum computing in feature extraction and the integration of residual connections. The model achieves quantum superiority, providing faster and more accurate computations compared to existing methodologies. The results suggest that the proposed approach offers a promising solution for facial expression recognition tasks, significantly improving both speed and accuracy.
{"title":"Facial emotion recognition using deep quantum and advanced transfer learning mechanism.","authors":"Shtwai Alsubai, Abdullah Alqahtani, Abed Alanazi, Mohemmed Sha, Abdu Gumaei","doi":"10.3389/fncom.2024.1435956","DOIUrl":"10.3389/fncom.2024.1435956","url":null,"abstract":"<p><strong>Introduction: </strong>Facial expressions have become a common way for interaction among humans. People cannot comprehend and predict the emotions or expressions of individuals through simple vision. Thus, in psychology, detecting facial expressions or emotion analysis demands an assessment and evaluation of decisions for identifying the emotions of a person or any group during communication. With the recent evolution of technology, AI (Artificial Intelligence) has gained significant usage, wherein DL (Deep Learning) based algorithms are employed for detecting facial expressions.</p><p><strong>Methods: </strong>The study proposes a system design that detects facial expressions by extracting relevant features using a Modified ResNet model. The proposed system stacks building-blocks with residual connections and employs an advanced extraction method with quantum computing, which significantly reduces computation time compared to conventional methods. The backbone stem utilizes a quantum convolutional layer comprised of several parameterized quantum-filters. Additionally, the research integrates residual connections in the ResNet-18 model with the Modified up Sampled Bottle Neck Process (MuS-BNP), retaining computational efficacy while benefiting from residual connections.</p><p><strong>Results: </strong>The proposed model demonstrates superior performance by overcoming the issue of maximum similarity within varied facial expressions. The system's ability to accurately detect and differentiate between expressions is measured using performance metrics such as accuracy, F1-score, recall, and precision.</p><p><strong>Discussion: </strong>This performance analysis confirms the efficacy of the proposed system, highlighting the advantages of quantum computing in feature extraction and the integration of residual connections. The model achieves quantum superiority, providing faster and more accurate computations compared to existing methodologies. The results suggest that the proposed approach offers a promising solution for facial expression recognition tasks, significantly improving both speed and accuracy.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1435956"},"PeriodicalIF":2.1,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11557492/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142617216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-24eCollection Date: 2024-01-01DOI: 10.3389/fncom.2024.1482994
Liao Xuanzhi, Abeer Hakeem, Linda Mohaisen, Muhammad Umer, Muhammad Attique Khan, Shrooq Alsenan, Shtwai Alsubai, Nisreen Innab
Brain stress monitoring has emerged as a critical research area for understanding and managing stress and neurological health issues. This burgeoning field aims to provide accurate information and prediction about individuals' stress levels by analyzing behavioral data and physiological signals. To address this emerging problem, this research study proposes an innovative approach that uses an attention mechanism-based XLNet model (called BrainNet) for continuous stress monitoring and stress level prediction. The proposed model analyzes streams of brain data, including behavioral and physiological signal patterns using Swell and WESAD datasets. Testing on the Swell multi-class dataset, the model achieves an impressive accuracy of 95.76%. Furthermore, when evaluated on the WESAD dataset, it demonstrates even higher accuracy, reaching 98.32%. When applied to the binary classification of stress and no stress using the Swell dataset, the model achieves an outstanding accuracy of 97.19%. Comparative analysis with other previously published research studies underscores the superior performance of the proposed approach. In addition, cross-validation confirms the significance, efficacy, and robustness of the model in brain stress level prediction and aligns with the goals of smart diagnostics for understanding neurological behaviors.
{"title":"BrainNet: an automated approach for brain stress prediction utilizing electrodermal activity signal with XLNet model.","authors":"Liao Xuanzhi, Abeer Hakeem, Linda Mohaisen, Muhammad Umer, Muhammad Attique Khan, Shrooq Alsenan, Shtwai Alsubai, Nisreen Innab","doi":"10.3389/fncom.2024.1482994","DOIUrl":"https://doi.org/10.3389/fncom.2024.1482994","url":null,"abstract":"<p><p>Brain stress monitoring has emerged as a critical research area for understanding and managing stress and neurological health issues. This burgeoning field aims to provide accurate information and prediction about individuals' stress levels by analyzing behavioral data and physiological signals. To address this emerging problem, this research study proposes an innovative approach that uses an attention mechanism-based XLNet model (called BrainNet) for continuous stress monitoring and stress level prediction. The proposed model analyzes streams of brain data, including behavioral and physiological signal patterns using Swell and WESAD datasets. Testing on the Swell multi-class dataset, the model achieves an impressive accuracy of 95.76%. Furthermore, when evaluated on the WESAD dataset, it demonstrates even higher accuracy, reaching 98.32%. When applied to the binary classification of stress and no stress using the Swell dataset, the model achieves an outstanding accuracy of 97.19%. Comparative analysis with other previously published research studies underscores the superior performance of the proposed approach. In addition, cross-validation confirms the significance, efficacy, and robustness of the model in brain stress level prediction and aligns with the goals of smart diagnostics for understanding neurological behaviors.</p>","PeriodicalId":12363,"journal":{"name":"Frontiers in Computational Neuroscience","volume":"18 ","pages":"1482994"},"PeriodicalIF":2.1,"publicationDate":"2024-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11540660/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142603906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}