Pub Date : 2025-12-01Epub Date: 2025-09-01DOI: 10.1007/s11571-025-10327-w
Bin Zhan, Ziwei Ren, Shuaixia Li, Yiwen Li, Mingming Zhang, Weiqi He
Facial expressions enable individuals to assess and understand emotions conveyed by others. Two crucial sources of expressive cues on the human face-the eyes and the mouth-capture attention and serve as reliable shortcuts for expression recognition. However, how the brain effectively extracts emotional information from these diagnostic features remains unknown. We investigated this issue using an electroencephalogram combined with a rapid serial visual presentation task in which participants were asked to recognize facial expressions (fear, happiness, and neutrality) from three formats (whole face, eye region, and mouth region). We found that participants recognized happy expressions from the mouth region more accurately than the other expressions, affirming the role of diagnostic features in facilitating bottom-up attentional capture. The isolated eye region with higher visual saliency induced the largest P1 component. Diagnostic features, such as a happy mouth and fearful eyes, elicited a larger N170 component compared to non-diagnostic features, such as a fearful mouth and happy eyes. Source analysis of N170 showed that the fusiform gyrus exhibited similar patterns in response to these emotional features. The P3 was effective in discriminating between different emotional content. When whole faces were visible, fearful and happy expressions were not distinguishable in the N170, while the P3 amplitude was larger when induced by fearful faces than by happy faces. Our study contributes to understanding how facial features play distinct roles in emotional perception, attention, and facial processing.
{"title":"Happy mouth and fearful eyes: insights into emotional facial features from ERP.","authors":"Bin Zhan, Ziwei Ren, Shuaixia Li, Yiwen Li, Mingming Zhang, Weiqi He","doi":"10.1007/s11571-025-10327-w","DOIUrl":"10.1007/s11571-025-10327-w","url":null,"abstract":"<p><p>Facial expressions enable individuals to assess and understand emotions conveyed by others. Two crucial sources of expressive cues on the human face-the eyes and the mouth-capture attention and serve as reliable shortcuts for expression recognition. However, how the brain effectively extracts emotional information from these diagnostic features remains unknown. We investigated this issue using an electroencephalogram combined with a rapid serial visual presentation task in which participants were asked to recognize facial expressions (fear, happiness, and neutrality) from three formats (whole face, eye region, and mouth region). We found that participants recognized happy expressions from the mouth region more accurately than the other expressions, affirming the role of diagnostic features in facilitating bottom-up attentional capture. The isolated eye region with higher visual saliency induced the largest P1 component. Diagnostic features, such as a happy mouth and fearful eyes, elicited a larger N170 component compared to non-diagnostic features, such as a fearful mouth and happy eyes. Source analysis of N170 showed that the fusiform gyrus exhibited similar patterns in response to these emotional features. The P3 was effective in discriminating between different emotional content. When whole faces were visible, fearful and happy expressions were not distinguishable in the N170, while the P3 amplitude was larger when induced by fearful faces than by happy faces. Our study contributes to understanding how facial features play distinct roles in emotional perception, attention, and facial processing.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"143"},"PeriodicalIF":3.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12401829/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144991649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-11-04DOI: 10.1007/s11571-025-10360-9
Xiaolong Wu, Dingfu Long, Jianhong Yang
Improving the accuracy of non-invasive brain-computer interface (BCI) and promoting their daily use can be achieved by developing an individualized model training framework, where individual training means that the model is based on small-sample learning from individual data. In the process of data augmentation through synthetic data, the criteria for data generation needs to be further specified according to the requirements. Therefore, in this study, the proposed BCI model utilizes dynamic networks to describe electroencephalogram (EEG) activity during the motor imagery (MI) task, innovatively generates individualized dynamic networks from individual data, and ultimately achieves EEG-controlled grasping through model training. Specifically, this study involves the EEG signals of the right-hand grasping movements of eight subjects and proposes using morphological pattern spectrum (MPS) to encode EEG potentials during MI processes. The MI condition representation was achieved by combining the dynamic networks with MPS encoding, and more dynamic network EEG encoding samples were synthesized through generative adversarial network (GAN) or variational autoencoder (VAE). The AUCs based on the long short-term memory (LSTM) architecture for generating and classifying can be improved by 0.003-0.07. The optimal BCI model based on the Wasserstein GAN and Granger causality (GC) dynamic network encoded by MPS achieved a mean true/false positive rate (TPR/FPR) of 90.0%/0.0%, far better than the 52.9%/4.4% achieved without individualized modeling. Moreover, the BCI establishment of handling multi-task and complex command outputs further demonstrates the reliability of MPS encoding of the GC dynamic network in BCI modeling. The advantage of this "generative-individual" approach is that it not only reduces the sample size requirement while ensuring accuracy but also avoids building models that are applicable to all individuals, which leads to difficult convergence.
{"title":"Generative motor imagery dynamic networks: EEG-controlled grasping via individualized model training.","authors":"Xiaolong Wu, Dingfu Long, Jianhong Yang","doi":"10.1007/s11571-025-10360-9","DOIUrl":"https://doi.org/10.1007/s11571-025-10360-9","url":null,"abstract":"<p><p>Improving the accuracy of non-invasive brain-computer interface (BCI) and promoting their daily use can be achieved by developing an individualized model training framework, where individual training means that the model is based on small-sample learning from individual data. In the process of data augmentation through synthetic data, the criteria for data generation needs to be further specified according to the requirements. Therefore, in this study, the proposed BCI model utilizes dynamic networks to describe electroencephalogram (EEG) activity during the motor imagery (MI) task, innovatively generates individualized dynamic networks from individual data, and ultimately achieves EEG-controlled grasping through model training. Specifically, this study involves the EEG signals of the right-hand grasping movements of eight subjects and proposes using morphological pattern spectrum (MPS) to encode EEG potentials during MI processes. The MI condition representation was achieved by combining the dynamic networks with MPS encoding, and more dynamic network EEG encoding samples were synthesized through generative adversarial network (GAN) or variational autoencoder (VAE). The AUCs based on the long short-term memory (LSTM) architecture for generating and classifying can be improved by 0.003-0.07. The optimal BCI model based on the Wasserstein GAN and Granger causality (GC) dynamic network encoded by MPS achieved a mean true/false positive rate (TPR/FPR) of 90.0%/0.0%, far better than the 52.9%/4.4% achieved without individualized modeling. Moreover, the BCI establishment of handling multi-task and complex command outputs further demonstrates the reliability of MPS encoding of the GC dynamic network in BCI modeling. The advantage of this \"generative-individual\" approach is that it not only reduces the sample size requirement while ensuring accuracy but also avoids building models that are applicable to all individuals, which leads to difficult convergence.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"174"},"PeriodicalIF":3.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12586843/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145457857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-10-03DOI: 10.1007/s11571-025-10333-y
Mohammad Shahed Masud, Danko Nikolić, Liz Stuart, Roman Borisyuk
This study explores graph theory methods for analyzing the functional connectivity of multiple spike trains. We study simultaneously recorded multiple spike trains recorded from the visual cortex of a cat under different visual stimuli. To find the functional connectivity for a given visual stimulus we use the Cox method (Masud and Borisyuk, J Neurosci Methods 196:201-219, 2011). The application of graph theory methods for analysing the connectivity circuit, revealed that the functional connectivity of multiple spike trains is characterized by low density, long communication distances, and weak interconnectivity. Nevertheless, some spike trains also exhibit high degrees of centrality, including betweenness centrality, expansiveness coefficient, and attractiveness coefficient. Additionally, the analysis also identified significant motifs within the functional connections. Thus, our approach allows to describe the correspondence between the stimulus and functional connectivity diagram and compare functional connections under different stimuli.
本研究探索图论方法分析多尖峰列车的功能连通性。我们同时研究了在不同视觉刺激下猫的视觉皮层记录的多个脉冲序列。为了找到给定视觉刺激的功能连接,我们使用Cox方法(Masud and Borisyuk, J Neurosci Methods 196:201-219, 2011)。应用图论方法对连接电路进行分析,发现多尖峰串的功能连接具有密度低、通信距离长、互联性弱的特点。然而,一些尖峰列车也表现出高度的中心性,包括中间中心性、扩张性系数和吸引力系数。此外,分析还发现了功能连接中的重要基序。因此,我们的方法可以描述刺激和功能连接图之间的对应关系,并比较不同刺激下的功能连接。
{"title":"Graph theory methods for analyzing functional connectivity in multiple spike trains: application to data recorded from the visual cortex of a cat.","authors":"Mohammad Shahed Masud, Danko Nikolić, Liz Stuart, Roman Borisyuk","doi":"10.1007/s11571-025-10333-y","DOIUrl":"10.1007/s11571-025-10333-y","url":null,"abstract":"<p><p>This study explores graph theory methods for analyzing the functional connectivity of multiple spike trains. We study simultaneously recorded multiple spike trains recorded from the visual cortex of a cat under different visual stimuli. To find the functional connectivity for a given visual stimulus we use the Cox method (Masud and Borisyuk, J Neurosci Methods 196:201-219, 2011). The application of graph theory methods for analysing the connectivity circuit, revealed that the functional connectivity of multiple spike trains is characterized by low density, long communication distances, and weak interconnectivity. Nevertheless, some spike trains also exhibit high degrees of centrality, including betweenness centrality, expansiveness coefficient, and attractiveness coefficient. Additionally, the analysis also identified significant motifs within the functional connections. Thus, our approach allows to describe the correspondence between the stimulus and functional connectivity diagram and compare functional connections under different stimuli.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"162"},"PeriodicalIF":3.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12495021/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145231601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Motor imagery (MI) is a fundamental paradigm in brain-computer interfaces (BCIs), extensively employed to assist individuals with disabilities to operate external devices. Accurate decoding of MI signals is essential for effective interaction. However, robust decoding remains a challenge due to the inherent complexity and variability of MI-EEG signals. To address this issue, we propose an innovative Dual-Branch Multi-Attention Temporal Convolutional Network (DBMATCN) to improve the performance of MI-EEG signal classification. First, the dual-branch structure extracts rich spatial-temporal features. Then, the channel attention enhances local channel feature extraction and calibrate feature mapping. Next, by combining a sliding window technique and multi-head locality self-attention improves the feature representation of MI-EEG signals by emphasizing the most relevant features. Finally, the temporal convolution fusion network decoding module is used to extensively capture comprehensive temporal features from MI data and carry out the classification task. DBMATCN achieves average accuracies of 88.08%, 96.83%, and 89.71% in inter-session validation on the BCI-IV-2a, HGD, and BCI-IV-2b datasets, respectively. In cross-validation, the model reaches an accuracy of 85.14%, and in the subject-independent scenario, it attains 71.78%. DBMATCN outperforms all baseline models in these cases. These results suggest that our model is effective in decoding MI signals.
{"title":"A dual-branch neural network and attention mechanism for decoding EEG-based motor imagery.","authors":"Yangchuang Wang, Hongli Yu, Xiuzhi Zhao, Xiaozhe Yin, Hongxin Li, Chunfang Wang","doi":"10.1007/s11571-025-10356-5","DOIUrl":"https://doi.org/10.1007/s11571-025-10356-5","url":null,"abstract":"<p><p>Motor imagery (MI) is a fundamental paradigm in brain-computer interfaces (BCIs), extensively employed to assist individuals with disabilities to operate external devices. Accurate decoding of MI signals is essential for effective interaction. However, robust decoding remains a challenge due to the inherent complexity and variability of MI-EEG signals. To address this issue, we propose an innovative Dual-Branch Multi-Attention Temporal Convolutional Network (DBMATCN) to improve the performance of MI-EEG signal classification. First, the dual-branch structure extracts rich spatial-temporal features. Then, the channel attention enhances local channel feature extraction and calibrate feature mapping. Next, by combining a sliding window technique and multi-head locality self-attention improves the feature representation of MI-EEG signals by emphasizing the most relevant features. Finally, the temporal convolution fusion network decoding module is used to extensively capture comprehensive temporal features from MI data and carry out the classification task. DBMATCN achieves average accuracies of 88.08%, 96.83%, and 89.71% in inter-session validation on the BCI-IV-2a, HGD, and BCI-IV-2b datasets, respectively. In cross-validation, the model reaches an accuracy of 85.14%, and in the subject-independent scenario, it attains 71.78%. DBMATCN outperforms all baseline models in these cases. These results suggest that our model is effective in decoding MI signals.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"177"},"PeriodicalIF":3.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12586268/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145457825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-07-23DOI: 10.1007/s11571-025-10301-6
Ziling Wang, Dandan Qian, Songting Li, Wei Lu, Douglas Zhou
Enhanced beta oscillations (12-25 Hz) within the cortico-basal ganglia-thalamic network are significantly associated with motor deficits and are a prominent characteristic of the neural dynamic pathology in Parkinson's disease. Although the striatum has been proposed as a promising origin for enhanced beta oscillations, the precise mechanism through which distinct striatal neurons collaborate to orchestrate beta oscillations remains elusive. This study constructs a biophysical neural network model of the striatum based on experimental constraints. The model faithfully reproduces various experimental observations, including dopamine-dependent beta oscillations and phase-locked firing patterns. Through both theoretical and numerical analysis, our analysis reveals that striatal beta oscillations emerge from interactions within the cellular architecture, particularly the somatostatin-expressing interneurons (SOM) driven choline acetyltransferase-expressing interneurons (ChAT)-indirect pathway striatal projection neurons (iSPN) loop. Our results underscore the critical role of ChATs in enhancing beta oscillations. ChATs, instead of passively providing excitatory drive, actively amplify beta oscillations by enhancing their excitation efficacy through a phase-locked mode. Additionally, the inhibitory interactions among iSPNs, with robust and slow inhibitory recovery dynamics within iSPNs, potentially result in beta oscillations. The slow inhibitory recovery is likely attributed to the slow dynamics of the KCNQ current. SOMs further modulate the beta oscillations by affecting their downstream ChAT-iSPN loop. These results provide novel insights into the mechanism underlying striatal beta oscillations, shedding light on the processes involved in beta oscillations generation during pathological states.
{"title":"Adaptive cholinergic feedback network oscillations: insights into striatal beta oscillations and circuit dynamics.","authors":"Ziling Wang, Dandan Qian, Songting Li, Wei Lu, Douglas Zhou","doi":"10.1007/s11571-025-10301-6","DOIUrl":"10.1007/s11571-025-10301-6","url":null,"abstract":"<p><p>Enhanced beta oscillations (12-25 Hz) within the cortico-basal ganglia-thalamic network are significantly associated with motor deficits and are a prominent characteristic of the neural dynamic pathology in Parkinson's disease. Although the striatum has been proposed as a promising origin for enhanced beta oscillations, the precise mechanism through which distinct striatal neurons collaborate to orchestrate beta oscillations remains elusive. This study constructs a biophysical neural network model of the striatum based on experimental constraints. The model faithfully reproduces various experimental observations, including dopamine-dependent beta oscillations and phase-locked firing patterns. Through both theoretical and numerical analysis, our analysis reveals that striatal beta oscillations emerge from interactions within the cellular architecture, particularly the somatostatin-expressing interneurons (SOM) driven choline acetyltransferase-expressing interneurons (ChAT)-indirect pathway striatal projection neurons (iSPN) loop. Our results underscore the critical role of ChATs in enhancing beta oscillations. ChATs, instead of passively providing excitatory drive, actively amplify beta oscillations by enhancing their excitation efficacy through a phase-locked mode. Additionally, the inhibitory interactions among iSPNs, with robust and slow inhibitory recovery dynamics within iSPNs, potentially result in beta oscillations. The slow inhibitory recovery is likely attributed to the slow dynamics of the KCNQ current. SOMs further modulate the beta oscillations by affecting their downstream ChAT-iSPN loop. These results provide novel insights into the mechanism underlying striatal beta oscillations, shedding light on the processes involved in beta oscillations generation during pathological states.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"117"},"PeriodicalIF":3.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12286910/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144728382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Accurate decoding and strong feature interpretability of Motor Imagery (MI) are expected to drive MI applications in stroke rehabilitation. However, the inherent nonstationarity and high intra-class variability of MI-EEG pose significant challenges in extracting reliable spatio-temporal features. We proposed the Dynamic Spatio-Temporal Feature Augmentation Network (DSTA-Net), which combines DSTA and the Spatio-Temporal Convolution (STC) modules. In DSTA module, multi-scale temporal convolutional kernels tailored to the α and β frequency bands of MI neurophysiological characteristics, while raw EEG serve as a baseline feature layer to retain original information. Next, Grouped Spatial Convolutions extract multi-level spatial features, combined with weight constraints to prevent overfitting. Spatial convolution kernels map EEG channel information into a new spatial domain, enabling further feature extraction through dimensional transformation. And STC module further extracts features and conducts classification. We evaluated DSTA-Net on three public datasets and applied it to a self-collected stroke dataset. In tenfold cross-validation, DSTA-Net achieved average accuracy improvements of 6.29% (p < 0.01), 3.05% (p < 0.01), 5.26% (p < 0.01), and 2.25% over the ShallowConvNet on the BCI-IV-2a, OpenBMI, CASIA, and stroke dataset, respectively. In hold-out validation, DSTA-Net achieved average accuracy improvements of 3.99% (p < 0.01) and 4.2% (p < 0.01) over the ShallowConvNet on the OpenBMI and CASIA datasets, respectively. Finally, we applied DeepLIFT, Common Spatial Pattern, and t-SNE to analyze the contributions of individual EEG channels, extract spatial patterns, and visualize features. The superiority of DSTA-Net offers new insights for further research and application in MI. The code is available in https://github.com/CL-Cloud-BCI/DSTANet-code.
运动意象的准确解码和较强的特征可解释性有望推动运动意象在脑卒中康复中的应用。然而,MI-EEG固有的非平稳性和高类内变异性给提取可靠的时空特征带来了重大挑战。本文提出了动态时空特征增强网络(DSTA- net),该网络将DSTA和时空卷积(STC)模块相结合。在DSTA模块中,根据MI神经生理特征的α和β频段定制多尺度时间卷积核,而原始EEG作为基线特征层保留原始信息。其次,分组空间卷积提取多层次空间特征,结合权约束防止过拟合。空间卷积核将脑电信号通道信息映射到一个新的空间域,通过维度变换进一步提取特征。STC模块进一步提取特征并进行分类。我们在三个公共数据集上评估了DSTA-Net,并将其应用于一个自收集的中风数据集。在十倍交叉验证中,DSTA-Net的准确率平均提高了6.29% (p p p p p p)
{"title":"DSTA-Net: dynamic spatio-temporal feature augmentation network for motor imagery classification.","authors":"Liang Chang, Banghua Yang, Jiayang Zhang, Tie Li, Juntao Feng, Wendong Xu","doi":"10.1007/s11571-025-10296-0","DOIUrl":"10.1007/s11571-025-10296-0","url":null,"abstract":"<p><p>Accurate decoding and strong feature interpretability of Motor Imagery (MI) are expected to drive MI applications in stroke rehabilitation. However, the inherent nonstationarity and high intra-class variability of MI-EEG pose significant challenges in extracting reliable spatio-temporal features. We proposed the Dynamic Spatio-Temporal Feature Augmentation Network (DSTA-Net), which combines DSTA and the Spatio-Temporal Convolution (STC) modules. In DSTA module, multi-scale temporal convolutional kernels tailored to the α and β frequency bands of MI neurophysiological characteristics, while raw EEG serve as a baseline feature layer to retain original information. Next, Grouped Spatial Convolutions extract multi-level spatial features, combined with weight constraints to prevent overfitting. Spatial convolution kernels map EEG channel information into a new spatial domain, enabling further feature extraction through dimensional transformation. And STC module further extracts features and conducts classification. We evaluated DSTA-Net on three public datasets and applied it to a self-collected stroke dataset. In tenfold cross-validation, DSTA-Net achieved average accuracy improvements of 6.29% (<i>p</i> < 0.01), 3.05% (<i>p</i> < 0.01), 5.26% (<i>p</i> < 0.01), and 2.25% over the ShallowConvNet on the BCI-IV-2a, OpenBMI, CASIA, and stroke dataset, respectively. In hold-out validation, DSTA-Net achieved average accuracy improvements of 3.99% (<i>p</i> < 0.01) and 4.2% (<i>p</i> < 0.01) over the ShallowConvNet on the OpenBMI and CASIA datasets, respectively. Finally, we applied DeepLIFT, Common Spatial Pattern, and t-SNE to analyze the contributions of individual EEG channels, extract spatial patterns, and visualize features. The superiority of DSTA-Net offers new insights for further research and application in MI. The code is available in https://github.com/CL-Cloud-BCI/DSTANet-code.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"118"},"PeriodicalIF":3.9,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12286908/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144728383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2024-12-31DOI: 10.1007/s11571-024-10185-y
Hui Wang, Xiaxia Xu, Zhuo Yang, Tao Zhang
Adolescent brain development is characterized by significant anatomical and physiological alterations, but little is known whether and how these alterations impact the neural network. Here we investigated the development of functional networks by measuring synaptic plasticity and neural synchrony of local filed potentials (LFPs), and further explored the underlying mechanisms. LFPs in the hippocampus were recorded in young (21 ~ 25 days), adolescent (1.5 months) and adult (3 months) rats. Long term potentiation (LTP) and neural synchrony were analyzed. The results showed that the LTP was the lowest in adolescent rats. During development, the theta coupling strength was increased progressively but there was no significant change of gamma coupling between young rats and adolescent rats. The density of dendrite spines was decreased progressively during development. The lowest levels of NR2A, NR2B and PSD95 were detected in adolescent rats. Importantly, it was found that the expression levels of autophagy markers were the highest during adolescent compared to that in other developmental stages. Moreover, there were more co-localization of autophagosome and PSD95 in adolescent rats. It suggests that autophagy is possibly involved in synaptic elimination during adolescence, and further impacts synaptic plasticity and neural synchrony.
{"title":"Alterations of synaptic plasticity and brain oscillation are associated with autophagy induced synaptic pruning during adolescence.","authors":"Hui Wang, Xiaxia Xu, Zhuo Yang, Tao Zhang","doi":"10.1007/s11571-024-10185-y","DOIUrl":"10.1007/s11571-024-10185-y","url":null,"abstract":"<p><p>Adolescent brain development is characterized by significant anatomical and physiological alterations, but little is known whether and how these alterations impact the neural network. Here we investigated the development of functional networks by measuring synaptic plasticity and neural synchrony of local filed potentials (LFPs), and further explored the underlying mechanisms. LFPs in the hippocampus were recorded in young (21 ~ 25 days), adolescent (1.5 months) and adult (3 months) rats. Long term potentiation (LTP) and neural synchrony were analyzed. The results showed that the LTP was the lowest in adolescent rats. During development, the theta coupling strength was increased progressively but there was no significant change of gamma coupling between young rats and adolescent rats. The density of dendrite spines was decreased progressively during development. The lowest levels of NR2A, NR2B and PSD95 were detected in adolescent rats. Importantly, it was found that the expression levels of autophagy markers were the highest during adolescent compared to that in other developmental stages. Moreover, there were more co-localization of autophagosome and PSD95 in adolescent rats. It suggests that autophagy is possibly involved in synaptic elimination during adolescence, and further impacts synaptic plasticity and neural synchrony.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"2"},"PeriodicalIF":3.1,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11688264/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142920782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2024-12-31DOI: 10.1007/s11571-024-10182-1
Xudong Zhao, Hualin Wang, Ke Li, Shanguang Chen, Lijuan Hou
Fatigue, a complex and multifaceted symptom, profoundly influences quality of life, particularly among individuals suffering from chronic medical conditions or neurological disorders. This symptom not only exacerbates existing conditions but also hinders daily functioning, thereby perpetuating a vicious cycle of worsening symptoms and reduced physical activity. Given the pivotal role of the motor cortex (M1) in coordinating and executing voluntary movements, understanding how the cortex regulates fatigue is crucial. Despite its importance, the neural mechanisms underlying fatigue remain inadequately explored. In this study, we employed electrophysiological recordings in the M1 region of mice to investigate how excitation-inhibition dynamics and neural oscillations are regulated during exercise-induced fatigue. We observed that fatigue led to decreased voluntary physical activity and cognitive performance, manifesting as reduced running wheel distance, mean speed, exercise intensity, and exploratory behaviour. At the neural level, we detected increased firing frequencies for M1 neurons, including both pyramidal neurons and interneurons, along with heightened beta-band oscillatory activity and stronger coupling between beta-band oscillations and interneurons. These findings enhance our understanding of the mechanisms underlying fatigue, offering insights into behavioural, excitability, and oscillatory changes. The results of this study could pave the way for the development of novel intervention strategies to combat fatigue.
{"title":"Beta-band oscillations and spike-local field potential synchronization in the motor cortex are correlated with movement deficits in an exercise-induced fatigue mouse model.","authors":"Xudong Zhao, Hualin Wang, Ke Li, Shanguang Chen, Lijuan Hou","doi":"10.1007/s11571-024-10182-1","DOIUrl":"10.1007/s11571-024-10182-1","url":null,"abstract":"<p><p>Fatigue, a complex and multifaceted symptom, profoundly influences quality of life, particularly among individuals suffering from chronic medical conditions or neurological disorders. This symptom not only exacerbates existing conditions but also hinders daily functioning, thereby perpetuating a vicious cycle of worsening symptoms and reduced physical activity. Given the pivotal role of the motor cortex (M1) in coordinating and executing voluntary movements, understanding how the cortex regulates fatigue is crucial. Despite its importance, the neural mechanisms underlying fatigue remain inadequately explored. In this study, we employed electrophysiological recordings in the M1 region of mice to investigate how excitation-inhibition dynamics and neural oscillations are regulated during exercise-induced fatigue. We observed that fatigue led to decreased voluntary physical activity and cognitive performance, manifesting as reduced running wheel distance, mean speed, exercise intensity, and exploratory behaviour. At the neural level, we detected increased firing frequencies for M1 neurons, including both pyramidal neurons and interneurons, along with heightened beta-band oscillatory activity and stronger coupling between beta-band oscillations and interneurons. These findings enhance our understanding of the mechanisms underlying fatigue, offering insights into behavioural, excitability, and oscillatory changes. The results of this study could pave the way for the development of novel intervention strategies to combat fatigue.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"3"},"PeriodicalIF":3.1,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11688262/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142920741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-01-03DOI: 10.1007/s11571-024-10196-9
Qiang Meng, Lan Tian, Guoyang Liu, Xue Zhang
Pitch plays an essential role in music perception and forms the fundamental component of melodic interpretation. However, objectively detecting and decoding brain responses to musical pitch perception across subjects remains to be explored. In this study, we employed electroencephalography (EEG) as an objective measure to obtain the neural responses of musical pitch perception. The EEG signals from 34 subjects under hearing violin sounds at pitches G3 and B6 were collected with an efficient passive Go/No-Go paradigm. The lightweight modified EEGNet model was proposed for EEG-based pitch classification. Specifically, within-subject modeling with the modified EEGNet model was performed to construct individually optimized models. Subsequently, based on the within-subject model pool, a classifier ensemble (CE) method was adopted to construct the cross-subject model. Additionally, we analyzed the optimal time window of brain decoding for pitch perception in the EEG data and discussed the interpretability of these models. The experiment results show that the modified EEGNet model achieved an average classification accuracy of 77% for within-subject modeling, significantly outperforming other compared methods. Meanwhile, the proposed CE method achieved an average accuracy of 74% for cross-subject modeling, significantly exceeding the chance-level accuracy of 50%. Furthermore, we found that the optimal EEG data window for the pitch perception lies 0.4 to 0.9 s onset. These promising results demonstrate that the proposed methods can be effectively used in the objective assessment of pitch perception and have generalization ability in cross-subject modeling.
{"title":"EEG-based cross-subject passive music pitch perception using deep learning models.","authors":"Qiang Meng, Lan Tian, Guoyang Liu, Xue Zhang","doi":"10.1007/s11571-024-10196-9","DOIUrl":"10.1007/s11571-024-10196-9","url":null,"abstract":"<p><p>Pitch plays an essential role in music perception and forms the fundamental component of melodic interpretation. However, objectively detecting and decoding brain responses to musical pitch perception across subjects remains to be explored. In this study, we employed electroencephalography (EEG) as an objective measure to obtain the neural responses of musical pitch perception. The EEG signals from 34 subjects under hearing violin sounds at pitches G3 and B6 were collected with an efficient passive Go/No-Go paradigm. The lightweight modified EEGNet model was proposed for EEG-based pitch classification. Specifically, within-subject modeling with the modified EEGNet model was performed to construct individually optimized models. Subsequently, based on the within-subject model pool, a classifier ensemble (CE) method was adopted to construct the cross-subject model. Additionally, we analyzed the optimal time window of brain decoding for pitch perception in the EEG data and discussed the interpretability of these models. The experiment results show that the modified EEGNet model achieved an average classification accuracy of 77% for within-subject modeling, significantly outperforming other compared methods. Meanwhile, the proposed CE method achieved an average accuracy of 74% for cross-subject modeling, significantly exceeding the chance-level accuracy of 50%. Furthermore, we found that the optimal EEG data window for the pitch perception lies 0.4 to 0.9 s onset. These promising results demonstrate that the proposed methods can be effectively used in the objective assessment of pitch perception and have generalization ability in cross-subject modeling.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"6"},"PeriodicalIF":3.1,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11699146/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142930820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-01-09DOI: 10.1007/s11571-024-10184-z
Qiang Li
In recent years, substantial strides have been made in the field of visual image reconstruction, particularly in its capacity to generate high-quality visual representations from human brain activity while considering semantic information. This advancement not only enables the recreation of visual content but also provides valuable insights into the intricate processes occurring within high-order functional brain regions, contributing to a deeper understanding of brain function. However, considering fusion semantics in reconstructing visual images from brain activity involves semantic-to-image guide reconstruction and may ignore underlying neural computational mechanisms, which does not represent true reconstruction from brain activity. In response to this limitation, our study introduces a novel approach that combines linear mapping with nonlinear noise suppression to reconstruct visual images perceived by subjects based on their brain activity patterns. The primary challenge associated with linear mapping lies in its susceptibility to noise interference. To address this issue, we leverage a flexible denoised deep convolutional neural network, which can suppress noise from linear mapping. Our investigation encompasses linear mapping as well as the training of shallow and deep autoencoder denoised neural networks, including a pre-trained, state-of-the-art denoised neural network. The outcome of our study reveals that combining linear image decoding with nonlinear noise reduction significantly enhances the quality of reconstructed images from human brain activity. This suggests that our methodology holds promise for decoding intricate perceptual experiences directly from brain activity patterns without semantic information. Moreover, the model has strong neural explanatory power because it shares structural and functional similarities with the visual brain.
{"title":"Visual image reconstructed without semantics from human brain activity using linear image decoders and nonlinear noise suppression.","authors":"Qiang Li","doi":"10.1007/s11571-024-10184-z","DOIUrl":"10.1007/s11571-024-10184-z","url":null,"abstract":"<p><p>In recent years, substantial strides have been made in the field of visual image reconstruction, particularly in its capacity to generate high-quality visual representations from human brain activity while considering semantic information. This advancement not only enables the recreation of visual content but also provides valuable insights into the intricate processes occurring within high-order functional brain regions, contributing to a deeper understanding of brain function. However, considering fusion semantics in reconstructing visual images from brain activity involves semantic-to-image guide reconstruction and may ignore underlying neural computational mechanisms, which does not represent true reconstruction from brain activity. In response to this limitation, our study introduces a novel approach that combines linear mapping with nonlinear noise suppression to reconstruct visual images perceived by subjects based on their brain activity patterns. The primary challenge associated with linear mapping lies in its susceptibility to noise interference. To address this issue, we leverage a flexible denoised deep convolutional neural network, which can suppress noise from linear mapping. Our investigation encompasses linear mapping as well as the training of shallow and deep autoencoder denoised neural networks, including a pre-trained, state-of-the-art denoised neural network. The outcome of our study reveals that combining linear image decoding with nonlinear noise reduction significantly enhances the quality of reconstructed images from human brain activity. This suggests that our methodology holds promise for decoding intricate perceptual experiences directly from brain activity patterns without semantic information. Moreover, the model has strong neural explanatory power because it shares structural and functional similarities with the visual brain.</p>","PeriodicalId":10500,"journal":{"name":"Cognitive Neurodynamics","volume":"19 1","pages":"20"},"PeriodicalIF":3.1,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11718044/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142969971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}