首页 > 最新文献

Neural Networks最新文献

英文 中文
StochCA: A novel approach for exploiting pretrained models with cross-attention StochCA:利用交叉关注预训练模型的新方法
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-23 DOI: 10.1016/j.neunet.2024.106663

Utilizing large-scale pretrained models is a well-known strategy to enhance performance on various target tasks. It is typically achieved through fine-tuning pretrained models on target tasks. However, naï ve fine-tuning may not fully leverage knowledge embedded in pretrained models. In this study, we introduce a novel fine-tuning method, called stochastic cross-attention (StochCA), specific to Transformer architectures. This method modifies the Transformer’s self-attention mechanism to selectively utilize knowledge from pretrained models during fine-tuning. Specifically, in each block, instead of self-attention, cross-attention is performed stochastically according to the predefined probability, where keys and values are extracted from the corresponding block of a pretrained model. By doing so, queries and channel-mixing multi-layer perceptron layers of a target model are fine-tuned to target tasks to learn how to effectively exploit rich representations of pretrained models. To verify the effectiveness of StochCA, extensive experiments are conducted on benchmarks in the areas of transfer learning and domain generalization, where the exploitation of pretrained models is critical. Our experimental results show the superiority of StochCA over state-of-the-art approaches in both areas. Furthermore, we demonstrate that StochCA is complementary to existing approaches, i.e., it can be combined with them to further improve performance. We release the code at https://github.com/daintlab/stochastic_cross_attention.

利用大规模预训练模型是一种众所周知的提高各种目标任务性能的策略。它通常是通过在目标任务上对预训练模型进行微调来实现的。然而,简单的微调可能无法充分利用预训练模型中蕴含的知识。在本研究中,我们引入了一种新颖的微调方法,称为随机交叉注意(StochCA),专门针对 Transformer 架构。这种方法修改了 Transformer 的自我注意机制,以便在微调过程中选择性地利用来自预训练模型的知识。具体来说,在每个区块中,根据预定义的概率随机执行交叉注意,而不是自注意,其中的键和值是从预训练模型的相应区块中提取的。这样,目标模型的查询和通道混合多层感知器层就能根据目标任务进行微调,从而学会如何有效利用预训练模型的丰富表征。为了验证StochCA的有效性,我们在迁移学习和领域泛化领域的基准上进行了大量实验,在这些领域,利用预训练模型至关重要。实验结果表明,StochCA 在这两个领域都优于最先进的方法。此外,我们还证明了 StochCA 与现有方法的互补性,也就是说,它可以与现有方法相结合,进一步提高性能。我们在 https://github.com/daintlab/stochastic_cross_attention 上发布了代码。
{"title":"StochCA: A novel approach for exploiting pretrained models with cross-attention","authors":"","doi":"10.1016/j.neunet.2024.106663","DOIUrl":"10.1016/j.neunet.2024.106663","url":null,"abstract":"<div><p>Utilizing large-scale pretrained models is a well-known strategy to enhance performance on various target tasks. It is typically achieved through fine-tuning pretrained models on target tasks. However, naï ve fine-tuning may not fully leverage knowledge embedded in pretrained models. In this study, we introduce a novel fine-tuning method, called stochastic cross-attention (StochCA), specific to Transformer architectures. This method modifies the Transformer’s self-attention mechanism to selectively utilize knowledge from pretrained models during fine-tuning. Specifically, in each block, instead of self-attention, cross-attention is performed stochastically according to the predefined probability, where keys and values are extracted from the corresponding block of a pretrained model. By doing so, queries and channel-mixing multi-layer perceptron layers of a target model are fine-tuned to target tasks to learn how to effectively exploit rich representations of pretrained models. To verify the effectiveness of StochCA, extensive experiments are conducted on benchmarks in the areas of transfer learning and domain generalization, where the exploitation of pretrained models is critical. Our experimental results show the superiority of StochCA over state-of-the-art approaches in both areas. Furthermore, we demonstrate that StochCA is complementary to existing approaches, i.e., it can be combined with them to further improve performance. We release the code at <span><span>https://github.com/daintlab/stochastic_cross_attention</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Inter-participant transfer learning with attention based domain adversarial training for P300 detection 利用基于注意力的领域对抗训练进行 P300 检测的参与者间迁移学习
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106655

A Brain-computer interface (BCI) system establishes a novel communication channel between the human brain and a computer. Most event related potential-based BCI applications make use of decoding models, which requires training. This training process is often time-consuming and inconvenient for new users. In recent years, deep learning models, especially participant-independent models, have garnered significant attention in the domain of ERP classification. However, individual differences in EEG signals hamper model generalization, as the ERP component and other aspects of the EEG signal vary across participants, even when they are exposed to the same stimuli. This paper proposes a novel One-source domain transfer learning method based Attention Domain Adversarial Neural Network (OADANN) to mitigate data distribution discrepancies for cross-participant classification tasks. We train and validate our proposed model on both a publicly available OpenBMI dataset and a Self-collected dataset, employing a leave one participant out cross validation scheme. Experimental results demonstrate that the proposed OADANN method achieves the highest and most robust classification performance and exhibits significant improvements when compared to baseline methods (CNN, EEGNet, ShallowNet, DeepCovNet) and domain generalization methods (ERM, Mixup, and Groupdro). These findings underscore the efficacy of our proposed method.

脑机接口(BCI)系统在人脑和计算机之间建立了一种新型通信渠道。大多数基于事件相关电位的脑机接口应用都使用解码模型,这需要培训。对于新用户来说,这种训练过程往往既耗时又不方便。近年来,深度学习模型,尤其是与参与者无关的模型,在 ERP 分类领域备受关注。然而,脑电信号的个体差异阻碍了模型的泛化,因为不同参与者的ERP成分和脑电信号的其他方面各不相同,即使他们暴露在相同的刺激下也是如此。本文提出了一种基于注意力域对抗神经网络(OADANN)的新型单源域转移学习方法,以减轻跨参与者分类任务的数据分布差异。我们在公开的 OpenBMI 数据集和自选数据集上训练和验证了我们提出的模型,并采用了排除一个参与者的交叉验证方案。实验结果表明,与基线方法(CNN、EEGNet、ShallowNet、DeepCovNet)和领域泛化方法(ERM、Mixup 和 Groupdro)相比,所提出的 OADANN 方法实现了最高和最稳健的分类性能,并表现出显著的改进。这些发现凸显了我们提出的方法的功效。
{"title":"Inter-participant transfer learning with attention based domain adversarial training for P300 detection","authors":"","doi":"10.1016/j.neunet.2024.106655","DOIUrl":"10.1016/j.neunet.2024.106655","url":null,"abstract":"<div><p>A Brain-computer interface (BCI) system establishes a novel communication channel between the human brain and a computer. Most event related potential-based BCI applications make use of decoding models, which requires training. This training process is often time-consuming and inconvenient for new users. In recent years, deep learning models, especially participant-independent models, have garnered significant attention in the domain of ERP classification. However, individual differences in EEG signals hamper model generalization, as the ERP component and other aspects of the EEG signal vary across participants, even when they are exposed to the same stimuli. This paper proposes a novel One-source domain transfer learning method based Attention Domain Adversarial Neural Network (OADANN) to mitigate data distribution discrepancies for cross-participant classification tasks. We train and validate our proposed model on both a publicly available OpenBMI dataset and a Self-collected dataset, employing a leave one participant out cross validation scheme. Experimental results demonstrate that the proposed OADANN method achieves the highest and most robust classification performance and exhibits significant improvements when compared to baseline methods (CNN, EEGNet, ShallowNet, DeepCovNet) and domain generalization methods (ERM, Mixup, and Groupdro). These findings underscore the efficacy of our proposed method.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142122537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reconstruct incomplete relation for incomplete modality brain tumor segmentation 为不完整模态脑肿瘤分割重建不完整关系。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106657

Different brain tumor magnetic resonance imaging (MRI) modalities provide diverse tumor-specific information. Previous works have enhanced brain tumor segmentation performance by integrating multiple MRI modalities. However, multi-modal MRI data are often unavailable in clinical practice. An incomplete modality leads to missing tumor-specific information, which degrades the performance of existing models. Various strategies have been proposed to transfer knowledge from a full modality network (teacher) to an incomplete modality one (student) to address this issue. However, they neglect the fact that brain tumor segmentation is a structural prediction problem that requires voxel semantic relations. In this paper, we propose a Reconstruct Incomplete Relation Network (RIRN) that transfers voxel semantic relational knowledge from the teacher to the student. Specifically, we propose two types of voxel relations to incorporate structural knowledge: Class-relative relations (CRR) and Class-agnostic relations (CAR). The CRR groups voxels into different tumor regions and constructs a relation between them. The CAR builds a global relation between all voxel features, complementing the local inter-region relation. Moreover, we use adversarial learning to align the holistic structural prediction between the teacher and the student. Extensive experimentation on both the BraTS 2018 and BraTS 2020 datasets establishes that our method outperforms all state-of-the-art approaches.

不同的脑肿瘤磁共振成像(MRI)模式可提供不同的肿瘤特异性信息。以往的研究通过整合多种磁共振成像模式提高了脑肿瘤的分割性能。然而,临床实践中往往无法获得多模态磁共振成像数据。不完整的模式会导致肿瘤特异性信息缺失,从而降低现有模型的性能。为了解决这个问题,人们提出了各种策略,将知识从完整模态网络(教师)转移到不完整模态网络(学生)。然而,它们忽略了脑肿瘤分割是一个结构预测问题,需要体素语义关系。在本文中,我们提出了一种重建不完整关系网络(RIRN),将体素语义关系知识从教师转移到学生。具体来说,我们提出了两种类型的体素关系,以纳入结构知识:类相关关系(CRR)和类无关关系(CAR)。CRR 将体素分为不同的肿瘤区域,并构建它们之间的关系。CAR在所有体素特征之间建立一种全局关系,对局部区域间关系进行补充。此外,我们还利用对抗学习来调整教师和学生之间的整体结构预测。在 BraTS 2018 和 BraTS 2020 数据集上进行的广泛实验证明,我们的方法优于所有最先进的方法。
{"title":"Reconstruct incomplete relation for incomplete modality brain tumor segmentation","authors":"","doi":"10.1016/j.neunet.2024.106657","DOIUrl":"10.1016/j.neunet.2024.106657","url":null,"abstract":"<div><p>Different brain tumor magnetic resonance imaging (MRI) modalities provide diverse tumor-specific information. Previous works have enhanced brain tumor segmentation performance by integrating multiple MRI modalities. However, multi-modal MRI data are often unavailable in clinical practice. An incomplete modality leads to missing tumor-specific information, which degrades the performance of existing models. Various strategies have been proposed to transfer knowledge from a full modality network (teacher) to an incomplete modality one (student) to address this issue. However, they neglect the fact that brain tumor segmentation is a structural prediction problem that requires voxel semantic relations. In this paper, we propose a Reconstruct Incomplete Relation Network (RIRN) that transfers voxel semantic relational knowledge from the teacher to the student. Specifically, we propose two types of voxel relations to incorporate structural knowledge: Class-relative relations (CRR) and Class-agnostic relations (CAR). The CRR groups voxels into different tumor regions and constructs a relation between them. The CAR builds a global relation between all voxel features, complementing the local inter-region relation. Moreover, we use adversarial learning to align the holistic structural prediction between the teacher and the student. Extensive experimentation on both the BraTS 2018 and BraTS 2020 datasets establishes that our method outperforms all state-of-the-art approaches.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142074383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A new hybrid learning control system for robots based on spiking neural networks 基于尖峰神经网络的新型机器人混合学习控制系统
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106656

This paper presents a new hybrid learning and control method that can tune their parameters based on reinforcement learning. In the new proposed method, nonlinear controllers are considered multi-input multi-output functions and then the functions are replaced with SNNs with reinforcement learning algorithms. Dopamine-modulated spike-timing-dependent plasticity (STDP) is used for reinforcement learning and manipulating the synaptic weights between the input and output of neuronal groups (for parameter adjustment). Details of the method are presented and some case studies are done on nonlinear controllers such as Fractional Order PID (FOPID) and Feedback Linearization. The structure and the dynamic equations for learning are presented, and the proposed algorithm is tested on robots and results are compared with other works. Moreover, to demonstrate the effectiveness of SNNFOPID, we conducted rigorous testing on a variety of systems including a two-wheel mobile robot, a double inverted pendulum, and a four-link manipulator robot. The results revealed impressively low errors of 0.01 m, 0.03 rad, and 0.03 rad for each system, respectively. The method is tested on another controller named Feedback Linearization, which provides acceptable results. Results show that the new method has better performance in terms of Integral Absolute Error (IAE) and is highly useful in hardware implementation due to its low energy consumption, high speed, and accuracy. The duration necessary for achieving full and stable proficiency in the control of various robotic systems using SNNFOPD, and SNNFL on an Asus Core i5 system within Simulink’s Simscape environment is as follows:

– Two-link robot manipulator with SNNFOPID: 19.85656 hours

– Two-link robot manipulator with SNNFL: 0.45828 hours

– Double inverted pendulum with SNNFOPID: 3.455 hours

– Mobile robot with SNNFOPID: 3.71948 hours

– Four-link robot manipulator with SNNFOPID: 16.6789 hours.

This method can be generalized to other controllers and systems like robots.

本文提出了一种新的混合学习和控制方法,该方法可以在强化学习的基础上调整参数。在新提出的方法中,非线性控制器被视为多输入多输出函数,然后通过强化学习算法用 SNNs 替代这些函数。多巴胺调节的尖峰计时可塑性(STDP)被用于强化学习和操纵神经元组输入和输出之间的突触权重(用于参数调整)。文中介绍了该方法的细节,并对分数阶 PID (FOPID) 和反馈线性化等非线性控制器进行了案例研究。介绍了学习的结构和动态方程,在机器人上测试了所提出的算法,并将结果与其他著作进行了比较。此外,为了证明 SNNFOPID 的有效性,我们对各种系统进行了严格测试,包括双轮移动机器人、双倒立摆和四连杆操纵器机器人。结果显示,每个系统的误差分别为 0.01 m、0.03 rad 和 0.03 rad,低得令人印象深刻。该方法在另一个名为 "反馈线性化 "的控制器上进行了测试,结果可以接受。结果表明,新方法在绝对整数误差(IAE)方面具有更好的性能,而且能耗低、速度快、精度高,非常适合硬件实施。在 Simulink 的 Simscape 环境中,在华硕酷睿 i5 系统上使用 SNNFOPD 和 SNNFL 对各种机器人系统进行完全、稳定的熟练控制所需的时间如下:- 使用 SNNFOPID 的双链路机器人机械手:19.85656 小时- 使用 SNNFL 的双链路机器人机械手:0.45828 小时- 使用 SNNFOPID 的双链路机器人机械手:0.45828 小时- 使用 SNNFL 的双链路机器人机械手:0.45828 小时0.45828 小时- 采用 SNNFOPID 的双倒立摆:3.455 小时- 采用 SNNFOPID 的移动机器人:3.71948 小时- 采用 SNNFOPID 的四连杆机器人机械手:16.6789 小时。
{"title":"A new hybrid learning control system for robots based on spiking neural networks","authors":"","doi":"10.1016/j.neunet.2024.106656","DOIUrl":"10.1016/j.neunet.2024.106656","url":null,"abstract":"<div><p>This paper presents a new hybrid learning and control method that can tune their parameters based on reinforcement learning. In the new proposed method, nonlinear controllers are considered multi-input multi-output functions and then the functions are replaced with SNNs with reinforcement learning algorithms. Dopamine-modulated spike-timing-dependent plasticity (STDP) is used for reinforcement learning and manipulating the synaptic weights between the input and output of neuronal groups (for parameter adjustment). Details of the method are presented and some case studies are done on nonlinear controllers such as Fractional Order PID (FOPID) and Feedback Linearization. The structure and the dynamic equations for learning are presented, and the proposed algorithm is tested on robots and results are compared with other works. Moreover, to demonstrate the effectiveness of SNNFOPID, we conducted rigorous testing on a variety of systems including a two-wheel mobile robot, a double inverted pendulum, and a four-link manipulator robot. The results revealed impressively low errors of 0.01 m, 0.03 rad, and 0.03 rad for each system, respectively. The method is tested on another controller named Feedback Linearization, which provides acceptable results. Results show that the new method has better performance in terms of Integral Absolute Error (IAE) and is highly useful in hardware implementation due to its low energy consumption, high speed, and accuracy. The duration necessary for achieving full and stable proficiency in the control of various robotic systems using SNNFOPD, and SNNFL on an Asus Core i5 system within Simulink’s Simscape environment is as follows:</p><p>– Two-link robot manipulator with SNNFOPID: 19.85656 hours</p><p>– Two-link robot manipulator with SNNFL: 0.45828 hours</p><p>– Double inverted pendulum with SNNFOPID: 3.455 hours</p><p>– Mobile robot with SNNFOPID: 3.71948 hours</p><p>– Four-link robot manipulator with SNNFOPID: 16.6789 hours.</p><p>This method can be generalized to other controllers and systems like robots.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S089360802400580X/pdfft?md5=52c30df5a4230d66467ce3fcee2b3a54&pid=1-s2.0-S089360802400580X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-based stackable graph convolutional network for multi-view learning 用于多视角学习的基于注意力的可堆叠图卷积网络
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106648

In multi-view learning, graph-based methods like Graph Convolutional Network (GCN) are extensively researched due to effective graph processing capabilities. However, most GCN-based methods often require complex preliminary operations such as sparsification, which may bring additional computation costs and training difficulties. Additionally, as the number of stacking layers increases in most GCN, over-smoothing problem arises, resulting in ineffective utilization of GCN capabilities. In this paper, we propose an attention-based stackable graph convolutional network that captures consistency across views and combines attention mechanism to exploit the powerful aggregation capability of GCN to effectively mitigate over-smoothing. Specifically, we introduce node self-attention to establish dynamic connections between nodes and generate view-specific representations. To maintain cross-view consistency, a data-driven approach is devised to assign attention weights to views, forming a common representation. Finally, based on residual connectivity, we apply an attention mechanism to the original projection features to generate layer-specific complementarity, which compensates for the information loss during graph convolution. Comprehensive experimental results demonstrate that the proposed method outperforms other state-of-the-art methods in multi-view semi-supervised tasks.

在多视图学习中,基于图的方法(如图卷积网络(GCN))因其有效的图处理能力而被广泛研究。然而,大多数基于 GCN 的方法往往需要复杂的初步操作,如稀疏化,这可能会带来额外的计算成本和训练困难。此外,随着大多数 GCN 堆叠层数的增加,会出现过度平滑问题,导致无法有效利用 GCN 的功能。在本文中,我们提出了一种基于注意力的可堆叠图卷积网络,它能捕捉不同视图之间的一致性,并结合注意力机制,利用 GCN 强大的聚合能力来有效缓解过平滑问题。具体来说,我们引入了节点自注意力来建立节点之间的动态连接,并生成特定视图的表示。为了保持跨视图的一致性,我们设计了一种数据驱动的方法来为视图分配注意力权重,从而形成一个共同的表征。最后,基于残余连接性,我们将注意力机制应用于原始投影特征,生成特定层的互补性,从而弥补图卷积过程中的信息损失。综合实验结果表明,在多视图半监督任务中,所提出的方法优于其他最先进的方法。
{"title":"Attention-based stackable graph convolutional network for multi-view learning","authors":"","doi":"10.1016/j.neunet.2024.106648","DOIUrl":"10.1016/j.neunet.2024.106648","url":null,"abstract":"<div><p>In multi-view learning, graph-based methods like Graph Convolutional Network (GCN) are extensively researched due to effective graph processing capabilities. However, most GCN-based methods often require complex preliminary operations such as sparsification, which may bring additional computation costs and training difficulties. Additionally, as the number of stacking layers increases in most GCN, over-smoothing problem arises, resulting in ineffective utilization of GCN capabilities. In this paper, we propose an attention-based stackable graph convolutional network that captures consistency across views and combines attention mechanism to exploit the powerful aggregation capability of GCN to effectively mitigate over-smoothing. Specifically, we introduce node self-attention to establish dynamic connections between nodes and generate view-specific representations. To maintain cross-view consistency, a data-driven approach is devised to assign attention weights to views, forming a common representation. Finally, based on residual connectivity, we apply an attention mechanism to the original projection features to generate layer-specific complementarity, which compensates for the information loss during graph convolution. Comprehensive experimental results demonstrate that the proposed method outperforms other state-of-the-art methods in multi-view semi-supervised tasks.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SFT-SGAT: A semi-supervised fine-tuning self-supervised graph attention network for emotion recognition and consciousness detection SFT-SGAT:用于情感识别和意识检测的半监督微调自监督图注意力网络
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106643

Emotional recognition is highly important in the field of brain-computer interfaces (BCIs). However, due to the individual variability in electroencephalogram (EEG) signals and the challenges in obtaining accurate emotional labels, traditional methods have shown poor performance in cross-subject emotion recognition. In this study, we propose a cross-subject EEG emotion recognition method based on a semi-supervised fine-tuning self-supervised graph attention network (SFT-SGAT). First, we model multi-channel EEG signals by constructing a graph structure that dynamically captures the spatiotemporal topological features of EEG signals. Second, we employ a self-supervised graph attention neural network to facilitate model training, mitigating the impact of signal noise on the model. Finally, a semi-supervised approach is used to fine-tune the model, enhancing its generalization ability in cross-subject classification. By combining supervised and unsupervised learning techniques, the SFT-SGAT maximizes the utility of limited labeled data in EEG emotion recognition tasks, thereby enhancing the model’s performance. Experiments based on leave-one-subject-out cross-validation demonstrate that SFT-SGAT achieves state-of-the-art cross-subject emotion recognition performance on the SEED and SEED-IV datasets, with accuracies of 92.04% and 82.76%, respectively. Furthermore, experiments conducted on a self-collected dataset comprising ten healthy subjects and eight patients with disorders of consciousness (DOCs) revealed that the SFT-SGAT attains high classification performance in healthy subjects (maximum accuracy of 95.84%) and was successfully applied to DOC patients, with four patients achieving emotion recognition accuracies exceeding 60%. The experiments demonstrate the effectiveness of the proposed SFT-SGAT model in cross-subject EEG emotion recognition and its potential for assessing levels of consciousness in patients with DOC.

情绪识别在脑机接口(BCI)领域非常重要。然而,由于脑电图(EEG)信号的个体差异性和获取准确情感标签的挑战,传统方法在跨主体情感识别方面表现不佳。在本研究中,我们提出了一种基于半监督微调自监督图注意网络(SFT-SGAT)的跨主体脑电图情感识别方法。首先,我们通过构建能动态捕捉脑电信号时空拓扑特征的图结构来建立多通道脑电信号模型。其次,我们采用自监督图注意神经网络来促进模型训练,减轻信号噪声对模型的影响。最后,我们采用半监督方法对模型进行微调,从而增强其在跨主体分类中的泛化能力。通过结合监督和非监督学习技术,SFT-SGAT 在脑电图情感识别任务中最大限度地利用了有限的标记数据,从而提高了模型的性能。基于留一主体交叉验证的实验表明,SFT-SGAT 在 SEED 和 SEED-IV 数据集上实现了最先进的跨主体情感识别性能,准确率分别为 92.04% 和 82.76%。此外,在一个由 10 名健康受试者和 8 名意识障碍(DOCs)患者组成的自收集数据集上进行的实验表明,SFT-SGAT 在健康受试者身上获得了很高的分类性能(最高准确率为 95.84%),并成功地应用于 DOC 患者身上,其中 4 名患者的情绪识别准确率超过了 60%。实验证明了所提出的 SFT-SGAT 模型在跨受试者脑电图情绪识别方面的有效性,以及其在评估 DOC 患者意识水平方面的潜力。
{"title":"SFT-SGAT: A semi-supervised fine-tuning self-supervised graph attention network for emotion recognition and consciousness detection","authors":"","doi":"10.1016/j.neunet.2024.106643","DOIUrl":"10.1016/j.neunet.2024.106643","url":null,"abstract":"<div><p>Emotional recognition is highly important in the field of brain-computer interfaces (BCIs). However, due to the individual variability in electroencephalogram (EEG) signals and the challenges in obtaining accurate emotional labels, traditional methods have shown poor performance in cross-subject emotion recognition. In this study, we propose a cross-subject EEG emotion recognition method based on a semi-supervised fine-tuning self-supervised graph attention network (SFT-SGAT). First, we model multi-channel EEG signals by constructing a graph structure that dynamically captures the spatiotemporal topological features of EEG signals. Second, we employ a self-supervised graph attention neural network to facilitate model training, mitigating the impact of signal noise on the model. Finally, a semi-supervised approach is used to fine-tune the model, enhancing its generalization ability in cross-subject classification. By combining supervised and unsupervised learning techniques, the SFT-SGAT maximizes the utility of limited labeled data in EEG emotion recognition tasks, thereby enhancing the model’s performance. Experiments based on leave-one-subject-out cross-validation demonstrate that SFT-SGAT achieves state-of-the-art cross-subject emotion recognition performance on the SEED and SEED-IV datasets, with accuracies of 92.04% and 82.76%, respectively. Furthermore, experiments conducted on a self-collected dataset comprising ten healthy subjects and eight patients with disorders of consciousness (DOCs) revealed that the SFT-SGAT attains high classification performance in healthy subjects (maximum accuracy of 95.84%) and was successfully applied to DOC patients, with four patients achieving emotion recognition accuracies exceeding 60%. The experiments demonstrate the effectiveness of the proposed SFT-SGAT model in cross-subject EEG emotion recognition and its potential for assessing levels of consciousness in patients with DOC.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142058046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DiagSWin: A multi-scale vision transformer with diagonal-shaped windows for object detection and segmentation DiagSWin:多尺度视觉转换器,带对角线形窗口,用于物体检测和分割
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106653

Recently, Vision Transformer and its variants have demonstrated remarkable performance on various computer vision tasks, thanks to its competence in capturing global visual dependencies through self-attention. However, global self-attention suffers from high computational cost due to quadratic computational overhead, especially for the high-resolution vision tasks (e.g., object detection and semantic segmentation). Many recent works have attempted to reduce the cost by applying fine-grained local attention, but these approaches cripple the long-range modeling power of the original self-attention mechanism. Furthermore, these approaches usually have similar receptive fields within each layer, thus limiting the ability of each self-attention layer to capture multi-scale features, resulting in performance degradation when handling images with objects of different scales. To address these issues, we develop the Diagonal-shaped Window (DiagSWin) attention mechanism for modeling attentions in diagonal regions at hybrid scales per attention layer. The key idea of DiagSWin attention is to inject multi-scale receptive field sizes into tokens: before computing the self-attention matrix, each token attends its closest surrounding tokens at fine granularity and the tokens far away at coarse granularity. This mechanism is able to effectively capture multi-scale context information while reducing computational complexity. With DiagSwin attention, we present a new variant of Vision Transformer models, called DiagSWin Transformers, and demonstrate their superiority in extensive experiments across various tasks. Specifically, the DiagSwin Transformer with a large size achieves 84.4% Top-1 accuracy and outperforms the SOTA CSWin Transformer on ImageNet with 40% fewer model size and computation cost. When employed as backbones, DiagSWin Transformers achieve significant improvements over the current SOTA modules. In addition, our DiagSWin-Base model yields 51.1 box mAP and 45.8 mask mAP on COCO for object detection and segmentation, and 52.3 mIoU on the ADE20K for semantic segmentation.

最近,Vision Transformer 及其变体通过自我注意捕捉全局视觉依赖关系的能力,在各种计算机视觉任务中表现出了不俗的性能。然而,全局自我注意因二次计算开销而导致计算成本过高,尤其是在高分辨率视觉任务(如物体检测和语义分割)中。最近的许多研究都试图通过应用细粒度的局部注意来降低成本,但这些方法削弱了原始自我注意机制的长程建模能力。此外,这些方法通常在每个层内都有相似的感受野,从而限制了每个自我注意层捕捉多尺度特征的能力,导致在处理具有不同尺度物体的图像时性能下降。为了解决这些问题,我们开发了对角线形窗口(DiagSWin)注意力机制,用于在每个注意力层的混合尺度对角线区域建立注意力模型。DiagSWin 注意力的关键理念是向标记注入多尺度感受野大小:在计算自我注意矩阵之前,每个标记以细粒度注意其周围最近的标记,以粗粒度注意其周围较远的标记。这种机制能够有效捕捉多尺度上下文信息,同时降低计算复杂度。借助 DiagSwin 注意力,我们提出了 Vision Transformer 模型的新变体,称为 DiagSWin Transformers,并在各种任务的广泛实验中证明了其优越性。具体而言,DiagSwin Transformer 的尺寸较大,在 ImageNet 上达到了 84.4% 的 Top-1 准确率,在模型尺寸和计算成本上都比 SOTA CSWin Transformer 少 40%。与当前的 SOTA 模块相比,DiagSWin 转换器在作为骨干模块使用时取得了显著的改进。此外,我们的 DiagSWin-Base 模型在 COCO 上的物体检测和分割中产生了 51.1 box mAP 和 45.8 mask mAP,在 ADE20K 上的语义分割中产生了 52.3 mIoU。
{"title":"DiagSWin: A multi-scale vision transformer with diagonal-shaped windows for object detection and segmentation","authors":"","doi":"10.1016/j.neunet.2024.106653","DOIUrl":"10.1016/j.neunet.2024.106653","url":null,"abstract":"<div><p>Recently, Vision Transformer and its variants have demonstrated remarkable performance on various computer vision tasks, thanks to its competence in capturing global visual dependencies through self-attention. However, global self-attention suffers from high computational cost due to quadratic computational overhead, especially for the high-resolution vision tasks (e.g., object detection and semantic segmentation). Many recent works have attempted to reduce the cost by applying fine-grained local attention, but these approaches cripple the long-range modeling power of the original self-attention mechanism. Furthermore, these approaches usually have similar receptive fields within each layer, thus limiting the ability of each self-attention layer to capture multi-scale features, resulting in performance degradation when handling images with objects of different scales. To address these issues, we develop the Diagonal-shaped Window (DiagSWin) attention mechanism for modeling attentions in diagonal regions at hybrid scales per attention layer. The key idea of DiagSWin attention is to inject multi-scale receptive field sizes into tokens: before computing the self-attention matrix, each token attends its closest surrounding tokens at fine granularity and the tokens far away at coarse granularity. This mechanism is able to effectively capture multi-scale context information while reducing computational complexity. With DiagSwin attention, we present a new variant of Vision Transformer models, called DiagSWin Transformers, and demonstrate their superiority in extensive experiments across various tasks. Specifically, the DiagSwin Transformer with a large size achieves 84.4% Top-1 accuracy and outperforms the SOTA CSWin Transformer on ImageNet with 40% fewer model size and computation cost. When employed as backbones, DiagSWin Transformers achieve significant improvements over the current SOTA modules. In addition, our DiagSWin-Base model yields 51.1 box mAP and 45.8 mask mAP on COCO for object detection and segmentation, and 52.3 mIoU on the ADE20K for semantic segmentation.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142076670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A surrogate-assisted extended generative adversarial network for parameter optimization in free-form metasurface design 用于自由曲面设计参数优化的代理辅助扩展生成式对抗网络
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106654

Metasurfaces have widespread applications in fifth-generation (5G) microwave communication. Among the metasurface family, free-form metasurfaces excel in achieving intricate spectral responses compared to regular-shape counterparts. However, conventional numerical methods for free-form metasurfaces are time-consuming and demand specialized expertise. Alternatively, recent studies demonstrate that deep learning has great potential to accelerate and refine metasurface designs. Here, we present XGAN, an extended generative adversarial network (GAN) with a surrogate for high-quality free-form metasurface designs. The proposed surrogate provides a physical constraint to XGAN so that XGAN can accurately generate metasurfaces monolithically from input spectral responses. In comparative experiments involving 20000 free-form metasurface designs, XGAN achieves 0.9734 average accuracy and is 500 times faster than the conventional methodology. This method facilitates the metasurface library building for specific spectral responses and can be extended to various inverse design problems, including optical metamaterials, nanophotonic devices, and drug discovery.

元表面在第五代(5G)微波通信中有着广泛的应用。在元表面家族中,自由形态元表面与规则形态元表面相比,在实现复杂频谱响应方面表现出色。然而,自由形态元曲面的传统数值方法耗时较长,而且需要专业的知识。另外,最近的研究表明,深度学习在加速和完善元曲面设计方面具有巨大潜力。在这里,我们提出了 XGAN,一种具有高质量自由曲面设计代理的扩展生成对抗网络(GAN)。所提出的替代物为 XGAN 提供了物理约束,因此 XGAN 可以从输入频谱响应中准确地单片生成元曲面。在涉及 20000 个自由形式元面设计的对比实验中,XGAN 达到了 0.9734 的平均精度,比传统方法快 500 倍。该方法有助于针对特定光谱响应建立元表面库,并可扩展到各种逆设计问题,包括光学超材料、纳米光子器件和药物发现。
{"title":"A surrogate-assisted extended generative adversarial network for parameter optimization in free-form metasurface design","authors":"","doi":"10.1016/j.neunet.2024.106654","DOIUrl":"10.1016/j.neunet.2024.106654","url":null,"abstract":"<div><p>Metasurfaces have widespread applications in fifth-generation (5G) microwave communication. Among the metasurface family, free-form metasurfaces excel in achieving intricate spectral responses compared to regular-shape counterparts. However, conventional numerical methods for free-form metasurfaces are time-consuming and demand specialized expertise. Alternatively, recent studies demonstrate that deep learning has great potential to accelerate and refine metasurface designs. Here, we present XGAN, an extended generative adversarial network (GAN) with a surrogate for high-quality free-form metasurface designs. The proposed surrogate provides a physical constraint to XGAN so that XGAN can accurately generate metasurfaces monolithically from input spectral responses. In comparative experiments involving 20000 free-form metasurface designs, XGAN achieves 0.9734 average accuracy and is 500 times faster than the conventional methodology. This method facilitates the metasurface library building for specific spectral responses and can be extended to various inverse design problems, including optical metamaterials, nanophotonic devices, and drug discovery.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-task heterogeneous graph learning on electronic health records 电子健康记录的多任务异构图学习
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106644

Learning electronic health records (EHRs) has received emerging attention because of its capability to facilitate accurate medical diagnosis. Since the EHRs contain enriched information specifying complex interactions between entities, modeling EHRs with graphs is shown to be effective in practice. The EHRs, however, present a great degree of heterogeneity, sparsity, and complexity, which hamper the performance of most of the models applied to them. Moreover, existing approaches modeling EHRs often focus on learning the representations for a single task, overlooking the multi-task nature of EHR analysis problems and resulting in limited generalizability across different tasks. In view of these limitations, we propose a novel framework for EHR modeling, namely MulT-EHR (Multi-Task EHR), which leverages a heterogeneous graph to mine the complex relations and model the heterogeneity in the EHRs. To mitigate the large degree of noise, we introduce a denoising module based on the causal inference framework to adjust for severe confounding effects and reduce noise in the EHR data. Additionally, since our model adopts a single graph neural network for simultaneous multi-task prediction, we design a multi-task learning module to leverage the inter-task knowledge to regularize the training process. Extensive empirical studies on MIMIC-III and MIMIC-IV datasets validate that the proposed method consistently outperforms the state-of-the-art designs in four popular EHR analysis tasks — drug recommendation, and predictions of the length of stay, mortality, and readmission. Thorough ablation studies demonstrate the robustness of our method upon variations to key components and hyperparameters.

电子健康记录(EHRs)能够促进准确的医疗诊断,因此学习电子健康记录受到越来越多的关注。由于电子病历包含丰富的信息,说明实体之间复杂的相互作用,因此用图对电子病历建模在实践中证明是有效的。然而,电子病历具有很大程度的异质性、稀疏性和复杂性,这阻碍了大多数应用于电子病历的模型的性能。此外,现有的 EHR 建模方法通常只关注学习单一任务的表征,忽视了 EHR 分析问题的多任务性质,导致在不同任务间的通用性有限。鉴于这些局限性,我们提出了一种新颖的电子病历建模框架,即 MulT-EHR(多任务电子病历),它利用异构图挖掘电子病历中的复杂关系并建立异构模型。为了减少大量噪声,我们在因果推理框架的基础上引入了去噪模块,以调整严重混杂效应并减少 EHR 数据中的噪声。此外,由于我们的模型采用单图神经网络同时进行多任务预测,因此我们设计了一个多任务学习模块,利用任务间知识来规范训练过程。在 MIMIC-III 和 MIMIC-IV 数据集上进行的广泛实证研究证实,在药物推荐、住院时间预测、死亡率预测和再入院预测这四项流行的 EHR 分析任务中,所提出的方法始终优于最先进的设计。彻底的消融研究证明了我们的方法在关键组件和超参数发生变化时的稳健性。
{"title":"Multi-task heterogeneous graph learning on electronic health records","authors":"","doi":"10.1016/j.neunet.2024.106644","DOIUrl":"10.1016/j.neunet.2024.106644","url":null,"abstract":"<div><p>Learning electronic health records (EHRs) has received emerging attention because of its capability to facilitate accurate medical diagnosis. Since the EHRs contain enriched information specifying complex interactions between entities, modeling EHRs with graphs is shown to be effective in practice. The EHRs, however, present a great degree of heterogeneity, sparsity, and complexity, which hamper the performance of most of the models applied to them. Moreover, existing approaches modeling EHRs often focus on learning the representations for a single task, overlooking the multi-task nature of EHR analysis problems and resulting in limited generalizability across different tasks. In view of these limitations, we propose a novel framework for EHR modeling, namely MulT-EHR (<u>Mul</u>ti-<u>T</u>ask EHR), which leverages a heterogeneous graph to mine the complex relations and model the heterogeneity in the EHRs. To mitigate the large degree of noise, we introduce a denoising module based on the causal inference framework to adjust for severe confounding effects and reduce noise in the EHR data. Additionally, since our model adopts a single graph neural network for simultaneous multi-task prediction, we design a multi-task learning module to leverage the inter-task knowledge to regularize the training process. Extensive empirical studies on MIMIC-III and MIMIC-IV datasets validate that the proposed method consistently outperforms the state-of-the-art designs in four popular EHR analysis tasks — drug recommendation, and predictions of the length of stay, mortality, and readmission. Thorough ablation studies demonstrate the robustness of our method upon variations to key components and hyperparameters.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142049202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning functional brain networks with heterogeneous connectivities for brain disease identification 学习具有异质连接性的脑功能网络,用于脑疾病识别
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-22 DOI: 10.1016/j.neunet.2024.106660

Functional brain networks (FBNs), which are used to portray interactions between different brain regions, have been widely used to identify potential biomarkers of neurological and mental disorders. The FBNs estimated using current methods tend to be homogeneous, indicating that different brain regions exhibit the same type of correlation. This homogeneity limits our ability to accurately encode complex interactions within the brain. Therefore, to the best of our knowledge, in the present study, for the first time, we propose the existence of heterogeneous FBNs and introduce a novel FBN estimation model that adaptively assigns heterogeneous connections to different pairs of brain regions, thereby effectively encoding the complex interaction patterns in the brain. Specifically, we first construct multiple types of candidate correlations from different views or based on different methods and then develop an improved orthogonal matching pursuit algorithm to select at most one correlation for each brain region pair under the guidance of label information. These adaptively estimated heterogeneous FBNs were then used to distinguish subjects with neurological/mental disorders from healthy controls and identify potential biomarkers related to these disorders. Experimental results on real datasets show that the proposed scheme improves classification performance by 7.07% and 7.58% at the two sites, respectively, compared with the baseline approaches. This emphasizes the plausibility of the heterogeneity hypothesis and effectiveness of the heterogeneous connection assignment algorithm.

脑功能网络(FBN)用于描述不同脑区之间的相互作用,已被广泛用于识别神经和精神疾病的潜在生物标志物。使用现有方法估算出的 FBN 往往是同质的,这表明不同的脑区表现出相同类型的相关性。这种同质性限制了我们准确编码大脑内部复杂相互作用的能力。因此,据我们所知,在本研究中,我们首次提出了异质 FBN 的存在,并引入了一种新的 FBN 估计模型,该模型能自适应地将异质连接分配给不同的脑区对,从而有效地编码大脑中复杂的交互模式。具体来说,我们首先从不同视图或基于不同方法构建多种类型的候选相关性,然后开发一种改进的正交匹配追寻算法,在标签信息的指导下为每对脑区选择最多一种相关性。然后,这些自适应估计的异质 FBN 被用于区分神经/精神疾病受试者与健康对照组,并识别与这些疾病相关的潜在生物标记物。在真实数据集上的实验结果表明,与基线方法相比,所提出的方案在两个地点的分类性能分别提高了 7.07% 和 7.58%。这强调了异质性假设的合理性和异质性连接分配算法的有效性。
{"title":"Learning functional brain networks with heterogeneous connectivities for brain disease identification","authors":"","doi":"10.1016/j.neunet.2024.106660","DOIUrl":"10.1016/j.neunet.2024.106660","url":null,"abstract":"<div><p>Functional brain networks (FBNs), which are used to portray interactions between different brain regions, have been widely used to identify potential biomarkers of neurological and mental disorders. The FBNs estimated using current methods tend to be homogeneous, indicating that different brain regions exhibit the same type of correlation. This homogeneity limits our ability to accurately encode complex interactions within the brain. Therefore, to the best of our knowledge, in the present study, for the first time, we propose the existence of heterogeneous FBNs and introduce a novel FBN estimation model that adaptively assigns heterogeneous connections to different pairs of brain regions, thereby effectively encoding the complex interaction patterns in the brain. Specifically, we first construct multiple types of candidate correlations from different views or based on different methods and then develop an improved orthogonal matching pursuit algorithm to select at most one correlation for each brain region pair under the guidance of label information. These adaptively estimated heterogeneous FBNs were then used to distinguish subjects with neurological/mental disorders from healthy controls and identify potential biomarkers related to these disorders. Experimental results on real datasets show that the proposed scheme improves classification performance by 7.07% and 7.58% at the two sites, respectively, compared with the baseline approaches. This emphasizes the plausibility of the heterogeneity hypothesis and effectiveness of the heterogeneous connection assignment algorithm.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1