首页 > 最新文献

IEEE Transactions on Cognitive and Developmental Systems最新文献

英文 中文
IEEE Transactions on Cognitive and Developmental Systems Information for Authors IEEE作者认知与发展系统信息汇刊
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-04 DOI: 10.1109/TCDS.2025.3553655
{"title":"IEEE Transactions on Cognitive and Developmental Systems Information for Authors","authors":"","doi":"10.1109/TCDS.2025.3553655","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3553655","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"C4-C4"},"PeriodicalIF":5.0,"publicationDate":"2025-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11024000","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Delving Deeper Into Astromorphic Transformers 更深入地研究星形变形金刚
IF 4.9 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-24 DOI: 10.1109/TCDS.2025.3564285
Md Zesun Ahmed Mia;Malyaban Bal;Abhronil Sengupta
Preliminary attempts at incorporating the critical role of astrocytes—cells that constitute more than 50% of human brain cells—in brain-inspired neuromorphic computing remain in infancy. This article seeks to delve deeper into various key aspects of neuron-synapse-astrocyte interactions to mimic self-attention mechanisms in Transformers. The cross-layer perspective explored in this work involves bioplausible modeling of Hebbian and presynaptic plasticities in neuron-astrocyte networks, incorporating effects of nonlinearities and feedback along with algorithmic formulations to map the neuron-astrocyte computations to self-attention mechanism and evaluating the impact of incorporating bio-realistic effects from the machine learning application side. Our analysis on sentiment and image classification tasks (IMDB and CIFAR10 datasets) highlights the advantages of Astromorphic Transformers, offering improved accuracy and learning speed. Furthermore, the model demonstrates strong natural language generation capabilities on the WikiText-2 dataset, achieving better perplexity compared with conventional models, thus showcasing enhanced generalization and stability across diverse machine learning tasks.
将星形胶质细胞——占人类脑细胞50%以上的细胞——纳入大脑启发的神经形态计算的关键作用的初步尝试仍处于起步阶段。本文旨在深入探讨神经元-突触-星形胶质细胞相互作用的各个关键方面,以模仿变形金刚中的自我注意机制。本研究探索的跨层视角涉及神经元-星形胶质细胞网络中Hebbian和突触前可塑性的生物似是而非的建模,结合非线性和反馈的影响以及将神经元-星形胶质细胞计算映射到自我注意机制的算法公式,并评估从机器学习应用方面结合生物逼真效果的影响。我们对情感和图像分类任务(IMDB和CIFAR10数据集)的分析突出了Astromorphic Transformers的优势,提供了更高的准确性和学习速度。此外,该模型在WikiText-2数据集上展示了强大的自然语言生成能力,与传统模型相比,实现了更好的困惑度,从而在不同的机器学习任务中展示了增强的泛化和稳定性。
{"title":"Delving Deeper Into Astromorphic Transformers","authors":"Md Zesun Ahmed Mia;Malyaban Bal;Abhronil Sengupta","doi":"10.1109/TCDS.2025.3564285","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3564285","url":null,"abstract":"Preliminary attempts at incorporating the critical role of astrocytes—cells that constitute more than 50% of human brain cells—in brain-inspired neuromorphic computing remain in infancy. This article seeks to delve deeper into various key aspects of neuron-synapse-astrocyte interactions to mimic self-attention mechanisms in Transformers. The cross-layer perspective explored in this work involves bioplausible modeling of Hebbian and presynaptic plasticities in neuron-astrocyte networks, incorporating effects of nonlinearities and feedback along with algorithmic formulations to map the neuron-astrocyte computations to self-attention mechanism and evaluating the impact of incorporating bio-realistic effects from the machine learning application side. Our analysis on sentiment and image classification tasks (IMDB and CIFAR10 datasets) highlights the advantages of Astromorphic Transformers, offering improved accuracy and learning speed. Furthermore, the model demonstrates strong natural language generation capabilities on the WikiText-2 dataset, achieving better perplexity compared with conventional models, thus showcasing enhanced generalization and stability across diverse machine learning tasks.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 6","pages":"1436-1446"},"PeriodicalIF":4.9,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145729446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing Representation for Abstractive Multidocument Summarization Based on Adversarial Learning Strategy 基于对抗学习策略的抽象多文档摘要优化表示
IF 4.9 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-21 DOI: 10.1109/TCDS.2025.3563357
Bin Cao;Xinxin Guan;Songlin Bao;Jiawei Wu;Jing Fan
Abstractive multi-document summarization (MDS) is a crucial technique in cognitive computing, enabling the efficient synthesis of a documents cluster into a concise and complete summary. Despite recent advances, existing approaches still face challenges in representation learning when processing large-scale documents clusters: 1) incomplete semantic learning caused by documents truncation or exclusion; 2) the incorporation of noise, such as irrelevant or redundant information from documents; and 3) the potential omission of critical content due to partial coverage of documents. These limitations collectively undermine the semantic integrity and conciseness of the generated summaries. To address these issues, we propose TALER, a two-stage representation architecture enhanced by adversarial learning for abstractive MDS, which reformulates the MDS task as a single-document optimization problem. In Stage I, TALER focuses on enhancing single-document representations by maximizing semantic learning from each document in the cluster and employing the adversarial learning to suppress the introduction of documents noise. In Stage II, TALER conducts multidocument semantic fusion and summary generation by aggregating the learned document embeddings based on Stage I into a cluster-level representation through a pooling mechanism, followed by a self-attention module to capture salient content and produce the final summary. Experimental results on the Multi-News, DUC04, and Multi-XScience datasets demonstrate that TALER consistently outperforms existing baseline models across multiple evaluation metrics.
抽象多文档摘要(MDS)是认知计算中的一项关键技术,它能够将一个文档簇高效地合成为一个简洁完整的摘要。尽管最近取得了一些进展,但现有的方法在处理大规模文档聚类时仍然面临着挑战:1)由于文档截断或排除导致的语义学习不完整;2)噪声的掺入,例如文件中不相关或冗余的信息;3)由于文件的部分覆盖,可能遗漏关键内容。这些限制共同破坏了生成摘要的语义完整性和简洁性。为了解决这些问题,我们提出了TALER,这是一种通过对抗性学习增强的两阶段表示架构,用于抽象MDS,它将MDS任务重新表述为单个文档优化问题。在第一阶段,TALER侧重于通过最大化集群中每个文档的语义学习和使用对抗性学习来抑制文档噪声的引入来增强单文档表示。在第二阶段,TALER通过池化机制将基于第一阶段的学习到的文档嵌入聚合成聚类级表示,进行多文档语义融合和摘要生成,然后通过自关注模块捕获突出内容并生成最终摘要。在Multi-News、DUC04和Multi-XScience数据集上的实验结果表明,TALER在多个评估指标上始终优于现有的基线模型。
{"title":"Optimizing Representation for Abstractive Multidocument Summarization Based on Adversarial Learning Strategy","authors":"Bin Cao;Xinxin Guan;Songlin Bao;Jiawei Wu;Jing Fan","doi":"10.1109/TCDS.2025.3563357","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3563357","url":null,"abstract":"Abstractive multi-document summarization (MDS) is a crucial technique in cognitive computing, enabling the efficient synthesis of a documents cluster into a concise and complete summary. Despite recent advances, existing approaches still face challenges in representation learning when processing large-scale documents clusters: 1) incomplete semantic learning caused by documents truncation or exclusion; 2) the incorporation of noise, such as irrelevant or redundant information from documents; and 3) the potential omission of critical content due to partial coverage of documents. These limitations collectively undermine the semantic integrity and conciseness of the generated summaries. To address these issues, we propose <italic>TALER</i>, a two-stage representation architecture enhanced by adversarial learning for abstractive MDS, which reformulates the MDS task as a single-document optimization problem. In Stage I, <italic>TALER</i> focuses on enhancing single-document representations by maximizing semantic learning from each document in the cluster and employing the adversarial learning to suppress the introduction of documents noise. In Stage II, <italic>TALER</i> conducts multidocument semantic fusion and summary generation by aggregating the learned document embeddings based on Stage I into a cluster-level representation through a pooling mechanism, followed by a self-attention module to capture salient content and produce the final summary. Experimental results on the Multi-News, DUC04, and Multi-XScience datasets demonstrate that <italic>TALER</i> consistently outperforms existing baseline models across multiple evaluation metrics.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 6","pages":"1426-1435"},"PeriodicalIF":4.9,"publicationDate":"2025-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145705910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised Hyperbolic Spectro-Temporal Graph Convolution Network for Early 3-D Behavior Prediction 早期三维行为预测的自监督双曲谱-时间图卷积网络
IF 4.9 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-16 DOI: 10.1109/TCDS.2025.3561422
Peng Liu;Qin Lai;Haibo Li;Chong Zhao;Qicong Wang;Hongying Meng
3-D human behavior is a highly nonlinear spatiotemporal interaction process. Therefore, early behavior prediction is a challenging task, especially prediction with low observation rates in unsupervised mode. To this end, we propose a novel self-supervised early 3-D behavior prediction framework that learns graph structures on hyperbolic manifold. First, we employ the sequence construction of multidynamic key information to enlarge the key details of spatiotemporal behavior sequences, addressing the high redundancy between frames of spatiotemporal interaction. Second, for capturing dependencies among long-distance joints, we explore a unique graph Laplacian on hyperbolic manifold to perceive the subtle local difference within frames. Finally, we leverage the learned spatiotemporal features under different observation rates for progressive contrast, forming self-supervised signals. This facilitates the extraction of more discriminative global and local spatiotemporal information from early behavior sequences in unsupervised mode. Extensive experiments on three behavior datasets have demonstrated the superiority of our approach at low to medium observation rates.
三维人类行为是一个高度非线性的时空交互过程。因此,早期行为预测是一项具有挑战性的任务,特别是在无监督模式下低观察率的预测。为此,我们提出了一种新的自监督早期三维行为预测框架,该框架学习双曲流形上的图结构。首先,采用多动态关键信息序列构建,扩大了时空行为序列的关键细节,解决了时空交互帧间的高冗余问题;其次,为了捕获长距离关节之间的依赖关系,我们在双曲流形上探索了一个独特的图拉普拉斯算子来感知帧内微妙的局部差异。最后,利用学习到的不同观测率下的时空特征进行递进对比,形成自监督信号。这有助于在无监督模式下从早期行为序列中提取更具判别性的全局和局部时空信息。在三个行为数据集上进行的大量实验证明了我们的方法在中低观察率下的优越性。
{"title":"Self-Supervised Hyperbolic Spectro-Temporal Graph Convolution Network for Early 3-D Behavior Prediction","authors":"Peng Liu;Qin Lai;Haibo Li;Chong Zhao;Qicong Wang;Hongying Meng","doi":"10.1109/TCDS.2025.3561422","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3561422","url":null,"abstract":"3-D human behavior is a highly nonlinear spatiotemporal interaction process. Therefore, early behavior prediction is a challenging task, especially prediction with low observation rates in unsupervised mode. To this end, we propose a novel self-supervised early 3-D behavior prediction framework that learns graph structures on hyperbolic manifold. First, we employ the sequence construction of multidynamic key information to enlarge the key details of spatiotemporal behavior sequences, addressing the high redundancy between frames of spatiotemporal interaction. Second, for capturing dependencies among long-distance joints, we explore a unique graph Laplacian on hyperbolic manifold to perceive the subtle local difference within frames. Finally, we leverage the learned spatiotemporal features under different observation rates for progressive contrast, forming self-supervised signals. This facilitates the extraction of more discriminative global and local spatiotemporal information from early behavior sequences in unsupervised mode. Extensive experiments on three behavior datasets have demonstrated the superiority of our approach at low to medium observation rates.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 6","pages":"1411-1425"},"PeriodicalIF":4.9,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145705917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improve Knowledge Distillation via Label Revision and Data Selection 通过标签修订和数据选择改进知识蒸馏
IF 4.9 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-11 DOI: 10.1109/TCDS.2025.3559881
Weichao Lan;Yiu-ming Cheung;Qing Xu;Buhua Liu;Zhikai Hu;Mengke Li;Zhenghua Chen
Knowledge distillation (KD) transferring knowledge from a large teacher model to a lightweight student one has received great attention in deep model compression. In addition to the supervision of ground truth, the vanilla KD method regards the predictions of the teacher as soft labels to supervise the training of the student model. Based on vanilla KD, various approaches have been developed to improve the performance of the student model further. However, few of these previous methods have considered the reliability of the supervision from teacher models. Supervision from erroneous predictions may mislead the training of the student model. This article therefore proposes to tackle this problem from two aspects: label revision to rectify the incorrect supervision and data selection to select appropriate samples for distillation to reduce the impact of erroneous supervision. In the former, we propose to rectify the teacher’s inaccurate predictions using the ground truth. In the latter, we introduce a data selection technique to choose suitable training samples to be supervised by the teacher, thereby reducing the impact of incorrect predictions to some extent. Experiment results demonstrate the effectiveness of the proposed method, which can be further combined with other distillation approaches to enhance their performance.
在深度模型压缩中,知识蒸馏(Knowledge distillation, KD)是一种将知识从大型教师模型转移到轻量级学生模型的方法。除了监督基础真理之外,香草KD方法还将教师的预测作为软标签来监督学生模型的训练。基于香草KD,已经开发了各种方法来进一步提高学生模型的性能。然而,这些先前的方法很少考虑到教师模型监督的可靠性。来自错误预测的监督可能会误导学生模型的训练。因此,本文建议从两个方面来解决这一问题:修改标签以纠正错误的监督,选择数据以选择合适的样本进行蒸馏,以减少错误监督的影响。在前者中,我们建议使用基本事实来纠正教师的不准确预测。在后者中,我们引入了一种数据选择技术,选择合适的训练样本由教师监督,从而在一定程度上减少了错误预测的影响。实验结果证明了该方法的有效性,该方法可以进一步与其他蒸馏方法相结合以提高其性能。
{"title":"Improve Knowledge Distillation via Label Revision and Data Selection","authors":"Weichao Lan;Yiu-ming Cheung;Qing Xu;Buhua Liu;Zhikai Hu;Mengke Li;Zhenghua Chen","doi":"10.1109/TCDS.2025.3559881","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3559881","url":null,"abstract":"Knowledge distillation (KD) transferring knowledge from a large teacher model to a lightweight student one has received great attention in deep model compression. In addition to the supervision of ground truth, the vanilla KD method regards the predictions of the teacher as soft labels to supervise the training of the student model. Based on vanilla KD, various approaches have been developed to improve the performance of the student model further. However, few of these previous methods have considered the reliability of the supervision from teacher models. Supervision from erroneous predictions may mislead the training of the student model. This article therefore proposes to tackle this problem from two aspects: label revision to rectify the incorrect supervision and data selection to select appropriate samples for distillation to reduce the impact of erroneous supervision. In the former, we propose to rectify the teacher’s inaccurate predictions using the ground truth. In the latter, we introduce a data selection technique to choose suitable training samples to be supervised by the teacher, thereby reducing the impact of incorrect predictions to some extent. Experiment results demonstrate the effectiveness of the proposed method, which can be further combined with other distillation approaches to enhance their performance.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 6","pages":"1377-1388"},"PeriodicalIF":4.9,"publicationDate":"2025-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10962557","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145705891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating the Tradeoff Between Analogical Reasoning Ability and Efficiency in Large Language Models 评估大型语言模型中类比推理能力和效率之间的权衡
IF 4.9 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-11 DOI: 10.1109/TCDS.2025.3559771
Kara L. Combs;Isaiah Goble;Spencer V. Howlett;Yuki B. Adams;Trevor J. Bihl
Recent advances in large language models (LLMs) have led to the general public’s assumption of human-equivalent logic and cognition. However, the research community is inconclusive, especially concerning LLM’s analogical reasoning abilities. Twenty-one proprietary and open-source LLMs were evaluated on two long-text/story analogy datasets. The LLMs produced mixed results on the four qualitative and seven quantitative metrics. LLMs performed well when tasked with determining the presence or absence of similar elements between stories based on the qualitative assessment of their outputs. However, despite this success, LLMs still struggled with the correct identification of the most analogous story to the base story. Further inspection indicates that the models struggled with recognizing high-order (similar to cause and effect) relationships associated with higher cognitive function(s). Regardless of the overall performance, there is a clear advantage that propriety has over open-source models concerning analogical reasoning. Last, this study suggests that LLM accuracy and their number of parameters explain over half of the variation in the energy consumed based on a statistically significant multivariate regression model. Future work may consider evaluating other types of reasoning and LLMs’ learning abilities by providing “correct” responses to guide future results.
最近在大型语言模型(llm)方面的进展导致了公众对人类等效逻辑和认知的假设。然而,研究界尚无定论,特别是关于法学硕士的类比推理能力。在两个长文本/故事类比数据集上评估了21个专有和开源法学硕士。法学硕士课程在四项定性指标和七项定量指标上的结果好坏参半。法学硕士在根据对其产出的定性评估确定故事之间是否存在类似元素的任务时表现良好。然而,尽管取得了这样的成功,法学硕士们仍然难以正确识别与基础故事最相似的故事。进一步的检查表明,这些模型难以识别与高级认知功能相关的高阶(类似于因果关系)关系。不管整体性能如何,在类比推理方面,专有模型比开源模型有一个明显的优势。最后,本研究表明,基于统计显著的多元回归模型,LLM精度及其参数数量解释了能源消耗变化的一半以上。未来的工作可能会考虑通过提供“正确”的回答来评估其他类型的推理和法学硕士的学习能力,以指导未来的结果。
{"title":"Evaluating the Tradeoff Between Analogical Reasoning Ability and Efficiency in Large Language Models","authors":"Kara L. Combs;Isaiah Goble;Spencer V. Howlett;Yuki B. Adams;Trevor J. Bihl","doi":"10.1109/TCDS.2025.3559771","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3559771","url":null,"abstract":"Recent advances in large language models (LLMs) have led to the general public’s assumption of human-equivalent logic and cognition. However, the research community is inconclusive, especially concerning LLM’s analogical reasoning abilities. Twenty-one proprietary and open-source LLMs were evaluated on two long-text/story analogy datasets. The LLMs produced mixed results on the four qualitative and seven quantitative metrics. LLMs performed well when tasked with determining the presence or absence of similar elements between stories based on the qualitative assessment of their outputs. However, despite this success, LLMs still struggled with the correct identification of the most analogous story to the base story. Further inspection indicates that the models struggled with recognizing high-order (similar to cause and effect) relationships associated with higher cognitive function(s). Regardless of the overall performance, there is a clear advantage that propriety has over open-source models concerning analogical reasoning. Last, this study suggests that LLM accuracy and their number of parameters explain over half of the variation in the energy consumed based on a statistically significant multivariate regression model. Future work may consider evaluating other types of reasoning and LLMs’ learning abilities by providing “correct” responses to guide future results.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 6","pages":"1401-1410"},"PeriodicalIF":4.9,"publicationDate":"2025-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10962558","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145705925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiscale Convolutional Transformer With Diverse-Aware Feature Learning for Motor Imagery EEG Decoding 基于多感知特征学习的多尺度卷积变压器用于运动意象脑电解码
IF 4.9 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-08 DOI: 10.1109/TCDS.2025.3559187
Wenlong Hang;Junliang Wang;Shuang Liang;Baiying Lei;Qiong Wang;Guanglin Li;Badong Chen;Jing Qin
Electroencephalogram (EEG)-based motor imagery (MI) brain–computer interfaces (BCIs) have significant potential in improving motor function for neurorehabilitation. Despite recent advancements, learning diversified EEG features across different frequency ranges remains a significant challenge, as the homogenization of feature representations often limits the generalization capabilities of EEG decoding models for BCIs. In this article, we propose a novel multiscale convolutional transformer framework for EEG decoding that integrates multiscale convolution, transformer, and diverse-aware feature learning scheme (MCTD) to tackle the above challenge. Specifically, we first capture multiple frequency features using dynamic one-dimensional temporal convolution with different kernel lengths. Subsequently, we incorporate convolutional layers and transformers with a contrastive learning scheme to extract discriminative local and global EEG features within a single frequency range. To mitigate the homogenization of features extracted from different frequency ranges, we propose a novel decorrelation regularization. It enables multiscale convolutional transformers to produce less correlated features with each other, thereby enhancing the overall expressiveness of EEG decoding model. The performance of MCTD is evaluated on four public MI-based EEG datasets, including the BCI competition III 3a and IV 2a, the BNCI 2015-001, and the OpenBMI. For the average Kappa/Accuracy scores, MCTD obtains improvements of 3.58%/2.68%, 3.09%/2.20%, 2.33%/1.54%, and 4.44%/2.22%, over the state-of-the-art method on four EEG datasets, respectively. Experimental results demonstrate that our method exhibits superior performance. Code is available at: https://github.com/kfhss/MCTD.
基于脑电图(EEG)的运动图像(MI)脑机接口(bci)在改善神经康复中的运动功能方面具有重要的潜力。尽管最近取得了一些进展,但学习不同频率范围内的多样化脑电特征仍然是一个重大挑战,因为特征表示的同质化往往限制了脑机接口的脑电解码模型的泛化能力。在本文中,我们提出了一种新的多尺度卷积变压器EEG解码框架,该框架集成了多尺度卷积、变压器和多元感知特征学习方案(MCTD)来解决上述挑战。具体来说,我们首先使用不同核长度的动态一维时间卷积捕获多个频率特征。随后,我们将卷积层和变压器与对比学习方案结合在一起,在单个频率范围内提取有区别的局部和全局脑电特征。为了减轻从不同频率范围提取的特征的均匀化,我们提出了一种新的去相关正则化方法。它使多尺度卷积变换产生的特征之间的相关性更小,从而增强了脑电信号解码模型的整体表达能力。MCTD的性能在4个公开的基于mi的EEG数据集上进行了评估,包括BCI competition III 3a和IV 2a、BNCI 2015-001和OpenBMI。对于Kappa/Accuracy的平均分数,MCTD在4个EEG数据集上分别比最先进的方法提高了3.58%/2.68%、3.09%/2.20%、2.33%/1.54%和4.44%/2.22%。实验结果表明,该方法具有良好的性能。代码可从https://github.com/kfhss/MCTD获得。
{"title":"Multiscale Convolutional Transformer With Diverse-Aware Feature Learning for Motor Imagery EEG Decoding","authors":"Wenlong Hang;Junliang Wang;Shuang Liang;Baiying Lei;Qiong Wang;Guanglin Li;Badong Chen;Jing Qin","doi":"10.1109/TCDS.2025.3559187","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3559187","url":null,"abstract":"Electroencephalogram (EEG)-based motor imagery (MI) brain–computer interfaces (BCIs) have significant potential in improving motor function for neurorehabilitation. Despite recent advancements, learning diversified EEG features across different frequency ranges remains a significant challenge, as the homogenization of feature representations often limits the generalization capabilities of EEG decoding models for BCIs. In this article, we propose a novel multiscale convolutional transformer framework for EEG decoding that integrates multiscale convolution, transformer, and diverse-aware feature learning scheme (MCTD) to tackle the above challenge. Specifically, we first capture multiple frequency features using dynamic one-dimensional temporal convolution with different kernel lengths. Subsequently, we incorporate convolutional layers and transformers with a contrastive learning scheme to extract discriminative local and global EEG features within a single frequency range. To mitigate the homogenization of features extracted from different frequency ranges, we propose a novel decorrelation regularization. It enables multiscale convolutional transformers to produce less correlated features with each other, thereby enhancing the overall expressiveness of EEG decoding model. The performance of MCTD is evaluated on four public MI-based EEG datasets, including the BCI competition III 3a and IV 2a, the BNCI 2015-001, and the OpenBMI. For the average Kappa/Accuracy scores, MCTD obtains improvements of 3.58%/2.68%, 3.09%/2.20%, 2.33%/1.54%, and 4.44%/2.22%, over the state-of-the-art method on four EEG datasets, respectively. Experimental results demonstrate that our method exhibits superior performance. Code is available at: <uri>https://github.com/kfhss/MCTD</uri>.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 6","pages":"1389-1400"},"PeriodicalIF":4.9,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145705867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Cognitive and Developmental Systems Publication Information IEEE认知与发展系统汇刊
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-02 DOI: 10.1109/TCDS.2025.3537202
{"title":"IEEE Transactions on Cognitive and Developmental Systems Publication Information","authors":"","doi":"10.1109/TCDS.2025.3537202","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3537202","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"C2-C2"},"PeriodicalIF":5.0,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10947666","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Cognitive and Developmental Systems Information for Authors IEEE作者认知与发展系统信息汇刊
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-02 DOI: 10.1109/TCDS.2025.3537206
{"title":"IEEE Transactions on Cognitive and Developmental Systems Information for Authors","authors":"","doi":"10.1109/TCDS.2025.3537206","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3537206","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"C4-C4"},"PeriodicalIF":5.0,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10947664","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Computational Intelligence Society Information IEEE计算智能学会信息
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-02 DOI: 10.1109/TCDS.2025.3537204
{"title":"IEEE Computational Intelligence Society Information","authors":"","doi":"10.1109/TCDS.2025.3537204","DOIUrl":"https://doi.org/10.1109/TCDS.2025.3537204","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"C3-C3"},"PeriodicalIF":5.0,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10947665","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Cognitive and Developmental Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1