首页 > 最新文献

Neural Networks最新文献

英文 中文
ReTri: Progressive domain bridging via representation disentanglement and triple-level consistency-driven feature alignment for unsupervised domain adaptive medical image segmentation 基于表示解纠缠和三级一致性驱动特征对齐的渐进式域桥接无监督域自适应医学图像分割
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-09 DOI: 10.1016/j.neunet.2026.108564
Xiaoru Gao , Guoyan Zheng
Unsupervised domain adaptation (UDA) in medical image segmentation presents significant challenges due to substantial cross-domain disparities and the inherent absence of target domain annotations. In this study, to address these challenges, we propose an end-to-end progressive domain bridging framework based on representation disentanglement and triple-level consistency-driven feature alignment, referred to as ReTri, that synergistically integrates a representation disentanglement-based image alignment (RDIA) module with a novel triple-level consistency-driven feature alignment (TCFA) module. In particular, the RDIA module aims to establish an initial domain bridge by decoupling and aligning fundamental visual disparities through disentangled representation learning, while the novel TCFA module hierarchically bridges remaining cross-domain semantic discrepancies and feature distribution disparities via two novel consistency-driven alignment mechanisms: 1) attention-guided semantics-level consistency alignment, where we purposely design a bi-attentive semantic feature extraction (BSFE) component coupled with an attention-adaptive semantic consistency (ASC) loss function, facilitating dynamic alignment of high-level semantic representations across domains, and 2) multi-view dual-level mixing consistency alignment, consisting of Feature-Cut consistent self-ensembling (FCCS) and Trans-Cut consistent self-ensembling (TCCS) components. These two components operate within intermediate mixing spaces to ensure robust knowledge transfer through complementary feature- and prediction-level consistency regularization. Extensive experimental evaluations are conducted on four challenging datasets (Lumbar Spine CT-MR, Cardiac CT-MR, Cross-domain Echocardiography, and Multi-center Prostate MR) across seven UDA-based segmentation scenarios and two external validation scenarios. Our framework achieves superior performance over the best state-of-the-art (SOTA) methods on following UDA-based segmentation scenarios: +2.9% DSC for spine CT → MR segmentation, +3.6% and +2.4% DSC for bidirectional cardiac CT↔MR segmentation, +1.7% and +2.3% DSC for bidirectional cross-center cross-vendor Echocardiography (CAMUS↔EchoNet-Dynamic) segmentation, and +12.2% and +12.0% DSC for bidirectional multi-center prostate MR segmentation. The source code and the datasets are publicly available at https://github.com/xiaorugao999/ReTri.
无监督域自适应(UDA)在医学图像分割中存在较大的跨域差异和缺乏目标域标注等问题。在本研究中,为了解决这些挑战,我们提出了一个基于表示解除纠缠和三级一致性驱动特征对齐的端到端渐进式领域桥接框架(ReTri),该框架将基于表示解除纠缠的图像对齐(RDIA)模块与新型三级一致性驱动的特征对齐(TCFA)模块协同集成。特别是,RDIA模块旨在通过解耦表示学习来解耦和对齐基本的视觉差异,从而建立一个初始的领域桥梁,而新颖的TCFA模块通过两种新颖的一致性驱动对齐机制分层地桥接剩余的跨领域语义差异和特征分布差异:1)注意引导语义级一致性对齐,其中设计了双注意语义特征提取(BSFE)组件和注意自适应语义一致性(ASC)损失函数,促进跨域高级语义表示的动态对齐;2)多视图双级别混合一致性对齐,由特征切割一致自集成(FCCS)和横切一致自集成(TCCS)组件组成。这两个组件在中间混合空间中运行,通过互补的特征级和预测级一致性正则化来确保稳健的知识转移。在四个具有挑战性的数据集(腰椎CT-MR,心脏CT-MR,跨域超声心动图和多中心前列腺MR)上进行了广泛的实验评估,包括七个基于uda的分割场景和两个外部验证场景。我们的框架在以下基于uda的分割方案上实现了比最先进的(SOTA)方法更优越的性能:脊柱CT → MR分割+2.9% DSC,双向心脏CT↔MR分割+3.6%和+2.4% DSC,双向跨中心跨供应商超声心动图(CAMUS↔EchoNet-Dynamic)分割+1.7%和+2.3% DSC,双向多中心前列腺MR分割+12.2%和+12.0% DSC。源代码和数据集可以在https://github.com/xiaorugao999/ReTri上公开获得。
{"title":"ReTri: Progressive domain bridging via representation disentanglement and triple-level consistency-driven feature alignment for unsupervised domain adaptive medical image segmentation","authors":"Xiaoru Gao ,&nbsp;Guoyan Zheng","doi":"10.1016/j.neunet.2026.108564","DOIUrl":"10.1016/j.neunet.2026.108564","url":null,"abstract":"<div><div>Unsupervised domain adaptation (UDA) in medical image segmentation presents significant challenges due to substantial cross-domain disparities and the inherent absence of target domain annotations. In this study, to address these challenges, we propose an end-to-end progressive domain bridging framework based on representation disentanglement and triple-level consistency-driven feature alignment, referred to as ReTri, that synergistically integrates a representation disentanglement-based image alignment (RDIA) module with a novel triple-level consistency-driven feature alignment (TCFA) module. In particular, the RDIA module aims to establish an initial domain bridge by decoupling and aligning fundamental visual disparities through disentangled representation learning, while the novel TCFA module hierarchically bridges remaining cross-domain semantic discrepancies and feature distribution disparities via two novel consistency-driven alignment mechanisms: 1) attention-guided semantics-level consistency alignment, where we purposely design a bi-attentive semantic feature extraction (BSFE) component coupled with an attention-adaptive semantic consistency (ASC) loss function, facilitating dynamic alignment of high-level semantic representations across domains, and 2) multi-view dual-level mixing consistency alignment, consisting of Feature-Cut consistent self-ensembling (FCCS) and Trans-Cut consistent self-ensembling (TCCS) components. These two components operate within intermediate mixing spaces to ensure robust knowledge transfer through complementary feature- and prediction-level consistency regularization. Extensive experimental evaluations are conducted on four challenging datasets (Lumbar Spine CT-MR, Cardiac CT-MR, Cross-domain Echocardiography, and Multi-center Prostate MR) across seven UDA-based segmentation scenarios and two external validation scenarios. Our framework achieves superior performance over the best state-of-the-art (SOTA) methods on following UDA-based segmentation scenarios: +2.9% DSC for spine CT → MR segmentation, +3.6% and +2.4% DSC for bidirectional cardiac CT↔MR segmentation, +1.7% and +2.3% DSC for bidirectional cross-center cross-vendor Echocardiography (CAMUS↔EchoNet-Dynamic) segmentation, and +12.2% and +12.0% DSC for bidirectional multi-center prostate MR segmentation. The source code and the datasets are publicly available at <span><span>https://github.com/xiaorugao999/ReTri</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108564"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An online forecasting-based fine-tuning pipeline for time-series anomaly prediction 基于在线预测的时间序列异常预测微调管道
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-10 DOI: 10.1016/j.neunet.2026.108568
Zhou Zhou , Van Hoan Trinh , Yuet Ming Joyce Yue , Dit-Yan Yeung , Ka-Hing Wong , Wai-Kin Wong
Time-series anomaly detection is critical for numerous real-world applications and has been extensively studied. However, existing methods are typically designed to identify anomalies within a complete time series. In other words, they rely on access to ground truth data to calculate anomaly scores and distinguish anomalous data from normal patterns. This reliance limits their applicability in scenarios where predicting future anomalies is required, as the ground truth is inherently unavailable. To address this gap, we introduce the concept of Time-Series Anomaly Prediction (TSAP), which focuses on forecasting the occurrence and progression of anomalies in time series simultaneously without relying on ground truth. In this paper, we propose a novel exemplar-based pre-training and fine-tuning pipeline tailored to this task, based on recent achievements in online time-series forecasting techniques. The pipeline begins with an offline pre-training phase, where a deep learning model is trained to capture the underlying temporal correlations in time-series data. During the online fine-tuning stage, a three-step process is employed to predict the timing and evolution of anomalies. This process includes prediction and anomaly detection, motif search for similar patterns, and fine-tuning using exemplars. These steps are repeated as new data arrives. We evaluate the proposed method against state-of-the-art approaches from various relevant categories on both real-world and synthetic datasets. Experimental results show that the proposed method improves anomaly detection accuracy by up to 53.8% in terms of F1 score and enhances time-series forecasting accuracy during and after anomaly periods by up to 82.4% and 49.1% in terms of MSE. Through analysing the results, we prove the proposed method’s effectiveness in addressing the new TSAP tasks, which are incapable of being handled by current time-series anomaly detection or online time-series forecasting methods.
时间序列异常检测对于许多实际应用至关重要,并且已经得到了广泛的研究。然而,现有的方法通常用于识别完整时间序列中的异常。换句话说,他们依赖于对地面真实数据的访问来计算异常分数,并将异常数据与正常模式区分开来。这种依赖限制了它们在需要预测未来异常情况的情况下的适用性,因为地面事实本质上是不可用的。为了解决这一差距,我们引入了时间序列异常预测(TSAP)的概念,其重点是在不依赖于地面事实的情况下同时预测时间序列中异常的发生和进展。在本文中,我们基于在线时间序列预测技术的最新成果,提出了一种新的基于样本的预训练和微调管道。该流程从离线预训练阶段开始,在该阶段训练深度学习模型以捕获时间序列数据中的潜在时间相关性。在在线微调阶段,采用三步过程来预测异常的时间和演化。这个过程包括预测和异常检测,相似模式的基序搜索,以及使用示例进行微调。当新数据到达时,重复这些步骤。我们在真实世界和合成数据集上对来自各种相关类别的最先进方法进行了评估。实验结果表明,该方法在F1分数方面的异常检测精度提高了53.8%,在MSE方面,异常期间和异常后的时间序列预测精度分别提高了82.4%和49.1%。通过对结果的分析,我们证明了该方法在解决当前时间序列异常检测或在线时间序列预测方法无法处理的新的TSAP任务方面的有效性。
{"title":"An online forecasting-based fine-tuning pipeline for time-series anomaly prediction","authors":"Zhou Zhou ,&nbsp;Van Hoan Trinh ,&nbsp;Yuet Ming Joyce Yue ,&nbsp;Dit-Yan Yeung ,&nbsp;Ka-Hing Wong ,&nbsp;Wai-Kin Wong","doi":"10.1016/j.neunet.2026.108568","DOIUrl":"10.1016/j.neunet.2026.108568","url":null,"abstract":"<div><div>Time-series anomaly detection is critical for numerous real-world applications and has been extensively studied. However, existing methods are typically designed to identify anomalies within a complete time series. In other words, they rely on access to ground truth data to calculate anomaly scores and distinguish anomalous data from normal patterns. This reliance limits their applicability in scenarios where predicting future anomalies is required, as the ground truth is inherently unavailable. To address this gap, we introduce the concept of <em>Time-Series Anomaly Prediction</em> (<em>TSAP</em>), which focuses on forecasting the occurrence and progression of anomalies in time series simultaneously without relying on ground truth. In this paper, we propose a novel exemplar-based pre-training and fine-tuning pipeline tailored to this task, based on recent achievements in online time-series forecasting techniques. The pipeline begins with an offline pre-training phase, where a deep learning model is trained to capture the underlying temporal correlations in time-series data. During the online fine-tuning stage, a three-step process is employed to predict the timing and evolution of anomalies. This process includes prediction and anomaly detection, motif search for similar patterns, and fine-tuning using exemplars. These steps are repeated as new data arrives. We evaluate the proposed method against state-of-the-art approaches from various relevant categories on both real-world and synthetic datasets. Experimental results show that the proposed method improves anomaly detection accuracy by up to 53.8% in terms of F1 score and enhances time-series forecasting accuracy during and after anomaly periods by up to 82.4% and 49.1% in terms of MSE. Through analysing the results, we prove the proposed method’s effectiveness in addressing the new <em>TSAP</em> tasks, which are incapable of being handled by current time-series anomaly detection or online time-series forecasting methods.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108568"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implicit neural network-based coal SEM super-resolution for enhancing micro-pores measurement tasks 基于隐式神经网络的煤SEM超分辨增强微孔测量任务
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-10 DOI: 10.1016/j.neunet.2026.108578
Xiaowei An , Shenghua Teng , Zhuopeng Wang , Quanquan Liang
Prolonged radiation exposure in coal Scanning Electron Microscopy (SEM) poses structural damage risks to specimens during high-resolution observation. To mitigate this situation, we propose an interactive-interpretable super-resolution (SR) framework that integrates implicit neural representation (INR) with model-driven Half-Quadratic Splitting (HQS) optimization. Specifically, the implicit neural representation employs a local window attention mechanism to capture contextual dependencies across reconstructed regions. Furthermore, an interactive dual-branch network decouples feature content and positional encoding, providing an initial solution for the subsequent HQS optimization. By unfolding the HQS algorithm into a deep network, each layer corresponds to an explicit and interpretable optimization step with the explicit mathmatical transparency. Experimental results demonstrate that our method outperforms the related state-of-the-art SR algorithms in visual fidelity and exhibits the applicability and stability in downstream geometry-sensitive measurement tasks.
煤炭扫描电镜(SEM)在高分辨率观测过程中,长时间的辐射暴露会对样品造成结构损伤。为了缓解这种情况,我们提出了一个交互式可解释的超分辨率(SR)框架,该框架将隐式神经表示(INR)与模型驱动的半二次分裂(HQS)优化相结合。具体而言,隐式神经表征采用局部窗口注意机制来捕获重建区域之间的上下文依赖关系。此外,交互式双分支网络解耦了特征内容和位置编码,为后续的HQS优化提供了初始解决方案。通过将HQS算法展开为一个深度网络,每一层都对应一个明确的、可解释的优化步骤,具有明确的数学透明性。实验结果表明,我们的方法在视觉保真度上优于相关的SR算法,并且在下游几何敏感测量任务中表现出适用性和稳定性。
{"title":"Implicit neural network-based coal SEM super-resolution for enhancing micro-pores measurement tasks","authors":"Xiaowei An ,&nbsp;Shenghua Teng ,&nbsp;Zhuopeng Wang ,&nbsp;Quanquan Liang","doi":"10.1016/j.neunet.2026.108578","DOIUrl":"10.1016/j.neunet.2026.108578","url":null,"abstract":"<div><div>Prolonged radiation exposure in coal Scanning Electron Microscopy (SEM) poses structural damage risks to specimens during high-resolution observation. To mitigate this situation, we propose an interactive-interpretable super-resolution (SR) framework that integrates implicit neural representation (INR) with model-driven Half-Quadratic Splitting (HQS) optimization. Specifically, the implicit neural representation employs a local window attention mechanism to capture contextual dependencies across reconstructed regions. Furthermore, an interactive dual-branch network decouples feature content and positional encoding, providing an initial solution for the subsequent HQS optimization. By unfolding the HQS algorithm into a deep network, each layer corresponds to an explicit and interpretable optimization step with the explicit mathmatical transparency. Experimental results demonstrate that our method outperforms the related state-of-the-art SR algorithms in visual fidelity and exhibits the applicability and stability in downstream geometry-sensitive measurement tasks.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108578"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Taming polarized fitting: BLINEX-Pcomp with asymmetric risk penalty for robust Pcomp classification 驯服极化拟合:具有不对称风险惩罚的BLINEX-Pcomp稳健Pcomp分类
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-09 DOI: 10.1016/j.neunet.2026.108580
Long Tang , Xin Si , Yingjie Tian , Panos M Pardalos
As a novel paradigm for learning with inexact supervision, Pcomp classification reduces the annotation costs of training a binary classifier by using ordered pairwise samples without requiring precise labels. However, existing methods fail to fully account for sign differences in empirical risk at the level of individual sample pairs, resulting in polarized fitting where the risks of overfitting and underfitting coexist. Actually, positive and negative empirical risks indicate varying degrees of training difficulty, necessitating differentiated treatments. In this work, we propose a BLINEX-Pcomp model that employs a bounded linear-exponential function to impose distinct penalties on positive and negative risks for each sample pair. The BLINEX-Pcomp model dynamically shifts the training focus toward challenging sample pairs, well balancing pairwise-level risks of overfitting and underfitting. Additionally, a multi-view version of BLINEX-Pcomp (MV-BLINEX-Pcomp) is developed to further enhance performance by integrating multi-view features. We have theoretically verified that MV-BLINEX-Pcomp degrades to BLINEX-Pcomp when only a single view of features is available. A dual-stage solver is designed to train the MV-BLINEX-Pcomp model. Exciting numerical results from comparative experiments validate the effectiveness of our methods in tackling Pcomp classification.
作为一种新的非精确监督学习范式,Pcomp分类通过使用有序成对样本而不需要精确标记来减少训练二分类器的标注成本。然而,现有方法未能充分考虑个体样本对水平上经验风险的符号差异,导致过度拟合和欠拟合风险并存的极化拟合。实际上,正、负经验风险表明了不同程度的训练难度,需要区别对待。在这项工作中,我们提出了一个BLINEX-Pcomp模型,该模型采用有界线性指数函数对每个样本对的正风险和负风险施加不同的惩罚。BLINEX-Pcomp模型动态地将训练重点转向具有挑战性的样本对,很好地平衡了过拟合和欠拟合的成对水平风险。此外,还开发了多视图版本的BLINEX-Pcomp (MV-BLINEX-Pcomp),通过集成多视图功能进一步提高性能。我们已经从理论上验证了当只有一个特征视图可用时,MV-BLINEX-Pcomp会降级为BLINEX-Pcomp。设计了一种双级求解器来训练MV-BLINEX-Pcomp模型。对比实验的令人振奋的数值结果验证了我们的方法在处理Pcomp分类方面的有效性。
{"title":"Taming polarized fitting: BLINEX-Pcomp with asymmetric risk penalty for robust Pcomp classification","authors":"Long Tang ,&nbsp;Xin Si ,&nbsp;Yingjie Tian ,&nbsp;Panos M Pardalos","doi":"10.1016/j.neunet.2026.108580","DOIUrl":"10.1016/j.neunet.2026.108580","url":null,"abstract":"<div><div>As a novel paradigm for learning with inexact supervision, Pcomp classification reduces the annotation costs of training a binary classifier by using ordered pairwise samples without requiring precise labels. However, existing methods fail to fully account for sign differences in empirical risk at the level of individual sample pairs, resulting in polarized fitting where the risks of overfitting and underfitting coexist. Actually, positive and negative empirical risks indicate varying degrees of training difficulty, necessitating differentiated treatments. In this work, we propose a BLINEX-Pcomp model that employs a bounded linear-exponential function to impose distinct penalties on positive and negative risks for each sample pair. The BLINEX-Pcomp model dynamically shifts the training focus toward challenging sample pairs, well balancing pairwise-level risks of overfitting and underfitting. Additionally, a multi-view version of BLINEX-Pcomp (MV-BLINEX-Pcomp) is developed to further enhance performance by integrating multi-view features. We have theoretically verified that MV-BLINEX-Pcomp degrades to BLINEX-Pcomp when only a single view of features is available. A dual-stage solver is designed to train the MV-BLINEX-Pcomp model. Exciting numerical results from comparative experiments validate the effectiveness of our methods in tackling Pcomp classification.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108580"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PMNO: A novel physics guided multi-step neural operator predictor for partial differential equations PMNO:一种新的物理导向多步神经算子预测偏微分方程
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-13 DOI: 10.1016/j.neunet.2026.108593
Jin Song , Kenji Kawaguchi , Zhenya Yan
Neural operators, which aim to approximate mappings between infinite-dimensional function spaces, have been widely applied in the simulation and prediction of physical systems. However, the limited representational capacity of network architectures, combined with their heavy reliance on large-scale data, often hinder effective training and result in poor extrapolation performance. In this paper, inspired by traditional numerical methods, we propose a novel Physics guided Multi-step Neural Operator (PMNO) architecture to address these challenges in long-horizon prediction of complex physical systems. Distinct from general operator learning methods, the PMNO framework replaces the single-step input with multi-step historical data in the forward pass and introduces an implicit time-stepping scheme based on the Backward Differentiation Formula (BDF) during backpropagation. This design not only strengthens the model’s extrapolation capacity but also facilitates more efficient and stable training with fewer data samples, especially for long-term predictions. Meanwhile, a causal training strategy is employed to circumvent the need for multi-stage training and to ensure efficient end-to-end optimization. The neural operator architecture possesses resolution-invariant properties, enabling the trained model to perform fast extrapolation on arbitrary spatial resolutions. We demonstrate the superior predictive performance of PMNO predictor across a diverse range of physical systems, including 2D linear system, modeling over irregular domain, complex-valued wave dynamics, and reaction-diffusion processes. Depending on the specific problem setting, various neural operator architectures, including FNO, DeepONet, and their variants, can be seamlessly integrated into the PMNO framework.
神经算子以逼近无限维函数空间之间的映射为目的,在物理系统的模拟和预测中得到了广泛的应用。然而,网络架构有限的表示能力,加上它们对大规模数据的严重依赖,往往会阻碍有效的训练,并导致较差的外推性能。本文在传统数值方法的启发下,提出了一种新的物理引导多步神经算子(PMNO)架构,以解决复杂物理系统长期预测中的这些挑战。与一般算子学习方法不同,PMNO框架在前传过程中用多步历史数据代替单步输入,并在反向传播过程中引入基于后向微分公式(BDF)的隐式时间步进方案。这种设计不仅增强了模型的外推能力,而且可以在更少的数据样本下实现更高效和稳定的训练,特别是对于长期预测。同时,采用因果训练策略,避免了多阶段训练的需要,保证了高效的端到端优化。神经算子结构具有分辨率不变性,使训练模型能够在任意空间分辨率上进行快速外推。我们证明了PMNO预测器在各种物理系统中的卓越预测性能,包括二维线性系统、不规则域建模、复值波动动力学和反应扩散过程。根据具体的问题设置,各种神经算子架构,包括FNO、DeepONet及其变体,可以无缝集成到PMNO框架中。
{"title":"PMNO: A novel physics guided multi-step neural operator predictor for partial differential equations","authors":"Jin Song ,&nbsp;Kenji Kawaguchi ,&nbsp;Zhenya Yan","doi":"10.1016/j.neunet.2026.108593","DOIUrl":"10.1016/j.neunet.2026.108593","url":null,"abstract":"<div><div>Neural operators, which aim to approximate mappings between infinite-dimensional function spaces, have been widely applied in the simulation and prediction of physical systems. However, the limited representational capacity of network architectures, combined with their heavy reliance on large-scale data, often hinder effective training and result in poor extrapolation performance. In this paper, inspired by traditional numerical methods, we propose a novel Physics guided Multi-step Neural Operator (PMNO) architecture to address these challenges in long-horizon prediction of complex physical systems. Distinct from general operator learning methods, the PMNO framework replaces the single-step input with multi-step historical data in the forward pass and introduces an implicit time-stepping scheme based on the Backward Differentiation Formula (BDF) during backpropagation. This design not only strengthens the model’s extrapolation capacity but also facilitates more efficient and stable training with fewer data samples, especially for long-term predictions. Meanwhile, a causal training strategy is employed to circumvent the need for multi-stage training and to ensure efficient end-to-end optimization. The neural operator architecture possesses resolution-invariant properties, enabling the trained model to perform fast extrapolation on arbitrary spatial resolutions. We demonstrate the superior predictive performance of PMNO predictor across a diverse range of physical systems, including 2D linear system, modeling over irregular domain, complex-valued wave dynamics, and reaction-diffusion processes. Depending on the specific problem setting, various neural operator architectures, including FNO, DeepONet, and their variants, can be seamlessly integrated into the PMNO framework.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108593"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Source Temporal-Depth fusion for robust end-to-End visual odometry 鲁棒端到端视觉里程计的多源时间深度融合
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-17 DOI: 10.1016/j.neunet.2026.108598
Sihang Zhang , Congqi Cao , Qiang Gao , Ganchao Liu
End-to-end visual odometry models have recently achieved localization accuracy on par with conventional techniques, while effectively reducing the occurrence of catastrophic failures. However, the relevant models cannot leverage the complete time-series data for pose adjustment and optimization. Moreover, these models are limited to using joint depth prediction tasks merely as a means of scale constraint, lacking effective utilization of depth information. In this paper, we propose an end-to-end multi-source visual odometry (MVO) model that dynamically integrates the key components of hybrid visual odometry pipelines into a unified, learnable deep framework. Specifically, we propose TimePoseNet to model the mapping relationship from time to pose, capturing temporal dependencies across the entire sequence. Additionally, a wavelet convolutional attention mechanism is employed to extract global depth information from the depth map, which is then directly embedded into the pose features to dynamically constrain scale ambiguity. Furthermore, temporal and depth cues are jointly incorporated into the post-processing stage of pose estimation. The proposed method attains state-of-the-art performance on both the KITTI benchmark and the newly introduced UAV-2025 dataset, while preserving computational efficiency during inference.
端到端视觉里程计模型最近实现了与传统技术相当的定位精度,同时有效地减少了灾难性故障的发生。然而,相关模型无法利用完整的时间序列数据进行位姿调整和优化。此外,这些模型仅将联合深度预测任务作为一种尺度约束手段,缺乏对深度信息的有效利用。在本文中,我们提出了一个端到端多源视觉里程计(MVO)模型,该模型将混合视觉里程计管道的关键组件动态集成到一个统一的、可学习的深度框架中。具体来说,我们提出了TimePoseNet来对从时间到姿势的映射关系进行建模,从而捕获整个序列的时间依赖性。此外,采用小波卷积注意机制从深度图中提取全局深度信息,然后将其直接嵌入到姿态特征中,动态约束尺度模糊度。此外,在姿态估计的后处理阶段,将时间线索和深度线索结合起来。所提出的方法在KITTI基准和新引入的UAV-2025数据集上都获得了最先进的性能,同时在推理过程中保持了计算效率。
{"title":"Multi-Source Temporal-Depth fusion for robust end-to-End visual odometry","authors":"Sihang Zhang ,&nbsp;Congqi Cao ,&nbsp;Qiang Gao ,&nbsp;Ganchao Liu","doi":"10.1016/j.neunet.2026.108598","DOIUrl":"10.1016/j.neunet.2026.108598","url":null,"abstract":"<div><div>End-to-end visual odometry models have recently achieved localization accuracy on par with conventional techniques, while effectively reducing the occurrence of catastrophic failures. However, the relevant models cannot leverage the complete time-series data for pose adjustment and optimization. Moreover, these models are limited to using joint depth prediction tasks merely as a means of scale constraint, lacking effective utilization of depth information. In this paper, we propose an end-to-end multi-source visual odometry (MVO) model that dynamically integrates the key components of hybrid visual odometry pipelines into a unified, learnable deep framework. Specifically, we propose TimePoseNet to model the mapping relationship from time to pose, capturing temporal dependencies across the entire sequence. Additionally, a wavelet convolutional attention mechanism is employed to extract global depth information from the depth map, which is then directly embedded into the pose features to dynamically constrain scale ambiguity. Furthermore, temporal and depth cues are jointly incorporated into the post-processing stage of pose estimation. The proposed method attains state-of-the-art performance on both the KITTI benchmark and the newly introduced UAV-2025 dataset, while preserving computational efficiency during inference.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108598"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CBAM-ST-GCN: An enhanced DRL-based end-to-end visual navigation framework for mobile robot CBAM-ST-GCN:一种基于drl的移动机器人端到端视觉导航框架
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-20 DOI: 10.1016/j.neunet.2026.108622
Mingyang Xie, Wei Yu, Huanyu Jin, Wei Li, Xin Chen
Visual-based navigation for mobile robot poses significant challenges due to limited visual perception and the presence of unforeseen dynamic obstacles. Deep reinforcement learning (DRL) provides an end-to-end solution by directly mapping raw sensor data to control commands, offering high adaptability and reduced reliance on handcrafted rules. However, high-dimensional visual inputs and the non-stationarity introduced by dynamic obstacles easily make the policy learning of DRL difficult to convergent and unstable. In this paper, an enhanced end-to-end visual navigation framework is proposed for mobile robot operating in dynamic environments, denoted as CBAM-ST-GCN. A convolutional block attention module (CBAM) is introduced into the framework to enhance visual perception by assigning attention weights across spatial and temporal dimensions. Furthermore, a spatio-temporal graph convolutional network (ST-GCN) is designed to capture the behavior features of moving obstacles. In addition, a velocity obstacle (VO) method-based penalty term is incorporated into the reward function for the enhancement of collision avoidance. Extensive simulation results demonstrate that the proposed method achieves superior success rates and significantly higher convergence speed. Real-world experiments further validate the effectiveness and adaptability of the proposed approach in practical scenarios.
由于视觉感知能力的限制和不可预见的动态障碍物的存在,移动机器人基于视觉的导航面临着巨大的挑战。深度强化学习(DRL)通过直接将原始传感器数据映射到控制命令,提供了一个端到端解决方案,具有高适应性,减少了对手工规则的依赖。然而,高维视觉输入和动态障碍物引入的非平稳性容易使DRL的策略学习难以收敛和不稳定。针对动态环境下的移动机器人,本文提出了一种增强的端到端视觉导航框架,记作CBAM-ST-GCN。在框架中引入了卷积块注意模块(CBAM),通过在空间和时间维度上分配注意权重来增强视觉感知。此外,设计了一个时空图卷积网络(ST-GCN)来捕捉移动障碍物的行为特征。此外,在奖励函数中引入基于速度障碍(VO)方法的惩罚项,增强了避碰性能。大量的仿真结果表明,该方法具有较高的成功率和显著的收敛速度。现实世界的实验进一步验证了该方法在实际场景中的有效性和适应性。
{"title":"CBAM-ST-GCN: An enhanced DRL-based end-to-end visual navigation framework for mobile robot","authors":"Mingyang Xie,&nbsp;Wei Yu,&nbsp;Huanyu Jin,&nbsp;Wei Li,&nbsp;Xin Chen","doi":"10.1016/j.neunet.2026.108622","DOIUrl":"10.1016/j.neunet.2026.108622","url":null,"abstract":"<div><div>Visual-based navigation for mobile robot poses significant challenges due to limited visual perception and the presence of unforeseen dynamic obstacles. Deep reinforcement learning (DRL) provides an end-to-end solution by directly mapping raw sensor data to control commands, offering high adaptability and reduced reliance on handcrafted rules. However, high-dimensional visual inputs and the non-stationarity introduced by dynamic obstacles easily make the policy learning of DRL difficult to convergent and unstable. In this paper, an enhanced end-to-end visual navigation framework is proposed for mobile robot operating in dynamic environments, denoted as CBAM-ST-GCN. A convolutional block attention module (CBAM) is introduced into the framework to enhance visual perception by assigning attention weights across spatial and temporal dimensions. Furthermore, a spatio-temporal graph convolutional network (ST-GCN) is designed to capture the behavior features of moving obstacles. In addition, a velocity obstacle (VO) method-based penalty term is incorporated into the reward function for the enhancement of collision avoidance. Extensive simulation results demonstrate that the proposed method achieves superior success rates and significantly higher convergence speed. Real-world experiments further validate the effectiveness and adaptability of the proposed approach in practical scenarios.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108622"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MambaFPN: A SSM-based feature pyramid network for object detection MambaFPN:一种基于ssm的目标检测特征金字塔网络
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-09 DOI: 10.1016/j.neunet.2026.108544
Le Liang , Cheng Wang , Lefei Zhang
Object detection is a fundamental task in computer vision, aiming to localize and classify objects within images. Feature pyramid networks (FPNs) play a crucial role in modern object detectors by constructing hierarchical multi-scale feature maps to effectively handle objects of varying sizes. However, most existing advanced FPN methods rely heavily on convolutional neural networks (CNNs), which struggle to capture global context information. To address this limitation, we propose leveraging vision mamba blocks to enhance global modeling capabilities. The vanilla vision mamba block, through its state space mechanism, enables global context modeling for every spatial pixel within a single feature map. Building on this, we first use vision mamba blocks to extract global information from individual feature maps in the hierarchy. Subsequently, additional vision mamba blocks facilitate inter-scale information exchange among multi-scale feature maps, ensuring comprehensive global context integration. The proposed method, termed MambaFPN, significantly enhances object detector performance. For instance, it improves the Average Precision (AP) of vanilla FPN from 38.6 to 39.4, with fewer parameters. This demonstrates the effectiveness and efficiency of MambaFPN in advancing object detection.
目标检测是计算机视觉中的一项基本任务,旨在对图像中的目标进行定位和分类。特征金字塔网络(FPNs)通过构建分层多尺度特征映射来有效处理不同大小的目标,在现代目标检测器中起着至关重要的作用。然而,大多数现有的先进FPN方法严重依赖于卷积神经网络(cnn),而卷积神经网络难以捕获全局上下文信息。为了解决这个限制,我们建议利用视觉曼巴块来增强全局建模能力。香草视觉曼巴块通过其状态空间机制,为单个特征图中的每个空间像素实现全局上下文建模。在此基础上,我们首先使用视觉曼巴块从层次结构中的单个特征映射中提取全局信息。随后,额外的视觉曼巴块促进了多尺度特征地图之间的尺度间信息交换,确保了全面的全球背景整合。所提出的方法,称为MambaFPN,显著提高了目标检测器的性能。例如,它使用更少的参数将vanilla FPN的平均精度(AP)从38.6提高到39.4。这证明了MambaFPN在推进目标检测方面的有效性和效率。
{"title":"MambaFPN: A SSM-based feature pyramid network for object detection","authors":"Le Liang ,&nbsp;Cheng Wang ,&nbsp;Lefei Zhang","doi":"10.1016/j.neunet.2026.108544","DOIUrl":"10.1016/j.neunet.2026.108544","url":null,"abstract":"<div><div>Object detection is a fundamental task in computer vision, aiming to localize and classify objects within images. Feature pyramid networks (FPNs) play a crucial role in modern object detectors by constructing hierarchical multi-scale feature maps to effectively handle objects of varying sizes. However, most existing advanced FPN methods rely heavily on convolutional neural networks (CNNs), which struggle to capture global context information. To address this limitation, we propose leveraging vision mamba blocks to enhance global modeling capabilities. The vanilla vision mamba block, through its state space mechanism, enables global context modeling for every spatial pixel within a single feature map. Building on this, we first use vision mamba blocks to extract global information from individual feature maps in the hierarchy. Subsequently, additional vision mamba blocks facilitate inter-scale information exchange among multi-scale feature maps, ensuring comprehensive global context integration. The proposed method, termed MambaFPN, significantly enhances object detector performance. For instance, it improves the Average Precision (AP) of vanilla FPN from 38.6 to 39.4, with fewer parameters. This demonstrates the effectiveness and efficiency of MambaFPN in advancing object detection.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108544"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hippocampus-centered structural covariance network reorganization in Alzheimer’s disease: An individualized graph-based biomarker validated by machine learning 阿尔茨海默病中以海马体为中心的结构协方差网络重组:通过机器学习验证的个体化基于图的生物标志物
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-08 DOI: 10.1016/j.neunet.2026.108542
Weiye Lu , Qian Gong , Yuna Chen , Shijun Qiu , Jie An
Alzheimer’s disease (AD) is characterized by progressive brain network disintegration, yet quantifying this process at an individual level remains challenging. This study explores the potential of an individualized differential structural covariance network (IDSCN) as a graph theory-based biomarker to capture disease-specific network reorganization. We found that throughout the AD spectrum, significant progressive atrophy occurred in multiple brain regions, especially the hippocampus. At the same time, the brain underwent a profound structural covariant reorganization, and this reorganization was significantly centered on the hippocampus. Graph theory analysis revealed a significant enhancement in nodal strength and nodal efficiency across widespread brain regions, with the hippocampus, amygdala, middle temporal gyrus, and entorhinal cortex emerging as core hubs of pathological impact. Importantly, betweenness centrality selectively increased only in the bilateral hippocampus, highlighting their critical role as bridges in the pathological propagation network. Machine learning validation confirmed that this individualized network biomarker performs excellently in distinguishing AD patients from cognitively normal individuals, demonstrates comparable efficacy to traditional morphological models in capturing early disease-related changes, and shows potential in differentiating between mild cognitive impairment converters and non-converters. Our study establishes the hippocampus-centered IDSCN as an effective, individualized graph theory-based biomarker, providing new insights into the network pathophysiology of AD and holding significant potential for early diagnosis and prognostic stratification.
阿尔茨海默病(AD)的特点是进行性脑网络解体,但在个体水平上量化这一过程仍然具有挑战性。本研究探讨了个体化差异结构协方差网络(IDSCN)作为基于图论的生物标志物捕捉疾病特异性网络重组的潜力。我们发现,在整个阿尔茨海默病谱系中,多个大脑区域,尤其是海马体,都出现了显著的进行性萎缩。与此同时,大脑发生了深刻的结构协变重组,这种重组主要集中在海马体。图论分析显示,在广泛的大脑区域中,节点强度和节点效率显著增强,海马、杏仁核、中颞回和内鼻皮层成为病理影响的核心枢纽。重要的是,中间性中心性仅在双侧海马中选择性增加,突出了它们在病理传播网络中作为桥梁的关键作用。机器学习验证证实,这种个性化的网络生物标志物在区分AD患者和认知正常个体方面表现出色,在捕捉早期疾病相关变化方面表现出与传统形态学模型相当的功效,并显示出区分轻度认知障碍转化者和非转化者的潜力。我们的研究确立了以海马体为中心的IDSCN是一种有效的、个性化的基于图论的生物标志物,为阿尔茨海默病的网络病理生理学提供了新的见解,并在早期诊断和预后分层方面具有重要潜力。
{"title":"Hippocampus-centered structural covariance network reorganization in Alzheimer’s disease: An individualized graph-based biomarker validated by machine learning","authors":"Weiye Lu ,&nbsp;Qian Gong ,&nbsp;Yuna Chen ,&nbsp;Shijun Qiu ,&nbsp;Jie An","doi":"10.1016/j.neunet.2026.108542","DOIUrl":"10.1016/j.neunet.2026.108542","url":null,"abstract":"<div><div>Alzheimer’s disease (AD) is characterized by progressive brain network disintegration, yet quantifying this process at an individual level remains challenging. This study explores the potential of an individualized differential structural covariance network (IDSCN) as a graph theory-based biomarker to capture disease-specific network reorganization. We found that throughout the AD spectrum, significant progressive atrophy occurred in multiple brain regions, especially the hippocampus. At the same time, the brain underwent a profound structural covariant reorganization, and this reorganization was significantly centered on the hippocampus. Graph theory analysis revealed a significant enhancement in nodal strength and nodal efficiency across widespread brain regions, with the hippocampus, amygdala, middle temporal gyrus, and entorhinal cortex emerging as core hubs of pathological impact. Importantly, betweenness centrality selectively increased only in the bilateral hippocampus, highlighting their critical role as bridges in the pathological propagation network. Machine learning validation confirmed that this individualized network biomarker performs excellently in distinguishing AD patients from cognitively normal individuals, demonstrates comparable efficacy to traditional morphological models in capturing early disease-related changes, and shows potential in differentiating between mild cognitive impairment converters and non-converters. Our study establishes the hippocampus-centered IDSCN as an effective, individualized graph theory-based biomarker, providing new insights into the network pathophysiology of AD and holding significant potential for early diagnosis and prognostic stratification.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108542"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145941312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAGCN: A syntactic aware multi-branch graph attention network with structural bias for aspect sentiment triplet extraction 面向面向情感三元组提取的结构偏差多分支图注意网络。
IF 6.3 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-06-01 Epub Date: 2026-01-13 DOI: 10.1016/j.neunet.2026.108596
Xin Xiao , Bin Gao , Zelong Su , Linlin Li , Shutian Liu , Zhengjun Liu
Aspect-based sentiment triplet extraction (ASTE), a newly developed and complex subtask within aspect-based sentiment analysis, focuses on detecting aspect terms, opinion terms, and establishing sentiment polarity from human language, thereby extracting triplets composed of these three elements. Although numerous methods have been developed in previous research to tackle this task, these ASTE methods exhibit weak interactions in constructing contextual representations and overlook the syntactic relationships between aspect terms and opinion terms. Therefore, this paper proposes a syntax-aware multi-branch graph attention network to address this issue. We have designed an efficient approach that integrates new structural biases into pre-trained language models through adapters to enhance the original mappings in self-attention, significantly reducing the parameter requirements and achieving lower latency. Simultaneously, we have devised a syntax-aware attention mechanism that not only discerns edges with varying dependency types as well as those with identical types, learning the representation of each edge in the graph relying on the dependency types of neighboring edges, thereby enabling more accurate graph propagation. Finally, we have designed a special fusion interaction layer that achieves the final text representation by merging different branch features with varying weights. Through a range of tests performed on four widely accessible datasets, it was demonstrated that the introduction of structural bias adapters is both effective and efficient. The proposed method improved the average F1 score by up to 4.11% compared to all baseline models, while also exhibiting good interpretability. Additionally, the experimental results highlighted the robustness and effectiveness of SAGCN, significantly outperforming the compared state-of-the-art baseline models.
基于方面的情感三联体提取(ASTE)是基于方面的情感分析中一个新发展的复杂子任务,其重点是从人类语言中检测方面术语、观点术语和建立情感极性,从而提取由这三个元素组成的三联体。尽管在之前的研究中已经开发了许多方法来解决这一任务,但这些方法在构建上下文表示时表现出弱交互,并且忽略了方面术语和意见术语之间的句法关系。为此,本文提出了一种语法感知的多分支图关注网络来解决这一问题。我们设计了一种有效的方法,通过适配器将新的结构偏差集成到预训练的语言模型中,以增强自注意中的原始映射,显著降低参数要求并实现更低的延迟。同时,我们设计了一种语法感知的注意机制,不仅可以区分不同依赖类型的边和相同类型的边,还可以根据相邻边的依赖类型学习图中每条边的表示,从而实现更准确的图传播。最后,我们设计了一个特殊的融合交互层,通过合并不同权重的分支特征来实现最终的文本表示。通过在四个可广泛访问的数据集上进行的一系列测试,证明了引入结构偏差适配器既有效又高效。与所有基线模型相比,该方法将F1平均得分提高了4.11%,同时也具有良好的可解释性。此外,实验结果突出了SAGCN的鲁棒性和有效性,显著优于最先进的基线模型。
{"title":"SAGCN: A syntactic aware multi-branch graph attention network with structural bias for aspect sentiment triplet extraction","authors":"Xin Xiao ,&nbsp;Bin Gao ,&nbsp;Zelong Su ,&nbsp;Linlin Li ,&nbsp;Shutian Liu ,&nbsp;Zhengjun Liu","doi":"10.1016/j.neunet.2026.108596","DOIUrl":"10.1016/j.neunet.2026.108596","url":null,"abstract":"<div><div>Aspect-based sentiment triplet extraction (ASTE), a newly developed and complex subtask within aspect-based sentiment analysis, focuses on detecting aspect terms, opinion terms, and establishing sentiment polarity from human language, thereby extracting triplets composed of these three elements. Although numerous methods have been developed in previous research to tackle this task, these ASTE methods exhibit weak interactions in constructing contextual representations and overlook the syntactic relationships between aspect terms and opinion terms. Therefore, this paper proposes a syntax-aware multi-branch graph attention network to address this issue. We have designed an efficient approach that integrates new structural biases into pre-trained language models through adapters to enhance the original mappings in self-attention, significantly reducing the parameter requirements and achieving lower latency. Simultaneously, we have devised a syntax-aware attention mechanism that not only discerns edges with varying dependency types as well as those with identical types, learning the representation of each edge in the graph relying on the dependency types of neighboring edges, thereby enabling more accurate graph propagation. Finally, we have designed a special fusion interaction layer that achieves the final text representation by merging different branch features with varying weights. Through a range of tests performed on four widely accessible datasets, it was demonstrated that the introduction of structural bias adapters is both effective and efficient. The proposed method improved the average F1 score by up to 4.11% compared to all baseline models, while also exhibiting good interpretability. Additionally, the experimental results highlighted the robustness and effectiveness of SAGCN, significantly outperforming the compared state-of-the-art baseline models.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"198 ","pages":"Article 108596"},"PeriodicalIF":6.3,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146031529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1