首页 > 最新文献

Complex & Intelligent Systems最新文献

英文 中文
Enhancing traffic flow prediction through multi-view attention mechanism and dilated convolutional networks 通过多视角注意机制和扩展卷积网络增强交通流预测
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-15 DOI: 10.1007/s40747-025-02146-7
Wei Li, Hao Wei, Xin Liu, Jialin Liu, Dazhi Zhan, Xiao Han, Wei Tao
Accurate traffic flow forecasting serves as a cornerstone for intelligent transportation systems, enabling proactive accident prevention and metropolitan mobility optimization. However, existing approaches face fundamental limitations in modeling the spatiotemporal heterogeneity of traffic dynamics, particularly in simultaneously addressing (1) the decaying significance of temporal dependencies across input sequences and prediction horizons, (2) multi-scale spatial interactions spanning local congestion patterns and global functional correlations, and (3) inter-sample temporal variance in evolving traffic states. To address these limitations, this paper proposes MVA-DCNet (Multi-View Attention Dilated Convolutional Network), a novel deep learning architecture incorporating a multidimensional temporal analysis framework that systematically examines temporal influence mechanisms through three complementary perspectives: inter-sample variance, intra-sequence temporal importance, and output sequence temporal propagation. The proposed model systematically addresses temporal data heterogeneity through three innovative mechanisms: variance-aware data augmentation, adaptive temporal attention, and decaying loss weighting. For enhanced spatial correlation modeling, we develop a dilated convolutional architecture with enhanced receptive field coverage and multi-scale spatial pattern recognition capabilities. Empirical validation on two urban traffic datasets demonstrates superior efficacy in capturing complex spatiotemporal evolution patterns, achieving relative reductions of 12.7% and 9.3% in Root Mean Square Error (RMSE) respectively compared with state-of-the-art benchmarks.
准确的交通流量预测是智能交通系统的基石,可以实现主动事故预防和城市交通优化。然而,现有的方法在模拟交通动态的时空异质性方面面临着根本性的局限性,特别是在同时解决(1)输入序列和预测范围之间的时间依赖性的衰减意义,(2)跨越局部拥堵模式和全局功能相关性的多尺度空间相互作用,以及(3)不断变化的交通状态的样本间时间方差。为了解决这些限制,本文提出了MVA-DCNet(多视图注意扩展卷积网络),这是一种新型的深度学习架构,包含一个多维时间分析框架,通过三个互补的角度系统地检查时间影响机制:样本间方差、序列内时间重要性和输出序列时间传播。该模型通过三种创新机制系统地解决了时间数据的异质性:方差感知数据增强、自适应时间关注和衰减损失加权。为了增强空间相关建模,我们开发了一个扩展的卷积架构,具有增强的感受野覆盖和多尺度空间模式识别能力。在两个城市交通数据集上的实证验证表明,该方法在捕获复杂时空演化模式方面具有卓越的效果,与最先进的基准相比,均方根误差(RMSE)分别相对降低了12.7%和9.3%。
{"title":"Enhancing traffic flow prediction through multi-view attention mechanism and dilated convolutional networks","authors":"Wei Li, Hao Wei, Xin Liu, Jialin Liu, Dazhi Zhan, Xiao Han, Wei Tao","doi":"10.1007/s40747-025-02146-7","DOIUrl":"https://doi.org/10.1007/s40747-025-02146-7","url":null,"abstract":"Accurate traffic flow forecasting serves as a cornerstone for intelligent transportation systems, enabling proactive accident prevention and metropolitan mobility optimization. However, existing approaches face fundamental limitations in modeling the spatiotemporal heterogeneity of traffic dynamics, particularly in simultaneously addressing (1) the decaying significance of temporal dependencies across input sequences and prediction horizons, (2) multi-scale spatial interactions spanning local congestion patterns and global functional correlations, and (3) inter-sample temporal variance in evolving traffic states. To address these limitations, this paper proposes MVA-DCNet (Multi-View Attention Dilated Convolutional Network), a novel deep learning architecture incorporating a multidimensional temporal analysis framework that systematically examines temporal influence mechanisms through three complementary perspectives: inter-sample variance, intra-sequence temporal importance, and output sequence temporal propagation. The proposed model systematically addresses temporal data heterogeneity through three innovative mechanisms: variance-aware data augmentation, adaptive temporal attention, and decaying loss weighting. For enhanced spatial correlation modeling, we develop a dilated convolutional architecture with enhanced receptive field coverage and multi-scale spatial pattern recognition capabilities. Empirical validation on two urban traffic datasets demonstrates superior efficacy in capturing complex spatiotemporal evolution patterns, achieving relative reductions of 12.7% and 9.3% in Root Mean Square Error (RMSE) respectively compared with state-of-the-art benchmarks.","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"17 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145752822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ReqNet: an LLM-driven computational framework for automated requirements extraction from unstructured documents ReqNet:一个llm驱动的计算框架,用于从非结构化文档中自动提取需求
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-15 DOI: 10.1007/s40747-025-02143-w
Summra Saleem, Muhammad Nabeel Asim, Andreas Dengel
Within software development life-cycle, requirements guide the entire development process from inception to completion by ensuring alignment between stakeholder expectations and the final product. Requirements extraction from miscellaneous information is a challenging and complex task. Manual extraction of requirements is not only prone to human error but also contributes to increased project costs and delayed project timelines. To automate the requirement extraction process, researchers have investigated the potential of deep learning architectures, large language models (LLM) and generative language models such as ChatGPT and Gemini. However, the performance of requirements extraction could be further enhanced through the development of predictive pipelines by utilizing the combined potential of language models and deep learning architectures. To develop a powerful AI application for requirements extraction by utilizing the combined potential of LLMs and DL architectures, this study presents ReqNet framework. The framework encompasses 7 most widely used LLMs variants (small, large, Xlarge, XXlarge) and 2 DL architectures (LSTM, GRU). The framework facilitates the development of three distinct types predictive pipelines, namely standalone LLMs, LLMs + external classifiers and an ensemble of multiple LLMs representation + external classifiers. Extensive experimentation of 48 predictive pipelines across 2 public core datasets and 1 independent test set, demonstrates that predictive pipelines made up from LLMs and DL architectures generally exhibited superior performance compared to pipelines solely reliant on LLMs. In addition, a ensemble of three distinct LLMs (ALBERT, BERT and XLNet) and LSTM classifier achieved a 3% improvement in F1-score over state-of-the-art predictors on the PURE dataset, a 10% improvement on the Dronology dataset and a 3% improvement on the RFI independent test set.
在软件开发生命周期中,需求通过确保涉众期望和最终产品之间的一致性来指导从开始到完成的整个开发过程。从繁杂的信息中提取需求是一项具有挑战性和复杂性的任务。手动提取需求不仅容易出现人为错误,而且还会增加项目成本和延迟项目时间表。为了自动化需求提取过程,研究人员已经研究了深度学习架构、大型语言模型(LLM)和生成语言模型(如ChatGPT和Gemini)的潜力。然而,需求提取的性能可以通过利用语言模型和深度学习架构的组合潜力来开发预测管道来进一步增强。为了利用llm和DL架构的组合潜力开发强大的需求提取AI应用程序,本研究提出了ReqNet框架。该框架包含7种最广泛使用的llm变体(small, large, Xlarge, XXlarge)和2种DL架构(LSTM, GRU)。该框架促进了三种不同类型预测管道的开发,即独立的llm、llm +外部分类器和多个llm表示+外部分类器的集成。在2个公共核心数据集和1个独立测试集上对48个预测管道进行了广泛的实验,结果表明,与仅依赖于llm的管道相比,由llm和DL架构组成的预测管道通常表现出更好的性能。此外,三个不同的llm (ALBERT、BERT和XLNet)和LSTM分类器的集合在PURE数据集上的f1分数比最先进的预测器提高了3%,在Dronology数据集上提高了10%,在RFI独立测试集上提高了3%。
{"title":"ReqNet: an LLM-driven computational framework for automated requirements extraction from unstructured documents","authors":"Summra Saleem, Muhammad Nabeel Asim, Andreas Dengel","doi":"10.1007/s40747-025-02143-w","DOIUrl":"https://doi.org/10.1007/s40747-025-02143-w","url":null,"abstract":"Within software development life-cycle, requirements guide the entire development process from inception to completion by ensuring alignment between stakeholder expectations and the final product. Requirements extraction from miscellaneous information is a challenging and complex task. Manual extraction of requirements is not only prone to human error but also contributes to increased project costs and delayed project timelines. To automate the requirement extraction process, researchers have investigated the potential of deep learning architectures, large language models (LLM) and generative language models such as ChatGPT and Gemini. However, the performance of requirements extraction could be further enhanced through the development of predictive pipelines by utilizing the combined potential of language models and deep learning architectures. To develop a powerful AI application for requirements extraction by utilizing the combined potential of LLMs and DL architectures, this study presents ReqNet framework. The framework encompasses 7 most widely used LLMs variants (small, large, Xlarge, XXlarge) and 2 DL architectures (LSTM, GRU). The framework facilitates the development of three distinct types predictive pipelines, namely standalone LLMs, LLMs + external classifiers and an ensemble of multiple LLMs representation + external classifiers. Extensive experimentation of 48 predictive pipelines across 2 public core datasets and 1 independent test set, demonstrates that predictive pipelines made up from LLMs and DL architectures generally exhibited superior performance compared to pipelines solely reliant on LLMs. In addition, a ensemble of three distinct LLMs (ALBERT, BERT and XLNet) and LSTM classifier achieved a 3% improvement in F1-score over state-of-the-art predictors on the PURE dataset, a 10% improvement on the Dronology dataset and a 3% improvement on the RFI independent test set.","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"1 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145753180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Seed perception learning for weakly supervised semantic segmentation 弱监督语义分割的种子感知学习
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-15 DOI: 10.1007/s40747-025-02152-9
Wanchun Sun, Shujia Li, Xinyu Duan
The core challenge in image-level weakly supervised semantic segmentation lies in generating high-quality object localization maps from simple image labels. Class Activation Map (CAM) produced by existing methods commonly suffer from two major flaws: incomplete coverage of target regions and severe background interference. To address these issues, we present a CAM-native perception-optimization framework for weakly supervised semantic segmentation. First, design a CAM generation mechanism guided by image-level weak supervision, which refines activated regions via discriminative region enhancement and spatial noise suppression. This process promotes fine-grained pixel clustering and improves the completeness of object localization. Second, introduce a spatial cue generator to enhance the adaptability of class representations, coupled with an inter-class relation propagation module that explicitly models inter-class relationships to suppress erroneous activations and significantly reduce spatial noise. Additionally, incorporate a dynamic contrastive matching strategy to eliminate background activations closely associated with the target object, ultimately producing class activation maps that are both complete and compact. Extensive experiments on PASCAL VOC 2012 and MS COCO 2014 show that our method substantially outperforms existing weakly supervised approaches, validating the effectiveness of class-aware guidance and inter-class relational modeling in improving segmentation accuracy.
图像级弱监督语义分割的核心挑战在于如何从简单的图像标签生成高质量的目标定位图。现有方法生成的类激活图(Class Activation Map, CAM)存在两个主要缺陷:目标区域覆盖不完全和背景干扰严重。为了解决这些问题,我们提出了一个cam原生感知优化框架,用于弱监督语义分割。首先,设计图像级弱监督引导下的CAM生成机制,通过判别区域增强和空间噪声抑制来细化激活区域;该过程促进了细粒度的像素聚类,提高了目标定位的完整性。其次,引入空间线索生成器来增强类表示的适应性,再加上明确建模类间关系的类间关系传播模块,以抑制错误激活并显著降低空间噪声。此外,结合动态对比匹配策略来消除与目标对象密切相关的背景激活,最终生成既完整又紧凑的类激活映射。在PASCAL VOC 2012和MS COCO 2014上的大量实验表明,我们的方法大大优于现有的弱监督方法,验证了类感知引导和类间关系建模在提高分割精度方面的有效性。
{"title":"Seed perception learning for weakly supervised semantic segmentation","authors":"Wanchun Sun, Shujia Li, Xinyu Duan","doi":"10.1007/s40747-025-02152-9","DOIUrl":"https://doi.org/10.1007/s40747-025-02152-9","url":null,"abstract":"The core challenge in image-level weakly supervised semantic segmentation lies in generating high-quality object localization maps from simple image labels. Class Activation Map (CAM) produced by existing methods commonly suffer from two major flaws: incomplete coverage of target regions and severe background interference. To address these issues, we present a CAM-native perception-optimization framework for weakly supervised semantic segmentation. First, design a CAM generation mechanism guided by image-level weak supervision, which refines activated regions via discriminative region enhancement and spatial noise suppression. This process promotes fine-grained pixel clustering and improves the completeness of object localization. Second, introduce a spatial cue generator to enhance the adaptability of class representations, coupled with an inter-class relation propagation module that explicitly models inter-class relationships to suppress erroneous activations and significantly reduce spatial noise. Additionally, incorporate a dynamic contrastive matching strategy to eliminate background activations closely associated with the target object, ultimately producing class activation maps that are both complete and compact. Extensive experiments on PASCAL VOC 2012 and MS COCO 2014 show that our method substantially outperforms existing weakly supervised approaches, validating the effectiveness of class-aware guidance and inter-class relational modeling in improving segmentation accuracy.","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"148 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145752823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Llm-ga: A gradient-based multi-label adversarial attack by large language models Llm-ga:基于梯度的大型语言模型的多标签对抗性攻击
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-12 DOI: 10.1007/s40747-025-02184-1
Yujiang Liu, Yamin Hu, Zhijian Chen, Shiyin Wang, Wenjian Luo
{"title":"Llm-ga: A gradient-based multi-label adversarial attack by large language models","authors":"Yujiang Liu, Yamin Hu, Zhijian Chen, Shiyin Wang, Wenjian Luo","doi":"10.1007/s40747-025-02184-1","DOIUrl":"https://doi.org/10.1007/s40747-025-02184-1","url":null,"abstract":"","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"20 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145752824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MAAN: multi-scale atrous attention network for skin lesion segmentation 基于多尺度亚属性关注网络的皮肤病变分割
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-10 DOI: 10.1007/s40747-025-02186-z
Yang Lian, Ruizhi Han, Shiyuan Han, Defu Qiu, Jin Zhou
Skin cancer research is essential to finding new treatments and improving survival rates in computer-aided medicine. Within this research, the accurate segmentation of skin lesion images is an important step for both early diagnosis and personalized treatment strategies. However, while current popular Transformer-based models have achieved competitive segmentation results, they often ignore the computational complexity and the high costs associated with their training. In this paper, we propose a lightweight network, a multi-scale atrous attention network for skin lesion segmentation (MAAN). Firstly, we optimize the residual basic block by constructing a dual-path framework with both high and low-resolution paths, which reduces the number of parameters while maintaining effective feature extraction capability. Secondly, to better capture the information in the skin lesion images and further improve the model performance, we design an adaptive multi-scale atrous attention module at the final stage of the low-resolution path. The experiments conducted on the ISIC 2017 and ISIC2018 datasets show that the proposed model MAAN achieves mIoU of 85.20 and 85.67% respectively, outperforming recent MHorNet while maintaining only 0.37M parameters and 0.23G FLOPs computational complexity. Additionally, through ablation studies, we demonstrate that the AMAA module can work as a plug-and-play module for performance improvement on CNN-based methods.
皮肤癌研究对于寻找新的治疗方法和提高计算机辅助医学的存活率至关重要。在本研究中,皮肤病变图像的准确分割是早期诊断和个性化治疗策略的重要步骤。然而,虽然目前流行的基于transformer的模型已经取得了有竞争力的分割结果,但它们往往忽略了与它们的训练相关的计算复杂性和高成本。在本文中,我们提出了一种轻量级的网络,即多尺度的皮肤病变分割亚属性关注网络(MAAN)。首先,通过构建高分辨率和低分辨率的双路径框架对残差基本块进行优化,在减少参数数量的同时保持有效的特征提取能力;其次,为了更好地捕获皮肤病变图像中的信息,进一步提高模型性能,我们在低分辨率路径的最后阶段设计了自适应多尺度属性关注模块。在ISIC 2017和ISIC2018数据集上进行的实验表明,该模型MAAN的mIoU分别达到85.20和85.67%,优于现有的MHorNet,同时仅保持0.37M参数和0.23G FLOPs的计算复杂度。此外,通过烧蚀研究,我们证明AMAA模块可以作为一个即插即用模块,用于改进基于cnn的方法的性能。
{"title":"MAAN: multi-scale atrous attention network for skin lesion segmentation","authors":"Yang Lian, Ruizhi Han, Shiyuan Han, Defu Qiu, Jin Zhou","doi":"10.1007/s40747-025-02186-z","DOIUrl":"https://doi.org/10.1007/s40747-025-02186-z","url":null,"abstract":"Skin cancer research is essential to finding new treatments and improving survival rates in computer-aided medicine. Within this research, the accurate segmentation of skin lesion images is an important step for both early diagnosis and personalized treatment strategies. However, while current popular Transformer-based models have achieved competitive segmentation results, they often ignore the computational complexity and the high costs associated with their training. In this paper, we propose a lightweight network, a multi-scale atrous attention network for skin lesion segmentation (MAAN). Firstly, we optimize the residual basic block by constructing a dual-path framework with both high and low-resolution paths, which reduces the number of parameters while maintaining effective feature extraction capability. Secondly, to better capture the information in the skin lesion images and further improve the model performance, we design an adaptive multi-scale atrous attention module at the final stage of the low-resolution path. The experiments conducted on the ISIC 2017 and ISIC2018 datasets show that the proposed model MAAN achieves mIoU of 85.20 and 85.67% respectively, outperforming recent MHorNet while maintaining only 0.37M parameters and 0.23G FLOPs computational complexity. Additionally, through ablation studies, we demonstrate that the AMAA module can work as a plug-and-play module for performance improvement on CNN-based methods.","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"22 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145711460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive exploration and temporal attention in reinforcement learning for autonomous air combat decision making 自主空战决策强化学习中的自适应探索和时间注意
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-09 DOI: 10.1007/s40747-025-02189-w
Xiang Wu, Junzhe Jiang, Zhihong Chen, Shaojie Wu, Chenghong Ye, Xueyun Chen
{"title":"Adaptive exploration and temporal attention in reinforcement learning for autonomous air combat decision making","authors":"Xiang Wu, Junzhe Jiang, Zhihong Chen, Shaojie Wu, Chenghong Ye, Xueyun Chen","doi":"10.1007/s40747-025-02189-w","DOIUrl":"https://doi.org/10.1007/s40747-025-02189-w","url":null,"abstract":"","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"5 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145711165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motion-temporal calibration network for continuous sign language recognition 连续手语识别的运动-时间校正网络
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-09 DOI: 10.1007/s40747-025-02156-5
Hongguan Hu, Jianjun Peng, Zhidong Xiao, Li Guo, Yi Hu, Di Wu
Continuous Sign Language Recognition (CSLR) is fundamental to bridging the communication gap between hearing-impaired individuals and the broader society. The primary challenge lies in effectively modeling the complex spatial-temporal dynamic features in sign language videos. Current approaches typically employ independent processing strategies for motion feature extraction and temporal modeling, which impedes the unified modeling of action continuity and semantic integrity in sign language sequences. To address these limitations, we propose the Motion-Temporal Calibration Network (MTCNet), a novel framework for continuous sign language recognition that integrates dynamic feature enhancement and temporal calibration. The framework consists of two key innovative modules. First, the Cross-Frame Motion Refinement (CFMR) module implements an inter-frame differential attention mechanism combined with residual learning strategies, enabling precise motion feature modeling and effective enhancement of dynamic information between adjacent frames. Second, the Temporal-Channel Adaptive Recalibration (TCAR) module utilizes adaptive convolution kernel design and a dual-branch feature extraction architecture, facilitating joint optimization in both temporal and channel dimensions. In experimental evaluations, our method demonstrates competitive performance on the widely-used PHOENIX-2014 and PHOENIX-2014-T datasets, achieving results comparable to leading unimodal approaches. Moreover, it achieves state-of-the-art performance on the Chinese Sign Language (CSL) dataset. Through comprehensive ablation studies and quantitative analysis, we validate the effectiveness of our proposed method in fine-grained dynamic feature modeling and long-term dependency capture while maintaining computational efficiency.
持续手语识别(CSLR)是弥合听障人士与更广泛社会之间沟通差距的基础。手语视频中复杂的时空动态特征是手语视频研究面临的主要挑战。目前的方法通常采用独立的处理策略进行动作特征提取和时间建模,这阻碍了手势语言序列动作连续性和语义完整性的统一建模。为了解决这些限制,我们提出了运动-时间校准网络(MTCNet),这是一个集成了动态特征增强和时间校准的连续手语识别框架。该框架由两个关键的创新模块组成。首先,跨帧运动细化(CFMR)模块实现了帧间差分注意机制,结合残差学习策略,实现了精确的运动特征建模和相邻帧间动态信息的有效增强。其次,时间通道自适应再校准(TCAR)模块采用自适应卷积核设计和双分支特征提取架构,促进了时间和通道维度的联合优化。在实验评估中,我们的方法在广泛使用的PHOENIX-2014和PHOENIX-2014- t数据集上展示了具有竞争力的性能,取得了与领先的单峰方法相当的结果。此外,它在中国手语(CSL)数据集上达到了最先进的性能。通过综合消融研究和定量分析,我们验证了该方法在保持计算效率的同时,在细粒度动态特征建模和长期依赖捕获方面的有效性。
{"title":"Motion-temporal calibration network for continuous sign language recognition","authors":"Hongguan Hu, Jianjun Peng, Zhidong Xiao, Li Guo, Yi Hu, Di Wu","doi":"10.1007/s40747-025-02156-5","DOIUrl":"https://doi.org/10.1007/s40747-025-02156-5","url":null,"abstract":"Continuous Sign Language Recognition (CSLR) is fundamental to bridging the communication gap between hearing-impaired individuals and the broader society. The primary challenge lies in effectively modeling the complex spatial-temporal dynamic features in sign language videos. Current approaches typically employ independent processing strategies for motion feature extraction and temporal modeling, which impedes the unified modeling of action continuity and semantic integrity in sign language sequences. To address these limitations, we propose the Motion-Temporal Calibration Network (MTCNet), a novel framework for continuous sign language recognition that integrates dynamic feature enhancement and temporal calibration. The framework consists of two key innovative modules. First, the Cross-Frame Motion Refinement (CFMR) module implements an inter-frame differential attention mechanism combined with residual learning strategies, enabling precise motion feature modeling and effective enhancement of dynamic information between adjacent frames. Second, the Temporal-Channel Adaptive Recalibration (TCAR) module utilizes adaptive convolution kernel design and a dual-branch feature extraction architecture, facilitating joint optimization in both temporal and channel dimensions. In experimental evaluations, our method demonstrates competitive performance on the widely-used PHOENIX-2014 and PHOENIX-2014-T datasets, achieving results comparable to leading unimodal approaches. Moreover, it achieves state-of-the-art performance on the Chinese Sign Language (CSL) dataset. Through comprehensive ablation studies and quantitative analysis, we validate the effectiveness of our proposed method in fine-grained dynamic feature modeling and long-term dependency capture while maintaining computational efficiency.","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"134 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145704005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic RBFN with vector attention-guided feature selection for spam detection in social media 基于矢量注意力引导特征选择的动态RBFN社交媒体垃圾邮件检测
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-09 DOI: 10.1007/s40747-025-02148-5
E Elakkiya, Sumalatha Saleti, Arunkumar Balakrishnan
{"title":"Dynamic RBFN with vector attention-guided feature selection for spam detection in social media","authors":"E Elakkiya, Sumalatha Saleti, Arunkumar Balakrishnan","doi":"10.1007/s40747-025-02148-5","DOIUrl":"https://doi.org/10.1007/s40747-025-02148-5","url":null,"abstract":"","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"11 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145704006","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FedMGKD: a multi-granularity trusted knowledge distillation framework for edge personalized federated learning FedMGKD:用于边缘个性化联邦学习的多粒度可信知识蒸馏框架
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-09 DOI: 10.1007/s40747-025-02142-x
Ping Zhang, Wenlong Lu, Xiaoyu Zhou, An Bao
{"title":"FedMGKD: a multi-granularity trusted knowledge distillation framework for edge personalized federated learning","authors":"Ping Zhang, Wenlong Lu, Xiaoyu Zhou, An Bao","doi":"10.1007/s40747-025-02142-x","DOIUrl":"https://doi.org/10.1007/s40747-025-02142-x","url":null,"abstract":"","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"38 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145704007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fishing vessel behavior pattern recognition using AIS sub-trajectory prototype learning based on Gramian Angular Field 基于格拉曼角场的AIS子轨迹原型学习的渔船行为模式识别
IF 5.8 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-08 DOI: 10.1007/s40747-025-02187-y
Songtao Hu, Guanyu Chen, Rui Zhou, Xinghan Qin, Xiaokang Wang
{"title":"Fishing vessel behavior pattern recognition using AIS sub-trajectory prototype learning based on Gramian Angular Field","authors":"Songtao Hu, Guanyu Chen, Rui Zhou, Xinghan Qin, Xiaokang Wang","doi":"10.1007/s40747-025-02187-y","DOIUrl":"https://doi.org/10.1007/s40747-025-02187-y","url":null,"abstract":"","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"11 1","pages":""},"PeriodicalIF":5.8,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145697070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Complex & Intelligent Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1