首页 > 最新文献

Pattern Recognition Letters最新文献

英文 中文
Identifying real changes for height displaced buildings to aid in deep learning training sample generation 识别高度移位建筑物的真实变化,以帮助深度学习训练样本生成
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-25 DOI: 10.1016/j.patrec.2025.11.035
Haiyan Xu , Min Wang , Gang Xu , Qian Shen
Deep learning-based change detection methods often rely on many annotations, and automated sample generation methods for change detection are usually implemented via pixelwise comparisons after performing bitemporal image classification. In bitemporal images, high-rise buildings have different directional height displacements caused by different viewing angles; this issue generally causes serious false alarms in the automatic samples generated by post-classification comparison (PCC). In this study, by utilizing features such as the roof textures and facade geometry features of bitemporal buildings, automatic high-rise building change discrimination is implemented by matching the features of the building roofs and conducting height displacement triangle comparisons, which eliminates the false changes caused by building height displacements and preserves the true changes. Furthermore, method validation experiments were conducted on high-resolution images of Nanjing and Suzhou, two Chinese cities, and the results verify that the proposed method can automatically generate high-quality building samples with height displacement, which facilitates the training of deep learning-based change detection models.
基于深度学习的变化检测方法通常依赖于许多注释,而用于变化检测的自动样本生成方法通常是在执行双时图像分类后通过像素比较实现的。在双时影像中,高层建筑因视角不同而产生不同的方向高度位移;这个问题通常会导致PCC自动生成的样本出现严重的误报。本研究利用双时态建筑的屋顶纹理、立面几何特征等特征,通过匹配建筑屋顶特征并进行高度位移三角比对,实现高层建筑变化自动判别,消除了建筑高度位移引起的虚假变化,保留了真实变化。在南京和苏州两个城市的高分辨率图像上进行了方法验证实验,结果验证了该方法能够自动生成高质量的具有高度位移的建筑样本,为基于深度学习的变化检测模型的训练提供了便利。
{"title":"Identifying real changes for height displaced buildings to aid in deep learning training sample generation","authors":"Haiyan Xu ,&nbsp;Min Wang ,&nbsp;Gang Xu ,&nbsp;Qian Shen","doi":"10.1016/j.patrec.2025.11.035","DOIUrl":"10.1016/j.patrec.2025.11.035","url":null,"abstract":"<div><div>Deep learning-based change detection methods often rely on many annotations, and automated sample generation methods for change detection are usually implemented via pixelwise comparisons after performing bitemporal image classification. In bitemporal images, high-rise buildings have different directional height displacements caused by different viewing angles; this issue generally causes serious false alarms in the automatic samples generated by post-classification comparison (PCC). In this study, by utilizing features such as the roof textures and facade geometry features of bitemporal buildings, automatic high-rise building change discrimination is implemented by matching the features of the building roofs and conducting height displacement triangle comparisons, which eliminates the false changes caused by building height displacements and preserves the true changes. Furthermore, method validation experiments were conducted on high-resolution images of Nanjing and Suzhou, two Chinese cities, and the results verify that the proposed method can automatically generate high-quality building samples with height displacement, which facilitates the training of deep learning-based change detection models.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 269-277"},"PeriodicalIF":3.3,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145684523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid CNN and SVM model for Alzheimer’s disease classification using categorical focal loss function 基于分类局灶损失函数的混合CNN和SVM模型用于阿尔茨海默病分类
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-24 DOI: 10.1016/j.patrec.2025.11.031
Wided Hechkel , Rim Missaoui , Abdelhamid Helali , Marco Leo
Alzheimer’s disease (AD) is the leading cause of dementia worldwide. It attacks the elderly population, causing a dangerous cognitive decline and memory loss due to the degeneration and atrophy of brain neurons. Recent developments in machine learning techniques for the detection and classification of AD boost the early diagnosis and enable slowing the disease by adopting preclinical treatments. However, a major defect of these techniques is their high complexity architectures and their less generalizability, which provokes difficulties in clinical integration. This paper presents a new approach that combines convolutional neural network (CNN) and support vector machines (SVM) for the detection of AD. CNN stage enhances the accuracy of the system because it is an excellent feature extractor. SVM stage handles classification performance by optimizing the decision boundaries; meanwhile, it requires fewer hyperparameter updates compared to end-to-end CNN with Softmax classifier. SVM reduces the computational cost of the training. Experiments are conducted on the Kaggle dataset for Magnetic Resonance Imaging (MRI) brain images of AD. The hybrid model achieved accuracy scores of 98.52 %, 97.71 %, and 97.58 % for the training set, validation set, and testing set respectively, inference times per sample of 0.0588s, 0.0586s, and 0.0592s on the above three sets respectively. Obtained results confirm high effectiveness and potential prospect of the developed CNN-SVM model in early diagnosis of AD with reduced implementation complexity.
阿尔茨海默病(AD)是全球痴呆症的主要原因。它攻击老年人,由于大脑神经元的退化和萎缩,导致危险的认知能力下降和记忆丧失。机器学习技术用于阿尔茨海默病的检测和分类的最新发展促进了早期诊断,并通过采用临床前治疗来减缓疾病。然而,这些技术的一个主要缺陷是它们的结构高度复杂和不太普遍,这给临床整合带来了困难。本文提出了一种将卷积神经网络(CNN)与支持向量机(SVM)相结合的AD检测方法。CNN stage是一种优秀的特征提取器,提高了系统的准确率。支持向量机阶段通过优化决策边界来处理分类性能;同时,与使用Softmax分类器的端到端CNN相比,它需要更少的超参数更新。支持向量机减少了训练的计算量。在Kaggle数据集上对AD的磁共振成像(MRI)脑图像进行实验。混合模型在训练集、验证集和测试集上的准确率分别为98.52%、97.71%和97.58%,每样本推理次数分别为0.0588s、0.0586s和0.0592s。研究结果证实了所建立的CNN-SVM模型在AD早期诊断中的有效性和潜在的应用前景。
{"title":"Hybrid CNN and SVM model for Alzheimer’s disease classification using categorical focal loss function","authors":"Wided Hechkel ,&nbsp;Rim Missaoui ,&nbsp;Abdelhamid Helali ,&nbsp;Marco Leo","doi":"10.1016/j.patrec.2025.11.031","DOIUrl":"10.1016/j.patrec.2025.11.031","url":null,"abstract":"<div><div>Alzheimer’s disease (AD) is the leading cause of dementia worldwide. It attacks the elderly population, causing a dangerous cognitive decline and memory loss due to the degeneration and atrophy of brain neurons. Recent developments in machine learning techniques for the detection and classification of AD boost the early diagnosis and enable slowing the disease by adopting preclinical treatments. However, a major defect of these techniques is their high complexity architectures and their less generalizability, which provokes difficulties in clinical integration. This paper presents a new approach that combines convolutional neural network (CNN) and support vector machines (SVM) for the detection of AD. CNN stage enhances the accuracy of the system because it is an excellent feature extractor. SVM stage handles classification performance by optimizing the decision boundaries; meanwhile, it requires fewer hyperparameter updates compared to end-to-end CNN with Softmax classifier. SVM reduces the computational cost of the training. Experiments are conducted on the Kaggle dataset for Magnetic Resonance Imaging (MRI) brain images of AD. The hybrid model achieved accuracy scores of 98.52 %, 97.71 %, and 97.58 % for the training set, validation set, and testing set respectively, inference times per sample of 0.0588s, 0.0586s, and 0.0592s on the above three sets respectively. Obtained results confirm high effectiveness and potential prospect of the developed CNN-SVM model in early diagnosis of AD with reduced implementation complexity.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 261-268"},"PeriodicalIF":3.3,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable multimodal brain imaging through a multiple-branch neural network 通过多分支神经网络可解释的多模态脑成像
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-22 DOI: 10.1016/j.patrec.2025.11.030
Giuseppe Placidi , Alessia Cipriani , Michele Nappi , Matteo Polsinelli
Brain studies require the use of several complementary imaging modalities. When some modality is unavailable, Artificial Intelligence (AI) has recently provided ways to estimate them. Radiologists modulate the use of the available modalities depending on the task they have to perform. We aim to trace artificially the radiological process through a multibranch neural network architecture, the StarNet. The goal is to explain how and where different imaging modalities, either really collected or artificially reconstructed, are used in different radiological tasks by reading inside the structure of the network. To do that, StarNet includes several satellite networks, one per source modality, connected at each layer by a central unit. This design enables us to assess the contribution of each imaging modality, identifying where the contribution occurs, and to quantify the variations if certain modalities are substituted with AI-generated counterparts. The ultimate goal is to enable data-related and task-related ablation studies through the complete explainability of StarNet, thus offering radiologists clear guidance on which imaging sequences contribute to the task, to what extent, and at which stages of the process. As an example, we applied the proposed architecture to the 2D slices extracted from 3D volumes acquired with multimodal magnetic resonance imaging (MRI), to assess: 1. The role of the used imaging modalities; 2. The change in role when the radiological task changes; 3. The effects of synthetic data on the process. The results are presented and discussed.
脑研究需要使用几种互补的成像方式。当某些形态不可用时,人工智能(AI)最近提供了评估它们的方法。放射科医生根据他们必须执行的任务调整可用模式的使用。我们的目标是通过一个多分支神经网络架构,即StarNet,人工地追踪放射过程。目的是通过解读神经网络的内部结构,解释不同的成像模式(无论是真实收集的还是人工重建的)如何以及在哪里被用于不同的放射学任务。为了做到这一点,StarNet包括几个卫星网络,每个源模式一个,每层由一个中心单元连接。这种设计使我们能够评估每种成像模式的贡献,确定贡献发生的位置,并量化某些模式被人工智能生成的对应模式所取代时的变化。最终目标是通过StarNet的完全可解释性来实现与数据相关和与任务相关的消融研究,从而为放射科医生提供明确的指导,说明哪些成像序列有助于任务,在多大程度上以及在过程的哪个阶段。作为一个例子,我们将所提出的架构应用于从多模态磁共振成像(MRI)获得的3D体中提取的2D切片,以评估:1。所使用的成像模式的作用;2. 放射任务变化时角色的变化;3. 合成数据对过程的影响。给出了实验结果并进行了讨论。
{"title":"Explainable multimodal brain imaging through a multiple-branch neural network","authors":"Giuseppe Placidi ,&nbsp;Alessia Cipriani ,&nbsp;Michele Nappi ,&nbsp;Matteo Polsinelli","doi":"10.1016/j.patrec.2025.11.030","DOIUrl":"10.1016/j.patrec.2025.11.030","url":null,"abstract":"<div><div>Brain studies require the use of several complementary imaging modalities. When some modality is unavailable, Artificial Intelligence (AI) has recently provided ways to estimate them. Radiologists modulate the use of the available modalities depending on the task they have to perform. We aim to trace artificially the radiological process through a multibranch neural network architecture, the StarNet. The goal is to explain how and where different imaging modalities, either really collected or artificially reconstructed, are used in different radiological tasks by reading inside the structure of the network. To do that, StarNet includes several satellite networks, one per source modality, connected at each layer by a central unit. This design enables us to assess the contribution of each imaging modality, identifying where the contribution occurs, and to quantify the variations if certain modalities are substituted with AI-generated counterparts. The ultimate goal is to enable data-related and task-related ablation studies through the complete explainability of StarNet, thus offering radiologists clear guidance on which imaging sequences contribute to the task, to what extent, and at which stages of the process. As an example, we applied the proposed architecture to the 2D slices extracted from 3D volumes acquired with multimodal magnetic resonance imaging (MRI), to assess: 1. The role of the used imaging modalities; 2. The change in role when the radiological task changes; 3. The effects of synthetic data on the process. The results are presented and discussed.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 254-260"},"PeriodicalIF":3.3,"publicationDate":"2025-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mitigating task randomness in graph few-shot learning 减少图少次学习中的任务随机性
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-20 DOI: 10.1016/j.patrec.2025.11.022
Shuzhen Rao , Jun Huang
In graph few-shot learning, meta-training tasks are sampled to improve the model’s ability to learn from limited nodes. Existing methods adapted from computer vision, generally employ random task sampling, which can lead to excessive task randomness. This hinders effective training on the graph as models struggle to adapt to tasks with substantial variations in classes and nodes. To address this issue, we propose a novel method called TRARM, i.e., Task RAndomness Reduced graph Meta-learning to mitigate adverse effects of excessive task randomness. Firstly, we design progressive grouping-based sampling to adjust combinations of classes and nodes by stages, thereby enabling more focused and efficient meta-training. Secondly, complementing sampling, a unified memory-based meta-update module is first deployed to effectively accumulate cross-task knowledge, improving both efficiency and stability of meta-learning. Despite its simplicity, comprehensive experiments demonstrate the superior performance of TRARM on four widely used benchmarks.
在图少射学习中,对元训练任务进行采样,以提高模型从有限节点学习的能力。现有的基于计算机视觉的方法通常采用随机任务抽样,这可能导致任务随机性过大。这阻碍了对图的有效训练,因为模型很难适应类和节点中存在大量变化的任务。为了解决这个问题,我们提出了一种名为TRARM的新方法,即任务随机性减少图元学习,以减轻过度任务随机性的不利影响。首先,我们设计了基于渐进式分组的采样,分阶段调整类和节点的组合,从而使元训练更加集中和高效。其次,在抽样的基础上,首先部署统一的基于记忆的元更新模块,有效地积累跨任务知识,提高元学习的效率和稳定性。尽管它很简单,但综合实验证明了TRARM在四个广泛使用的基准测试上的优越性能。
{"title":"Mitigating task randomness in graph few-shot learning","authors":"Shuzhen Rao ,&nbsp;Jun Huang","doi":"10.1016/j.patrec.2025.11.022","DOIUrl":"10.1016/j.patrec.2025.11.022","url":null,"abstract":"<div><div>In graph few-shot learning, meta-training tasks are sampled to improve the model’s ability to learn from limited nodes. Existing methods adapted from computer vision, generally employ random task sampling, which can lead to excessive task randomness. This hinders effective training on the graph as models struggle to adapt to tasks with substantial variations in classes and nodes. To address this issue, we propose a novel method called TRARM, i.e., <strong>T</strong>ask <strong>RA</strong>ndomness <strong>R</strong>educed graph <strong>M</strong>eta-learning to mitigate adverse effects of excessive task randomness. Firstly, we design progressive grouping-based sampling to adjust combinations of classes and nodes by stages, thereby enabling more focused and efficient meta-training. Secondly, complementing sampling, a unified memory-based meta-update module is first deployed to effectively accumulate cross-task knowledge, improving both efficiency and stability of meta-learning. Despite its simplicity, comprehensive experiments demonstrate the superior performance of TRARM on four widely used benchmarks.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 232-238"},"PeriodicalIF":3.3,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAM-guided prompt learning for Multiple Sclerosis lesion segmentation sam引导下的多发性硬化症病灶分割提示学习
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-17 DOI: 10.1016/j.patrec.2025.11.018
Federica Proietto Salanitri , Giovanni Bellitto , Salvatore Calcagno , Ulas Bagci , Concetto Spampinato , Manuela Pennisi
Accurate segmentation of Multiple Sclerosis (MS) lesions remains a critical challenge in medical image analysis due to their small size, irregular shape, and sparse distribution. Despite recent progress in vision foundation models — such as SAM and its medical variant MedSAM — these models have not yet been explored in the context of MS lesion segmentation. Moreover, their reliance on manually crafted prompts and high inference-time computational cost limits their applicability in clinical workflows, especially in resource-constrained environments. In this work, we introduce a novel training-time framework for effective and efficient MS lesion segmentation. Our method leverages SAM solely during training to guide a prompt learner that automatically discovers task-specific embeddings. At inference, SAM is replaced by a lightweight convolutional aggregator that maps the learned embeddings directly into segmentation masks—enabling fully automated, low-cost deployment. We show that our approach significantly outperforms existing specialized methods on the public MSLesSeg dataset, establishing new performance benchmarks in a domain where foundation models had not previously been applied. To assess generalizability, we also evaluate our method on pancreas and prostate segmentation tasks, where it achieves competitive accuracy while requiring an order of magnitude fewer parameters and computational resources compared to SAM-based pipelines. By eliminating the need for foundation models at inference time, our framework enables efficient segmentation without sacrificing accuracy. This design bridges the gap between large-scale pretraining and real-world clinical deployment, offering a scalable and practical solution for MS lesion segmentation and beyond. Code is available at https://github.com/perceivelab/MS-SAM-LESS.
由于多发性硬化症(MS)病变体积小、形状不规则、分布稀疏,其准确分割一直是医学图像分析中的一个关键挑战。尽管最近在视觉基础模型(如SAM及其医学变体MedSAM)方面取得了进展,但这些模型尚未在MS病变分割的背景下进行探索。此外,它们对手工制作提示和高推断时间计算成本的依赖限制了它们在临床工作流程中的适用性,特别是在资源受限的环境中。在这项工作中,我们引入了一种新的训练时间框架,用于有效和高效的MS病变分割。我们的方法仅在训练期间利用SAM来指导快速学习者自动发现特定于任务的嵌入。在推理中,SAM被一个轻量级的卷积聚合器取代,该聚合器将学习到的嵌入直接映射到分割掩码中,从而实现全自动、低成本的部署。我们表明,我们的方法在公共MSLesSeg数据集上显著优于现有的专门方法,在以前没有应用基础模型的领域建立了新的性能基准。为了评估泛化性,我们还在胰腺和前列腺分割任务中评估了我们的方法,与基于sam的管道相比,它在需要更少的参数和计算资源的同时达到了相当的准确性。通过在推理时消除对基础模型的需求,我们的框架可以在不牺牲准确性的情况下实现有效的分割。该设计弥合了大规模预训练和实际临床部署之间的差距,为MS病变分割等提供了可扩展和实用的解决方案。代码可从https://github.com/perceivelab/MS-SAM-LESS获得。
{"title":"SAM-guided prompt learning for Multiple Sclerosis lesion segmentation","authors":"Federica Proietto Salanitri ,&nbsp;Giovanni Bellitto ,&nbsp;Salvatore Calcagno ,&nbsp;Ulas Bagci ,&nbsp;Concetto Spampinato ,&nbsp;Manuela Pennisi","doi":"10.1016/j.patrec.2025.11.018","DOIUrl":"10.1016/j.patrec.2025.11.018","url":null,"abstract":"<div><div>Accurate segmentation of Multiple Sclerosis (MS) lesions remains a critical challenge in medical image analysis due to their small size, irregular shape, and sparse distribution. Despite recent progress in vision foundation models — such as SAM and its medical variant MedSAM — these models have not yet been explored in the context of MS lesion segmentation. Moreover, their reliance on manually crafted prompts and high inference-time computational cost limits their applicability in clinical workflows, especially in resource-constrained environments. In this work, we introduce a novel training-time framework for effective and efficient MS lesion segmentation. Our method leverages SAM solely during training to guide a prompt learner that automatically discovers task-specific embeddings. At inference, SAM is replaced by a lightweight convolutional aggregator that maps the learned embeddings directly into segmentation masks—enabling fully automated, low-cost deployment. We show that our approach significantly outperforms existing specialized methods on the public MSLesSeg dataset, establishing new performance benchmarks in a domain where foundation models had not previously been applied. To assess generalizability, we also evaluate our method on pancreas and prostate segmentation tasks, where it achieves competitive accuracy while requiring an order of magnitude fewer parameters and computational resources compared to SAM-based pipelines. By eliminating the need for foundation models at inference time, our framework enables efficient segmentation without sacrificing accuracy. This design bridges the gap between large-scale pretraining and real-world clinical deployment, offering a scalable and practical solution for MS lesion segmentation and beyond. Code is available at <span><span>https://github.com/perceivelab/MS-SAM-LESS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 205-211"},"PeriodicalIF":3.3,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Additive decomposition of one-dimensional signals using Transformers 基于变压器的一维信号加性分解
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-17 DOI: 10.1016/j.patrec.2025.11.002
Samuele Salti , Andrea Pinto , Alessandro Lanza , Serena Morigi
One-dimensional signal decomposition is a well-established and widely used technique across various scientific fields. It serves as a highly valuable pre-processing step for data analysis. While traditional decomposition techniques often rely on mathematical models, recent research suggests that applying the latest deep learning models to this very ill-posed inverse problem represents an exciting, unexplored area with promising potential. This work presents a novel method for the additive decomposition of one-dimensional signals. We leverage the Transformer architecture to decompose signals into their constituent components: piecewise constant, smooth (trend), highly-oscillatory, and noise components. Our model, trained on synthetic data, achieves excellent accuracy in modeling and decomposing input signals from the same distribution, as demonstrated by the experimental results.
一维信号分解是一种成熟且广泛应用于各个科学领域的技术。它是数据分析中非常有价值的预处理步骤。虽然传统的分解技术通常依赖于数学模型,但最近的研究表明,将最新的深度学习模型应用于这个非常不适定的逆问题代表了一个令人兴奋的、尚未开发的领域,具有很大的潜力。本文提出了一种一维信号加性分解的新方法。我们利用Transformer架构将信号分解为它们的组成组件:分段常量、平滑(趋势)、高振荡和噪声组件。实验结果表明,我们的模型在模拟和分解来自相同分布的输入信号方面取得了优异的精度。
{"title":"Additive decomposition of one-dimensional signals using Transformers","authors":"Samuele Salti ,&nbsp;Andrea Pinto ,&nbsp;Alessandro Lanza ,&nbsp;Serena Morigi","doi":"10.1016/j.patrec.2025.11.002","DOIUrl":"10.1016/j.patrec.2025.11.002","url":null,"abstract":"<div><div>One-dimensional signal decomposition is a well-established and widely used technique across various scientific fields. It serves as a highly valuable pre-processing step for data analysis. While traditional decomposition techniques often rely on mathematical models, recent research suggests that applying the latest deep learning models to this very ill-posed inverse problem represents an exciting, unexplored area with promising potential. This work presents a novel method for the additive decomposition of one-dimensional signals. We leverage the Transformer architecture to decompose signals into their constituent components: piecewise constant, smooth (trend), highly-oscillatory, and noise components. Our model, trained on synthetic data, achieves excellent accuracy in modeling and decomposing input signals from the same distribution, as demonstrated by the experimental results.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 239-245"},"PeriodicalIF":3.3,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAMIRO: Spatial Attention Mutual Information Regularization with a pre-trained model as Oracle for lane detection SAMIRO:基于预训练模型的空间注意互信息正则化,用于车道检测
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-17 DOI: 10.1016/j.patrec.2025.10.013
Hyunjong Lee , Jangho Lee , Jaekoo Lee
Lane detection is an important topic in the future mobility solutions. Real-world environmental challenges such as background clutter, varying illumination, and occlusions pose significant obstacles to effective lane detection, particularly when relying on data-driven approaches that require substantial effort and cost for data collection and annotation. To address these issues, lane detection methods must leverage contextual and global information from surrounding lanes and objects. In this paper, we propose a Spatial Attention Mutual Information Regularization with a pre-trained model as an Oracle, called SAMIRO. SAMIRO enhances lane detection performance by transferring knowledge from a pre-trained model while preserving domain-agnostic spatial information. Leveraging SAMIRO’s plug-and-play characteristic, we integrate it into various state-of-the-art lane detection approaches and conduct extensive experiments on major benchmarks such as CULane, Tusimple, and LLAMAS. The results demonstrate that SAMIRO consistently improves performance across different models and datasets. The code will be made available upon publication.
车道检测是未来交通解决方案中的一个重要课题。现实世界的环境挑战,如背景杂乱、光照变化和遮挡,对有效的车道检测构成了重大障碍,特别是当依赖于数据驱动的方法时,需要大量的努力和成本来收集和注释数据。为了解决这些问题,车道检测方法必须利用来自周围车道和物体的上下文和全局信息。在本文中,我们提出了一种空间注意互信息正则化方法,将预训练模型作为Oracle,称为SAMIRO。SAMIRO通过从预训练模型转移知识来增强车道检测性能,同时保留与领域无关的空间信息。利用SAMIRO的即插即用特性,我们将其集成到各种最先进的车道检测方法中,并在CULane, Tusimple和LLAMAS等主要基准上进行了广泛的实验。结果表明,SAMIRO在不同的模型和数据集上一致地提高了性能。该准则将在出版后提供。
{"title":"SAMIRO: Spatial Attention Mutual Information Regularization with a pre-trained model as Oracle for lane detection","authors":"Hyunjong Lee ,&nbsp;Jangho Lee ,&nbsp;Jaekoo Lee","doi":"10.1016/j.patrec.2025.10.013","DOIUrl":"10.1016/j.patrec.2025.10.013","url":null,"abstract":"<div><div>Lane detection is an important topic in the future mobility solutions. Real-world environmental challenges such as background clutter, varying illumination, and occlusions pose significant obstacles to effective lane detection, particularly when relying on data-driven approaches that require substantial effort and cost for data collection and annotation. To address these issues, lane detection methods must leverage contextual and global information from surrounding lanes and objects. In this paper, we propose a <em>Spatial Attention Mutual Information Regularization with a pre-trained model as an Oracle</em>, called <em>SAMIRO</em>. SAMIRO enhances lane detection performance by transferring knowledge from a pre-trained model while preserving domain-agnostic spatial information. Leveraging SAMIRO’s plug-and-play characteristic, we integrate it into various state-of-the-art lane detection approaches and conduct extensive experiments on major benchmarks such as CULane, Tusimple, and LLAMAS. The results demonstrate that SAMIRO consistently improves performance across different models and datasets. The code will be made available upon publication.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 198-204"},"PeriodicalIF":3.3,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tadmo: A tabular distance measure with move operations 带移动操作的表格式距离测量
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-15 DOI: 10.1016/j.patrec.2025.11.009
Dirko Coetsee , Steve Kroon , Ralf Kistner , Adem Kikaj , McElory Hoffmann , Luc De Raedt
Tabular data is ubiquitous in pattern recognition, yet accurately measuring differences between tables remains challenging. Conventional methods rely on cell substitutions and row/column insertions and deletions, often overestimating the difference when cells are simply repositioned. We propose a distance metric that considers move operations, capturing structural changes more faithfully. Although exact computation is NP-complete, a greedy approach computes an effective approximation in practice. Experimental results on real-world datasets demonstrate that our approach yields a more compact and intuitive measure of table dissimilarity, enhancing applications such as clustering, table extraction evaluation, and version history recovery.
表格数据在模式识别中无处不在,但准确测量表之间的差异仍然具有挑战性。传统的方法依赖于细胞替换和行/列插入和删除,当细胞只是重新定位时,往往高估了差异。我们提出了一个考虑移动操作的距离度量,更忠实地捕捉结构变化。虽然精确计算是np完全的,但贪婪方法在实际中计算出一个有效的近似。在真实数据集上的实验结果表明,我们的方法产生了更紧凑和直观的表不相似性度量,增强了诸如聚类、表提取评估和版本历史恢复等应用程序。
{"title":"Tadmo: A tabular distance measure with move operations","authors":"Dirko Coetsee ,&nbsp;Steve Kroon ,&nbsp;Ralf Kistner ,&nbsp;Adem Kikaj ,&nbsp;McElory Hoffmann ,&nbsp;Luc De Raedt","doi":"10.1016/j.patrec.2025.11.009","DOIUrl":"10.1016/j.patrec.2025.11.009","url":null,"abstract":"<div><div>Tabular data is ubiquitous in pattern recognition, yet accurately measuring differences between tables remains challenging. Conventional methods rely on cell substitutions and row/column insertions and deletions, often overestimating the difference when cells are simply repositioned. We propose a distance metric that considers move operations, capturing structural changes more faithfully. Although exact computation is NP-complete, a greedy approach computes an effective approximation in practice. Experimental results on real-world datasets demonstrate that our approach yields a more compact and intuitive measure of table dissimilarity, enhancing applications such as clustering, table extraction evaluation, and version history recovery.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 212-218"},"PeriodicalIF":3.3,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning and multi-modal MRI for the segmentation of sub-acute and chronic stroke lesions 深度学习和多模态MRI对亚急性和慢性脑卒中病变的分割
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-14 DOI: 10.1016/j.patrec.2025.11.017
Alessandro Di Matteo , Youwan Mahé , Stéphanie Leplaideur , Isabelle Bonan , Elise Bannier , Francesca Galassi
Stroke is a leading cause of morbidity and mortality worldwide. Accurate segmentation of post-stroke lesions on MRI is crucial for assessing brain damage and informing rehabilitation. Manual segmentation, however, is time-consuming and prone to error, motivating the development of automated approaches. This study investigates how deep learning with multimodal MRI can improve automated lesion segmentation in sub-acute and chronic stroke. A single-modality baseline was trained on the public ATLAS v2.0 dataset (655 T1-w scans) using the nnU-Net v2 framework and evaluated on an independent clinical cohort (45 patients with paired T1-w and FLAIR MRI). On this internal dataset, we conducted a systematic ablation comparing (i) direct transfer of the ATLAS baseline, (ii) fine-tuning using T1-w only, and (iii) fusion of T1-w and FLAIR inputs through early, mid, and late fusion strategies, each tested with metric averaging and ensembling.
The ATLAS baseline model achieved a mean Dice score of 0.64 and a lesion-wise F1 score of 0.67. On the clinical dataset, ensembling improved performance (Dice 0.70 vs. 0.68; F1 0.79 vs. 0.73), while fine-tuning on T1-w data further increased accuracy (Dice 0.72; F1 0.78). The best overall results were obtained with a T1+FLAIR late-fusion ensemble (Dice 0.75; F1 0.80; Average Surface Distance (ASD) 2.94 mm), with statistically significant improvements, especially for small and medium lesions.
These results show that fine-tuning and multimodal fusion — particularly late fusion — improve generalization for post-stroke lesion segmentation, supporting robust, reproducible quantification in clinical settings.
中风是全世界发病率和死亡率的主要原因。脑卒中后MRI病变的准确分割对于评估脑损伤和告知康复至关重要。然而,手工分割既耗时又容易出错,这促使了自动化方法的发展。本研究探讨了多模态MRI的深度学习如何改善亚急性和慢性中风的自动病灶分割。使用nnU-Net v2框架在公共ATLAS v2.0数据集(655个T1-w扫描)上训练单模态基线,并在独立临床队列(45例配对T1-w和FLAIR MRI患者)中进行评估。在这个内部数据集上,我们进行了系统的消融比较(i) ATLAS基线的直接转移,(ii)仅使用T1-w进行微调,以及(iii)通过早期、中期和后期融合策略融合T1-w和FLAIR输入,每种策略都使用度量平均和集合进行测试。ATLAS基线模型的平均Dice评分为0.64,逐病变F1评分为0.67。在临床数据集上,集成提高了性能(Dice 0.70 vs. 0.68; F1 0.79 vs. 0.73),而在T1-w数据上的微调进一步提高了准确性(Dice 0.72; F1 0.78)。T1+FLAIR晚期融合整体效果最好(Dice 0.75; F1 0.80;平均表面距离(ASD) 2.94 mm),具有统计学上显著的改善,特别是对于中小型病变。这些结果表明,微调和多模态融合-特别是后期融合-提高了脑卒中后病变分割的泛化,支持临床环境中稳健、可重复的量化。
{"title":"Deep learning and multi-modal MRI for the segmentation of sub-acute and chronic stroke lesions","authors":"Alessandro Di Matteo ,&nbsp;Youwan Mahé ,&nbsp;Stéphanie Leplaideur ,&nbsp;Isabelle Bonan ,&nbsp;Elise Bannier ,&nbsp;Francesca Galassi","doi":"10.1016/j.patrec.2025.11.017","DOIUrl":"10.1016/j.patrec.2025.11.017","url":null,"abstract":"<div><div>Stroke is a leading cause of morbidity and mortality worldwide. Accurate segmentation of post-stroke lesions on MRI is crucial for assessing brain damage and informing rehabilitation. Manual segmentation, however, is time-consuming and prone to error, motivating the development of automated approaches. This study investigates how deep learning with multimodal MRI can improve automated lesion segmentation in sub-acute and chronic stroke. A single-modality baseline was trained on the public ATLAS v2.0 dataset (655 T1-w scans) using the nnU-Net v2 framework and evaluated on an independent clinical cohort (45 patients with paired T1-w and FLAIR MRI). On this internal dataset, we conducted a systematic ablation comparing (i) direct transfer of the ATLAS baseline, (ii) fine-tuning using T1-w only, and (iii) fusion of T1-w and FLAIR inputs through early, mid, and late fusion strategies, each tested with metric averaging and ensembling.</div><div>The ATLAS baseline model achieved a mean Dice score of 0.64 and a lesion-wise F1 score of 0.67. On the clinical dataset, ensembling improved performance (Dice 0.70 vs. 0.68; F1 0.79 vs. 0.73), while fine-tuning on T1-w data further increased accuracy (Dice 0.72; F1 0.78). The best overall results were obtained with a T1+FLAIR late-fusion ensemble (Dice 0.75; F1 0.80; Average Surface Distance (ASD) 2.94 mm), with statistically significant improvements, especially for small and medium lesions.</div><div>These results show that fine-tuning and multimodal fusion — particularly late fusion — improve generalization for post-stroke lesion segmentation, supporting robust, reproducible quantification in clinical settings.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 225-231"},"PeriodicalIF":3.3,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Regional patch-based MRI brain age modeling with an interpretable cognitive reserve proxy 基于区域斑块的MRI脑年龄模型与可解释的认知储备代理
IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-11-14 DOI: 10.1016/j.patrec.2025.11.027
Samuel Maddox , Lemuel Puglisi , Fatemeh Darabifard , Alzheimer’s Disease Neuroimaging Initiative , Australian Imaging Biomarkers and Lifestyle flagship study of aging , Saber Sami , Daniele Ravi
Accurate brain age prediction from MRI is a promising biomarker for brain health and neurodegenerative disease risk, but current deep learning models often lack anatomical specificity and clinical insight. We present a regional patch-based ensemble framework that uses 3D Convolutional Neural Networks (CNNs) trained on bilateral patches from ten subcortical structures, enhancing anatomical sensitivity. Ensemble predictions are combined with cognitive assessments to derive a cognitively informed proxy for cognitive reserve (CR-Proxy), quantifying resilience to age-related brain changes. We train our framework on a large, multi-cohort dataset of healthy controls and test it on independent samples that include individuals with Alzheimer’s disease and mild cognitive impairment. The results demonstrate that our method achieves robust brain age prediction and provides a practical, interpretable CR-Proxy capable of distinguishing diagnostic groups and identifying individuals with high or low cognitive reserve. This pipeline offers a scalable, clinically accessible tool for early risk assessment and personalized brain health monitoring.
从MRI中准确预测脑年龄是一种很有前途的脑健康和神经退行性疾病风险的生物标志物,但目前的深度学习模型往往缺乏解剖学特异性和临床洞察力。我们提出了一个基于区域斑块的集成框架,该框架使用3D卷积神经网络(cnn)对来自10个皮层下结构的双侧斑块进行训练,提高了解剖灵敏度。集合预测与认知评估相结合,得出认知储备的认知知情代理(CR-Proxy),量化与年龄相关的大脑变化的弹性。我们在健康对照的大型多队列数据集上训练我们的框架,并在包括患有阿尔茨海默病和轻度认知障碍的个体在内的独立样本上进行测试。结果表明,我们的方法实现了稳健的脑年龄预测,并提供了一个实用的、可解释的CR-Proxy,能够区分诊断组和识别具有高或低认知储备的个体。该管道为早期风险评估和个性化大脑健康监测提供了可扩展的、临床可访问的工具。
{"title":"Regional patch-based MRI brain age modeling with an interpretable cognitive reserve proxy","authors":"Samuel Maddox ,&nbsp;Lemuel Puglisi ,&nbsp;Fatemeh Darabifard ,&nbsp;Alzheimer’s Disease Neuroimaging Initiative ,&nbsp;Australian Imaging Biomarkers and Lifestyle flagship study of aging ,&nbsp;Saber Sami ,&nbsp;Daniele Ravi","doi":"10.1016/j.patrec.2025.11.027","DOIUrl":"10.1016/j.patrec.2025.11.027","url":null,"abstract":"<div><div>Accurate brain age prediction from MRI is a promising biomarker for brain health and neurodegenerative disease risk, but current deep learning models often lack anatomical specificity and clinical insight. We present a regional patch-based ensemble framework that uses 3D Convolutional Neural Networks (CNNs) trained on bilateral patches from ten subcortical structures, enhancing anatomical sensitivity. Ensemble predictions are combined with cognitive assessments to derive a cognitively informed proxy for cognitive reserve (CR-Proxy), quantifying resilience to age-related brain changes. We train our framework on a large, multi-cohort dataset of healthy controls and test it on independent samples that include individuals with Alzheimer’s disease and mild cognitive impairment. The results demonstrate that our method achieves robust brain age prediction and provides a practical, interpretable CR-Proxy capable of distinguishing diagnostic groups and identifying individuals with high or low cognitive reserve. This pipeline offers a scalable, clinically accessible tool for early risk assessment and personalized brain health monitoring.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"199 ","pages":"Pages 219-224"},"PeriodicalIF":3.3,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Pattern Recognition Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1