首页 > 最新文献

Biomedical Signal Processing and Control最新文献

英文 中文
DBMAF: Dual-branch multimodal attention-based feature fusion network for fusing histopathology and radiology images DBMAF:用于融合组织病理学和放射学图像的双分支多模式基于注意力的特征融合网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-07 DOI: 10.1016/j.bspc.2026.109739
Yingfa Li , Jialin Shi , Yufei Wang , Jiping Wei , Yaru Wei , Liang Wu , Meihao Wang , Zhifang Pan
Integrating radiology and histopathology images provides critical complementary perspectives for cancer survival prediction. However, current research faces two main challenges: (1) significant discrepancies in spatial scale and feature dimensionality between modalities; and (2) limited clinical generalizability due to existing methods being restricted to single cancer types or tasks. To overcome these barriers, we propose the Dual-Branch Multimodal Attention-based Feature Fusion Network (DBMAF). This framework employs an enhanced multi-scale channel attention mechanism for intra-modal feature extraction and an attention-guided cross-modal module to learn discriminative correlations between modalities. We validated DBMAF on four cancer cohorts, comprising three public datasets (TCGA-OV, TCGA-KIRC, TCGA-LIHC) and one private institutional dataset (WMU-CRC). Quantitative evaluations demonstrate that our method consistently outperforms all compared methods, achieving a maximum C-index of 0.910 on the TCGA-LIHC cohort. Furthermore, DBMAF showed robust performance across multiple survival endpoints (OS, TTP, and PFS) on the TCGA-OV dataset, highlighting its clinical utility for precise treatment stratification.
整合放射学和组织病理学图像为癌症生存预测提供了关键的互补视角。然而,目前的研究面临着两个主要挑战:(1)模态之间的空间尺度和特征维度存在显著差异;(2)由于现有方法仅限于单一癌症类型或任务,临床推广能力有限。为了克服这些障碍,我们提出了双分支多模态基于注意力的特征融合网络(DBMAF)。该框架采用增强的多尺度通道注意机制进行模态内特征提取,并采用注意引导的跨模态模块学习模态间的判别相关性。我们在四个癌症队列中验证了DBMAF,包括三个公共数据集(TCGA-OV, TCGA-KIRC, TCGA-LIHC)和一个私人机构数据集(WMU-CRC)。定量评估表明,我们的方法始终优于所有比较的方法,在TCGA-LIHC队列中实现了最大的c指数0.910。此外,在TCGA-OV数据集上,DBMAF在多个生存终点(OS, TTP和PFS)上表现出稳健的性能,突出了其在精确治疗分层方面的临床应用。
{"title":"DBMAF: Dual-branch multimodal attention-based feature fusion network for fusing histopathology and radiology images","authors":"Yingfa Li ,&nbsp;Jialin Shi ,&nbsp;Yufei Wang ,&nbsp;Jiping Wei ,&nbsp;Yaru Wei ,&nbsp;Liang Wu ,&nbsp;Meihao Wang ,&nbsp;Zhifang Pan","doi":"10.1016/j.bspc.2026.109739","DOIUrl":"10.1016/j.bspc.2026.109739","url":null,"abstract":"<div><div>Integrating radiology and histopathology images provides critical complementary perspectives for cancer survival prediction. However, current research faces two main challenges: (1) significant discrepancies in spatial scale and feature dimensionality between modalities; and (2) limited clinical generalizability due to existing methods being restricted to single cancer types or tasks. To overcome these barriers, we propose the Dual-Branch Multimodal Attention-based Feature Fusion Network (DBMAF). This framework employs an enhanced multi-scale channel attention mechanism for intra-modal feature extraction and an attention-guided cross-modal module to learn discriminative correlations between modalities. We validated DBMAF on four cancer cohorts, comprising three public datasets (TCGA-OV, TCGA-KIRC, TCGA-LIHC) and one private institutional dataset (WMU-CRC). Quantitative evaluations demonstrate that our method consistently outperforms all compared methods, achieving a maximum C-index of 0.910 on the TCGA-LIHC cohort. Furthermore, DBMAF showed robust performance across multiple survival endpoints (OS, TTP, and PFS) on the TCGA-OV dataset, highlighting its clinical utility for precise treatment stratification.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109739"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146191910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Refined myocardium segmentation from CT using a Hybrid-Fusion transformer 利用Hybrid-Fusion变压器对CT进行精细心肌分割
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-11 DOI: 10.1016/j.bspc.2026.109712
Shihua Qin , Fangxu Xing , Jihoon Cho , Jinah Park , Xiaofeng Liu , Amir Rouhollahi , Elias J. Bou Farhat , Hoda Javadikasgari , Ashraf Sabe , Farhad R. Nezami , Jonghye Woo , Iman Aganj
Accurate segmentation of the left ventricle (LV) in cardiac CT images is crucial for assessing ventricular function and diagnosing cardiovascular diseases. Creating a sufficiently large training set with accurate manual labels of LV can be cumbersome. More efficient semi-automatic segmentation, however, often includes unwanted structures, such as papillary muscles, due to low contrast between the LV wall and surrounding tissues. This study introduces a two-input-channel method within a Hybrid-Fusion Transformer deep-learning framework to produce refined LV labels from a combination of CT images and semi-automatic rough labels, effectively removing papillary muscles. By leveraging the efficiency of semi-automatic LV segmentation, we train an automatic refined segmentation model on a small set of images with both refined manual and rough semi-automatic labels. Evaluated through quantitative cross-validation, our method outperformed models that used only either CT images or rough masks as input.
心脏CT图像中左心室(LV)的准确分割对于评估心室功能和诊断心血管疾病至关重要。创建一个足够大的训练集并使用准确的LV手动标签是很麻烦的。然而,由于左室壁和周围组织的对比度较低,更有效的半自动分割通常包括不需要的结构,如乳头肌。本研究在Hybrid-Fusion Transformer深度学习框架中引入了一种双输入通道方法,从CT图像和半自动粗糙标记的组合中生成精细的LV标记,有效地去除乳头状肌肉。利用半自动LV分割的效率,我们在一小组图像上训练了一个自动精细分割模型,其中包括精细的手动和粗糙的半自动标签。通过定量交叉验证评估,我们的方法优于仅使用CT图像或粗糙掩模作为输入的模型。
{"title":"Refined myocardium segmentation from CT using a Hybrid-Fusion transformer","authors":"Shihua Qin ,&nbsp;Fangxu Xing ,&nbsp;Jihoon Cho ,&nbsp;Jinah Park ,&nbsp;Xiaofeng Liu ,&nbsp;Amir Rouhollahi ,&nbsp;Elias J. Bou Farhat ,&nbsp;Hoda Javadikasgari ,&nbsp;Ashraf Sabe ,&nbsp;Farhad R. Nezami ,&nbsp;Jonghye Woo ,&nbsp;Iman Aganj","doi":"10.1016/j.bspc.2026.109712","DOIUrl":"10.1016/j.bspc.2026.109712","url":null,"abstract":"<div><div>Accurate segmentation of the left ventricle (LV) in cardiac CT images is crucial for assessing ventricular function and diagnosing cardiovascular diseases. Creating a sufficiently large training set with accurate manual labels of LV can be cumbersome. More efficient semi-automatic segmentation, however, often includes unwanted structures, such as papillary muscles, due to low contrast between the LV wall and surrounding tissues. This study introduces a two-input-channel method within a Hybrid-Fusion Transformer deep-learning framework to produce refined LV labels from a combination of CT images and semi-automatic rough labels, effectively removing papillary muscles. By leveraging the efficiency of semi-automatic LV segmentation, we train an automatic refined segmentation model on a small set of images with both refined manual and rough semi-automatic labels. Evaluated through quantitative cross-validation, our method outperformed models that used only either CT images or rough masks as input.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109712"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Brain tumor classification method based on segmented uniformity measure and spatial shift information fusion 基于分割均匀度测度和空间偏移信息融合的脑肿瘤分类方法
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-12 DOI: 10.1016/j.bspc.2026.109705
Xiaorui Zhang , Peisen Lu , Wei Sun , Rui Jiang
As a common malignant tumor, the accurate classification of brain tumors is crucial for early diagnosis and prevention. Appropriate feature extraction and classification methods can help significantly to achieve this goal. Traditional methods like Local Ternary Patterns (LTP) are suitable for extracting complex texture features of brain tumors, despite deep learning’s excellent results in classifying brain tumor datasets. S2-MLP, while effective in processing brain tumor features extracted by LTP, lacks uniformity and discriminative power, despite enhancing correlation between input features and achieving excellent classification results when combined with LTP. The present research proposes a probability feature expression method based on partition uniformity measure, which reduces computational complexity by transforming three-dimensional coding into two-dimensional coding through partitioning. Regions are labeled based on uniformity measure, with non-uniform regions receiving different labels and uniform regions receiving the same. These labels are converted into features using occurrence probabilities. Additionally, a method for multi-spatial shift segmented attention information fusion is proposed. The classifier is redesigned by expanding the feature maps multiple times, applying spatial shifts in different directions to each feature map, and using a split attention module to fuse the shifted feature maps, enhancing the correlation between features. The internal nodes of the MLP are also optimized to improve the model’s generalization performance. The experiments achieved the highest classification accuracy on the Sa and SfB datasets, achieving 95.32% and 97.26%, respectively, indicating that this method has significant potential applications in brain tumor classification.
脑肿瘤作为一种常见的恶性肿瘤,准确分类对早期诊断和预防至关重要。适当的特征提取和分类方法可以显著帮助实现这一目标。尽管深度学习在脑肿瘤数据集分类方面取得了优异的成绩,但局部三元模式(LTP)等传统方法适用于提取脑肿瘤的复杂纹理特征。S2-MLP虽然对LTP提取的脑肿瘤特征进行了有效的处理,但与LTP结合后,虽然增强了输入特征之间的相关性,取得了很好的分类效果,但统一性和判别能力不足。本研究提出了一种基于分区均匀度测度的概率特征表达方法,通过分区将三维编码转化为二维编码,降低了计算复杂度。根据均匀性度量对区域进行标记,不均匀区域的标签不同,均匀区域的标签相同。使用发生概率将这些标签转换为特征。此外,提出了一种多空间移位分段注意信息融合方法。对分类器进行了重新设计,对特征图进行多次扩展,对每个特征图进行不同方向的空间位移,并使用分裂注意模块对位移后的特征图进行融合,增强特征之间的相关性。为了提高模型的泛化性能,还对MLP的内部节点进行了优化。实验在Sa和SfB数据集上的分类准确率最高,分别达到95.32%和97.26%,表明该方法在脑肿瘤分类中具有重要的潜在应用前景。
{"title":"Brain tumor classification method based on segmented uniformity measure and spatial shift information fusion","authors":"Xiaorui Zhang ,&nbsp;Peisen Lu ,&nbsp;Wei Sun ,&nbsp;Rui Jiang","doi":"10.1016/j.bspc.2026.109705","DOIUrl":"10.1016/j.bspc.2026.109705","url":null,"abstract":"<div><div>As a common malignant tumor, the accurate classification of brain tumors is crucial for early diagnosis and prevention. Appropriate feature extraction and classification methods can help significantly to achieve this goal. Traditional methods like Local Ternary Patterns (LTP) are suitable for extracting complex texture features of brain tumors, despite deep learning’s excellent results in classifying brain tumor datasets. S<sup>2</sup>-MLP, while effective in processing brain tumor features extracted by LTP, lacks uniformity and discriminative power, despite enhancing correlation between input features and achieving excellent classification results when combined with LTP. The present research proposes a probability feature expression method based on partition uniformity measure, which reduces computational complexity by transforming three-dimensional coding into two-dimensional coding through partitioning. Regions are labeled based on uniformity measure, with non-uniform regions receiving different labels and uniform regions receiving the same. These labels are converted into features using occurrence probabilities. Additionally, a method for multi-spatial shift segmented attention information fusion is proposed. The classifier is redesigned by expanding the feature maps multiple times, applying spatial shifts in different directions to each feature map, and using a split attention module to fuse the shifted feature maps, enhancing the correlation between features. The internal nodes of the MLP are also optimized to improve the model’s generalization performance. The experiments achieved the highest classification accuracy on the Sa and SfB datasets, achieving 95.32% and 97.26%, respectively, indicating that this method has significant potential applications in brain tumor classification.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109705"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MAR-GCNet: Multi-label abnormal detection of electrocardiograms by combining multiscale features and graph convolutional networks MAR-GCNet:结合多尺度特征和图卷积网络的多标签心电图异常检测
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-12 DOI: 10.1016/j.bspc.2026.109841
Kan Luo , Haixin He , Yu Chen , Lu You , Jiajia Yang , Dengke Hong , Jianxing Li , Chitin Hon
Cardiovascular diseases (CVDs) are the leading cause of global mortality, and accurate electrocardiogram (ECG) diagnoses are essential for effective clinical interventions. This paper introduces MAR-GCNet, a novel deep learning framework for multi-label ECG anomaly detection that integrates multi-scale feature extraction and inter-class correlation modeling. It combines multi-attention residual networks (MARNs), graph convolutional networks (GCNs) and a weighted fusion strategy. The MARNs incorporate ECA-ResNet blocks with convolutional kernels of sizes 3, 5, and 7 to capture both local and global temporal characteristics in 12-lead ECG signals. The GCNs use a conditional probability matrix (CPM) and a multi-label feature matrix (MLFM) to model inter-class dependencies and mutual exclusivity among cardiac abnormalities. A weighted fusion loss function is employed to integrate the outputs of the MARNs and GCNs branches, enabling optimal multi-label predictions. Experiments on the PTB-XL dataset show that MAR-GCNet outperforms several state-of-the-art (SOTA) models across various annotation levels, achieving the F1 scores of 72.68%, 66.80%, 69.46%, 76.84%, 52.06%, and 90.97% in the “all”, “diag.”, “sub-diag.”, “super-diag.”, “form”, and “rhythm” tasks, respectively. Ablation studies confirm that the integration of multi-scale feature extraction and the two-layer GCN configuration significantly enhance the model performance. These results indicate that MAR-GCNet is a promising approach for accurate and robust automated ECG analysis.
心血管疾病(cvd)是全球死亡的主要原因,准确的心电图(ECG)诊断对于有效的临床干预至关重要。本文介绍了一种新的多标签心电异常检测深度学习框架MAR-GCNet,该框架集成了多尺度特征提取和类间相关建模。它结合了多注意残差网络(marn)、图卷积网络(GCNs)和加权融合策略。marn将ECA-ResNet块与大小为3、5和7的卷积核结合起来,以捕获12导联心电信号的局部和全局时间特征。GCNs使用条件概率矩阵(CPM)和多标签特征矩阵(MLFM)来模拟心脏异常之间的类间依赖性和互斥性。采用加权融合损失函数对marn和GCNs分支的输出进行整合,实现最优的多标签预测。在PTB-XL数据集上的实验表明,MAR-GCNet在不同标注级别上的表现优于几种最先进的(SOTA)模型,在“all”、“diag.”、“sub-diag”上的F1得分分别为72.68%、66.80%、69.46%、76.84%、52.06%和90.97%。”、“super-diag。,“形式”和“节奏”任务。烧蚀研究证实,将多尺度特征提取与两层GCN配置相结合可以显著提高模型的性能。这些结果表明,MAR-GCNet是一种有前途的准确和鲁棒的自动心电分析方法。
{"title":"MAR-GCNet: Multi-label abnormal detection of electrocardiograms by combining multiscale features and graph convolutional networks","authors":"Kan Luo ,&nbsp;Haixin He ,&nbsp;Yu Chen ,&nbsp;Lu You ,&nbsp;Jiajia Yang ,&nbsp;Dengke Hong ,&nbsp;Jianxing Li ,&nbsp;Chitin Hon","doi":"10.1016/j.bspc.2026.109841","DOIUrl":"10.1016/j.bspc.2026.109841","url":null,"abstract":"<div><div>Cardiovascular diseases (CVDs) are the leading cause of global mortality, and accurate electrocardiogram (ECG) diagnoses are essential for effective clinical interventions. This paper introduces MAR-GCNet, a novel deep learning framework for multi-label ECG anomaly detection that integrates multi-scale feature extraction and inter-class correlation modeling. It combines multi-attention residual networks (MARNs), graph convolutional networks (GCNs) and a weighted fusion strategy. The MARNs incorporate ECA-ResNet blocks with convolutional kernels of sizes 3, 5, and 7 to capture both local and global temporal characteristics in 12-lead ECG signals. The GCNs use a conditional probability matrix (CPM) and a multi-label feature matrix (MLFM) to model inter-class dependencies and mutual exclusivity among cardiac abnormalities. A weighted fusion loss function is employed to integrate the outputs of the MARNs and GCNs branches, enabling optimal multi-label predictions. Experiments on the PTB-XL dataset show that MAR-GCNet outperforms several state-of-the-art (SOTA) models across various annotation levels, achieving the F1 scores of 72.68%, 66.80%, 69.46%, 76.84%, 52.06%, and 90.97% in the “all”, “diag.”, “sub-diag.”, “super-diag.”, “form”, and “rhythm” tasks, respectively. Ablation studies confirm that the integration of multi-scale feature extraction and the two-layer GCN configuration significantly enhance the model performance. These results indicate that MAR-GCNet is a promising approach for accurate and robust automated ECG analysis.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109841"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pain intensity classification and evaluation of individual differences in subjects based on hybrid CNN–BiLSTM approach 基于CNN-BiLSTM混合方法的受试者疼痛强度分类及个体差异评估
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-12 DOI: 10.1016/j.bspc.2026.109815
Mingxuan Sun , Yang Liu , Daoshuang Geng , Xiaobang Wu , Daoguo Yang
Pain is a complex subjective experience requiring objective assessment methods for precise diagnosis and treatment. Current approaches relying on self-reports are susceptible to bias and individual variability. This study proposes a cross-mixed model combining a convolutional neural network (CNN) and a bidirectional long short-term memory (BiLSTM) network (hybrid CNN–BiLSTM framework). It classifies pain intensity on the basis of electroencephalography (EEG) signals while explicitly modeling interindividual differences. We introduce a quantitative pain sensitivity index derived from pain threshold and tolerance measurements during cold pressor tests. It facilitates the categorization of subjects into high- and low-sensitivity groups. The CNN component extracts spatial features from EEG time–frequency representations, while the BiLSTM with self-attention captures the temporal dynamics of pain evolution. Subject-independent evaluation was performed using a Leave-One-Subject-Out (LOSO) cross-validation (LOSOCV) strategy. The model achieves accuracies of 88.64% (no pain), 95.80% (mild pain), 99.75% (moderate pain), and 82.96% (severe pain) in the undivided group. When individual sensitivity differences revealed through group-stratified training were considered, the overall accuracy increases to 93.98%, accompanied by increases in Recall and F1-scores increase. Ablation studies confirm the contributions of each architectural component (CNN: spatial feature extraction; BiLSTM: temporal modeling; attention: salient segment weighting; LOSOCV: generalization). Statistical analysis reveals significant correlation between intersubject pain score differences and prediction loss (R2 = 0.45, p < 0.01), validating the effect of individual variability. The proposed framework provides not only accurate pain classification but also a methodology for personalizing pain assessment based on individual sensitivity profiles, showing potential for precise clinical pain management.
疼痛是一种复杂的主观体验,需要客观的评估方法来精确诊断和治疗。目前依赖自我报告的方法容易受到偏见和个体差异的影响。本研究提出了一种结合卷积神经网络(CNN)和双向长短期记忆(BiLSTM)网络(混合CNN - BiLSTM框架)的交叉混合模型。它根据脑电图(EEG)信号对疼痛强度进行分类,同时明确地模拟个体间的差异。我们介绍了一个定量的疼痛敏感性指数,从疼痛阈值和耐受性测量在冷压试验。它有助于将受试者分为高敏感组和低敏感组。CNN组件从EEG时频表征中提取空间特征,而自注意BiLSTM捕获疼痛演化的时间动态。受试者独立评估采用留一受试者(LOSO)交叉验证(LOSOCV)策略进行。在未划分组中,模型的准确率分别为88.64%(无疼痛)、95.80%(轻度疼痛)、99.75%(中度疼痛)和82.96%(重度疼痛)。当考虑群体分层训练的个体敏感性差异时,总体准确率提高到93.98%,同时召回率和f1分数也有所提高。消融研究证实了每个建筑成分的贡献(CNN:空间特征提取;BiLSTM:时间建模;注意力:显著段加权;LOSOCV:泛化)。统计分析显示受试者间疼痛评分差异与预测损失之间存在显著相关性(R2 = 0.45, p < 0.01),验证了个体差异的影响。该框架不仅提供了准确的疼痛分类,还提供了基于个体敏感性特征的个性化疼痛评估方法,显示了精确临床疼痛管理的潜力。
{"title":"Pain intensity classification and evaluation of individual differences in subjects based on hybrid CNN–BiLSTM approach","authors":"Mingxuan Sun ,&nbsp;Yang Liu ,&nbsp;Daoshuang Geng ,&nbsp;Xiaobang Wu ,&nbsp;Daoguo Yang","doi":"10.1016/j.bspc.2026.109815","DOIUrl":"10.1016/j.bspc.2026.109815","url":null,"abstract":"<div><div>Pain is a complex subjective experience requiring objective assessment methods for precise diagnosis and treatment. Current approaches relying on self-reports are susceptible to bias and individual variability. This study proposes a cross-mixed model combining a convolutional neural network (CNN) and a bidirectional long short-term memory (BiLSTM) network (hybrid CNN–BiLSTM framework). It classifies pain intensity on the basis of electroencephalography (EEG) signals while explicitly modeling interindividual differences. We introduce a quantitative pain sensitivity index derived from pain threshold and tolerance measurements during cold pressor tests. It facilitates the categorization of subjects into high- and low-sensitivity groups. The CNN component extracts spatial features from EEG time–frequency representations, while the BiLSTM with self-attention captures the temporal dynamics of pain evolution. Subject-independent evaluation was performed using a Leave-One-Subject-Out (LOSO) cross-validation (LOSOCV) strategy. The model achieves accuracies of 88.64% (no pain), 95.80% (mild pain), 99.75% (moderate pain), and 82.96% (severe pain) in the undivided group. When individual sensitivity differences revealed through group-stratified training were considered, the overall accuracy increases to 93.98%, accompanied by increases in Recall and F1-scores increase. Ablation studies confirm the contributions of each architectural component (CNN: spatial feature extraction; BiLSTM: temporal modeling; attention: salient segment weighting; LOSOCV: generalization). Statistical analysis reveals significant correlation between intersubject pain score differences and prediction loss (<em>R</em><sup>2</sup> = 0.45, <em>p</em> &lt; 0.01), validating the effect of individual variability. The proposed framework provides not only accurate pain classification but also a methodology for personalizing pain assessment based on individual sensitivity profiles, showing potential for precise clinical pain management.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109815"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diffusion model-based medical ultrasound segmentation network in ultrasound image 基于弥散模型的医学超声图像分割网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-14 DOI: 10.1016/j.bspc.2026.109709
Miao Li , Jing Lian , Jizhao Liu , Huaikun Zhang , Bin Shi , Qidong Liu
In medical ultrasound image segmentation, lesion areas are often blurred, making it difficult to distinguish them from the background, thereby complicating segmentation tasks. In the past decade, deep convolutional neural networks have proven effective for medical image segmentation. However, the inductive biases in convolutional architectures limit their ability to capture long-range dependencies. Recently, denoising diffusion probabilistic models (DDPMs) have emerged as powerful generative frameworks in computer vision. Yet, many diffusion-based segmentation approaches overlook the semantic relationships between lesion regions (foreground) and surrounding normal tissues (background), often resulting in distorted segmentation outputs. To address these limitations, we propose DMUS-Net, a diffusion model-based network for medical ultrasound segmentation. DMUS-Net integrates a Multi-Scale Conditional Guidance Network (MSCGN) and Adaptive Detail-Oriented Attention (AODA) modules. By Leveraging the Transformer network’s global relational capabilities, DMUS-Net effectively balances attention between global context and fine-grained features. Subsequently, it dynamically integrates rich image prior information, enhancing semantic correlations between foreground and background. Additionally, we introduce Context-Aware Cross-Decoding layers (CACD) to capture global features and inter-channel correlations, thereby improving both segmentation accuracy and efficiency. DMUS-Net is applied to ultrasound segmentation tasks, including breast, thyroid, and gallbladder stones, achieving superior results, in comparative experiments. These findings highlight DMUS-Net’s robust generalization ability and potential for practical clinical applications.
在医学超声图像分割中,病变区域往往模糊不清,难以从背景中区分出来,从而使分割任务复杂化。在过去的十年中,深度卷积神经网络被证明是有效的医学图像分割。然而,卷积架构中的归纳偏差限制了它们捕获远程依赖关系的能力。近年来,去噪扩散概率模型(ddpm)作为一种强大的生成框架在计算机视觉领域崭露头角。然而,许多基于扩散的分割方法忽略了病变区域(前景)和周围正常组织(背景)之间的语义关系,往往导致分割结果失真。为了解决这些限制,我们提出了DMUS-Net,一个基于扩散模型的医学超声分割网络。DMUS-Net集成了多尺度条件制导网络(MSCGN)和自适应细节导向注意(AODA)模块。通过利用Transformer网络的全局关系功能,DMUS-Net有效地平衡了对全局上下文和细粒度特性的关注。然后,动态整合丰富的图像先验信息,增强前景和背景之间的语义相关性。此外,我们引入上下文感知交叉解码层(CACD)来捕获全局特征和信道间相关性,从而提高分割精度和效率。DMUS-Net应用于超声分割任务,包括乳腺、甲状腺和胆囊结石,在对比实验中取得了优异的效果。这些发现突出了DMUS-Net强大的泛化能力和实际临床应用的潜力。
{"title":"Diffusion model-based medical ultrasound segmentation network in ultrasound image","authors":"Miao Li ,&nbsp;Jing Lian ,&nbsp;Jizhao Liu ,&nbsp;Huaikun Zhang ,&nbsp;Bin Shi ,&nbsp;Qidong Liu","doi":"10.1016/j.bspc.2026.109709","DOIUrl":"10.1016/j.bspc.2026.109709","url":null,"abstract":"<div><div>In medical ultrasound image segmentation, lesion areas are often blurred, making it difficult to distinguish them from the background, thereby complicating segmentation tasks. In the past decade, deep convolutional neural networks have proven effective for medical image segmentation. However, the inductive biases in convolutional architectures limit their ability to capture long-range dependencies. Recently, denoising diffusion probabilistic models (DDPMs) have emerged as powerful generative frameworks in computer vision. Yet, many diffusion-based segmentation approaches overlook the semantic relationships between lesion regions (foreground) and surrounding normal tissues (background), often resulting in distorted segmentation outputs. To address these limitations, we propose DMUS-Net, a diffusion model-based network for medical ultrasound segmentation. DMUS-Net integrates a Multi-Scale Conditional Guidance Network (MSCGN) and Adaptive Detail-Oriented Attention (AODA) modules. By Leveraging the Transformer network’s global relational capabilities, DMUS-Net effectively balances attention between global context and fine-grained features. Subsequently, it dynamically integrates rich image prior information, enhancing semantic correlations between foreground and background. Additionally, we introduce Context-Aware Cross-Decoding layers (CACD) to capture global features and inter-channel correlations, thereby improving both segmentation accuracy and efficiency. DMUS-Net is applied to ultrasound segmentation tasks, including breast, thyroid, and gallbladder stones, achieving superior results, in comparative experiments. These findings highlight DMUS-Net’s robust generalization ability and potential for practical clinical applications.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109709"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unbiased diagnostic report generation via multi-modal counterfactual inference 通过多模态反事实推理生成无偏诊断报告
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-06 DOI: 10.1016/j.bspc.2026.109639
Yuting Guo , Shuai Li , Wenfeng Song , Aimin Hao
Automated diagnostic report generation is a challenging vision-and-language bridging task aimed at accurately describing medical images and performing cross-modal causal inference. Despite its significant clinical importance, widespread application remains challenging. Existing methods often rely on pre-trained models with large-scale medical report datasets, leading to data shifts between training and testing sets, resulting in irrelevant contextual biases in the visual domain and correlation biases within the knowledge graph. To address these issues, we propose a novel multimodal causal inference approach called Multimodal Counterfactual Unbiased Report Generation (MCURG), which incorporates causal inference to exploit invariant rationales. Our key innovation lies in leveraging counterfactual inference to reduce visual and knowledge biases. MCURG employs a Structural Causal Model (SCM) to elucidate the complex relationships among images, knowledge graphs, reports, confounders, and personalized features. We design two multimodal debiasing modules: a visual debiasing module and a knowledge graph debiasing module. The visual debiasing module focuses on the Total Direct Effect of image features, mitigating confounding factors, while the knowledge graph debiasing module identifies individualized treatments within the graph, reducing spurious generations. We conducted extensive experiments and comprehensive evaluations on multiple datasets, demonstrating that MCURG effectively reduces bias and improves the accuracy of generated reports. This multimodal causal inference approach, through the use of SCM and counterfactual reasoning, successfully addresses bias in automated diagnostic report generation, marking a significant innovation in the field. The codes are available at https://github.com/stellating/MCURG.
自动诊断报告生成是一项具有挑战性的视觉和语言桥接任务,旨在准确描述医学图像并执行跨模态因果推理。尽管其具有重要的临床意义,但广泛应用仍然具有挑战性。现有的方法通常依赖于大规模医疗报告数据集的预训练模型,导致数据在训练集和测试集之间发生偏移,从而导致视觉域中的无关上下文偏差和知识图中的相关偏差。为了解决这些问题,我们提出了一种新的多模态因果推理方法,称为多模态反事实无偏报告生成(MCURG),它结合了因果推理来利用不变的基本原理。我们的关键创新在于利用反事实推理来减少视觉和知识偏见。MCURG采用结构因果模型(SCM)来阐明图像、知识图、报告、混杂因素和个性化特征之间的复杂关系。我们设计了两个多模态去偏模块:视觉去偏模块和知识图去偏模块。视觉去偏模块侧重于图像特征的总直接效应,减轻混淆因素,而知识图去偏模块识别图内的个性化治疗,减少虚假生成。我们对多个数据集进行了广泛的实验和综合评估,证明MCURG有效地减少了偏差,提高了生成报告的准确性。这种多模态因果推理方法,通过使用SCM和反事实推理,成功地解决了自动诊断报告生成中的偏见,标志着该领域的重大创新。代码可在https://github.com/stellating/MCURG上获得。
{"title":"Unbiased diagnostic report generation via multi-modal counterfactual inference","authors":"Yuting Guo ,&nbsp;Shuai Li ,&nbsp;Wenfeng Song ,&nbsp;Aimin Hao","doi":"10.1016/j.bspc.2026.109639","DOIUrl":"10.1016/j.bspc.2026.109639","url":null,"abstract":"<div><div>Automated diagnostic report generation is a challenging vision-and-language bridging task aimed at accurately describing medical images and performing cross-modal causal inference. Despite its significant clinical importance, widespread application remains challenging. Existing methods often rely on pre-trained models with large-scale medical report datasets, leading to data shifts between training and testing sets, resulting in irrelevant contextual biases in the visual domain and correlation biases within the knowledge graph. To address these issues, we propose a novel multimodal causal inference approach called Multimodal Counterfactual Unbiased Report Generation (MCURG), which incorporates causal inference to exploit invariant rationales. Our key innovation lies in leveraging counterfactual inference to reduce visual and knowledge biases. MCURG employs a Structural Causal Model (SCM) to elucidate the complex relationships among images, knowledge graphs, reports, confounders, and personalized features. We design two multimodal debiasing modules: a visual debiasing module and a knowledge graph debiasing module. The visual debiasing module focuses on the Total Direct Effect of image features, mitigating confounding factors, while the knowledge graph debiasing module identifies individualized treatments within the graph, reducing spurious generations. We conducted extensive experiments and comprehensive evaluations on multiple datasets, demonstrating that MCURG effectively reduces bias and improves the accuracy of generated reports. This multimodal causal inference approach, through the use of SCM and counterfactual reasoning, successfully addresses bias in automated diagnostic report generation, marking a significant innovation in the field. The codes are available at <span><span>https://github.com/stellating/MCURG</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109639"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193097","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GMMA-Net: A CCTA image segmentation algorithm based on grouped multi-path feature fusion and multi-scale attention GMMA-Net:一种基于分组多路径特征融合和多尺度关注的CCTA图像分割算法
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-06 DOI: 10.1016/j.bspc.2026.109726
Yi Wang , Pei Deng , Tinghui Zheng , Haoyao Cao
Automatic and accurate segmentation of coronary arteries (CA) is a prerequisite for high-precision reconstruction of three-dimensional CA models. However, the complex structure of CA, including low contrast, significant variation in vessel diameter, and high curvature, poses significant challenges for segmentation and reconstruction. In addition, coronary computed tomography angiography (CCTA) images contain abundant background information (such as other tissues, organs, or vessels), further increasing the difficulty of segmentation. These factors often lead to vessel discontinuity and incomplete segmentation. Therefore, accurate CA segmentation remains a challenging task. In this study, we propose the GMMA-Net network to improve the continuity, robustness, and noise resistance of CA segmentation. GMMA-Net employs a grouped multi-path feature fusion module (GMFFM) in the encoder to capture richer multi-scale feature information. Furthermore, by introducing a multi-scale attention module (MAM) into the bottleneck layer of GMMA-Net, we achieve dynamic weight adjustment, capture long-range dependencies, and suppress redundant features. Experimental results show that GMMA-Net outperforms existing methods in the task of CA segmentation from CCTA images, effectively overcoming challenges caused by scale sensitivity and noise interference. GMMA-Net demonstrates superior performance on metrics such as IoU, Dice coefficient, recall rate, and HD95, especially exhibiting stronger segmentation capability when handling cases with poor image quality and large variations in vessel diameter. The code of the proposed method is available at https://github.com/DengPei-C/GMMA-Net.
自动准确分割冠状动脉是实现冠状动脉三维模型高精度重建的前提。然而,CA结构复杂,对比度低,血管直径变化大,曲率大,给分割和重建带来了很大的挑战。此外,冠状动脉ct血管造影(CCTA)图像包含丰富的背景信息(如其他组织、器官或血管),进一步增加了分割的难度。这些因素往往导致血管不连续性和不完全分割。因此,准确的CA分割仍然是一项具有挑战性的任务。在本研究中,我们提出了GMMA-Net网络来提高CA分割的连续性、鲁棒性和抗噪声性。GMMA-Net在编码器中采用分组多路径特征融合模块(GMFFM)来捕获更丰富的多尺度特征信息。此外,通过在GMMA-Net的瓶颈层中引入多尺度注意力模块(MAM),实现了动态权值调整、远程依赖关系捕获和冗余特征抑制。实验结果表明,GMMA-Net在CCTA图像的CA分割任务中优于现有方法,有效克服了尺度敏感性和噪声干扰带来的挑战。GMMA-Net在IoU、Dice系数、召回率和HD95等指标上表现优异,特别是在处理图像质量差和血管直径变化大的情况下表现出更强的分割能力。所提出的方法的代码可在https://github.com/DengPei-C/GMMA-Net上获得。
{"title":"GMMA-Net: A CCTA image segmentation algorithm based on grouped multi-path feature fusion and multi-scale attention","authors":"Yi Wang ,&nbsp;Pei Deng ,&nbsp;Tinghui Zheng ,&nbsp;Haoyao Cao","doi":"10.1016/j.bspc.2026.109726","DOIUrl":"10.1016/j.bspc.2026.109726","url":null,"abstract":"<div><div>Automatic and accurate segmentation of coronary arteries (CA) is a prerequisite for high-precision reconstruction of three-dimensional CA models. However, the complex structure of CA, including low contrast, significant variation in vessel diameter, and high curvature, poses significant challenges for segmentation and reconstruction. In addition, coronary computed tomography angiography (CCTA) images contain abundant background information (such as other tissues, organs, or vessels), further increasing the difficulty of segmentation. These factors often lead to vessel discontinuity and incomplete segmentation. Therefore, accurate CA segmentation remains a challenging task. In this study, we propose the GMMA-Net network to improve the continuity, robustness, and noise resistance of CA segmentation. GMMA-Net employs a grouped multi-path feature fusion module (GMFFM) in the encoder to capture richer multi-scale feature information. Furthermore, by introducing a multi-scale attention module (MAM) into the bottleneck layer of GMMA-Net, we achieve dynamic weight adjustment, capture long-range dependencies, and suppress redundant features. Experimental results show that GMMA-Net outperforms existing methods in the task of CA segmentation from CCTA images, effectively overcoming challenges caused by scale sensitivity and noise interference. GMMA-Net demonstrates superior performance on metrics such as IoU, Dice coefficient, recall rate, and <span><math><mrow><mi>H</mi><msub><mrow><mi>D</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span>, especially exhibiting stronger segmentation capability when handling cases with poor image quality and large variations in vessel diameter. The code of the proposed method is available at <span><span>https://github.com/DengPei-C/GMMA-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109726"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic regulation of brain network and muscle activity in upper limb force generation among older adults: A temporal dynamic graph Fourier transform approach 老年人上肢力量产生中脑网络和肌肉活动的动态调节:时间动态图傅立叶变换方法
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-13 DOI: 10.1016/j.bspc.2026.109829
Mingxia Zhang , Huijing Hu , Di Ao , Li Yan , Qinghua Huang , Zhengxiang Zhang , Le Li
This study investigates the neural mechanisms underlying age-related declines in motor control by proposing a novel Temporal Dynamic Graph Fourier Transform (TDGFT) method. TDGFT integrates graph signal processing with dynamic brain networks analysis to characterize time-varying corticomuscular interactions in the spectral domain, thereby linking global and local brain connectivity patterns to motor behavior. Integrating functional near-infrared spectroscopy (fNIRS) and electromyography (EMG), we systematically examine the dynamic regulation of brain network and muscle activity in older adults and younger adults during elbow flexion tasks at 30% and 70% of maximum voluntary contraction (MVC). Sixteen older adults and sixteen younger adults are recruited for the study. Our findings reveal that older adults exhibit weaker dynamic regulation of brain regions during high-load tasks, accompanied by significantly increased constraints of structural brain networks on functional activity, reflecting a decline in cognitive control. Additionally, older adults rely on multi-regional brain coordination for motor control during low-intensity tasks, while reducing cognitive load to enhance motor efficiency during high-intensity tasks. By providing an interpretable spectral representation of corticomuscular dynamics, TDGFT advances the understanding of how aging reshapes motor-related brain connectivity. These findings may help identify changes of age-related motor decline and facilitate the design of individualized motor rehabilitation strategies for older adults.
本研究通过提出一种新的时间动态图傅立叶变换(TDGFT)方法来研究运动控制能力与年龄相关衰退的神经机制。TDGFT将图形信号处理与动态脑网络分析相结合,以表征谱域中随时间变化的皮质肌肉相互作用,从而将全球和局部大脑连接模式与运动行为联系起来。结合功能性近红外光谱(fNIRS)和肌电图(EMG),我们系统地研究了老年人和年轻人在最大自愿收缩(MVC)的30%和70%时屈肘任务时脑网络和肌肉活动的动态调节。这项研究招募了16名老年人和16名年轻人。我们的研究结果表明,老年人在高负荷任务中表现出较弱的大脑区域动态调节,伴随着大脑结构网络对功能活动的限制显著增加,反映了认知控制的下降。此外,老年人在低强度任务中依靠多区域大脑协调来控制运动,而在高强度任务中减少认知负荷以提高运动效率。通过提供皮质肌肉动力学的可解释的光谱表示,TDGFT促进了对衰老如何重塑运动相关大脑连接的理解。这些发现可能有助于识别与年龄相关的运动衰退的变化,并促进老年人个性化运动康复策略的设计。
{"title":"Dynamic regulation of brain network and muscle activity in upper limb force generation among older adults: A temporal dynamic graph Fourier transform approach","authors":"Mingxia Zhang ,&nbsp;Huijing Hu ,&nbsp;Di Ao ,&nbsp;Li Yan ,&nbsp;Qinghua Huang ,&nbsp;Zhengxiang Zhang ,&nbsp;Le Li","doi":"10.1016/j.bspc.2026.109829","DOIUrl":"10.1016/j.bspc.2026.109829","url":null,"abstract":"<div><div>This study investigates the neural mechanisms underlying age-related declines in motor control by proposing a novel Temporal Dynamic Graph Fourier Transform (TDGFT) method. TDGFT integrates graph signal processing with dynamic brain networks analysis to characterize time-varying corticomuscular interactions in the spectral domain, thereby linking global and local brain connectivity patterns to motor behavior. Integrating functional near-infrared spectroscopy (fNIRS) and electromyography (EMG), we systematically examine the dynamic regulation of brain network and muscle activity in older adults and younger adults during elbow flexion tasks at 30% and 70% of maximum voluntary contraction (MVC). Sixteen older adults and sixteen younger adults are recruited for the study. Our findings reveal that older adults exhibit weaker dynamic regulation of brain regions during high-load tasks, accompanied by significantly increased constraints of structural brain networks on functional activity, reflecting a decline in cognitive control. Additionally, older adults rely on multi-regional brain coordination for motor control during low-intensity tasks, while reducing cognitive load to enhance motor efficiency during high-intensity tasks. By providing an interpretable spectral representation of corticomuscular dynamics, TDGFT advances the understanding of how aging reshapes motor-related brain connectivity. These findings may help identify changes of age-related motor decline and facilitate the design of individualized motor rehabilitation strategies for older adults.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109829"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DeepOsteoCls: Deep learning-based framework for Knee Osteoarthritis Classification with qualitative explanations from radiographs and MRI volumes DeepOsteoCls:基于深度学习的膝关节骨关节炎分类框架,从x线片和MRI体积中进行定性解释
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-10 DOI: 10.1016/j.bspc.2026.109819
Akshay Daydar , Arijit Sur , Subramani Kanagaraj , Hanif Laskar
Knee Osteoarthritis (KOA) is a degenerative joint disorder affecting middle-aged and elderly individuals, with its diagnosis facing challenges in achieving objective, transparent quantification and incorporating clinical manifestations, despite advances in deep-learning for medical imaging. To address these issues, in this paper, a deep learning-based hybrid (Convolutional Neural Network (CNN)-Transformer encoder) classification framework, DeepOsteoCls, is proposed to perform binary and multi-class classification of KOA from X-rays and MRI scans from OsteoXRNet and OsteoMRNet models separately, with Gradient-weighted Class Activation Mappings (Grad-CAMs). The Osteoarthritis Edge Detection (OAED) and Multi-Resolution Feature Integration (MRFI) modules are also introduced in the proposed framework to facilitate the extraction of edge-based features from X-ray images and multi-scale regional features from the MRI volume, respectively. Furthermore, a disorder-aware weakly supervised training scheme—Domain Knowledge Transfer and Entropy Regularization (DoKTER) is proposed to enhance the explainability of Radiological KOA (RKOA) diagnosis by predicting the region score and GradCAMs of MRI scans. Comprehensive experiments on the Osteoarthritis Initiative (OAI) dataset demonstrated that the proposed framework achieved a classification accuracy of 72.10% for X-ray and 53.16% for MRI in a multi-class classification task, and 85.74% for X-ray and 81.04% for MRI in a binary classification task, outperforming state-of-the-art models. The DoKTER scheme is found to accurately classify the affected region with 65.15% and 62.5% for the OAI and Multi-Hospital Knee Osteoarthritis (MHKOA) datasets, respectively. Additionally, Femoral Cartilage Thickness (FCT) in non-RKOA subjects can be effectively monitored using the region score, with distinct cut-offs values. The code is available at: https://github.com/adaydar/Deep-OsteoCls
膝关节骨性关节炎(KOA)是一种影响中老年人的退行性关节疾病,尽管在医学成像的深度学习方面取得了进展,但其诊断在实现客观、透明的量化和纳入临床表现方面面临挑战。为了解决这些问题,本文提出了一种基于深度学习的混合(卷积神经网络(CNN)-变压器编码器)分类框架DeepOsteoCls,该框架使用梯度加权类激活映射(梯度- cams)分别对来自骨oxrnet和骨omrnet模型的x射线和MRI扫描的KOA进行二元和多类分类。在该框架中还引入了骨关节炎边缘检测(OAED)和多分辨率特征集成(MRFI)模块,分别用于从x射线图像中提取基于边缘的特征和从MRI体积中提取多尺度区域特征。在此基础上,提出了一种无序感知的弱监督训练方案——领域知识转移和熵正则化(DoKTER),通过预测MRI扫描的区域评分和梯度梯度,提高放射KOA (RKOA)诊断的可解释性。在Osteoarthritis Initiative (OAI)数据集上进行的综合实验表明,该框架在多类别分类任务中,x射线和MRI的分类准确率分别为72.10%和53.16%,在二元分类任务中,x射线和MRI的分类准确率分别为85.74%和81.04%,优于目前最先进的模型。发现DoKTER方案对OAI和多医院膝关节骨关节炎(MHKOA)数据集的影响区域分类准确率分别为65.15%和62.5%。此外,非rkoa受试者的股骨软骨厚度(FCT)可以使用区域评分有效监测,具有不同的截止值。代码可从https://github.com/adaydar/Deep-OsteoCls获得
{"title":"DeepOsteoCls: Deep learning-based framework for Knee Osteoarthritis Classification with qualitative explanations from radiographs and MRI volumes","authors":"Akshay Daydar ,&nbsp;Arijit Sur ,&nbsp;Subramani Kanagaraj ,&nbsp;Hanif Laskar","doi":"10.1016/j.bspc.2026.109819","DOIUrl":"10.1016/j.bspc.2026.109819","url":null,"abstract":"<div><div>Knee Osteoarthritis (KOA) is a degenerative joint disorder affecting middle-aged and elderly individuals, with its diagnosis facing challenges in achieving objective, transparent quantification and incorporating clinical manifestations, despite advances in deep-learning for medical imaging. To address these issues, in this paper, a deep learning-based hybrid (Convolutional Neural Network (CNN)-Transformer encoder) classification framework, DeepOsteoCls, is proposed to perform binary and multi-class classification of KOA from X-rays and MRI scans from OsteoXRNet and OsteoMRNet models separately, with Gradient-weighted Class Activation Mappings (Grad-CAMs). The Osteoarthritis Edge Detection (OAED) and Multi-Resolution Feature Integration (MRFI) modules are also introduced in the proposed framework to facilitate the extraction of edge-based features from X-ray images and multi-scale regional features from the MRI volume, respectively. Furthermore, a disorder-aware weakly supervised training scheme—Domain Knowledge Transfer and Entropy Regularization (DoKTER) is proposed to enhance the explainability of Radiological KOA (RKOA) diagnosis by predicting the region score and GradCAMs of MRI scans. Comprehensive experiments on the Osteoarthritis Initiative (OAI) dataset demonstrated that the proposed framework achieved a classification accuracy of 72.10% for X-ray and 53.16% for MRI in a multi-class classification task, and 85.74% for X-ray and 81.04% for MRI in a binary classification task, outperforming state-of-the-art models. The DoKTER scheme is found to accurately classify the affected region with 65.15% and 62.5% for the OAI and Multi-Hospital Knee Osteoarthritis (MHKOA) datasets, respectively. Additionally, Femoral Cartilage Thickness (FCT) in non-RKOA subjects can be effectively monitored using the region score, with distinct cut-offs values. The code is available at: <span><span>https://github.com/adaydar/Deep-OsteoCls</span><svg><path></path></svg></span></div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109819"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Biomedical Signal Processing and Control
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1