首页 > 最新文献

Biomedical Signal Processing and Control最新文献

英文 中文
Clustering-enhanced active learning with dynamic sampling for brain tumor classification 基于动态采样的聚类增强主动学习脑肿瘤分类
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-31 DOI: 10.1016/j.bspc.2026.109715
Yawen Fan , Xiang Wang , Zhen Yue , Xinchen Zhang , Mingkai Chen , Jianxin Chen
Automated classification of brain tumors is essential for reliable diagnosis and effective treatment planning. However, deep learning-based methods require large, well-labeled MRI datasets, which can be expensive, time-consuming, and challenging to obtain in clinical settings. Moreover, real-world datasets often exhibit severe class imbalance and inter-subject variability, both of which can compromise model robustness and limit generalization to unseen cases. In this paper, we introduce a novel dynamic active learning framework enhanced by clustering for brain tumor classification. First, the proposed framework extracts high-level features of MRI images by a self-supervised learning method, which are then clustered to form a multi-class data pool, providing a pre-classification of the samples. To reduce annotation effort while maintaining model performance, the framework dynamically selects the most informative samples from each cluster by jointly considering prediction uncertainty and cluster diversity. Additionally, we have constructed a high-quality brain tumor MRI dataset that includes three tumor types: glioma, metastatic tumor, and diffuse large B-cell lymphoma. Notably, the latter is scarce in existing public datasets. Extensive experiments on both public and private datasets show that the proposed method achieves competitive performance using only a small portion of labeled data. Also, on an external test set, the method obtained an average accuracy of 0.92. All these results suggest that our method offers a practical and efficient solution for MRI-based brain tumor classification in real-world clinical settings.
脑肿瘤的自动分类对于可靠的诊断和有效的治疗计划至关重要。然而,基于深度学习的方法需要大量的、标记良好的MRI数据集,这可能是昂贵的、耗时的,并且在临床环境中难以获得。此外,现实世界的数据集经常表现出严重的类别不平衡和学科间的可变性,这两者都会损害模型的鲁棒性,并限制对未知情况的泛化。本文提出了一种新的基于聚类的动态主动学习框架,用于脑肿瘤分类。首先,提出的框架通过自监督学习方法提取MRI图像的高级特征,然后将其聚类形成多类数据池,提供样本的预分类。为了在保持模型性能的同时减少标注工作量,该框架通过综合考虑预测不确定性和聚类多样性,从每个聚类中动态选择信息量最大的样本。此外,我们还构建了一个高质量的脑肿瘤MRI数据集,其中包括三种肿瘤类型:胶质瘤、转移性肿瘤和弥漫性大b细胞淋巴瘤。值得注意的是,后者在现有的公共数据集中是稀缺的。在公共和私有数据集上进行的大量实验表明,该方法仅使用一小部分标记数据就能获得具有竞争力的性能。此外,在外部测试集上,该方法的平均精度为0.92。所有这些结果表明,我们的方法在现实世界的临床环境中为基于mri的脑肿瘤分类提供了一种实用有效的解决方案。
{"title":"Clustering-enhanced active learning with dynamic sampling for brain tumor classification","authors":"Yawen Fan ,&nbsp;Xiang Wang ,&nbsp;Zhen Yue ,&nbsp;Xinchen Zhang ,&nbsp;Mingkai Chen ,&nbsp;Jianxin Chen","doi":"10.1016/j.bspc.2026.109715","DOIUrl":"10.1016/j.bspc.2026.109715","url":null,"abstract":"<div><div>Automated classification of brain tumors is essential for reliable diagnosis and effective treatment planning. However, deep learning-based methods require large, well-labeled MRI datasets, which can be expensive, time-consuming, and challenging to obtain in clinical settings. Moreover, real-world datasets often exhibit severe class imbalance and inter-subject variability, both of which can compromise model robustness and limit generalization to unseen cases. In this paper, we introduce a novel dynamic active learning framework enhanced by clustering for brain tumor classification. First, the proposed framework extracts high-level features of MRI images by a self-supervised learning method, which are then clustered to form a multi-class data pool, providing a pre-classification of the samples. To reduce annotation effort while maintaining model performance, the framework dynamically selects the most informative samples from each cluster by jointly considering prediction uncertainty and cluster diversity. Additionally, we have constructed a high-quality brain tumor MRI dataset that includes three tumor types: glioma, metastatic tumor, and diffuse large B-cell lymphoma. Notably, the latter is scarce in existing public datasets. Extensive experiments on both public and private datasets show that the proposed method achieves competitive performance using only a small portion of labeled data. Also, on an external test set, the method obtained an average accuracy of 0.92. All these results suggest that our method offers a practical and efficient solution for MRI-based brain tumor classification in real-world clinical settings.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109715"},"PeriodicalIF":4.9,"publicationDate":"2026-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Detection of Epileptic Seizures via EEG Signals and Texture Analysis of Recurrence Plots 基于脑电图信号和递归图纹理分析的癫痫发作改进检测
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1016/j.bspc.2026.109675
Saleh Lashkari , Mohammad Ali Khalilzadeh , Seyyed Ali Zendehbad , Mohammad Reza Khakzad , Shahryar Salmani Bajestani , Elias Mazrooei Rad

Introduction

It is thus essential to describe dynamic changes in brain signals, especially in the Electroencephalogram (EEG) data, to analyze nonlinear and chaotic patterns, particularly in epilepsy. This work presents an approach that seeks to incorporate recurrence plots with textural features in epilepsy seizure detection. This method provides a detailed picture of brain activity and can be helpful for clinical neuroscience.

Materials and methods

In this study, Gray-Level Co-occurrence Matrix (GLCM) texture features were computed from recurrence plots of EEG signals to characterize neural dynamics. GLCM measures spatial relationships in the data, providing detailed insights into temporal and spatial patterns of brain activity. The method was initially validated on chaotic systems, demonstrating its ability to capture nonlinear behaviors. It was then applied to EEG data to detect seizures, highlighting its potential in clinical settings.

Results

The proposed framework outperformed traditional Recurrence Quantification Analysis (RQA) and other methods in detecting epileptic seizures. The GLCM-enhanced recurrence plots provided a more accurate and sensitive representation of brain dynamics, allowing for earlier and more reliable seizure detection. This method shows strong potential for clinical applications, enhancing the ability to detect seizures early. On the Bonn EEG corpus, the proposed GRP–GLCM + SVM pipeline achieved 98.6 % (Case 1: AB/CD/E), 99.6 % (Case 2: ABCD/E), and 100 % (Case 3: D/E) Accuracy under nested cross-validation. Precision, Recall, and F1 were ≥ 0.98 in Cases 1–2 and 1.00 in Case 3 (zero FP/FN). Compared with an RQA baseline (76.8 %, 95.6 %, 91.0 %), these results reflect + 21.8, +4.0, and + 9.0 percentage-point gains while remaining competitive with recent CNN-based approaches and preserving interpretability.

Conclusion

This study demonstrates that textural analysis of recurrence plots, particularly using GLCM features, provides a robust and efficient tool for epileptic seizure detection. By capturing subtle changes in brain activity, the framework offers a promising approach for improving early detection and intervention in clinical neuroscience.
因此,描述脑信号的动态变化,特别是脑电图(EEG)数据,分析非线性和混沌模式,特别是癫痫,是必不可少的。这项工作提出了一种方法,旨在结合复发图与纹理特征在癫痫发作检测。这种方法提供了大脑活动的详细图像,可以帮助临床神经科学。材料与方法本研究利用脑电信号的递归图计算灰度共生矩阵(GLCM)纹理特征来表征神经动力学。GLCM测量数据中的空间关系,为大脑活动的时空模式提供详细的见解。该方法在混沌系统上进行了初步验证,证明了其捕获非线性行为的能力。然后将其应用于脑电图数据以检测癫痫发作,突出其在临床环境中的潜力。结果该框架在癫痫发作检测方面优于传统的复发量化分析(RQA)等方法。glcm增强的复发图提供了更准确和敏感的脑动力学表征,允许更早和更可靠的癫痫发作检测。该方法具有很强的临床应用潜力,提高了早期发现癫痫发作的能力。在波恩脑电图语料库上,本文提出的GRP-GLCM + SVM管道在嵌套交叉验证下达到98.6%(案例1:AB/CD/E)、99.6%(案例2:ABCD/E)和100%(案例3:D/E)的准确率。病例1-2的Precision、Recall和F1≥0.98,病例3的F1≥1.00 (FP/FN为零)。与RQA基线(76.8%,95.6%,91.0%)相比,这些结果反映了+ 21.8%,+4.0和+ 9.0个百分点的增长,同时与最近基于cnn的方法保持竞争力并保持可解释性。结论本研究表明,复发图的纹理分析,特别是使用GLCM特征,为癫痫发作检测提供了一种强大而有效的工具。通过捕捉大脑活动的细微变化,该框架为改善临床神经科学的早期检测和干预提供了一种有希望的方法。
{"title":"Improved Detection of Epileptic Seizures via EEG Signals and Texture Analysis of Recurrence Plots","authors":"Saleh Lashkari ,&nbsp;Mohammad Ali Khalilzadeh ,&nbsp;Seyyed Ali Zendehbad ,&nbsp;Mohammad Reza Khakzad ,&nbsp;Shahryar Salmani Bajestani ,&nbsp;Elias Mazrooei Rad","doi":"10.1016/j.bspc.2026.109675","DOIUrl":"10.1016/j.bspc.2026.109675","url":null,"abstract":"<div><h3>Introduction</h3><div>It is thus essential to describe dynamic changes in brain signals, especially in the Electroencephalogram (EEG) data, to analyze nonlinear and chaotic patterns, particularly in epilepsy. This work presents an approach that seeks to incorporate recurrence plots with textural features in epilepsy seizure detection. This method provides a detailed picture of brain activity and can be helpful for clinical neuroscience.</div></div><div><h3>Materials and methods</h3><div>In this study, Gray-Level Co-occurrence Matrix (GLCM) texture features were computed from recurrence plots of EEG signals to characterize neural dynamics. GLCM measures spatial relationships in the data, providing detailed insights into temporal and spatial patterns of brain activity. The method was initially validated on chaotic systems, demonstrating its ability to capture nonlinear behaviors. It was then applied to EEG data to detect seizures, highlighting its potential in clinical settings.</div></div><div><h3>Results</h3><div>The proposed framework outperformed traditional Recurrence Quantification Analysis (RQA) and other methods in detecting epileptic seizures. The GLCM-enhanced recurrence plots provided a more accurate and sensitive representation of brain dynamics, allowing for earlier and more reliable seizure detection. This method shows strong potential for clinical applications, enhancing the ability to detect seizures early. On the Bonn EEG corpus, the proposed GRP–GLCM + SVM pipeline achieved 98.6 % (Case 1: AB/CD/E), 99.6 % (Case 2: ABCD/E), and 100 % (Case 3: D/E) Accuracy under nested cross-validation. Precision, Recall, and F1 were ≥ 0.98 in Cases 1–2 and 1.00 in Case 3 (zero FP/FN). Compared with an RQA baseline (76.8 %, 95.6 %, 91.0 %), these results reflect + 21.8, +4.0, and + 9.0 percentage-point gains while remaining competitive with recent CNN-based approaches and preserving interpretability.</div></div><div><h3>Conclusion</h3><div>This study demonstrates that textural analysis of recurrence plots, particularly using GLCM features, provides a robust and efficient tool for epileptic seizure detection. By capturing subtle changes in brain activity, the framework offers a promising approach for improving early detection and intervention in clinical neuroscience.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109675"},"PeriodicalIF":4.9,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AXNet: Attention-enhanced X-ray network for pneumonia detection AXNet:用于肺炎检测的注意力增强x射线网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1016/j.bspc.2026.109618
Mojtaba Jahanian , Abbas Karimi , Nafiseh Osati Eraghi , Faraneh Zarafshan

Background:

Pneumonia remains one of the leading causes of childhood mortality worldwide, especially in low-resource clinical settings where access to expert radiologists is limited. Automated and interpretable deep learning models can provide rapid and reliable diagnostic support.

Objective:

This study introduces AXNet+ECA, a lightweight attention-augmented convolutional neural network, designed to improve pneumonia detection from pediatric chest X-ray (CXR) images while ensuring computational efficiency and interpretability. The novelty of AXNet+ECA lies in the dual-attention integration of Convolutional Block Attention Module (CBAM) and Efficient Channel Attention (ECA) mechanisms within a lightweight backbone, jointly enhancing diagnostic accuracy and model interpretability while maintaining computational frugality.

Methods:

The proposed model builds upon the ResNet-18 backbone by embedding CBAM blocks within each residual stage and appending an ECA head for fine-grained channel calibration. AXNet+ECA was trained and evaluated on 5863 pediatric chest X-ray images from the publicly available Kaggle pneumonia dataset, using an 80–10–10 train/validation/test split. Evaluation encompassed baseline comparisons, ablation studies, robustness analysis, and statistical significance testing.

Results:

AXNet+ECA achieved a test accuracy of 93.6%, F1-score of 93.1%, and AUC of 0.964, outperforming or matching CNN baselines (ResNet-18, DenseNet-121, VGG-16, CheXNet) and recent transformer-based models (ViT-B/16, Swin-T). Despite competitive performance, AXNet+ECA requires only 13.1M parameters and 4.7 ms/image inference time, highlighting its computational efficiency. Visual interpretability via CBAM and Grad-CAM revealed 86.7% alignment with radiologist-annotated abnormalities.

Conclusion:

By integrating dual-path attention within a compact architecture, AXNet+ECA achieves an effective balance between diagnostic accuracy, interpretability, and efficiency. These characteristics underline its potential for real-time clinical deployment in resource-constrained healthcare environments and large-scale screening initiatives.
背景:肺炎仍然是世界范围内儿童死亡的主要原因之一,特别是在资源匮乏的临床环境中,获得放射科专家的机会有限。自动化和可解释的深度学习模型可以提供快速可靠的诊断支持。目的:本研究介绍了AXNet+ECA,一种轻量级的注意力增强卷积神经网络,旨在提高儿童胸部x射线(CXR)图像的肺炎检测,同时确保计算效率和可解释性。AXNet+ECA的新颖之处在于在轻量级主干内集成了卷积块注意模块(CBAM)和高效通道注意(ECA)机制的双注意,在保持计算节约的同时,共同提高了诊断准确性和模型可解释性。方法:提出的模型建立在ResNet-18主干上,通过在每个剩余阶段嵌入CBAM块并附加ECA头进行细粒度通道校准。AXNet+ECA对来自公开可用的Kaggle肺炎数据集的5863张儿科胸部x射线图像进行训练和评估,采用80-10-10训练/验证/测试分割。评估包括基线比较、消融研究、稳健性分析和统计显著性检验。结果:AXNet+ECA的测试准确率为93.6%,f1得分为93.1%,AUC为0.964,优于或匹配CNN基线(ResNet-18、DenseNet-121、VGG-16、CheXNet)和最新的基于变压器的模型(vitb /16、swun - t)。尽管具有相当的性能,AXNet+ECA只需要131m参数和4.7 ms/图像推理时间,突出了其计算效率。通过CBAM和Grad-CAM的视觉可解释性显示86.7%的异常与放射科医生注释的异常一致。结论:通过在紧凑的体系结构中集成双路径关注,AXNet+ECA在诊断准确性、可解释性和效率之间实现了有效的平衡。这些特点强调了它在资源有限的医疗环境和大规模筛查活动中进行实时临床部署的潜力。
{"title":"AXNet: Attention-enhanced X-ray network for pneumonia detection","authors":"Mojtaba Jahanian ,&nbsp;Abbas Karimi ,&nbsp;Nafiseh Osati Eraghi ,&nbsp;Faraneh Zarafshan","doi":"10.1016/j.bspc.2026.109618","DOIUrl":"10.1016/j.bspc.2026.109618","url":null,"abstract":"<div><h3>Background:</h3><div>Pneumonia remains one of the leading causes of childhood mortality worldwide, especially in low-resource clinical settings where access to expert radiologists is limited. Automated and interpretable deep learning models can provide rapid and reliable diagnostic support.</div></div><div><h3>Objective:</h3><div>This study introduces <strong>AXNet+ECA</strong>, a lightweight attention-augmented convolutional neural network, designed to improve pneumonia detection from pediatric chest X-ray (CXR) images while ensuring computational efficiency and interpretability. The novelty of AXNet+ECA lies in the dual-attention integration of <em>Convolutional Block Attention Module (CBAM)</em> and <em>Efficient Channel Attention (ECA)</em> mechanisms within a lightweight backbone, jointly enhancing diagnostic accuracy and model interpretability while maintaining computational frugality.</div></div><div><h3>Methods:</h3><div>The proposed model builds upon the ResNet-18 backbone by embedding CBAM blocks within each residual stage and appending an ECA head for fine-grained channel calibration. AXNet+ECA was trained and evaluated on 5863 pediatric chest X-ray images from the publicly available Kaggle pneumonia dataset, using an 80–10–10 train/validation/test split. Evaluation encompassed baseline comparisons, ablation studies, robustness analysis, and statistical significance testing.</div></div><div><h3>Results:</h3><div>AXNet+ECA achieved a test accuracy of <strong>93.6%</strong>, F1-score of <strong>93.1%</strong>, and AUC of <strong>0.964</strong>, outperforming or matching CNN baselines (ResNet-18, DenseNet-121, VGG-16, CheXNet) and recent transformer-based models (ViT-B/16, Swin-T). Despite competitive performance, AXNet+ECA requires only <strong>13.1M parameters</strong> and <strong>4.7 ms/image</strong> inference time, highlighting its computational efficiency. Visual interpretability via CBAM and Grad-CAM revealed <strong>86.7%</strong> alignment with radiologist-annotated abnormalities.</div></div><div><h3>Conclusion:</h3><div>By integrating dual-path attention within a compact architecture, AXNet+ECA achieves an effective balance between diagnostic accuracy, interpretability, and efficiency. These characteristics underline its potential for real-time clinical deployment in resource-constrained healthcare environments and large-scale screening initiatives.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109618"},"PeriodicalIF":4.9,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifying relevant EEG features for personalized emotional videos: a cross-population analysis in healthy controls and patients with disorders of consciousness 识别个性化情感视频的相关EEG特征:健康对照和意识障碍患者的跨人群分析
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1016/j.bspc.2026.109669
Anny Maza , Sandra Goizueta , María Dolores Navarro , Enrique Noé , Joan Ferri , Valery Naranjo , Roberto Llorens
Emerging evidence suggests that personalized emotional stimuli can elicit measurable brain responses indicative of cognitive processing. This could be particularly relevant in patients with disorders of consciousness (DOC), where accurate assessment remains remarkably challenging. Despite this, the electroencephalography (EEG) features most sensitive to these personalized emotional stimuli, and their generalizability from healthy individuals to patients with DOC, remain underexplored. This study aimed to identify EEG features distinguishing brain responses to familiar versus non-familiar audiovisual stimuli in healthy controls and to assess their applicability in patients with DOC. Nineteen healthy controls and nineteen patients with DOC viewed personalized emotional videos featuring either familiar or unfamiliar individuals while EEG data were recorded. Seventeen EEG features across various domains were compared using subject-independent machine-learning models in healthy controls, and the top-performing features were validated in patients with DOC. Results indicated fuzzy entropy, common spatial pattern (CSP), and Hjorth activity were the most discriminative features. Applying models trained on healthy individuals to patients with DOC revealed statistically significant performances in 60% of patients in minimally conscious state and 33% of patients with unresponsive wakefulness syndrome. Topographical analyses identified prominent differences in temporal, parietal, and frontal regions within beta and gamma bands for healthy controls, partially replicated in responsive patients. These findings underscore fuzzy entropy, CSP, and Hjorth activity as sensitive EEG markers for detecting emotional responses to personalized videos in healthy controls. Their partial generalization to patients with DOC highlights potential clinical utility in assessing residual cognitive processing and consciousness, particularly in responsive states.
新出现的证据表明,个性化的情绪刺激可以引发可测量的大脑反应,表明认知过程。这在意识障碍(DOC)患者中尤为重要,在这些患者中,准确的评估仍然非常具有挑战性。尽管如此,脑电图(EEG)对这些个性化情绪刺激最敏感,其从健康个体到DOC患者的普遍性仍未得到充分探索。本研究旨在确定健康对照者对熟悉与不熟悉的视听刺激的脑电图特征,并评估其在DOC患者中的适用性。19名健康对照者和19名DOC患者观看了个性化的情绪视频,其中包括熟悉或不熟悉的个体,同时记录了脑电图数据。使用独立于受试者的机器学习模型对健康对照中不同领域的17个EEG特征进行了比较,并在DOC患者中验证了表现最好的特征。结果表明,模糊熵、共同空间格局(CSP)和Hjorth活动是最具判别性的特征。将健康个体训练的模型应用于DOC患者,60%的最低意识状态患者和33%的无反应性清醒综合征患者的表现具有统计学意义。地形分析发现,健康对照在β和γ波段内的颞、顶叶和额叶区域存在显著差异,在反应性患者中也有部分相同。这些发现强调模糊熵、CSP和Hjorth活动是检测健康人对个性化视频的情绪反应的敏感EEG标记。他们对DOC患者的部分推广强调了评估残余认知加工和意识的潜在临床应用,特别是在反应状态下。
{"title":"Identifying relevant EEG features for personalized emotional videos: a cross-population analysis in healthy controls and patients with disorders of consciousness","authors":"Anny Maza ,&nbsp;Sandra Goizueta ,&nbsp;María Dolores Navarro ,&nbsp;Enrique Noé ,&nbsp;Joan Ferri ,&nbsp;Valery Naranjo ,&nbsp;Roberto Llorens","doi":"10.1016/j.bspc.2026.109669","DOIUrl":"10.1016/j.bspc.2026.109669","url":null,"abstract":"<div><div>Emerging evidence suggests that personalized emotional stimuli can elicit measurable brain responses indicative of cognitive processing. This could be particularly relevant in patients with disorders of consciousness (DOC), where accurate assessment remains remarkably challenging. Despite this, the electroencephalography (EEG) features most sensitive to these personalized emotional stimuli, and their generalizability from healthy individuals to patients with DOC, remain underexplored. This study aimed to identify EEG features distinguishing brain responses to familiar versus non-familiar audiovisual stimuli in healthy controls and to assess their applicability in patients with DOC. Nineteen healthy controls and nineteen patients with DOC viewed personalized emotional videos featuring either familiar or unfamiliar individuals while EEG data were recorded. Seventeen EEG features across various domains were compared using subject-independent machine-learning models in healthy controls, and the top-performing features were validated in patients with DOC. Results indicated fuzzy entropy, common spatial pattern (CSP), and Hjorth activity were the most discriminative features. Applying models trained on healthy individuals to patients with DOC revealed statistically significant performances in 60% of patients in minimally conscious state and 33% of patients with unresponsive wakefulness syndrome. Topographical analyses identified prominent differences in temporal, parietal, and frontal regions within beta and gamma bands for healthy controls, partially replicated in responsive patients. These findings underscore fuzzy entropy, CSP, and Hjorth activity as sensitive EEG markers for detecting emotional responses to personalized videos in healthy controls. Their partial generalization to patients with DOC highlights potential clinical utility in assessing residual cognitive processing and consciousness, particularly in responsive states.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109669"},"PeriodicalIF":4.9,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic cross-modal spatio-temporal graph attention network: Multimodal coupling analysis in sleep stage classification 动态跨模态时空图注意网络:睡眠阶段分类中的多模态耦合分析
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1016/j.bspc.2026.109687
Xiaolin Wang , Xiaowei Li , Jing Li , Yunfa Fu , Dan Zhang , Yan Peng
The lack of dynamic coupling relationships between multimodal signals has limited research into the neural mechanisms of sleep disorders. However, existing studies primarily focus on intra-modal or cross-frequency coupling analysis, lacking methods that can simultaneously quantify nonlinear, time-varying cross-modal coupling and be applied to sleep staging, thereby failing to reveal cross-modal coupling relationships effectively. To address this issue, a novel end-to-end dynamic graph constructor-spatio-temporal graph attention network (DGC-STGAT) is proposed for modeling cross-modal dynamic coupling relationships among electroencephalogram (EEG), electrocardiogram (ECG), and electrooculogram (EOG) signals. The DGC module maps multimodal signals into graph node representations through node feature encoding and adaptively constructs weighted adjacency matrices based on intermodal similarity, thereby generating dynamic graph sequences. The ST-GAT integrates a spatial graph attention mechanism with bidirectional long short-term memory (Bi-LSTM) network to jointly model the spatial dependency structure and temporal evolution of dynamic graph sequences. This extracts spatio-temporal features for analyzing cross-modal coupling relationships and automatically classifying five sleep stages, namely wakefulness (WAKE), non-rapid eye movement (NREM) stages N1, N2, and N3, and rapid eye movement (REM) sleep. Experimental results on a private polysomnography (PSG) dataset involving 50 subjects and the public ISRUC-Sleep dataset demonstrate the effectiveness of the proposed framework. The EEG-EOG modality pair dominates cross-modal coupling, accounting for approximately 80% of the coupling ratio, while modality pairs incorporating ECG contribute significantly less, highlighting the asymmetry in interaction patterns across modalities. In the five-class sleep staging task, DGC-STGAT achieved classification accuracies of 89.1% and 88.6% on the private and ISRUC-Sleep datasets, respectively, marking an improvement of 1.6% over the best-performing baseline model ST-GCN. The overall classification performance of DGC-STGAT outperforms six representative baseline methods DeepSleepNet, AttnSleep, SleepTransformer, ST-GCN, Sleep-CLIP, and MVF-SleepNet. By modeling dynamic cross-modal coupling relationships and applying them to sleep staging, this study not only provides interpretable coupling patterns and achieves high overall classification performance but also offers new insights into the synergistic mechanisms of multimodal physiological signals.
由于缺乏多模态信号之间的动态耦合关系,限制了对睡眠障碍神经机制的研究。然而,现有研究主要集中于模态内耦合或跨频耦合分析,缺乏同时量化非线性时变跨模态耦合并应用于睡眠分期的方法,未能有效揭示跨模态耦合关系。为了解决这一问题,提出了一种新的端到端动态图构造器-时空图注意网络(DGC-STGAT),用于模拟脑电图(EEG)、心电图(ECG)和眼电(EOG)信号之间的跨模态动态耦合关系。DGC模块通过节点特征编码将多模态信号映射为图节点表示,并根据多模态相似度自适应构造加权邻接矩阵,生成动态图序列。ST-GAT将空间图注意机制与双向长短期记忆(Bi-LSTM)网络相结合,共同模拟动态图序列的空间依赖结构和时间演化。通过提取时空特征,分析跨模态耦合关系,自动划分醒觉(WAKE)、非快速眼动(NREM)阶段N1、N2、N3和快速眼动(REM) 5个睡眠阶段。在包含50个受试者的私人多导睡眠图(PSG)数据集和公共ISRUC-Sleep数据集上的实验结果证明了所提出框架的有效性。EEG-EOG模态对主导跨模态耦合,约占耦合比例的80%,而合并ECG的模态对贡献明显较小,突出了跨模态交互模式的不对称性。在五类睡眠分期任务中,DGC-STGAT在private和ISRUC-Sleep数据集上分别实现了89.1%和88.6%的分类准确率,比表现最好的基线模型ST-GCN提高了1.6%。DGC-STGAT的总体分类性能优于6种代表性基线方法DeepSleepNet、AttnSleep、SleepTransformer、ST-GCN、Sleep-CLIP和MVF-SleepNet。通过建立动态跨模态耦合关系模型并将其应用于睡眠分期,本研究不仅提供了可解释的耦合模式,实现了较高的整体分类性能,而且为多模态生理信号的协同机制提供了新的见解。
{"title":"Dynamic cross-modal spatio-temporal graph attention network: Multimodal coupling analysis in sleep stage classification","authors":"Xiaolin Wang ,&nbsp;Xiaowei Li ,&nbsp;Jing Li ,&nbsp;Yunfa Fu ,&nbsp;Dan Zhang ,&nbsp;Yan Peng","doi":"10.1016/j.bspc.2026.109687","DOIUrl":"10.1016/j.bspc.2026.109687","url":null,"abstract":"<div><div>The lack of dynamic coupling relationships between multimodal signals has limited research into the neural mechanisms of sleep disorders. However, existing studies primarily focus on intra-modal or cross-frequency coupling analysis, lacking methods that can simultaneously quantify nonlinear, time-varying cross-modal coupling and be applied to sleep staging, thereby failing to reveal cross-modal coupling relationships effectively. To address this issue, a novel end-to-end dynamic graph constructor-spatio-temporal graph attention network (DGC-STGAT) is proposed for modeling cross-modal dynamic coupling relationships among electroencephalogram (EEG), electrocardiogram (ECG), and electrooculogram (EOG) signals. The DGC module maps multimodal signals into graph node representations through node feature encoding and adaptively constructs weighted adjacency matrices based on intermodal similarity, thereby generating dynamic graph sequences. The ST-GAT integrates a spatial graph attention mechanism with bidirectional long short-term memory (Bi-LSTM) network to jointly model the spatial dependency structure and temporal evolution of dynamic graph sequences. This extracts spatio-temporal features for analyzing cross-modal coupling relationships and automatically classifying five sleep stages, namely wakefulness (WAKE), non-rapid eye movement (NREM) stages N1, N2, and N3, and rapid eye movement (REM) sleep. Experimental results on a private polysomnography (PSG) dataset involving 50 subjects and the public ISRUC-Sleep dataset demonstrate the effectiveness of the proposed framework. The EEG-EOG modality pair dominates cross-modal coupling, accounting for approximately 80% of the coupling ratio, while modality pairs incorporating ECG contribute significantly less, highlighting the asymmetry in interaction patterns across modalities. In the five-class sleep staging task, DGC-STGAT achieved classification accuracies of 89.1% and 88.6% on the private and ISRUC-Sleep datasets, respectively, marking an improvement of 1.6% over the best-performing baseline model ST-GCN. The overall classification performance of DGC-STGAT outperforms six representative baseline methods DeepSleepNet, AttnSleep, SleepTransformer, ST-GCN, Sleep-CLIP, and MVF-SleepNet. By modeling dynamic cross-modal coupling relationships and applying them to sleep staging, this study not only provides interpretable coupling patterns and achieves high overall classification performance but also offers new insights into the synergistic mechanisms of multimodal physiological signals.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109687"},"PeriodicalIF":4.9,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning for lumbar spine segmentation in magnetic resonance imaging—A systematic review 磁共振成像中腰椎分割的深度学习系统综述
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1016/j.bspc.2026.109700
Diogo Mendes , João Manuel R.S. Tavares
Lumbar spine disorders are a major cause of disability worldwide, often requiring magnetic resonance imaging (MRI) for accurate diagnosis and treatment planning. Segmentation of spinal structures in MRI is a critical yet time-consuming task that is traditionally performed manually by experts. In recent years, deep learning (DL) has emerged as a powerful tool to automate this process, offering improvements in efficiency, consistency, and accuracy. This systematic review presents a comprehensive analysis of DL-based methods for lumbar spine segmentation in MRI, conducted by searching the Scopus and PubMed databases for peer-reviewed journal articles and reviews. Studies were included only if they employed MRI as the imaging modality, applied deep learning techniques for segmentation, and explicitly reported both the model architecture and segmentation results. A total of 56 studies met these criteria, comprising 49 original research articles and 7 review articles. The selected works were analyzed in two main dimensions: (1) dataset characteristics, including image orientation, modality, annotation methods, and data availability; and (2) deep learning frameworks, covering preprocessing, data augmentation, network architectures, and evaluation metrics. Convolutional neural networks (CNNs), particularly U-Net and its variants, are the most used architectures, often enhanced with residual blocks, attention mechanisms, and multi-scale feature extraction. Despite promising results, most studies relied on private datasets and lacked external validation, highlighting challenges to reproducibility.
腰椎疾病是世界范围内致残的主要原因,通常需要磁共振成像(MRI)来准确诊断和治疗计划。MRI中脊柱结构的分割是一项关键而耗时的任务,传统上由专家手工完成。近年来,深度学习(DL)已经成为自动化这一过程的强大工具,提高了效率、一致性和准确性。本系统综述通过检索Scopus和PubMed数据库中同行评议的期刊文章和评论,对基于dl的MRI腰椎分割方法进行了全面分析。只有采用MRI作为成像方式,应用深度学习技术进行分割,并明确报告模型架构和分割结果的研究才被纳入。共有56项研究符合这些标准,包括49篇原创研究文章和7篇综述文章。从两个主要维度对入选作品进行分析:(1)数据集特征,包括图像方向、模态、标注方法和数据可用性;(2)深度学习框架,包括预处理、数据增强、网络架构和评估指标。卷积神经网络(cnn),特别是U-Net及其变体,是最常用的体系结构,通常通过残差块、注意机制和多尺度特征提取进行增强。尽管结果令人鼓舞,但大多数研究依赖于私人数据集,缺乏外部验证,突出了可重复性的挑战。
{"title":"Deep learning for lumbar spine segmentation in magnetic resonance imaging—A systematic review","authors":"Diogo Mendes ,&nbsp;João Manuel R.S. Tavares","doi":"10.1016/j.bspc.2026.109700","DOIUrl":"10.1016/j.bspc.2026.109700","url":null,"abstract":"<div><div>Lumbar spine disorders are a major cause of disability worldwide, often requiring magnetic resonance imaging (MRI) for accurate diagnosis and treatment planning. Segmentation of spinal structures in MRI is a critical yet time-consuming task that is traditionally performed manually by experts. In recent years, deep learning (DL) has emerged as a powerful tool to automate this process, offering improvements in efficiency, consistency, and accuracy. This systematic review presents a comprehensive analysis of DL-based methods for lumbar spine segmentation in MRI, conducted by searching the Scopus and PubMed databases for peer-reviewed journal articles and reviews. Studies were included only if they employed MRI as the imaging modality, applied deep learning techniques for segmentation, and explicitly reported both the model architecture and segmentation results. A total of 56 studies met these criteria, comprising 49 original research articles and 7 review articles. The selected works were analyzed in two main dimensions: (1) dataset characteristics, including image orientation, modality, annotation methods, and data availability; and (2) deep learning frameworks, covering preprocessing, data augmentation, network architectures, and evaluation metrics. Convolutional neural networks (CNNs), particularly U-Net and its variants, are the most used architectures, often enhanced with residual blocks, attention mechanisms, and multi-scale feature extraction. Despite promising results, most studies relied on private datasets and lacked external validation, highlighting challenges to reproducibility.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109700"},"PeriodicalIF":4.9,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A deep learning-based multimodal data fusion algorithm for predicting checkpoint inhibitors related pneumonia 基于深度学习的多模态数据融合算法预测检查点抑制剂相关肺炎
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1016/j.bspc.2026.109688
Yuncheng Jin , Jing Zheng , BinSun , Kaidi Fu , Jingjing Qu , Chenyi Ren , Ting Wang , Yuncui Gan , Binggen Wu , Xinyu Jin , Jianya Zhou

Background and Objective

Due to the underutilization of advanced deep learning techniques in the diagnosis and treatment of checkpoint inhibitor-related pneumonia (CIP), this study proposed a CIP prediction algorithm based on multimodal data fusion.

Methods

Specifically, the algorithm constructed a model using patients’ computed tomography (CT) imaging, electronic medical records, and physiological examination reports. First, feature extraction modules were developed to process each data modality. Subsequently, multimodal features were fused using a cross-attention approach. Finally, these features were input into a classifier for classification.

Results

Experimental results demonstrated that multimodal cross-attention network (MMCA-Net) significantly outperformed single-modality models and traditional fusion methods. In10-fold patient-level cross-validation, the proposed model achieved an average accuracy of 87.12% (±0.83%) and an area under the curve (AUC) of 0.8981 (±0.007). Furthermore, the algorithm showed excellent reproducibility, with performance deviation of less than 0.2% in independent replication trials. Quantitative analysis of attention weights confirmed that the model effectively integrated clinical context to resolve ambiguous radiological patterns.

Conclusions

The proposed deep learning-based multimodal method provides a stable and highly accurate tool for predicting CIP. By integrating information from imaging, textual data, and laboratory results, MMCA-Net offers a valuable clinical reference for physicians, with the potential to enhance patient safety and improve treatment outcomes in the management of cancer immunotherapy.
背景与目的针对目前先进的深度学习技术在检查点抑制剂相关性肺炎(CIP)诊断与治疗中的应用不足,本研究提出了一种基于多模态数据融合的CIP预测算法。方法利用患者CT影像、电子病历和生理检查报告构建模型。首先,开发了特征提取模块对各数据模态进行处理。随后,使用交叉注意方法融合多模态特征。最后,将这些特征输入到分类器中进行分类。结果实验结果表明,多模态交叉注意网络(MMCA-Net)显著优于单模态模型和传统融合方法。在10倍患者水平的交叉验证中,该模型的平均准确率为87.12%(±0.83%),曲线下面积(AUC)为0.8981(±0.007)。此外,该算法具有良好的再现性,在独立的重复试验中,性能偏差小于0.2%。定量分析的注意力权重证实,该模型有效地整合临床背景,以解决模棱两可的放射模式。结论基于深度学习的多模态方法为CIP预测提供了一种稳定、高精度的工具。通过整合影像、文本数据和实验室结果的信息,MMCA-Net为医生提供了有价值的临床参考,在癌症免疫治疗管理中具有增强患者安全性和改善治疗结果的潜力。
{"title":"A deep learning-based multimodal data fusion algorithm for predicting checkpoint inhibitors related pneumonia","authors":"Yuncheng Jin ,&nbsp;Jing Zheng ,&nbsp;BinSun ,&nbsp;Kaidi Fu ,&nbsp;Jingjing Qu ,&nbsp;Chenyi Ren ,&nbsp;Ting Wang ,&nbsp;Yuncui Gan ,&nbsp;Binggen Wu ,&nbsp;Xinyu Jin ,&nbsp;Jianya Zhou","doi":"10.1016/j.bspc.2026.109688","DOIUrl":"10.1016/j.bspc.2026.109688","url":null,"abstract":"<div><h3>Background and Objective</h3><div>Due to the underutilization of advanced deep learning techniques in the diagnosis and treatment of checkpoint inhibitor-related pneumonia (CIP), this study proposed a CIP prediction algorithm based on multimodal data fusion.</div></div><div><h3>Methods</h3><div>Specifically, the algorithm constructed a model using patients’ computed tomography (CT) imaging, electronic medical records, and physiological examination reports. First, feature extraction modules were developed to process each data modality. Subsequently, multimodal features were fused using a cross-attention approach. Finally, these features were input into a classifier for classification.</div></div><div><h3>Results</h3><div>Experimental results demonstrated that multimodal cross-attention network (MMCA-Net) significantly outperformed single-modality models and traditional fusion methods. In10-fold patient-level cross-validation, the proposed model achieved an average accuracy of 87.12% (±0.83%) and an area under the curve (AUC) of 0.8981 (±0.007). Furthermore, the algorithm showed excellent reproducibility, with performance deviation of less than 0.2% in independent replication trials. Quantitative analysis of attention weights confirmed that the model effectively integrated clinical context to resolve ambiguous radiological patterns.</div></div><div><h3>Conclusions</h3><div>The proposed deep learning-based multimodal method provides a stable and highly accurate tool for predicting CIP. By integrating information from imaging, textual data, and laboratory results, MMCA-Net offers a valuable clinical reference for physicians, with the potential to enhance patient safety and improve treatment outcomes in the management of cancer immunotherapy.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109688"},"PeriodicalIF":4.9,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated measurement of aortic parameters using deep learning and computer vision 使用深度学习和计算机视觉自动测量主动脉参数
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-29 DOI: 10.1016/j.bspc.2026.109673
Ivan Blekanov , Gleb Kim , Fedor Ezhov , Evgenii Larin , Lev Kovalenko , Anthony Nwohiri , Egor Razumilov
Advancements in artificial intelligence are rapidly transforming healthcare, including the diagnosis of aortic aneurysms, which relies on precise measurement of aortic parameters from CT scans. Current manual methods are time-consuming and require expert surgeons, making automation essential. Accurate automation depends on robust aortic semantic segmentation, cross-section reconstruction, and parameter extraction. Existing 2D segmentation models achieve Dice similarity coefficients (DSC) of 0.842–0.890, while 3D models reach 0.750–0.950. Despite the generally high segmentation accuracy, 3D models require substantial computational resources for both training and inference. This presents a substantial challenge for clinical deployment, especially in developing countries. Our research bridges this gap by advancing state-of-the-art 2D deep learning techniques for aortic semantic segmentation on CT scans. In this regard, we developed a pipeline leveraging novel neural network (NN) architectures and computer vision (CV) techniques. Various high-performing semantic segmentation NNs were rigorously compared. The best NNs (such as VAN-S-UNet, rViT-UNet (TransUNet), MiT-B2-UNet) achieved a DSC of 0.938–0.976 for open datasets, and 0.912 for our dataset of 50 aortic CT scans. The proposed pipeline automates the main stages of CT image processing, from raw CT scan data to quantitative aortic assessment, extracting clinically relevant parameters such as cross-sectional area, border length, and major and minor diameters for subsequent pathology diagnosis and informed clinical decision-making. Case study experiments show minor deviations between the results of the proposed method and expert assessments: approximately 5% for perimeter, 6% for major diameter, 10% for minor diameter, and 15% for cross-sectional area measurement.
人工智能的进步正在迅速改变医疗保健,包括主动脉瘤的诊断,这依赖于从CT扫描中精确测量主动脉参数。目前的手工方法既耗时又需要专业的外科医生,因此自动化是必不可少的。准确的自动化依赖于强大的主动脉语义分割、横截面重建和参数提取。现有二维分割模型的Dice相似系数(DSC)为0.842-0.890,三维模型达到0.750-0.950。尽管分割精度普遍较高,但3D模型的训练和推理都需要大量的计算资源。这对临床部署提出了重大挑战,特别是在发展中国家。我们的研究通过推进最先进的2D深度学习技术,在CT扫描上进行主动脉语义分割,弥合了这一差距。在这方面,我们开发了一个利用新型神经网络(NN)架构和计算机视觉(CV)技术的管道。对各种高性能的语义分割神经网络进行了严格的比较。最好的神经网络(如VAN-S-UNet, rvitt - unet (TransUNet), MiT-B2-UNet)在开放数据集上的DSC为0.938-0.976,在我们的50个主动脉CT扫描数据集上的DSC为0.912。提出的管道自动化了CT图像处理的主要阶段,从原始CT扫描数据到定量主动脉评估,提取临床相关参数,如横断面积、边界长度、主要和次要直径,用于后续病理诊断和知情的临床决策。案例研究实验表明,所提出方法的结果与专家评估之间的偏差很小:周长约为5%,大直径约为6%,小直径约为10%,横截面积测量约为15%。
{"title":"Automated measurement of aortic parameters using deep learning and computer vision","authors":"Ivan Blekanov ,&nbsp;Gleb Kim ,&nbsp;Fedor Ezhov ,&nbsp;Evgenii Larin ,&nbsp;Lev Kovalenko ,&nbsp;Anthony Nwohiri ,&nbsp;Egor Razumilov","doi":"10.1016/j.bspc.2026.109673","DOIUrl":"10.1016/j.bspc.2026.109673","url":null,"abstract":"<div><div>Advancements in artificial intelligence are rapidly transforming healthcare, including the diagnosis of aortic aneurysms, which relies on precise measurement of aortic parameters from CT scans. Current manual methods are time-consuming and require expert surgeons, making automation essential. Accurate automation depends on robust aortic semantic segmentation, cross-section reconstruction, and parameter extraction. Existing 2D segmentation models achieve Dice similarity coefficients (DSC) of 0.842–0.890, while 3D models reach 0.750–0.950. Despite the generally high segmentation accuracy, 3D models require substantial computational resources for both training and inference. This presents a substantial challenge for clinical deployment, especially in developing countries. Our research bridges this gap by advancing state-of-the-art 2D deep learning techniques for aortic semantic segmentation on CT scans. In this regard, we developed a pipeline leveraging novel neural network (NN) architectures and computer vision (CV) techniques. Various high-performing semantic segmentation NNs were rigorously compared. The best NNs (such as VAN-S-UNet, rViT-UNet (TransUNet), MiT-B2-UNet) achieved a DSC of 0.938–0.976 for open datasets, and 0.912 for our dataset of 50 aortic CT scans. The proposed pipeline automates the main stages of CT image processing, from raw CT scan data to quantitative aortic assessment, extracting clinically relevant parameters such as cross-sectional area, border length, and major and minor diameters for subsequent pathology diagnosis and informed clinical decision-making. Case study experiments show minor deviations between the results of the proposed method and expert assessments: approximately 5% for perimeter, 6% for major diameter, 10% for minor diameter, and 15% for cross-sectional area measurement.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109673"},"PeriodicalIF":4.9,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EEG-based depression detection using a local–global feature fusion deep learning network 基于脑电图的局部-全局特征融合深度学习网络抑郁检测
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-29 DOI: 10.1016/j.bspc.2026.109681
Xugang Li , Guanghao Huang , Yinhua Liu , Keum-Shik Hong
Electroencephalography (EEG) provides a non-invasive, cost-effective, and objective means for detecting depression, yet existing methods often fail to integrate the full spectrum of EEG information, including time–frequency, spatial, and connectivity features. This paper proposes a multidimensional deep learning framework, termed the Local-Global Feature Fusion Network (LGFF-Net), which unifies local brain-region dynamics with whole-brain functional connectivity (FC) patterns. In LGFF-Net, local time–frequency features are extracted from eight functional brain regions using a multi-branch residual neural network enhanced with multi-head attention, highlighting the most discriminative regions. Simultaneously, global connectivity features are captured by phase-locking-value-based FC matrices across five frequency bands, combined with a learnable band-weighting mechanism and a convolutional neural network. A global-to-local fusion mechanism further allows whole-brain connectivity to adaptively modulate the contribution of each brain region before classification, thereby strengthening the complementarity between local and global features. On the MODMA dataset, LGFF-Net achieves an accuracy of 97.64 ± 0.21%, outperforming state-of-the-art approaches, and it also maintains strong performance on an additional independent EEG dataset collected in Malaysia, indicating good cross-dataset generalization. Interpretability analyses highlight the frontal and temporal lobes, together with theta and alpha band connectivity, as key markers differentiating patients with depression from healthy controls, confirming both the robustness and interpretability of the proposed framework.
脑电图(EEG)为检测抑郁症提供了一种无创、经济、客观的手段,但现有方法往往无法整合脑电图的全谱信息,包括时间频率、空间和连通性特征。本文提出了一种多维深度学习框架,称为局部-全局特征融合网络(LGFF-Net),该框架将局部脑区域动态与全脑功能连接(FC)模式相结合。在LGFF-Net中,利用头部注意增强的多分支残差神经网络从8个脑功能区域提取局部时频特征,突出最具判别性的区域。同时,结合可学习的频带加权机制和卷积神经网络,通过基于锁相值的FC矩阵在五个频带上捕获全局连接特征。全局到局部融合机制进一步允许全脑连接在分类前自适应调节每个脑区域的贡献,从而加强局部特征和全局特征之间的互补性。在MODMA数据集上,LGFF-Net达到了97.64±0.21%的准确率,优于最先进的方法,并且在马来西亚收集的另一个独立脑电图数据集上也保持了良好的表现,表明了良好的跨数据集泛化。可解释性分析强调了额叶和颞叶,以及θ和α波段的连通性,作为区分抑郁症患者和健康对照的关键标志,证实了所提出框架的稳健性和可解释性。
{"title":"EEG-based depression detection using a local–global feature fusion deep learning network","authors":"Xugang Li ,&nbsp;Guanghao Huang ,&nbsp;Yinhua Liu ,&nbsp;Keum-Shik Hong","doi":"10.1016/j.bspc.2026.109681","DOIUrl":"10.1016/j.bspc.2026.109681","url":null,"abstract":"<div><div>Electroencephalography (EEG) provides a non-invasive, cost-effective, and objective means for detecting depression, yet existing methods often fail to integrate the full spectrum of EEG information, including time–frequency, spatial, and connectivity features. This paper proposes a multidimensional deep learning framework, termed the Local-Global Feature Fusion Network (LGFF-Net), which unifies local brain-region dynamics with whole-brain functional connectivity (FC) patterns. In LGFF-Net, local time–frequency features are extracted from eight functional brain regions using a multi-branch residual neural network enhanced with multi-head attention, highlighting the most discriminative regions. Simultaneously, global connectivity features are captured by phase-locking-value-based FC matrices across five frequency bands, combined with a learnable band-weighting mechanism and a convolutional neural network. A global-to-local fusion mechanism further allows whole-brain connectivity to adaptively modulate the contribution of each brain region before classification, thereby strengthening the complementarity between local and global features. On the MODMA dataset, LGFF-Net achieves an accuracy of 97.64 ± 0.21%, outperforming state-of-the-art approaches, and it also maintains strong performance on an additional independent EEG dataset collected in Malaysia, indicating good cross-dataset generalization. Interpretability analyses highlight the frontal and temporal lobes, together with theta and alpha band connectivity, as key markers differentiating patients with depression from healthy controls, confirming both the robustness and interpretability of the proposed framework.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109681"},"PeriodicalIF":4.9,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An empirical study for breast cancer detection using MRI images MRI影像检测乳腺癌的实证研究
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-29 DOI: 10.1016/j.bspc.2026.109640
D.E. Martina Jaincy, Pattabiraman Venkatasubbu
Tumor is a dreadful disease faced by human beings and can lead to death. Thus, ultimate methods must be applied to these diseases and save human beings from them. One of the dangerous kinds of tumor faced by women is breast cancer (BC). Earlier diagnosis increases the survival rate, and it should save more lives by protecting women from these dangerous diseases. Different types of images are there for detecting BC. Magnetic resonance imaging (MRI) is the most essential imaging method in predicting BC. Various survey papers are reviewed in this survey for detecting BC using MRI images. In this review, 50 research papers are analyzed regarding breast cancer (BC) detection using MRI images. Moreover, 50 research papers are reviewed in this study regarding breast cancer (BC) detection using MRI images. The technique-related overviews are categorized as follows: machine learning (ML)-enabled approaches, including deep learning (DL) techniques as a specialized subset; clustering-based approaches; segmentation-based methods; and hybrid techniques. MRI is the most accomplished imaging for detecting BC compared to other images like mammography, ultrasound and so on. This survey is comprised of categorization research approaches, toolset, year of publication, datasets and performance metrics to detect BC. Finally, the limitations of the investigated approaches are explained, which motivates these researchers to develop the latest effectual approaches for detecting BC by wielding MRI images.
肿瘤是人类面临的一种可怕的疾病,可以导致死亡。因此,必须对这些疾病采取终极方法,使人类免于疾病。女性面临的危险肿瘤之一是乳腺癌(BC)。早期诊断可以提高生存率,并通过保护妇女免受这些危险疾病的侵害来挽救更多的生命。有不同类型的图像用于检测BC。磁共振成像(MRI)是预测BC最重要的成像方法。在本调查中回顾了使用MRI图像检测BC的各种调查论文。本文回顾了50篇关于乳腺癌(BC) MRI检测的研究论文。此外,本研究还回顾了50篇关于使用MRI图像检测乳腺癌的研究论文。与技术相关的概述分类如下:支持机器学习(ML)的方法,包括作为专门子集的深度学习(DL)技术;clustering-based方法;有效方法;还有混合技术。与乳房x线摄影、超声等影像相比,MRI是检测BC最成熟的影像学手段。这项调查包括分类研究方法、工具集、出版年份、数据集和检测BC的性能指标。最后,所调查的方法的局限性进行了解释,这促使这些研究人员开发最新的有效的方法来检测BC挥舞MRI图像。
{"title":"An empirical study for breast cancer detection using MRI images","authors":"D.E. Martina Jaincy,&nbsp;Pattabiraman Venkatasubbu","doi":"10.1016/j.bspc.2026.109640","DOIUrl":"10.1016/j.bspc.2026.109640","url":null,"abstract":"<div><div>Tumor is a dreadful disease faced by human beings and can lead to death. Thus, ultimate methods must be applied to these diseases and save human beings from them. One of the dangerous kinds of tumor faced by women is breast cancer (BC). Earlier diagnosis increases the survival rate, and it should save more lives by protecting women from these dangerous diseases. Different types of images are there for detecting BC. Magnetic resonance imaging (MRI) is the most essential imaging method in predicting BC. Various survey papers are reviewed in this survey for detecting BC using MRI images. In this review, 50 research papers are analyzed regarding breast cancer (BC) detection using MRI images. Moreover, 50 research papers are reviewed in this study regarding breast cancer (BC) detection using MRI images. The technique-related overviews are categorized as follows: machine learning (ML)-enabled approaches, including deep learning (DL) techniques as a specialized subset; clustering-based approaches; segmentation-based methods; and hybrid techniques. MRI is the most accomplished imaging for detecting BC compared to other images like mammography, ultrasound and so on. This survey is comprised of categorization research approaches, toolset, year of publication, datasets and performance metrics to detect BC. Finally, the limitations of the investigated approaches are explained, which motivates these researchers to develop the latest effectual approaches for detecting BC by wielding MRI images.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109640"},"PeriodicalIF":4.9,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Biomedical Signal Processing and Control
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1