首页 > 最新文献

Biomedical Signal Processing and Control最新文献

英文 中文
Component-wise score diffusion model with momentum-accelerated updates for low-dose CT reconstruction 基于动量加速更新的低剂量CT重建成分评分扩散模型
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-11 DOI: 10.1016/j.bspc.2026.109765
Dalin Wang , Xuemei Wu , Chao He
Low-dose CT reconstruction remains challenging because dose reduction amplifies noise and streak artifacts, while strong priors risk removing subtle anatomical details. Score-based diffusion models provide a flexible way to model CT image distributions, yet pixel-domain diffusion couples low-frequency structure and high-frequency texture within a single score function and the sampling process can be slow when data consistency is enforced with generic updates. We present a component-wise score diffusion model that performs diffusion on wavelet subbands and interleaves reverse sampling with a momentum-accelerated OS-SART projection step. This design decouples structural and textural priors in the wavelet domain and enforces projection fidelity throughout sampling. Experiments on the AAPM-Mayo dataset show consistent improvements over competitive baselines in both low-dose full-view and sparse-view settings, achieving 41.03 dB PSNR and 0.965 SSIM at 10 percent dose and 39.26 dB PSNR at 96 views while reducing inference time relative to other score-based methods.
低剂量CT重建仍然具有挑战性,因为剂量降低会放大噪声和条纹伪影,而强先验可能会去除细微的解剖细节。基于分数的扩散模型提供了一种灵活的方法来模拟CT图像分布,然而像素域扩散在单个分数函数中耦合了低频结构和高频纹理,并且当使用通用更新强制数据一致性时,采样过程可能很慢。我们提出了一种基于分量的分数扩散模型,该模型在小波子带上进行扩散,并通过动量加速的OS-SART投影步骤交织反向采样。该设计解耦了小波域的结构和纹理先验,并在整个采样过程中增强了投影保真度。在AAPM-Mayo数据集上的实验表明,在低剂量全视图和稀疏视图设置下,与竞争基线相比,该方法在低剂量全视图和稀疏视图设置下均有一致的改进,在10%剂量下达到41.03 dB PSNR和0.965 SSIM,在96视图下达到39.26 dB PSNR,同时相对于其他基于分数的方法减少了推理时间。
{"title":"Component-wise score diffusion model with momentum-accelerated updates for low-dose CT reconstruction","authors":"Dalin Wang ,&nbsp;Xuemei Wu ,&nbsp;Chao He","doi":"10.1016/j.bspc.2026.109765","DOIUrl":"10.1016/j.bspc.2026.109765","url":null,"abstract":"<div><div>Low-dose CT reconstruction remains challenging because dose reduction amplifies noise and streak artifacts, while strong priors risk removing subtle anatomical details. Score-based diffusion models provide a flexible way to model CT image distributions, yet pixel-domain diffusion couples low-frequency structure and high-frequency texture within a single score function and the sampling process can be slow when data consistency is enforced with generic updates. We present a component-wise score diffusion model that performs diffusion on wavelet subbands and interleaves reverse sampling with a momentum-accelerated OS-SART projection step. This design decouples structural and textural priors in the wavelet domain and enforces projection fidelity throughout sampling. Experiments on the AAPM-Mayo dataset show consistent improvements over competitive baselines in both low-dose full-view and sparse-view settings, achieving 41.03 dB PSNR and 0.965 SSIM at 10 percent dose and 39.26 dB PSNR at 96 views while reducing inference time relative to other score-based methods.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109765"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sympathetic nerve activity recovery from the skin recording using the modern optimal shrinkage technique 现代最佳收缩技术记录皮肤交感神经活动恢复
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-10 DOI: 10.1016/j.bspc.2026.109710
Pei-Chun Su , Chao-Yi Chen , Chia-Hao Kuo , Wei-Chung Tsai , Hau-Tieng Wu

Objective

The widely used bandpass filter (BPF)-based algorithm for recovering sympathetic nerve activity (SNA) from the skin sympathetic nerve activity (SKNA −I) signal, recorded via electrocardiogram electrodes or subcutaneous sympathetic nerve activity (SCNA-I) in a lead I setup, has limitations. It excludes spectral information outside the BPF range and may retain artifacts, such as cardiac activity or pacemaker interference, in the recovered SNA (rSNA) signal. This study aims to develop an algorithm that recovers the full spectral SNA information as comprehensively as possible for evaluating the autonomic nervous system (ANS).

Methods

We propose a novel algorithm, S3 (SNA from Shrink and Subtraction), which integrates the optimal shrinkage algorithm (eOptShrink) with the template subtraction (TS) method, and make the Matlab code publicly available. The performance of S3 was evaluated against other algorithms using semi-real simulated SKNA-I data, a human SKNA-I database including subjects with pacemakers or atrial fibrillation (Af), and a mice SCNA-I database.

Results

The S3 algorithm demonstrated numerical efficiency and outperformed existing approaches, including traditional TS, BPF and other methods, in both time and frequency domains. Notably, in addition to the traditional 500–1000 Hz spectral band, S3 effectively recovers spectral information across the 50–300 Hz and 300–500 Hz frequency bands. All quantitative results are supported by the rSNA tracing for visual inspections.

Conclusion

S3 overcomes key limitations of existing methods and accurately recovers full-spectrum SNA from human SKNA-I, including cases with pacemaker and AF, as well as from mouse SCNA-I, with both theoretical justification and numerical validation. Since S3 can recover spectral information across the 50–300 Hz and 300–500 Hz frequency bands, and ECG signals in the homecare environments are typically sampled at 1–2 kHz, S3 is potentially suitable for home-based ANS evaluation.

Significance

S3 enables exploration of the entire SNA spectrum and shows strong potential for ANS evaluation in homecare settings.
目的广泛使用的基于带通滤波器(BPF)的算法从皮肤交感神经活动(SKNA -I)信号中恢复交感神经活动(SNA),这些信号是通过心电图电极或在导联I装置中记录的皮下交感神经活动(SCNA-I),但存在局限性。它排除了BPF范围外的频谱信息,并可能保留伪象,如心脏活动或起搏器干扰,在恢复的SNA (rSNA)信号中。本研究旨在开发一种尽可能全面地恢复全谱SNA信息的算法,用于评估自主神经系统(ANS)。方法我们提出了一种新的算法S3 (SNA from Shrink and Subtraction),它将最优收缩算法(eOptShrink)与模板减法(TS)方法相结合,并公开了Matlab代码。通过半真实的模拟SKNA-I数据、包括心脏起搏器或心房颤动(Af)受试者的人类SKNA-I数据库和小鼠SCNA-I数据库,对S3的性能与其他算法进行了评估。结果S3算法在时域和频域均优于传统TS、BPF等方法。值得注意的是,除了传统的500-1000 Hz频段外,S3还可以有效地恢复50-300 Hz和300-500 Hz频段的频谱信息。所有定量结果均由目视检查的rSNA追踪支持。结论s3克服了现有方法的主要局限性,准确地恢复了包括起搏器和心房颤动病例在内的人SCNA-I以及小鼠SCNA-I的全谱SNA,具有理论依据和数值验证。由于S3可以恢复50-300 Hz和300-500 Hz频段的频谱信息,并且家庭护理环境中的心电信号通常以1-2 kHz采样,因此S3可能适用于基于家庭的ANS评估。意义3能够探索整个SNA谱,并显示出在家庭护理环境中进行ANS评估的强大潜力。
{"title":"Sympathetic nerve activity recovery from the skin recording using the modern optimal shrinkage technique","authors":"Pei-Chun Su ,&nbsp;Chao-Yi Chen ,&nbsp;Chia-Hao Kuo ,&nbsp;Wei-Chung Tsai ,&nbsp;Hau-Tieng Wu","doi":"10.1016/j.bspc.2026.109710","DOIUrl":"10.1016/j.bspc.2026.109710","url":null,"abstract":"<div><h3>Objective</h3><div>The widely used bandpass filter (BPF)-based algorithm for recovering sympathetic nerve activity (SNA) from the skin sympathetic nerve activity (SKNA −I) signal, recorded via electrocardiogram electrodes or subcutaneous sympathetic nerve activity (SCNA-I) in a lead I setup, has limitations. It excludes spectral information outside the BPF range and may retain artifacts, such as cardiac activity or pacemaker interference, in the recovered SNA (rSNA) signal. This study aims to develop an algorithm that recovers the full spectral SNA information as comprehensively as possible for evaluating the autonomic nervous system (ANS).</div></div><div><h3>Methods</h3><div>We propose a novel algorithm, S3 (<em>SNA from Shrink and Subtraction</em>), which integrates the optimal shrinkage algorithm (eOptShrink) with the template subtraction (TS) method, and make the Matlab code publicly available. The performance of S3 was evaluated against other algorithms using semi-real simulated SKNA-I data, a human SKNA-I database including subjects with pacemakers or atrial fibrillation (Af), and a mice SCNA-I database.</div></div><div><h3>Results</h3><div>The S3 algorithm demonstrated numerical efficiency and outperformed existing approaches, including traditional TS, BPF and other methods, in both time and frequency domains. Notably, in addition to the traditional 500–1000 Hz spectral band, S3 effectively recovers spectral information across the 50–300 Hz and 300–500 Hz frequency bands. All quantitative results are supported by the rSNA tracing for visual inspections.</div></div><div><h3>Conclusion</h3><div>S3 overcomes key limitations of existing methods and accurately recovers full-spectrum SNA from human SKNA-I, including cases with pacemaker and AF, as well as from mouse SCNA-I, with both theoretical justification and numerical validation. Since S3 can recover spectral information across the 50–300 Hz and 300–500 Hz frequency bands, and ECG signals in the homecare environments are typically sampled at 1–2 kHz, S3 is potentially suitable for home-based ANS evaluation.</div></div><div><h3>Significance</h3><div>S3 enables exploration of the entire SNA spectrum and shows strong potential for ANS evaluation in homecare settings.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109710"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Accuracy enhancement in melanoma diagnosis: A comparative study of residual networks and visual geometry group architectures 提高黑色素瘤诊断的准确性:残差网络和视觉几何群结构的比较研究
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-09 DOI: 10.1016/j.bspc.2026.109635
Shisheng Chen , Wenjing Pan , Tongyao Chen , Xinxin Xie , Yi Zhang
Melanoma is a highly aggressive skin malignancy, and its early detection is critical for reducing mortality. With the rapid progress of deep learning in medical imaging, convolutional neural networks (CNNs) have become powerful tools for automated pathological image analysis. This study aimed to systematically evaluate the performance, interpretability, and potential clinical utility of different deep learning models in classifying melanoma on H&E-stained pathology images. A total of 312 clinical H&E whole-slide images (210 normal skin and 102 melanoma) were acquired and preprocessed through resizing, normalization, and data augmentation. Four CNN architectures—ResNet, VGG, MobileNetV2, and DenseNet121—were constructed for classification, and five-fold cross-validation was used for performance evaluation based on accuracy, sensitivity, specificity, F1-score, and AUC. Grad-CAM was further applied for model interpretability, with pathological verification by experienced dermatopathologists. All four models successfully differentiated melanoma from normal skin tissue, with ResNet achieving the highest mean accuracy (96.10%) and the best F1-score and AUC. VGG exhibited strong stability, while MobileNetV2 and DenseNet121 provided higher computational efficiency but slightly lower diagnostic performance. Statistical analysis confirmed that ResNet outperformed the other models significantly (p < 0.05). Grad-CAM visualization demonstrated that the highlighted regions corresponded closely to key histopathological features of melanoma, indicating that the model’s decision-making process is pathologically plausible.
黑色素瘤是一种高度侵袭性的皮肤恶性肿瘤,早期发现对降低死亡率至关重要。随着深度学习在医学成像领域的快速发展,卷积神经网络(cnn)已经成为病理图像自动分析的有力工具。本研究旨在系统评估不同深度学习模型在H&; e染色病理图像上对黑色素瘤进行分类的性能、可解释性和潜在的临床应用。共获得312张临床H&;E全片图像(210张正常皮肤和102张黑色素瘤),并通过调整大小、归一化和数据增强进行预处理。构建了4种CNN架构(resnet、VGG、MobileNetV2和densenet121)进行分类,并基于准确性、灵敏度、特异性、f1评分和AUC进行了5次交叉验证。Grad-CAM进一步应用于模型的可解释性,由经验丰富的皮肤病理学家进行病理验证。四种模型均成功将黑色素瘤与正常皮肤组织区分开来,其中ResNet的平均准确率最高(96.10%),f1评分和AUC最高。VGG表现出很强的稳定性,而MobileNetV2和DenseNet121提供了更高的计算效率,但诊断性能略低。统计分析证实,ResNet的表现明显优于其他模型(p < 0.05)。Grad-CAM可视化显示,突出显示的区域与黑色素瘤的关键组织病理学特征密切相关,表明该模型的决策过程在病理学上是合理的。
{"title":"Accuracy enhancement in melanoma diagnosis: A comparative study of residual networks and visual geometry group architectures","authors":"Shisheng Chen ,&nbsp;Wenjing Pan ,&nbsp;Tongyao Chen ,&nbsp;Xinxin Xie ,&nbsp;Yi Zhang","doi":"10.1016/j.bspc.2026.109635","DOIUrl":"10.1016/j.bspc.2026.109635","url":null,"abstract":"<div><div>Melanoma is a highly aggressive skin malignancy, and its early detection is critical for reducing mortality. With the rapid progress of deep learning in medical imaging, convolutional neural networks (CNNs) have become powerful tools for automated pathological image analysis. This study aimed to systematically evaluate the performance, interpretability, and potential clinical utility of different deep learning models in classifying melanoma on H&amp;E-stained pathology images. A total of 312 clinical H&amp;E whole-slide images (210 normal skin and 102 melanoma) were acquired and preprocessed through resizing, normalization, and data augmentation. Four CNN architectures—ResNet, VGG, MobileNetV2, and DenseNet121—were constructed for classification, and five-fold cross-validation was used for performance evaluation based on accuracy, sensitivity, specificity, F1-score, and AUC. Grad-CAM was further applied for model interpretability, with pathological verification by experienced dermatopathologists. All four models successfully differentiated melanoma from normal skin tissue, with ResNet achieving the highest mean accuracy (96.10%) and the best F1-score and AUC. VGG exhibited strong stability, while MobileNetV2 and DenseNet121 provided higher computational efficiency but slightly lower diagnostic performance. Statistical analysis confirmed that ResNet outperformed the other models significantly (<em>p</em> &lt; 0.05). Grad-CAM visualization demonstrated that the highlighted regions corresponded closely to key histopathological features of melanoma, indicating that the model’s decision-making process is pathologically plausible.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109635"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-assisted ensemble CNN–MobileNetV2–transformer architecture for automated TB diagnosis 用于自动结核病诊断的注意力辅助集成CNN-MobileNetV2-transformer架构
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-11 DOI: 10.1016/j.bspc.2026.109778
Beaulah Jeyavathana R , Kalaivani Chellappan , M. Sai Ganeshan
Tuberculosis (TB) is considered an airborne disease, causing a high death rate globally by affecting the lungs. The early detection of TB remains a challenge owing to the lack of screening facilities. The availability of public datasets, the advancement of artificial intelligence (AI), computerized systems have enabled the automatic diagnosis of tuberculosis using chest X-rays. The existing AI algorithms utilize complicated architectures; therefore, the procedure is time-consuming and costly. To overcome these limitations, Research proposes a novel lightweight ensemble deep learning (DL) model for detecting and localizing TB from Chest X-Ray (CXR) images. The proposed research gathered the CXR images from the Kaggle repository. The images are standardized before being input to the DL models to ensure enhanced learning and stability. Further, make the image dimensions suited for DL models, the images are resized. The lung regions are segmented from the pre-processed images using a spatial attention based residual U-net model (SA-Res-UNet) to ensure accurate detection of TB. Finally, the CXR images are classified as normal and TB using the ensemble model. The proposed model has an ensemble of custom convolutional neural network (CNN), MobileNetV2, and Swin Transformer (ST) models. Individual model predictions are combined based on majority voting. Finally, the classified images are explained and interpreted by providing visualization through self-attention-based class activation mapping (SA-CAM). The experiments are conducted in the Python programming language. The proposed model attained 99% accuracy in detecting TB disease. The proposed model’s exceptional results demonstrate its efficacy in TB detection, allowing for practical application.
结核病被认为是一种空气传播疾病,通过影响肺部在全球造成高死亡率。由于缺乏筛查设施,结核病的早期发现仍然是一项挑战。公共数据集的可用性、人工智能(AI)的进步和计算机化系统使利用胸部x光自动诊断结核病成为可能。现有的人工智能算法使用复杂的架构;因此,这个过程既耗时又昂贵。为了克服这些限制,研究人员提出了一种新的轻量级集成深度学习(DL)模型,用于从胸部x射线(CXR)图像中检测和定位结核病。拟议的研究从Kaggle存储库中收集了CXR图像。图像在输入到DL模型之前经过标准化处理,以确保增强的学习性和稳定性。此外,使图像尺寸适合深度学习模型,图像被调整大小。使用基于空间注意的残差U-net模型(SA-Res-UNet)从预处理图像中分割肺区域,以确保准确检测结核病。最后,使用集成模型将CXR图像分类为normal和TB。该模型集成了自定义卷积神经网络(CNN)、MobileNetV2和Swin Transformer (ST)模型。单个模型的预测是基于多数投票组合的。最后,通过基于自我注意的类激活映射(SA-CAM)提供可视化来解释和解释分类图像。实验是用Python编程语言进行的。该模型检测结核病的准确率达到99%。该模型的优异结果证明了其在结核病检测中的有效性,为实际应用提供了可能。
{"title":"Attention-assisted ensemble CNN–MobileNetV2–transformer architecture for automated TB diagnosis","authors":"Beaulah Jeyavathana R ,&nbsp;Kalaivani Chellappan ,&nbsp;M. Sai Ganeshan","doi":"10.1016/j.bspc.2026.109778","DOIUrl":"10.1016/j.bspc.2026.109778","url":null,"abstract":"<div><div>Tuberculosis (TB) is considered an airborne disease, causing a high death rate globally by affecting the lungs. The early detection of TB remains a challenge owing to the lack of screening facilities. The availability of public datasets, the advancement of artificial intelligence (AI), computerized systems have enabled the automatic diagnosis of tuberculosis using chest X-rays. The existing AI algorithms utilize complicated architectures; therefore, the procedure is time-consuming and costly. To overcome these limitations, Research proposes a novel lightweight ensemble deep learning (DL) model for detecting and localizing TB from Chest X-Ray (CXR) images. The proposed research gathered the CXR images from the Kaggle repository. The images are standardized before being input to the DL models to ensure enhanced learning and stability. Further, make the image dimensions suited for DL models, the images are resized. The lung regions are segmented from the pre-processed images using a spatial attention based residual U-net model (SA-Res-UNet) to ensure accurate detection of TB. Finally, the CXR images are classified as normal and TB using the ensemble model. The proposed model has an ensemble of custom convolutional neural network (CNN), MobileNetV2, and Swin Transformer (ST) models. Individual model predictions are combined based on majority voting. Finally, the classified images are explained and interpreted by providing visualization through self-attention-based class activation mapping (SA-CAM). The experiments are conducted in the Python programming language. The proposed model attained 99% accuracy in detecting TB disease. The proposed model’s exceptional results demonstrate its efficacy in TB detection, allowing for practical application.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109778"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic histopathology with controllable class distribution: A dual-GAN framework for melanoma segmentation 具有可控类分布的合成组织病理学:黑色素瘤分割的双gan框架
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-09 DOI: 10.1016/j.bspc.2026.109741
Ziad Elshaer , Ahmed Jamal , Essam A. Rashed
Tumor-infiltrating lymphocytes (TILs) assessment in melanoma histopathology is critical for predicting immunotherapy response and improving patient outcomes, yet current automated segmentation methods are severely constrained by limited datasets and pronounced class imbalance. We present a novel dual-generator adversarial framework that revolutionizes synthetic histopathology data generation by decomposing the complex synthesis problem into two specialized sequential tasks: controllable mask generation with user-specified class distributions, followed by high-fidelity histopathology image synthesis. This innovative approach enables precise dataset augmentation with any desired number of tissue classes per image, fundamentally addressing the scarcity of balanced training data. Leveraging the PUMA Grand Challenge dataset, we systematically generated two complementary datasets and evaluated them using a custom U-Net architecture that integrates a powerful MedSAM encoder with a specialized decoder optimized for fine-grained tissue segmentation. Our dual-GAN framework demonstrates exceptional capability in generating photorealistic histopathology images while maintaining precise control over tissue class distributions and spatial relationships. The proposed architecture achieved outstanding performance with an F1 score of 0.91 on the PUMA dataset and new data from the three-class per-image dataset, significantly advancing the state-of-the-art in melanoma tissue segmentation. This scalable framework establishes a new paradigm for computational pathology, enabling robust TIL assessment and enhanced clinical decision-making in melanoma management.
黑色素瘤组织病理学中肿瘤浸润淋巴细胞(til)的评估对于预测免疫治疗反应和改善患者预后至关重要,但目前的自动分割方法受到有限的数据集和明显的分类不平衡的严重限制。我们提出了一种新的双生成器对抗框架,通过将复杂的合成问题分解为两个专门的顺序任务,彻底改变了合成组织病理学数据的生成:具有用户指定类分布的可控掩膜生成,然后是高保真的组织病理学图像合成。这种创新的方法可以通过每张图像的任意数量的组织类来精确地增强数据集,从根本上解决了平衡训练数据的稀缺性。利用PUMA Grand Challenge数据集,我们系统地生成了两个互补的数据集,并使用定制的U-Net架构对它们进行了评估,该架构集成了强大的MedSAM编码器和针对细粒度组织分割优化的专用解码器。我们的双gan框架在生成逼真的组织病理学图像同时保持对组织类分布和空间关系的精确控制方面表现出卓越的能力。该架构在PUMA数据集和来自每幅图像三级数据集的新数据上取得了0.91的F1分数,显著推进了黑色素瘤组织分割的最新技术。这个可扩展的框架为计算病理学建立了一个新的范例,使TIL评估和增强黑色素瘤管理的临床决策成为可能。
{"title":"Synthetic histopathology with controllable class distribution: A dual-GAN framework for melanoma segmentation","authors":"Ziad Elshaer ,&nbsp;Ahmed Jamal ,&nbsp;Essam A. Rashed","doi":"10.1016/j.bspc.2026.109741","DOIUrl":"10.1016/j.bspc.2026.109741","url":null,"abstract":"<div><div>Tumor-infiltrating lymphocytes (TILs) assessment in melanoma histopathology is critical for predicting immunotherapy response and improving patient outcomes, yet current automated segmentation methods are severely constrained by limited datasets and pronounced class imbalance. We present a novel dual-generator adversarial framework that revolutionizes synthetic histopathology data generation by decomposing the complex synthesis problem into two specialized sequential tasks: controllable mask generation with user-specified class distributions, followed by high-fidelity histopathology image synthesis. This innovative approach enables precise dataset augmentation with any desired number of tissue classes per image, fundamentally addressing the scarcity of balanced training data. Leveraging the PUMA Grand Challenge dataset, we systematically generated two complementary datasets and evaluated them using a custom U-Net architecture that integrates a powerful MedSAM encoder with a specialized decoder optimized for fine-grained tissue segmentation. Our dual-GAN framework demonstrates exceptional capability in generating photorealistic histopathology images while maintaining precise control over tissue class distributions and spatial relationships. The proposed architecture achieved outstanding performance with an F1 score of 0.91 on the PUMA dataset and new data from the three-class per-image dataset, significantly advancing the state-of-the-art in melanoma tissue segmentation. This scalable framework establishes a new paradigm for computational pathology, enabling robust TIL assessment and enhanced clinical decision-making in melanoma management.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109741"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146191954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-knowledge feature fusion for PPG-based blood pressure prediction: Low-dimensional extraction via functional data analysis and knowledge constraint 基于ppg的血压预测数据-知识特征融合:基于功能数据分析和知识约束的低维提取
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-07 DOI: 10.1016/j.bspc.2026.109754
Qingfeng Tang , Huihui Hu , Chao Tao , Pengcheng Ding , Guowei Dai , Guangjun Wang , Xiaojuan Hu , Benyue Su , Jiatuo Xu , Hui An
Although concatenating knowledge features (KF) and data features (DF) of photoplethysmography (PPG) can improve the predictive performance of blood pressure monitoring models, this approach inevitably increases the dimensionality of feature space. To address this limitation, we propose an innovative feature extraction method that deeply integrate KF and DF, rather than simply concatenating them.
Our method employs functional data analysis to extract DF by treating PPG as continuously functional curve. Subsequently, the distribution patterns of KF are thoroughly analyzed to construct a KF-based constrained space, which serves as a guide during DF extraction, to yield novel data-knowledge features (DKF).
The experimental results on blood pressure prediction showed that, without the need for additional dimensions, 9-dimensional DKF delivered superior predictive performance compared to both 9-dimensional DF and 8-dimensional KF. Specifically, for systolic blood pressure prediction, DKF reduces the mean absolute error (MAE) to 11.41, outperforming KF (MAE=12.11) and DF (MAE=13.24). Similarly, for diastolic blood pressure, DKF achieves an MAE of 7.27, lower than that of KF (7.41) and DF (7.84).
The proposed feature extraction method effectively overcomes the drawbacks of feature concatenation, offering a novel and effective approach to extracting low-dimensional, highly discriminative features from PPG for accurate blood pressure estimation.
虽然将photoplethysmography (PPG)的知识特征(KF)和数据特征(DF)串联可以提高血压监测模型的预测性能,但这种方法不可避免地增加了特征空间的维数。为了解决这一限制,我们提出了一种创新的特征提取方法,该方法将KF和DF深度集成,而不是简单地将它们连接起来。我们的方法采用功能数据分析,将PPG作为连续功能曲线来提取DF。然后,深入分析KF的分布模式,构建基于KF的约束空间,并在DF提取过程中作为指导,生成新的数据-知识特征(DKF)。血压预测的实验结果表明,在不需要额外维度的情况下,与9维DF和8维KF相比,9维DKF具有更好的预测性能。具体而言,对于收缩压预测,DKF将平均绝对误差(MAE)降低至11.41,优于KF (MAE=12.11)和DF (MAE=13.24)。同样,对于舒张压,DKF的MAE为7.27,低于KF(7.41)和DF(7.84)。所提出的特征提取方法有效地克服了特征拼接的缺点,为从PPG中提取低维、高判别性的特征以实现准确的血压估计提供了一种新颖有效的方法。
{"title":"Data-knowledge feature fusion for PPG-based blood pressure prediction: Low-dimensional extraction via functional data analysis and knowledge constraint","authors":"Qingfeng Tang ,&nbsp;Huihui Hu ,&nbsp;Chao Tao ,&nbsp;Pengcheng Ding ,&nbsp;Guowei Dai ,&nbsp;Guangjun Wang ,&nbsp;Xiaojuan Hu ,&nbsp;Benyue Su ,&nbsp;Jiatuo Xu ,&nbsp;Hui An","doi":"10.1016/j.bspc.2026.109754","DOIUrl":"10.1016/j.bspc.2026.109754","url":null,"abstract":"<div><div>Although concatenating knowledge features (KF) and data features (DF) of photoplethysmography (PPG) can improve the predictive performance of blood pressure monitoring models, this approach inevitably increases the dimensionality of feature space. To address this limitation, we propose an innovative feature extraction method that deeply integrate KF and DF, rather than simply concatenating them.</div><div>Our method employs functional data analysis to extract DF by treating PPG as continuously functional curve. Subsequently, the distribution patterns of KF are thoroughly analyzed to construct a KF-based constrained space, which serves as a guide during DF extraction, to yield novel data-knowledge features (DKF).</div><div>The experimental results on blood pressure prediction showed that, without the need for additional dimensions, 9-dimensional DKF delivered superior predictive performance compared to both 9-dimensional DF and 8-dimensional KF. Specifically, for systolic blood pressure prediction, DKF reduces the mean absolute error (MAE) to 11.41, outperforming KF (MAE=12.11) and DF (MAE=13.24). Similarly, for diastolic blood pressure, DKF achieves an MAE of 7.27, lower than that of KF (7.41) and DF (7.84).</div><div>The proposed feature extraction method effectively overcomes the drawbacks of feature concatenation, offering a novel and effective approach to extracting low-dimensional, highly discriminative features from PPG for accurate blood pressure estimation.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109754"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146191946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RA2M-UNet: Efficient medical image segmentation via reparameterized convolution, dual-domain attention and 2D state–space modeling RA2M-UNet:通过重参数化卷积、双域关注和二维状态空间建模的高效医学图像分割
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-06 DOI: 10.1016/j.bspc.2026.109733
Chao Zhang , Lei Yang , Sai Zhang , Hongliang Duan , Jingjing Guo
Deep learning has made remarkable progress across various domains, particularly in medical image segmentation. However, a persistent challenge remains in balancing accuracy and computational efficiency, as current state-of-the-art models often sacrifice one aspect to enhance the other. Here, we propose RA2M-UNet, a novel network that addresses this trade-off through key innovations: (1) a feature fusion module that integrates multi-scale dilated convolutions with 2D selective scan module (2D-SSM); (2) an enhanced 2D-SSM for better spatial and semantic dependency capture; (3) parameter-efficient structural re-parameterization; (4) multi-output supervision for further refined segmentation. Comprehensive experiments demonstrate that our approach outperforms existing methods while maintaining parameter efficiency, effectively resolving the accuracy-efficiency dilemma in medical image segmentation.
深度学习在各个领域都取得了显著的进展,特别是在医学图像分割方面。然而,平衡准确性和计算效率仍然是一个持久的挑战,因为目前最先进的模型经常牺牲一个方面来增强另一个方面。在这里,我们提出了RA2M-UNet,这是一种通过关键创新解决这种权衡的新型网络:(1)将多尺度扩展卷积与2D选择性扫描模块(2D- ssm)集成在一起的特征融合模块;(2)增强2D-SSM,更好地捕获空间和语义依赖性;(3)参数高效结构再参数化;(4)多输出监督,进一步细化细分。综合实验表明,我们的方法在保持参数效率的前提下优于现有方法,有效解决了医学图像分割中精度-效率的难题。
{"title":"RA2M-UNet: Efficient medical image segmentation via reparameterized convolution, dual-domain attention and 2D state–space modeling","authors":"Chao Zhang ,&nbsp;Lei Yang ,&nbsp;Sai Zhang ,&nbsp;Hongliang Duan ,&nbsp;Jingjing Guo","doi":"10.1016/j.bspc.2026.109733","DOIUrl":"10.1016/j.bspc.2026.109733","url":null,"abstract":"<div><div>Deep learning has made remarkable progress across various domains, particularly in medical image segmentation. However, a persistent challenge remains in balancing accuracy and computational efficiency, as current state-of-the-art models often sacrifice one aspect to enhance the other. Here, we propose RA2M-UNet, a novel network that addresses this trade-off through key innovations: (1) a feature fusion module that integrates multi-scale dilated convolutions with 2D selective scan module (2D-SSM); (2) an enhanced 2D-SSM for better spatial and semantic dependency capture; (3) parameter-efficient structural re-parameterization; (4) multi-output supervision for further refined segmentation. Comprehensive experiments demonstrate that our approach outperforms existing methods while maintaining parameter efficiency, effectively resolving the accuracy-efficiency dilemma in medical image segmentation.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109733"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GCN-multiDA: A multi-source personalized domain adaptation model based on a novel streamlined GCN for motor imagery classification GCN- multida:一种基于新型流线型GCN的多源个性化领域自适应运动图像分类模型
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-06 DOI: 10.1016/j.bspc.2026.109773
Zhenxi Zhao , Yingyu Cao , Hongbin Yu , Huixian Yu , Junfen Huang
Brain–computer interfaces (BCIs) play a pivotal role in facilitating human–machine interaction and elucidating brain mechanisms, with motor imagery (MI) being one of the most widely studied paradigms due to its substantial potential. However, inherent inter-subject variability in physiological structures often constrains the accuracy of MI decoding models. To address this challenge, we construct a streamlined graph convolutional network (GCN) and develop an MI decoding model, termed GCN-multiDA. Specifically, the model employs a GCN to capture spatial dependencies in EEG signals and incorporates a graph pruning strategy based on the task-frequency index (TF), region-of-interest index (ROI), and topological index (Topo) to streamline the network. This design preserves neurophysiological relevance while enhancing decoding accuracy and reducing model complexity. Furthermore, drawing inspiration from multi-source personalized domain adaptation, we introduce a domain bias assessment measurement (DBAM) to align cross-domain feature distributions and mitigate inter-domain discrepancies, along with a classifier alignment module to enforce prediction consistency across domains, thereby enabling robust MI classification. Comprehensive experiments conducted on four datasets, including BCI competition IV 2a and 2b, OpenBMI, and PhysioNet, demonstrate that GCN-multiDA consistently outperforms baseline models, improving mean accuracy by 2.66%, 2.53%, 1.32%, and 3.55%, respectively, and achieving the best performance in terms of Kappa and rRMSE metrics. Ablation and sensitivity analyses further confirm that the pruning algorithm contributes substantially to performance improvements across all datasets.
脑机接口(bci)在促进人机交互和阐明脑机制方面发挥着关键作用,其中运动意象(MI)因其巨大的潜力而成为研究最广泛的范式之一。然而,生理结构中固有的主体间可变性往往限制了MI解码模型的准确性。为了解决这一挑战,我们构建了一个流线型图卷积网络(GCN),并开发了一个MI解码模型,称为GCN- multida。具体而言,该模型采用GCN捕获EEG信号中的空间依赖关系,并结合基于任务频率指数(TF)、感兴趣区域指数(ROI)和拓扑指数(Topo)的图修剪策略来简化网络。该设计保留了神经生理学相关性,同时提高了解码精度并降低了模型复杂性。此外,从多源个性化领域自适应中获得灵感,我们引入了一个领域偏差评估测量(DBAM)来校准跨领域的特征分布并减轻领域间的差异,以及一个分类器校准模块来强制跨领域的预测一致性,从而实现鲁棒的MI分类。在BCI competition IV 2a和2b、OpenBMI和PhysioNet四个数据集上进行的综合实验表明,GCN-multiDA持续优于基线模型,平均准确率分别提高了2.66%、2.53%、1.32%和3.55%,并且在Kappa和rRMSE指标方面取得了最佳性能。消融和敏感性分析进一步证实,修剪算法对所有数据集的性能改进都有很大的贡献。
{"title":"GCN-multiDA: A multi-source personalized domain adaptation model based on a novel streamlined GCN for motor imagery classification","authors":"Zhenxi Zhao ,&nbsp;Yingyu Cao ,&nbsp;Hongbin Yu ,&nbsp;Huixian Yu ,&nbsp;Junfen Huang","doi":"10.1016/j.bspc.2026.109773","DOIUrl":"10.1016/j.bspc.2026.109773","url":null,"abstract":"<div><div>Brain–computer interfaces (BCIs) play a pivotal role in facilitating human–machine interaction and elucidating brain mechanisms, with motor imagery (MI) being one of the most widely studied paradigms due to its substantial potential. However, inherent inter-subject variability in physiological structures often constrains the accuracy of MI decoding models. To address this challenge, we construct a streamlined graph convolutional network (GCN) and develop an MI decoding model, termed GCN-multiDA. Specifically, the model employs a GCN to capture spatial dependencies in EEG signals and incorporates a graph pruning strategy based on the task-frequency index (TF), region-of-interest index (ROI), and topological index (Topo) to streamline the network. This design preserves neurophysiological relevance while enhancing decoding accuracy and reducing model complexity. Furthermore, drawing inspiration from multi-source personalized domain adaptation, we introduce a domain bias assessment measurement (DBAM) to align cross-domain feature distributions and mitigate inter-domain discrepancies, along with a classifier alignment module to enforce prediction consistency across domains, thereby enabling robust MI classification. Comprehensive experiments conducted on four datasets, including BCI competition IV 2a and 2b, OpenBMI, and PhysioNet, demonstrate that GCN-multiDA consistently outperforms baseline models, improving mean accuracy by 2.66%, 2.53%, 1.32%, and 3.55%, respectively, and achieving the best performance in terms of Kappa and rRMSE metrics. Ablation and sensitivity analyses further confirm that the pruning algorithm contributes substantially to performance improvements across all datasets.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109773"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SEDP-SegResnet for human eyeball and lens segmentation SEDP-SegResnet用于人眼和晶状体分割
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-13 DOI: 10.1016/j.bspc.2026.109838
Li Ning , Yepei Qin , Wendong Zhao , Yangjiarui Yu , Qingcheng Yang , Chenxi Guo , Xuedian Zhang , Hui Chen , Yinghong Ji , Pei Ma
Accurate segmentation and quantification of the eyeball and lens from MRI images are crucial for clinical diagnosis and treatment planning of ocular diseases. Traditional methods for analyzing eye structures in MRI have drawbacks including low segmentation accuracy and reliance on laborious, time-consuming manual processes. To solve these problems, we propose a SEDP-SegResnet model for segmentation of the eyeball and lens structures from 3D MRI images. The framework takes SegResnet as its backbone network and incorporates a 3D-SE layer to handle deep features from decoder, 3D-SE layer assigns different weight information to the feature map channels through squeeze and excitation mechanism. Moreover, skip connections in the U-shaped architecture model are replaced with Dynamic Deep Feature Prefusion (DDFP) modules. The DDFP can achieve in-depth fusion of encoder and decoder features based on global information, thereby enhancing 3D image context comprehension of the model. The performance of SEDP-SegResnet is evaluated through a series of experiments using a proprietary dataset of orbital MRI scans. The results show that SEDP-SegResnet outperforms current mainstream 3D deep-learning-based segmentation models across multiple evaluation metrics including the Dice Similarity Coefficient (DSC) and Intersection over Union (IoU). The model achieves robust performances in segmenting margin of eyeballs and blur-edge lenses. SEDP-SegResnet achieves a DSC of 96.81% for eyeball segmentation and 90.57% for lens segmentation, superior than a variety of commonly used segmentation models. It provides a more accurate, automated and robust method for the segmentation and quantification of eyeball and lens in MRI, offering an advanced computer-aided diagnosis tool.
MRI图像中眼球和晶状体的准确分割和量化对眼科疾病的临床诊断和治疗计划至关重要。传统的MRI眼睛结构分析方法存在分割精度低、依赖于费力、耗时的人工处理等缺点。为了解决这些问题,我们提出了一种SEDP-SegResnet模型,用于从3D MRI图像中分割眼球和晶状体结构。该框架以SegResnet为骨干网络,结合3D-SE层处理来自解码器的深度特征,3D-SE层通过挤压和激励机制为特征图通道分配不同的权重信息。此外,用动态深度特征预融合(Dynamic Deep Feature Prefusion, DDFP)模块代替u型结构模型中的跳过连接。DDFP可以基于全局信息实现编码器和解码器特征的深度融合,从而增强模型对三维图像上下文的理解能力。SEDP-SegResnet的性能通过使用专有的眼眶MRI扫描数据集进行一系列实验来评估。结果表明,SEDP-SegResnet在多个评估指标上优于当前主流的基于3D深度学习的分割模型,包括Dice Similarity Coefficient (DSC)和Intersection over Union (IoU)。该模型在眼球边缘分割和模糊边缘镜头分割方面取得了较好的效果。SEDP-SegResnet对眼球分割的DSC为96.81%,对晶体分割的DSC为90.57%,优于多种常用分割模型。它为MRI中眼球和晶状体的分割和定量提供了一种更加准确、自动化和稳健的方法,提供了一种先进的计算机辅助诊断工具。
{"title":"SEDP-SegResnet for human eyeball and lens segmentation","authors":"Li Ning ,&nbsp;Yepei Qin ,&nbsp;Wendong Zhao ,&nbsp;Yangjiarui Yu ,&nbsp;Qingcheng Yang ,&nbsp;Chenxi Guo ,&nbsp;Xuedian Zhang ,&nbsp;Hui Chen ,&nbsp;Yinghong Ji ,&nbsp;Pei Ma","doi":"10.1016/j.bspc.2026.109838","DOIUrl":"10.1016/j.bspc.2026.109838","url":null,"abstract":"<div><div>Accurate segmentation and quantification of the eyeball and lens from MRI images are crucial for clinical diagnosis and treatment planning of ocular diseases. Traditional methods for analyzing eye structures in MRI have drawbacks including low segmentation accuracy and reliance on laborious, time-consuming manual processes. To solve these problems, we propose a SEDP-SegResnet model for segmentation of the eyeball and lens structures from 3D MRI images. The framework takes SegResnet as its backbone network and incorporates a 3D-SE layer to handle deep features from decoder, 3D-SE layer assigns different weight information to the feature map channels through squeeze and excitation mechanism. Moreover, skip connections in the U-shaped architecture model are replaced with Dynamic Deep Feature Prefusion (DDFP) modules. The DDFP can achieve in-depth fusion of encoder and decoder features based on global information, thereby enhancing 3D image context comprehension of the model. The performance of SEDP-SegResnet is evaluated through a series of experiments using a proprietary dataset of orbital MRI scans. The results show that SEDP-SegResnet outperforms current mainstream 3D deep-learning-based segmentation models across multiple evaluation metrics including the Dice Similarity Coefficient (DSC) and Intersection over Union (IoU). The model achieves robust performances in segmenting margin of eyeballs and blur-edge lenses. SEDP-SegResnet achieves a DSC of 96.81% for eyeball segmentation and 90.57% for lens segmentation, superior than a variety of commonly used segmentation models. It provides a more accurate, automated and robust method for the segmentation and quantification of eyeball and lens in MRI, offering an advanced computer-aided diagnosis tool.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109838"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An enhanced deep learning model for breast cancer histopathological grading based on Selective Kernel network 基于选择性核网络的乳腺癌组织病理学分级的增强深度学习模型
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2026-06-15 Epub Date: 2026-02-13 DOI: 10.1016/j.bspc.2026.109833
Yuandi Sun
<div><h3>Background</h3><div>Breast cancer is one of the most common malignant tumors in women worldwide. Its early detection and accurate grading are crucial for developing individualized treatment plans and improving patient prognosis. Pathological image grading is a key step in breast cancer diagnosis, but due to the high heterogeneity of tumor cell and tissue morphology, traditional manual image reading methods have subjective bias and low efficiency. Therefore, developing an automated and accurate breast cancer pathological image grading model has important clinical value for improving diagnostic efficiency and accuracy.</div></div><div><h3>Methods</h3><div>This study proposed a deep learning model based on the combination of DenseNet and Selective Kernel Block (SKBlock) − SKDenseNet for the automatic grading of breast cancer pathology images. DenseNet enhances feature reuse and gradient propagation efficiency through dense connection mechanism, while SKBlock realizes dynamic extraction and fusion of pathological features of different scales through multi-scale convolution operation and channel attention mechanism. The model was trained on the TCGA dataset and independently tested on the CHTN dataset to evaluate the generalization ability and stability of the model in cross-center tasks. The model parameters were optimized, and the classification performance was evaluated by accuracy (ACC), precision (PRE), recall (REC) and F1 score (F1), and the discriminability and interpretability of the model were analyzed by confusion matrix and activation heat map .</div></div><div><h3>Results</h3><div>The experimental results on the test set (CHTN dataset) showed that SKDenseNet significantly outperformed the baseline model in all key classification indicators. The average accuracy of SKDenseNet was 86.58%, the average precision was 87.91%, the average recall was 87.97%, and the F1 score was 86.71%, which were 7.84, 3.18, 6.24, and 7.32 percentage points higher than DenseNet121, respectively. The confusion matrix showed that SKDenseNet showed good discrimination and stability in the classification tasks of high- and low-grade breast cancer and stromal tissue. In addition, SKDenseNet also has the highest AUC, reaching 0.9693. The activation heat map generated by Grad-CAM further verified that the key areas of the model’s attention in the pathological images were highly consistent with the actual pathological features (such as nuclear morphology and glandular duct structure), which enhanced the interpretability of the model.</div></div><div><h3>Conclusion</h3><div>SKDenseNet model proposed in this study combines the global feature expression ability of DenseNet with the dynamic receptive field adjustment mechanism of SKBlock, and shows excellent classification performance and cross-center adaptability in the task of breast cancer pathology image grading . The model can reduce the misdiagnosis rate and missed diagnosis rate while maintaining high accurac
背景乳腺癌是世界范围内女性最常见的恶性肿瘤之一。早期发现和准确分级对制定个体化治疗方案和改善患者预后至关重要。病理图像分级是乳腺癌诊断的关键步骤,但由于肿瘤细胞和组织形态的高度异质性,传统的人工图像读取方法存在主观偏差,效率较低。因此,开发一种自动化、准确的乳腺癌病理图像分级模型对于提高诊断效率和准确率具有重要的临床价值。方法提出了一种基于DenseNet和选择性核块(SKBlock)−SKDenseNet相结合的深度学习模型,用于乳腺癌病理图像的自动分级。DenseNet通过密集连接机制提高特征重用和梯度传播效率,SKBlock通过多尺度卷积运算和通道关注机制实现不同尺度病理特征的动态提取和融合。在TCGA数据集上对模型进行训练,并在CHTN数据集上进行独立测试,以评估模型在跨中心任务中的泛化能力和稳定性。对模型参数进行优化,通过准确率(ACC)、精密度(PRE)、召回率(REC)和F1分数(F1)评价分类性能,并通过混淆矩阵和激活热图分析模型的可分辨性和可解释性。结果在测试集(CHTN数据集)上的实验结果表明,SKDenseNet在所有关键分类指标上都明显优于基线模型。SKDenseNet的平均准确率为86.58%,平均准确率为87.91%,平均召回率为87.97%,F1得分为86.71%,分别比DenseNet121高7.84、3.18、6.24和7.32个百分点。混淆矩阵显示,SKDenseNet在高分级和低分级乳腺癌及间质组织的分类任务中具有良好的鉴别性和稳定性。此外,SKDenseNet的AUC也最高,达到0.9693。通过Grad-CAM生成的激活热图进一步验证了病理图像中模型关注的关键区域与实际病理特征(如核形态、腺管结构等)高度一致,增强了模型的可解释性。结论本研究提出的skdensenet模型结合了DenseNet的全局特征表达能力和SKBlock的动态感受野调节机制,在乳腺癌病理图像分级任务中表现出优异的分类性能和跨中心适应性。该模型在保持较高准确率的同时降低了误诊率和漏诊率,具有良好的推广能力和临床应用潜力。
{"title":"An enhanced deep learning model for breast cancer histopathological grading based on Selective Kernel network","authors":"Yuandi Sun","doi":"10.1016/j.bspc.2026.109833","DOIUrl":"10.1016/j.bspc.2026.109833","url":null,"abstract":"&lt;div&gt;&lt;h3&gt;Background&lt;/h3&gt;&lt;div&gt;Breast cancer is one of the most common malignant tumors in women worldwide. Its early detection and accurate grading are crucial for developing individualized treatment plans and improving patient prognosis. Pathological image grading is a key step in breast cancer diagnosis, but due to the high heterogeneity of tumor cell and tissue morphology, traditional manual image reading methods have subjective bias and low efficiency. Therefore, developing an automated and accurate breast cancer pathological image grading model has important clinical value for improving diagnostic efficiency and accuracy.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;h3&gt;Methods&lt;/h3&gt;&lt;div&gt;This study proposed a deep learning model based on the combination of DenseNet and Selective Kernel Block (SKBlock) − SKDenseNet for the automatic grading of breast cancer pathology images. DenseNet enhances feature reuse and gradient propagation efficiency through dense connection mechanism, while SKBlock realizes dynamic extraction and fusion of pathological features of different scales through multi-scale convolution operation and channel attention mechanism. The model was trained on the TCGA dataset and independently tested on the CHTN dataset to evaluate the generalization ability and stability of the model in cross-center tasks. The model parameters were optimized, and the classification performance was evaluated by accuracy (ACC), precision (PRE), recall (REC) and F1 score (F1), and the discriminability and interpretability of the model were analyzed by confusion matrix and activation heat map .&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;h3&gt;Results&lt;/h3&gt;&lt;div&gt;The experimental results on the test set (CHTN dataset) showed that SKDenseNet significantly outperformed the baseline model in all key classification indicators. The average accuracy of SKDenseNet was 86.58%, the average precision was 87.91%, the average recall was 87.97%, and the F1 score was 86.71%, which were 7.84, 3.18, 6.24, and 7.32 percentage points higher than DenseNet121, respectively. The confusion matrix showed that SKDenseNet showed good discrimination and stability in the classification tasks of high- and low-grade breast cancer and stromal tissue. In addition, SKDenseNet also has the highest AUC, reaching 0.9693. The activation heat map generated by Grad-CAM further verified that the key areas of the model’s attention in the pathological images were highly consistent with the actual pathological features (such as nuclear morphology and glandular duct structure), which enhanced the interpretability of the model.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;h3&gt;Conclusion&lt;/h3&gt;&lt;div&gt;SKDenseNet model proposed in this study combines the global feature expression ability of DenseNet with the dynamic receptive field adjustment mechanism of SKBlock, and shows excellent classification performance and cross-center adaptability in the task of breast cancer pathology image grading . The model can reduce the misdiagnosis rate and missed diagnosis rate while maintaining high accurac","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109833"},"PeriodicalIF":4.9,"publicationDate":"2026-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Biomedical Signal Processing and Control
全部 Asia-Pac. J. Atmos. Sci. ARCHAEOMETRY "Laboratorio;" analisis clinicos, bacteriologia, inmunologia, parasitologia, hematologia, anatomia patologica, quimica clinica BIOGEOSCIENCES Appl. Geochem. Appl. Clay Sci. Org. Geochem. ARCT ANTARCT ALP RES Big Earth Data ENTROPY-SWITZ COMP BIOCHEM PHYS C EUR PHYS J-SPEC TOP Commun. Phys. Geochem. J. Geol. Ore Deposits INDIAN J PURE AP PHY Communications Earth & Environment Acta Oceanolog. Sin. J. Earth Sci. J. Atmos. Sol. Terr. Phys. Ecol. Indic. ECOSYSTEMS J. Afr. Earth. Sci. ICARUS 2013 Abstracts IEEE International Conference on Plasma Science (ICOPS) ASTRON ASTROPHYS Adv. Atmos. Sci. EPL-EUROPHYS LETT Geosci. Model Dev. 2005 Asian Conference on Sensors and the International Conference on New Techniques in Pharmaceutical and Biomedical Research Chin. Phys. B J. Clim. Nat. Clim. Change Energy Ecol Environ Jpn. J. Appl. Phys. J. Phys. Soc. Jpn. Int. J. Earth Sci. Nat. Photonics Laser Phys. Lett. npj Clim. Atmos. Sci. AAPS PHARMSCITECH TERR ATMOS OCEAN SCI [1993] Proceedings Eighth Annual IEEE Symposium on Logic in Computer Science Int. J. Paleopathol. J. Electron. Spectrosc. Relat. Phenom. J OPT SOC AM A 2009 International Workshop on Intelligent Systems and Applications J NONLINEAR OPT PHYS J. Synchrotron Radiat. Nat. Hazards Earth Syst. Sci. Ore Geol. Rev. Acta Geophys. Carbon Balance Manage. J. Atmos. Chem. Geostand. Geoanal. Res. Am. J. Phys. Anthropol. Adv. Meteorol. Contrib. Mineral. Petrol. IZV-PHYS SOLID EART+ AAPG Bull. Geobiology Atmos. Meas. Tech. Basin Res. ACTA PETROL SIN Ann. Glaciol. Am. J. Sci. Acta Pharmacol. Sin. GEOLOGY ACTA GEOL SIN-ENGL J. Hydrol. Atmos. Chem. Phys. Am. Mineral. Clean Technol. Environ. Policy 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT) Aust. J. Earth Sci. Geochim. Cosmochim. Acta Astrophys. J. Suppl. Ser. ACTA GEOL POL Acta Geochimica Phys. Chem. Miner. Int. J. Biometeorol. Geosci. Front. Geosci. J. Geol. J. Annu. Rev. Earth Planet. Sci. Clean-Soil Air Water Aquat. Geochem. CRIT REV ENV SCI TEC Chem. Ecol. Astrophys. Space Sci. Hydrol. Processes Atmos. Res. Appl. Phys. Rev. INT J MOD PHYS B 国际生物医学工程杂志 ACTA MED OKAYAMA Classical Quantum Gravity 2011 International Conference on Electrical and Control Engineering Chin. J. Phys. ASTROBIOLOGY
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1