CephTransXnet: An attention enhanced feature fusion network leveraging neighborhood rough set approach for cephalometric landmark prediction

IF 6.3 2区 医学 Q1 BIOLOGY Computers in biology and medicine Pub Date : 2025-04-01 Epub Date: 2025-02-25 DOI:10.1016/j.compbiomed.2025.109891
R. Neeraja, L. Jani Anbarasi
{"title":"CephTransXnet: An attention enhanced feature fusion network leveraging neighborhood rough set approach for cephalometric landmark prediction","authors":"R. Neeraja,&nbsp;L. Jani Anbarasi","doi":"10.1016/j.compbiomed.2025.109891","DOIUrl":null,"url":null,"abstract":"<div><div>The convergence of medical imaging, computer vision, and orthodontics has made automatic cephalometric landmark detection a pivotal area of research. Accurate cephalometric analysis is crucial in orthodontics, orthognathic and maxillofacial surgery for diagnosis, treatment planning, and monitoring craniofacial growth. In this research study, a multi-branch fused feature extraction network titled <span><math><mrow><msub><mrow><mi>C</mi><mi>e</mi><mi>p</mi><mi>h</mi><mi>T</mi><mi>r</mi><mi>a</mi><mi>n</mi><mi>s</mi><mi>X</mi></mrow><mrow><mi>n</mi><mi>e</mi><mi>t</mi></mrow></msub></mrow></math></span> is proposed to automatically predict landmark coordinates from cephalometric radiographs. The initial sequential branch enhances discriminative local feature learning and feature extraction through parallel feature fusion by integrating Convolved Pooled Normalized (<span><math><mrow><msub><mrow><mi>C</mi><mi>P</mi><mi>N</mi></mrow><mi>B</mi></msub></mrow></math></span>) and Gradient Optimized Multi-Path Bottleneck (<span><math><mrow><msub><mrow><mi>G</mi><mi>M</mi><mi>B</mi></mrow><mi>B</mi></msub></mrow></math></span>) blocks with Channel and Spatial Attention (<span><math><mrow><msub><mrow><mi>C</mi><mi>S</mi><mi>A</mi><mi>T</mi></mrow><mi>M</mi></msub></mrow></math></span>) module. The Swin Transformer (<span><math><mrow><msub><mrow><mi>S</mi><mi>T</mi></mrow><mi>B</mi></msub><mo>)</mo></mrow></math></span> branch efficiently handles long-range dependencies and extracts global features in cephalometric radiographs. The multi-branch fused features along with features from skip connections of <span><math><mrow><msub><mrow><mi>C</mi><mi>P</mi><mi>N</mi></mrow><mi>B</mi></msub></mrow></math></span> and <span><math><mrow><msub><mrow><mi>G</mi><mi>M</mi><mi>B</mi></mrow><mi>B</mi></msub></mrow></math></span> blocks are concatenated using a Coordinate Attention module <span><math><mrow><mo>(</mo><msub><mrow><mi>C</mi><mi>o</mi><mi>A</mi><mi>T</mi></mrow><mi>M</mi></msub><mo>)</mo></mrow></math></span> to captures the positional relationships between various landmark features. A Landmark Discriminative Deviation Factor <span><math><mrow><mo>(</mo><mrow><mi>L</mi><mi>D</mi><mi>D</mi><mi>F</mi></mrow><mo>)</mo></mrow></math></span> is determined by applying the Neighborhood Rough Set <span><math><mrow><mo>(</mo><mrow><mi>N</mi><mi>R</mi><mi>S</mi></mrow><mo>)</mo></mrow></math></span> approach to analyse the surrounding features of each landmark by considering spatial relationships or similarity measures between the landmarks and neighboring regions. The Spatial Pyramid Pooling (<span><math><mrow><msub><mrow><mi>S</mi><mi>P</mi><mi>P</mi></mrow><mi>L</mi></msub></mrow></math></span>) layer incorporated in the final phase of <span><math><mrow><msub><mrow><mi>C</mi><mi>e</mi><mi>p</mi><mi>h</mi><mi>T</mi><mi>r</mi><mi>a</mi><mi>n</mi><mi>s</mi><mi>X</mi></mrow><mrow><mi>n</mi><mi>e</mi><mi>t</mi></mrow></msub></mrow></math></span> model extracts multi-scale features by pooling over sub-regions of varying sizes, enabling the network to capture both local and global context for precise cephalometric landmark identification. The <span><math><mrow><msub><mrow><mi>C</mi><mi>e</mi><mi>p</mi><mi>h</mi><mi>T</mi><mi>r</mi><mi>a</mi><mi>n</mi><mi>s</mi><mi>X</mi></mrow><mrow><mi>n</mi><mi>e</mi><mi>t</mi></mrow></msub></mrow></math></span> framework achieved an average Successful Detection Rates <span><math><mrow><mo>(</mo><msub><mrow><mi>S</mi><mi>D</mi><mi>R</mi></mrow><mi>s</mi></msub><mo>)</mo></mrow></math></span> of 88.71 % and 79.05 % in 2 mm using the 2015 International Symposium on Biomedical Imaging (ISBI) grand challenge dental X-ray analysis dataset. The effectiveness of the <span><math><mrow><msub><mrow><mi>C</mi><mi>e</mi><mi>p</mi><mi>h</mi><mi>T</mi><mi>r</mi><mi>a</mi><mi>n</mi><mi>s</mi><mi>X</mi></mrow><mrow><mi>n</mi><mi>e</mi><mi>t</mi></mrow></msub></mrow></math></span> model is evaluated using a private clinical dataset obtained from Solanki Dental Care Clinic in Sharjah, UAE, and attained an average <span><math><mrow><msub><mrow><mi>S</mi><mi>D</mi><mi>R</mi></mrow><mi>s</mi></msub></mrow></math></span> of 74.38 % in 2 mm precision range.</div></div>","PeriodicalId":10578,"journal":{"name":"Computers in biology and medicine","volume":"188 ","pages":"Article 109891"},"PeriodicalIF":6.3000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers in biology and medicine","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0010482525002422","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/25 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"BIOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

The convergence of medical imaging, computer vision, and orthodontics has made automatic cephalometric landmark detection a pivotal area of research. Accurate cephalometric analysis is crucial in orthodontics, orthognathic and maxillofacial surgery for diagnosis, treatment planning, and monitoring craniofacial growth. In this research study, a multi-branch fused feature extraction network titled CephTransXnet is proposed to automatically predict landmark coordinates from cephalometric radiographs. The initial sequential branch enhances discriminative local feature learning and feature extraction through parallel feature fusion by integrating Convolved Pooled Normalized (CPNB) and Gradient Optimized Multi-Path Bottleneck (GMBB) blocks with Channel and Spatial Attention (CSATM) module. The Swin Transformer (STB) branch efficiently handles long-range dependencies and extracts global features in cephalometric radiographs. The multi-branch fused features along with features from skip connections of CPNB and GMBB blocks are concatenated using a Coordinate Attention module (CoATM) to captures the positional relationships between various landmark features. A Landmark Discriminative Deviation Factor (LDDF) is determined by applying the Neighborhood Rough Set (NRS) approach to analyse the surrounding features of each landmark by considering spatial relationships or similarity measures between the landmarks and neighboring regions. The Spatial Pyramid Pooling (SPPL) layer incorporated in the final phase of CephTransXnet model extracts multi-scale features by pooling over sub-regions of varying sizes, enabling the network to capture both local and global context for precise cephalometric landmark identification. The CephTransXnet framework achieved an average Successful Detection Rates (SDRs) of 88.71 % and 79.05 % in 2 mm using the 2015 International Symposium on Biomedical Imaging (ISBI) grand challenge dental X-ray analysis dataset. The effectiveness of the CephTransXnet model is evaluated using a private clinical dataset obtained from Solanki Dental Care Clinic in Sharjah, UAE, and attained an average SDRs of 74.38 % in 2 mm precision range.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
CephTransXnet:利用邻域粗糙集方法进行头颅测量地标预测的注意力增强特征融合网络
医学影像、计算机视觉和正畸学的融合使得自动头侧标记检测成为一个关键的研究领域。准确的头颅测量分析对于正畸、正颌和颌面外科的诊断、治疗计划和监测颅面生长至关重要。在本研究中,提出了一种名为cepphtransxnet的多分支融合特征提取网络,用于自动预测头颅x线片的地标坐标。初始序列分支通过将卷积池化归一化(CPNB)和梯度优化多路径瓶颈(GMBB)块与通道和空间注意(CSATM)模块集成,通过并行特征融合增强判别局部特征学习和特征提取。Swin Transformer (STB)分支有效地处理长程依赖关系并提取头颅x线片的全局特征。利用坐标注意模块(CoATM)将多分支融合特征与CPNB和GMBB块的跳过连接特征连接起来,以捕获各种地标特征之间的位置关系。采用邻域粗糙集(NRS)方法,通过考虑地标与相邻区域之间的空间关系或相似性度量,对每个地标的周边特征进行分析,确定地标判别偏差因子(LDDF)。在cepphtransxnet模型的最后阶段,空间金字塔池化(SPPL)层通过池化不同大小的子区域来提取多尺度特征,使网络能够捕获局部和全局背景,以精确识别头部测量地标。使用2015年国际生物医学成像研讨会(ISBI)大挑战牙科x射线分析数据集,cepphtransxnet框架在2mm内实现了88.71%和79.05%的平均成功检出率(sdr)。使用来自阿联酋沙迦Solanki牙科诊所的私人临床数据集对cepphtransxnet模型的有效性进行了评估,并在2mm精度范围内获得了74.38%的平均特别分配率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Computers in biology and medicine
Computers in biology and medicine 工程技术-工程:生物医学
CiteScore
11.70
自引率
10.40%
发文量
1086
审稿时长
74 days
期刊介绍: Computers in Biology and Medicine is an international forum for sharing groundbreaking advancements in the use of computers in bioscience and medicine. This journal serves as a medium for communicating essential research, instruction, ideas, and information regarding the rapidly evolving field of computer applications in these domains. By encouraging the exchange of knowledge, we aim to facilitate progress and innovation in the utilization of computers in biology and medicine.
期刊最新文献
Forecasting-based biomedical time-series data synthesis for open data and robust AI Precise oxygen therapy to emphysema patients by fuzzy-based gain tuning control of set-point regulated MRAC MIPHEI-ViT: Multiplex immunofluorescence prediction from H&E images using ViT foundation models Deep learning with limited data: a transfer learning approach for transcriptomic survival prediction Noninvasive heart rate estimation using semantic segmentation and parameter optimization on 4K UAV videos
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1