首页 > 最新文献

Physical and Engineering Sciences in Medicine最新文献

英文 中文
An open-source tool for converting 3D mesh volumes into synthetic DICOM CT images for medical physics research. 一个开源工具,用于将3D网格体积转换为医学物理研究的合成DICOM CT图像。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-07-24 DOI: 10.1007/s13246-025-01599-x
Michael John James Douglass

Access to medical imaging data is crucial for research, training, and treatment planning in medical imaging and radiation therapy. However, ethical constraints and time-consuming approval processes often limit the availability of such data for research. This study introduces DICOMator, an open-source Blender add-on designed to address this challenge by enabling the creation of synthetic CT datasets from 3D mesh objects. DICOMator aims to provide researchers and medical professionals with a flexible tool for generating customisable and semi-realistic synthetic CT data, including 4D CT datasets from user defined static or animated 3D mesh objects. The add-on leverages Blender's powerful 3D modelling environment, utilising its mesh manipulation, animation and rendering capabilities to create synthetic data ranging from simple phantoms to accurate anatomical models. DICOMator incorporates various features to simulate common CT imaging artefacts, bridging the gap between 3D modelling and medical imaging. DICOMator voxelises 3D mesh objects, assigns appropriate Hounsfield Unit values, and applies artefact simulations. These simulations include detector noise, metal artefacts and partial volume effects. By incorporating these artefacts, DICOMator produces synthetic CT data that more closely resembles real CT scans. The resulting data is then exported in DICOM format, ensuring compatibility with existing medical imaging workflows and treatment planning systems. To demonstrate DICOMator's capabilities, three synthetic CT datasets were created: a simple lung phantom to illustrate basic functionality, a more realistic cranial CT scan to demonstrate dose calculations and CT image registration on synthetic data in treatment planning systems. Finally, a thoracic 4D CT scan featuring multiple breathing phases was created to demonstrate the dynamic imaging capabilities and the quantitative accuracy of the synthetic datasets. These examples were chosen to highlight DICOMator's versatility in generating diverse and complex synthetic CT data suitable for various research and educational purposes, from basic quality assurance to advanced motion management studies. DICOMator offers a promising solution to the limitations of patient CT data availability in medical physics research. By providing a user-friendly interface for creating customisable synthetic datasets from 3D meshes, it has the potential to accelerate research, validate treatment planning tools such as deformable image registration, and enhance educational resources in the field of radiation oncology medical physics. Future developments may include incorporation of other imaging modalities, such as MRI or PET, further expanding its utility in multi-modal imaging research.

获得医学成像数据对于医学成像和放射治疗的研究、培训和治疗计划至关重要。然而,伦理约束和耗时的审批过程往往限制了这些数据用于研究的可用性。本研究介绍了DICOMator,这是一个开源的Blender插件,旨在通过3D网格对象创建合成CT数据集来解决这一挑战。DICOMator旨在为研究人员和医疗专业人员提供一个灵活的工具,用于生成可定制和半逼真的合成CT数据,包括来自用户定义的静态或动画3D网格对象的4D CT数据集。该附加组件利用Blender强大的3D建模环境,利用其网格操作,动画和渲染功能来创建从简单的幻影到精确的解剖模型的合成数据。DICOMator结合了各种功能来模拟常见的CT成像伪影,弥合了3D建模和医学成像之间的差距。DICOMator将3D网格对象体素化,分配适当的Hounsfield单位值,并应用人工模拟。这些模拟包括探测器噪声、金属伪影和部分体积效应。通过合并这些伪影,DICOMator生成的合成CT数据更接近于真实的CT扫描。结果数据然后以DICOM格式导出,确保与现有的医学成像工作流程和治疗计划系统兼容。为了演示DICOMator的功能,创建了三个合成CT数据集:一个简单的肺幻象来说明基本功能,一个更真实的颅脑CT扫描来演示剂量计算,以及治疗计划系统中合成数据的CT图像配准。最后,创建了具有多个呼吸期的胸部4D CT扫描,以展示动态成像能力和合成数据集的定量准确性。选择这些例子是为了突出DICOMator在生成各种复杂的合成CT数据方面的多功能性,适用于从基本质量保证到高级运动管理研究的各种研究和教育目的。DICOMator为医学物理研究中患者CT数据可用性的限制提供了一个有希望的解决方案。通过提供一个用户友好的界面,从3D网格创建可定制的合成数据集,它有可能加速研究,验证治疗计划工具,如可变形图像配准,并增强放射肿瘤学医学物理领域的教育资源。未来的发展可能包括纳入其他成像模式,如MRI或PET,进一步扩大其在多模态成像研究中的应用。
{"title":"An open-source tool for converting 3D mesh volumes into synthetic DICOM CT images for medical physics research.","authors":"Michael John James Douglass","doi":"10.1007/s13246-025-01599-x","DOIUrl":"10.1007/s13246-025-01599-x","url":null,"abstract":"<p><p>Access to medical imaging data is crucial for research, training, and treatment planning in medical imaging and radiation therapy. However, ethical constraints and time-consuming approval processes often limit the availability of such data for research. This study introduces DICOMator, an open-source Blender add-on designed to address this challenge by enabling the creation of synthetic CT datasets from 3D mesh objects. DICOMator aims to provide researchers and medical professionals with a flexible tool for generating customisable and semi-realistic synthetic CT data, including 4D CT datasets from user defined static or animated 3D mesh objects. The add-on leverages Blender's powerful 3D modelling environment, utilising its mesh manipulation, animation and rendering capabilities to create synthetic data ranging from simple phantoms to accurate anatomical models. DICOMator incorporates various features to simulate common CT imaging artefacts, bridging the gap between 3D modelling and medical imaging. DICOMator voxelises 3D mesh objects, assigns appropriate Hounsfield Unit values, and applies artefact simulations. These simulations include detector noise, metal artefacts and partial volume effects. By incorporating these artefacts, DICOMator produces synthetic CT data that more closely resembles real CT scans. The resulting data is then exported in DICOM format, ensuring compatibility with existing medical imaging workflows and treatment planning systems. To demonstrate DICOMator's capabilities, three synthetic CT datasets were created: a simple lung phantom to illustrate basic functionality, a more realistic cranial CT scan to demonstrate dose calculations and CT image registration on synthetic data in treatment planning systems. Finally, a thoracic 4D CT scan featuring multiple breathing phases was created to demonstrate the dynamic imaging capabilities and the quantitative accuracy of the synthetic datasets. These examples were chosen to highlight DICOMator's versatility in generating diverse and complex synthetic CT data suitable for various research and educational purposes, from basic quality assurance to advanced motion management studies. DICOMator offers a promising solution to the limitations of patient CT data availability in medical physics research. By providing a user-friendly interface for creating customisable synthetic datasets from 3D meshes, it has the potential to accelerate research, validate treatment planning tools such as deformable image registration, and enhance educational resources in the field of radiation oncology medical physics. Future developments may include incorporation of other imaging modalities, such as MRI or PET, further expanding its utility in multi-modal imaging research.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1525-1538"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12738608/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144709487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dosimetric evaluation of synthetic kilo-voltage CT images generated from megavoltage CT for head and neck tomotherapy using a conditional GAN network. 使用条件GAN网络对巨压CT生成的用于头颈部断层治疗的合成千伏CT图像进行剂量学评估。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-07-28 DOI: 10.1007/s13246-025-01603-4
Yazdan Choghazardi, Mohamad Bagher Tavakoli, Iraj Abedi, Mahnaz Roayaei, Simin Hemati, Ahmad Shanei

The lower image contrast of megavoltage computed tomography (MVCT), which corresponds to kilovoltage computed tomography (kVCT), can inhibit accurate dosimetric assessments. This study proposes a deep learning approach, specifically the pix2pix network, to generate high-quality synthetic kVCT (skVCT) images from MVCT data. The model was trained on a dataset of 25 paired patient images and evaluated on a test set of 15 paired images. We performed visual inspections to assess the quality of the generated skVCT images and calculated the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). Dosimetric equivalence was evaluated by comparing the gamma pass rates of treatment plans derived from skVCT and kVCT images. Results showed that skVCT images exhibited significantly higher quality than MVCT images, with PSNR and SSIM values of 31.9 ± 1.1 dB and 94.8% ± 1.3%, respectively, compared to 26.8 ± 1.7 dB and 89.5% ± 1.5% for MVCT-to-kVCT comparisons. Furthermore, treatment plans based on skVCT images achieved excellent gamma pass rates of 99.78 ± 0.14% and 99.82 ± 0.20% for 2 mm/2% and 3 mm/3% criteria, respectively, comparable to those obtained from kVCT-based plans (99.70 ± 0.31% and 99.79 ± 1.32%). This study demonstrates the potential of pix2pix models for generating high-quality skVCT images, which could significantly enhance Adaptive Radiation Therapy (ART).

与千伏计算机断层扫描(kVCT)相对应的巨电压计算机断层扫描(MVCT)图像对比度较低,可能会抑制准确的剂量学评估。本研究提出了一种深度学习方法,特别是pix2pix网络,从MVCT数据中生成高质量的合成kVCT (skVCT)图像。该模型在25个配对患者图像的数据集上进行训练,并在15个配对图像的测试集上进行评估。我们通过目视检查来评估生成的skVCT图像的质量,并计算峰值信噪比(PSNR)和结构相似指数(SSIM)。通过比较来自skVCT和kVCT图像的治疗方案的伽马通过率来评估剂量学等效性。结果显示,skVCT图像质量明显高于MVCT图像,PSNR和SSIM值分别为31.9±1.1 dB和94.8%±1.3%,而MVCT与kvct的PSNR和SSIM值分别为26.8±1.7 dB和89.5%±1.5%。此外,基于skVCT图像的治疗方案在2 mm/2%和3 mm/3%的标准下分别获得了99.78±0.14%和99.82±0.20%的优异伽马及格率,与基于kvct的方案(99.70±0.31%和99.79±1.32%)相当。这项研究证明了pix2pix模型在生成高质量skVCT图像方面的潜力,这可以显著增强适应性放射治疗(ART)。
{"title":"Dosimetric evaluation of synthetic kilo-voltage CT images generated from megavoltage CT for head and neck tomotherapy using a conditional GAN network.","authors":"Yazdan Choghazardi, Mohamad Bagher Tavakoli, Iraj Abedi, Mahnaz Roayaei, Simin Hemati, Ahmad Shanei","doi":"10.1007/s13246-025-01603-4","DOIUrl":"10.1007/s13246-025-01603-4","url":null,"abstract":"<p><p>The lower image contrast of megavoltage computed tomography (MVCT), which corresponds to kilovoltage computed tomography (kVCT), can inhibit accurate dosimetric assessments. This study proposes a deep learning approach, specifically the pix2pix network, to generate high-quality synthetic kVCT (skVCT) images from MVCT data. The model was trained on a dataset of 25 paired patient images and evaluated on a test set of 15 paired images. We performed visual inspections to assess the quality of the generated skVCT images and calculated the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). Dosimetric equivalence was evaluated by comparing the gamma pass rates of treatment plans derived from skVCT and kVCT images. Results showed that skVCT images exhibited significantly higher quality than MVCT images, with PSNR and SSIM values of 31.9 ± 1.1 dB and 94.8% ± 1.3%, respectively, compared to 26.8 ± 1.7 dB and 89.5% ± 1.5% for MVCT-to-kVCT comparisons. Furthermore, treatment plans based on skVCT images achieved excellent gamma pass rates of 99.78 ± 0.14% and 99.82 ± 0.20% for 2 mm/2% and 3 mm/3% criteria, respectively, comparable to those obtained from kVCT-based plans (99.70 ± 0.31% and 99.79 ± 1.32%). This study demonstrates the potential of pix2pix models for generating high-quality skVCT images, which could significantly enhance Adaptive Radiation Therapy (ART).</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1589-1600"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144734067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating frequency and dynamic characteristics of EMG signals as a new inter-muscular coordination feature. 肌电信号的频率与动态特征集成作为一种新的肌间协调特征。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-21 DOI: 10.1007/s13246-025-01620-3
Shaghayegh Hassanzadeh Khanmiri, Peyvand Ghaderyan, Alireza Hashemi Oskouei

The impairment of inter-muscular coordination and changes in frequency components are two major pathological symptoms associated with knee injuries; however, an effective method to simultaneously quantify these changes has yet to be developed. Moreover, there is a need to propose a reliable automated system for identifying knee injuries to eliminate human errors and enhance reliability and consistency. Hence, this study introduces two novel inter-muscular coordination features: Dynamic Time Warping (DTW) and Dynamic Frequency Warping (DFW), which integrate time and frequency characteristics with a dynamic matching procedure. The support vector machine classifier and two types of dynamic neural network classifiers have also been used to evaluate the effectiveness of the proposed features. The proposed system has been tested using a public dataset that includes five channels of electromyogram (EMG) signals from 33 uninjured subjects and 28 individuals with various types of knee injuries. The experimental results have demonstrated the superiority of DFW and cascade forward neural network, achieving an accuracy rate of 92.03% for detection and 94.42% for categorization of different types of knee injuries. The reliability of the proposed feature has been confirmed in identifying knee injuries using both inter-limb and intra-limb EMG channels. This highlights the potential to offer a trade-off between high detection performance and cost-effective procedures by utilizing fewer channels.

肌间协调功能障碍和频率成分改变是膝关节损伤的两大病理症状;然而,一种同时量化这些变化的有效方法还有待开发。此外,有必要提出一种可靠的自动化系统来识别膝关节损伤,以消除人为错误,提高可靠性和一致性。因此,本研究引入了两种新的肌肉间协调特征:动态时间扭曲(Dynamic Time Warping, DTW)和动态频率扭曲(Dynamic Frequency Warping, DFW),它们将时间和频率特征与动态匹配过程相结合。支持向量机分类器和两种动态神经网络分类器也被用来评估所提出特征的有效性。该系统已经使用公共数据集进行了测试,该数据集包括来自33名未受伤受试者和28名不同类型膝盖损伤个体的5个肌电图(EMG)信号通道。实验结果证明了DFW和级联前向神经网络的优越性,对不同类型膝关节损伤的检测准确率为92.03%,分类准确率为94.42%。所提出的特征的可靠性已被证实在识别膝关节损伤时使用了肢间和肢内肌电图通道。这突出了通过使用更少的通道在高检测性能和成本效益之间提供权衡的潜力。
{"title":"Integrating frequency and dynamic characteristics of EMG signals as a new inter-muscular coordination feature.","authors":"Shaghayegh Hassanzadeh Khanmiri, Peyvand Ghaderyan, Alireza Hashemi Oskouei","doi":"10.1007/s13246-025-01620-3","DOIUrl":"10.1007/s13246-025-01620-3","url":null,"abstract":"<p><p>The impairment of inter-muscular coordination and changes in frequency components are two major pathological symptoms associated with knee injuries; however, an effective method to simultaneously quantify these changes has yet to be developed. Moreover, there is a need to propose a reliable automated system for identifying knee injuries to eliminate human errors and enhance reliability and consistency. Hence, this study introduces two novel inter-muscular coordination features: Dynamic Time Warping (DTW) and Dynamic Frequency Warping (DFW), which integrate time and frequency characteristics with a dynamic matching procedure. The support vector machine classifier and two types of dynamic neural network classifiers have also been used to evaluate the effectiveness of the proposed features. The proposed system has been tested using a public dataset that includes five channels of electromyogram (EMG) signals from 33 uninjured subjects and 28 individuals with various types of knee injuries. The experimental results have demonstrated the superiority of DFW and cascade forward neural network, achieving an accuracy rate of 92.03% for detection and 94.42% for categorization of different types of knee injuries. The reliability of the proposed feature has been confirmed in identifying knee injuries using both inter-limb and intra-limb EMG channels. This highlights the potential to offer a trade-off between high detection performance and cost-effective procedures by utilizing fewer channels.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1775-1789"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144974752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparative study of multi-headed and baseline deep learning models for ADHD classification from EEG signals. 多头与基线深度学习模型在ADHD脑电信号分类中的比较研究。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-26 DOI: 10.1007/s13246-025-01609-y
Lamiaa A Amar, Ahmed M Otifi, Shimaa A Mohamed

The prevalence of Attention-Deficit/Hyperactivity Disorder among children is rising, emphasizing the need for early and accurate diagnostic methods to address associated academic and behavioral challenges. Electroencephalography-based analysis has emerged as a promising noninvasive approach for detecting Attention-Deficit/Hyperactivity Disorder; however, utilizing the full range of electroencephalography channels often results in high computational complexity and an increased risk of model overfitting. This study presents a comparative investigation between a proposed multi-headed deep learning framework and a traditional baseline single-model approach for classifying Attention-Deficit/Hyperactivity Disorder using electroencephalography signals. Electroencephalography data were collected from 79 participants (42 healthy adults and 37 diagnosed with Attention-Deficit/Hyperactivity Disorder) across four cognitive states: resting with eyes open, resting with eyes closed, performing cognitive tasks, and listening to omniarmonic sounds. To reduce complexity, signals from only five strategically selected electroencephalography channels were used. The multi-headed approach employed parallel deep learning branches-comprising combinations of Bidirectional Long Short-Term Memory, Long Short-Term Memory, and Gated Recurrent Unit architectures-to capture inter-channel relationships and extract richer temporal features. Comparative analysis revealed that the combination of Long Short-Term Memory and Bidirectional Long Short-Term Memory within the multi-headed framework achieved the highest classification accuracy of 89.87%, significantly outperforming all baseline configurations. These results demonstrate the effectiveness of integrating multiple deep learning architectures and highlight the potential of multi-headed models for enhancing electroencephalography-based Attention-Deficit/Hyperactivity Disorder diagnosis.

儿童中注意力缺陷/多动障碍的患病率正在上升,强调需要早期和准确的诊断方法来解决相关的学术和行为挑战。基于脑电图的分析已经成为一种很有前途的检测注意缺陷/多动障碍的无创方法;然而,利用全范围的脑电图通道通常会导致高计算复杂性和模型过拟合的风险增加。本研究提出了一个多头深度学习框架和传统的基线单模型方法之间的比较研究,用于使用脑电图信号对注意缺陷/多动障碍进行分类。从79名参与者(42名健康成年人和37名被诊断为注意力缺陷/多动障碍)中收集了四种认知状态的脑电图数据:睁眼休息、闭眼休息、执行认知任务和听全谐波声音。为了减少复杂性,只使用了5个策略性选择的脑电图通道的信号。多头方法采用并行深度学习分支——包括双向长短期记忆、长短期记忆和门控循环单元架构的组合——来捕获通道间关系并提取更丰富的时间特征。对比分析发现,多头框架下长短期记忆和双向长短期记忆组合的分类准确率最高,达到89.87%,显著优于所有基线配置。这些结果证明了整合多个深度学习架构的有效性,并突出了多头模型在增强基于脑电图的注意缺陷/多动障碍诊断方面的潜力。
{"title":"Comparative study of multi-headed and baseline deep learning models for ADHD classification from EEG signals.","authors":"Lamiaa A Amar, Ahmed M Otifi, Shimaa A Mohamed","doi":"10.1007/s13246-025-01609-y","DOIUrl":"10.1007/s13246-025-01609-y","url":null,"abstract":"<p><p>The prevalence of Attention-Deficit/Hyperactivity Disorder among children is rising, emphasizing the need for early and accurate diagnostic methods to address associated academic and behavioral challenges. Electroencephalography-based analysis has emerged as a promising noninvasive approach for detecting Attention-Deficit/Hyperactivity Disorder; however, utilizing the full range of electroencephalography channels often results in high computational complexity and an increased risk of model overfitting. This study presents a comparative investigation between a proposed multi-headed deep learning framework and a traditional baseline single-model approach for classifying Attention-Deficit/Hyperactivity Disorder using electroencephalography signals. Electroencephalography data were collected from 79 participants (42 healthy adults and 37 diagnosed with Attention-Deficit/Hyperactivity Disorder) across four cognitive states: resting with eyes open, resting with eyes closed, performing cognitive tasks, and listening to omniarmonic sounds. To reduce complexity, signals from only five strategically selected electroencephalography channels were used. The multi-headed approach employed parallel deep learning branches-comprising combinations of Bidirectional Long Short-Term Memory, Long Short-Term Memory, and Gated Recurrent Unit architectures-to capture inter-channel relationships and extract richer temporal features. Comparative analysis revealed that the combination of Long Short-Term Memory and Bidirectional Long Short-Term Memory within the multi-headed framework achieved the highest classification accuracy of 89.87%, significantly outperforming all baseline configurations. These results demonstrate the effectiveness of integrating multiple deep learning architectures and highlight the potential of multi-headed models for enhancing electroencephalography-based Attention-Deficit/Hyperactivity Disorder diagnosis.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1657-1665"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144974728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A full-scale attention-augmented CNN-transformer model for segmentation of oropharyngeal mucosa organs-at-risk in radiotherapy. 用于放疗中口咽粘膜危险器官分割的全尺寸注意力增强CNN-transformer模型。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-09-11 DOI: 10.1007/s13246-025-01614-1
Lian He, Jianda Sun, Shanfu Lu, Jingyang Li, Xiaoqing Wang, Ziye Yan, Jian Guan
<p><p>Radiation-induced oropharyngeal mucositis (ROM) is a common and severe side effect of radiotherapy in nasopharyngeal cancer patients, leading to significant clinical complications such as malnutrition, infections, and treatment interruptions. Accurate delineation of the oropharyngeal mucosa (OPM) as an organ-at-risk (OAR) is crucial to minimizing radiation exposure and preventing ROM. This study aims to develop and validate an advanced automatic segmentation model, attention-augmented Swin U-Net transformer (AA-Swin UNETR), for accurate delineation of OPM to improve radiotherapy planning and reduce the incidence of ROM. We proposed a hybrid CNN-transformer model, AA-Swin UNETR, based on the Swin UNETR framework, which integrates hierarchical feature extraction with full-scale attention mechanisms. The model includes a Swin Transformer-based encoder and a CNN-based decoder with residual blocks, connected via a full-scale feature connection scheme. The full-scale attention mechanism enables the model to capture long-range dependencies and multi-level features effectively, enhancing the segmentation accuracy. The model was trained on a dataset of 202 CT scans from Nanfang Hospital, using expert manual delineations as the gold standard. We evaluated the performance of AA-Swin UNETR against state-of-the-art (SOTA) segmentation models, including Swin UNETR, nnUNet, and 3D UX-Net, using geometric and dosimetric evaluation parameters. The geometric metrics include Dice similarity coefficient (DSC), surface DSC (sDSC), volume similarity (VS), Hausdorff distance (HD), precision, and recall. The dosimetric metrics include changes of D<sub>0.1 cc</sub> and D<sub>mean</sub> between results derived from manually delineated OPM and auto-segmentation models. The AA-Swin UNETR model achieved the highest mean DSC of 87.72 ± 1.98%, significantly outperforming Swin UNETR (83.53 ± 2.59%), nnUNet (85.48%± 2.68), and 3D UX-Net (80.04 ± 3.76%). The model also showed superior mean sDSC (98.44 ± 1.08%), mean VS (97.86 ± 1.43%), mean precision (87.60 ± 3.06%) and mean recall (89.22 ± 2.70%), with a competitive mean HD of 9.03 ± 2.79 mm. For dosimetric evaluation, the proposed model generates smallest mean [Formula: see text] (0.46 ± 4.92 cGy) and mean [Formula: see text] (6.26 ± 24.90 cGY) relative to manual delineation compared with other auto-segmentation results (mean [Formula: see text] of Swin UNETR = -0.56 ± 7.28 cGy, nnUNet = 0.99 ± 4.73 cGy, 3D UX-Net = -0.65 ± 8.05 cGy; mean [Formula: see text] of Swin UNETR = 7.46 ± 43.37, nnUNet = 21.76 ± 37.86 and 3D UX-Net = 44.61 ± 62.33). In this paper, we proposed a transformer and CNN hybrid deep-learning based model AA-Swin UNETR for automatic segmentation of OPM as an OAR structure in radiotherapy planning. Evaluations with geometric and dosimetric parameters demonstrated AA-Swin UNETR can generate delineations close to a manual reference, both in terms of geometry and dose-volume metrics. The proposed model out-pe
辐射诱发口咽黏膜炎(ROM)是鼻咽癌放疗患者常见且严重的副作用,可导致严重的临床并发症,如营养不良、感染和治疗中断。准确描绘口咽粘膜(OPM)作为危险器官(OAR)对于减少辐射暴露和预防ROM至关重要。本研究旨在开发和验证一种先进的自动分割模型,即注意力增强Swin U-Net变压器(AA-Swin UNETR),用于准确描绘OPM,以改善放疗计划并降低ROM的发生率。我们提出了一种基于Swin UNETR框架的混合CNN-transformer模型AA-Swin UNETR。它将分层特征提取与全尺度注意机制相结合。该模型包括一个基于Swin变压器的编码器和一个基于cnn的残差块解码器,通过全尺寸特征连接方案连接。全尺度注意机制使模型能够有效地捕捉远程依赖关系和多层次特征,提高了分割精度。该模型在南方医院的202个CT扫描数据集上进行训练,使用专家手动划定作为金标准。我们使用几何和剂量学评估参数,对最先进的(SOTA)分割模型(包括Swin UNETR, nnUNet和3D UX-Net)进行了AA-Swin UNETR的性能评估。几何指标包括Dice similarity coefficient (DSC)、surface DSC (sDSC)、volume similarity (VS)、Hausdorff distance (HD)、precision(精密度)和recall(召回率)。剂量学指标包括人工划定的OPM和自动分割模型得出的结果之间D0.1 cc和Dmean的变化。AA-Swin UNETR模型的平均DSC最高,为87.72±1.98%,显著优于Swin UNETR(83.53±2.59%)、nnUNet(85.48%±2.68)和3D UX-Net(80.04±3.76%)。平均sDSC(98.44±1.08%)、平均VS(97.86±1.43%)、平均精密度(87.60±3.06%)和平均召回率(89.22±2.70%)均优于模型,平均高清(HD)为9.03±2.79 mm。最小剂量测定的评价,该模型生成的意思是[公式:看到文本](0.46±4.92 cGy),意思是[公式:看到文本](6.26±24.90 cGy)相对于手动描述与其他auto-segmentation相比结果(意味着[公式:看到文本]斯温UNETR = -0.56±7.28 cGy nnUNet cGy = 0.99±4.73,3 d UX-Net = -0.65±8.05 cGy;意思是[公式:看到文本]斯温UNETR = 7.46±43.37,nnUNet = 21.76±37.86和3 d UX-Net = 44.61±62.33)。本文提出了一种基于transformer和CNN混合深度学习的模型AA-Swin UNETR,用于OPM的自动分割,作为放疗规划中的桨结构。利用几何和剂量学参数进行的评价表明,在几何和剂量-体积指标方面,AA-Swin UNETR可以产生接近人工参考的圈定。所提出的模型在两个评估指标上都优于现有的SOTA模型,并证明了其准确分割OPM复杂解剖结构的能力,为加强放疗计划提供了可靠的工具。
{"title":"A full-scale attention-augmented CNN-transformer model for segmentation of oropharyngeal mucosa organs-at-risk in radiotherapy.","authors":"Lian He, Jianda Sun, Shanfu Lu, Jingyang Li, Xiaoqing Wang, Ziye Yan, Jian Guan","doi":"10.1007/s13246-025-01614-1","DOIUrl":"10.1007/s13246-025-01614-1","url":null,"abstract":"&lt;p&gt;&lt;p&gt;Radiation-induced oropharyngeal mucositis (ROM) is a common and severe side effect of radiotherapy in nasopharyngeal cancer patients, leading to significant clinical complications such as malnutrition, infections, and treatment interruptions. Accurate delineation of the oropharyngeal mucosa (OPM) as an organ-at-risk (OAR) is crucial to minimizing radiation exposure and preventing ROM. This study aims to develop and validate an advanced automatic segmentation model, attention-augmented Swin U-Net transformer (AA-Swin UNETR), for accurate delineation of OPM to improve radiotherapy planning and reduce the incidence of ROM. We proposed a hybrid CNN-transformer model, AA-Swin UNETR, based on the Swin UNETR framework, which integrates hierarchical feature extraction with full-scale attention mechanisms. The model includes a Swin Transformer-based encoder and a CNN-based decoder with residual blocks, connected via a full-scale feature connection scheme. The full-scale attention mechanism enables the model to capture long-range dependencies and multi-level features effectively, enhancing the segmentation accuracy. The model was trained on a dataset of 202 CT scans from Nanfang Hospital, using expert manual delineations as the gold standard. We evaluated the performance of AA-Swin UNETR against state-of-the-art (SOTA) segmentation models, including Swin UNETR, nnUNet, and 3D UX-Net, using geometric and dosimetric evaluation parameters. The geometric metrics include Dice similarity coefficient (DSC), surface DSC (sDSC), volume similarity (VS), Hausdorff distance (HD), precision, and recall. The dosimetric metrics include changes of D&lt;sub&gt;0.1 cc&lt;/sub&gt; and D&lt;sub&gt;mean&lt;/sub&gt; between results derived from manually delineated OPM and auto-segmentation models. The AA-Swin UNETR model achieved the highest mean DSC of 87.72 ± 1.98%, significantly outperforming Swin UNETR (83.53 ± 2.59%), nnUNet (85.48%± 2.68), and 3D UX-Net (80.04 ± 3.76%). The model also showed superior mean sDSC (98.44 ± 1.08%), mean VS (97.86 ± 1.43%), mean precision (87.60 ± 3.06%) and mean recall (89.22 ± 2.70%), with a competitive mean HD of 9.03 ± 2.79 mm. For dosimetric evaluation, the proposed model generates smallest mean [Formula: see text] (0.46 ± 4.92 cGy) and mean [Formula: see text] (6.26 ± 24.90 cGY) relative to manual delineation compared with other auto-segmentation results (mean [Formula: see text] of Swin UNETR = -0.56 ± 7.28 cGy, nnUNet = 0.99 ± 4.73 cGy, 3D UX-Net = -0.65 ± 8.05 cGy; mean [Formula: see text] of Swin UNETR = 7.46 ± 43.37, nnUNet = 21.76 ± 37.86 and 3D UX-Net = 44.61 ± 62.33). In this paper, we proposed a transformer and CNN hybrid deep-learning based model AA-Swin UNETR for automatic segmentation of OPM as an OAR structure in radiotherapy planning. Evaluations with geometric and dosimetric parameters demonstrated AA-Swin UNETR can generate delineations close to a manual reference, both in terms of geometry and dose-volume metrics. The proposed model out-pe","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1703-1714"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145034335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Impact of differences in computed tomography value-electron density/physical density conversion tables on calculate dose in low-density areas. 低密度地区计算机断层扫描值-电子密度/物理密度转换表差异对计算剂量的影响。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-07-23 DOI: 10.1007/s13246-025-01611-4
Mia Nomura, Shunsuke Goto, Mizuki Yoshioka, Yuiko Kato, Ayaka Tsunoda, Kunio Nishioka, Yoshinori Tanabe

In radiotherapy treatment planning, the extrapolation of computed tomography (CT) values for low-density areas without known materials may differ between CT scanners, resulting in different calculated doses. We evaluated the differences in the percentage depth dose (PDD) calculated using eight CT scanners. Heterogeneous virtual phantoms were created using LN-300 lung and - 900 HU. For the two types of virtual phantoms, the PDD on the central axis was calculated using five energies, two irradiation field sizes, and two calculation algorithms (the anisotropic analytical algorithm and Acuros XB). For the LN-300 lung, the maximum CT value difference between the eight CT scanners was 51 HU for an electron density (ED) of 0.29 and 8.8 HU for an extrapolated ED of 0.05. The LN-300 lung CT values showed little variation in the CT-ED/physical density data among CT scanners. The difference in the point depth for the PDD in the LN-300 lung between the CT scanners was < 0.5% for all energies and calculation algorithms. Using Acuros XB, the PDD at - 900 HU had a maximum difference between facilities of > 5%, and the dose difference corresponding to an LN-300 lung CT value difference of > 20 HU was > 1% at a field size of 2 × 2 cm2. The study findings suggest that the calculated dose of low-density regions without known materials in the CT-ED conversion table introduces a risk of dose differences between facilities because of the calibration of the CT values, even when the same CT-ED phantom radiation treatment planning and treatment devices are used.

在放射治疗计划中,不同的CT扫描仪对未知物质的低密度区域的计算机断层扫描(CT)值的外推可能不同,从而导致不同的计算剂量。我们评估了使用8台CT扫描仪计算的百分比深度剂量(PDD)的差异。采用LN-300肺和- 900 HU制作异质虚拟幻象。对于两种类型的虚拟幻影,采用五种能量、两种辐照场大小和两种计算算法(各向异性解析算法和acros XB)计算中轴上的PDD。对于LN-300肺,8台CT扫描仪在电子密度(ED)为0.29时的最大CT值差为51 HU,而外推ED为0.05时的最大CT值差为8.8 HU。LN-300肺CT值显示不同CT扫描仪的CT- ed /物理密度数据差异不大。LN-300肺部PDD的点深在CT扫描仪之间的差异为5%,在场大小为2 × 2 cm2的情况下,LN-300肺部CT值差bb0 20 HU对应的剂量差为> 1%。研究结果表明,在CT- ed转换表中没有已知材料的低密度区域,即使使用相同的CT- ed虚辐射治疗计划和治疗设备,由于CT值的校准,计算出的剂量在设施之间存在剂量差异的风险。
{"title":"Impact of differences in computed tomography value-electron density/physical density conversion tables on calculate dose in low-density areas.","authors":"Mia Nomura, Shunsuke Goto, Mizuki Yoshioka, Yuiko Kato, Ayaka Tsunoda, Kunio Nishioka, Yoshinori Tanabe","doi":"10.1007/s13246-025-01611-4","DOIUrl":"10.1007/s13246-025-01611-4","url":null,"abstract":"<p><p>In radiotherapy treatment planning, the extrapolation of computed tomography (CT) values for low-density areas without known materials may differ between CT scanners, resulting in different calculated doses. We evaluated the differences in the percentage depth dose (PDD) calculated using eight CT scanners. Heterogeneous virtual phantoms were created using LN-300 lung and - 900 HU. For the two types of virtual phantoms, the PDD on the central axis was calculated using five energies, two irradiation field sizes, and two calculation algorithms (the anisotropic analytical algorithm and Acuros XB). For the LN-300 lung, the maximum CT value difference between the eight CT scanners was 51 HU for an electron density (ED) of 0.29 and 8.8 HU for an extrapolated ED of 0.05. The LN-300 lung CT values showed little variation in the CT-ED/physical density data among CT scanners. The difference in the point depth for the PDD in the LN-300 lung between the CT scanners was < 0.5% for all energies and calculation algorithms. Using Acuros XB, the PDD at - 900 HU had a maximum difference between facilities of > 5%, and the dose difference corresponding to an LN-300 lung CT value difference of > 20 HU was > 1% at a field size of 2 × 2 cm<sup>2</sup>. The study findings suggest that the calculated dose of low-density regions without known materials in the CT-ED conversion table introduces a risk of dose differences between facilities because of the calibration of the CT values, even when the same CT-ED phantom radiation treatment planning and treatment devices are used.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1679-1689"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12738602/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144692087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A review of image processing and analysis of computed tomography images using deep learning methods. 使用深度学习方法的图像处理和计算机断层扫描图像分析综述。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-09-03 DOI: 10.1007/s13246-025-01635-w
Darcie Anderson, Prabhakar Ramachandran, Jamie Trapp, Andrew Fielding

The use of machine learning has seen extraordinary growth since the development of deep learning techniques, notably the deep artificial neural network. Deep learning methodology excels in addressing complicated problems such as image classification, object detection, and natural language processing. A key feature of these networks is the capability to extract useful patterns from vast quantities of complex data, including images. As many branches of healthcare revolves around the generation, processing, and analysis of images, these techniques have become increasingly commonplace. This is especially true for radiotherapy, which relies on the use of anatomical and functional images from a range of imaging modalities, such as Computed Tomography (CT). The aim of this review is to provide an understanding of deep learning methodologies, including neural network types and structure, as well as linking these general concepts to medical CT image processing for radiotherapy. Specifically, it focusses on the stages of enhancement and analysis, incorporating image denoising, super-resolution, generation, registration, and segmentation, supported by examples of recent literature.

自从深度学习技术,特别是深度人工神经网络的发展以来,机器学习的使用已经有了惊人的增长。深度学习方法擅长解决复杂的问题,如图像分类、目标检测和自然语言处理。这些网络的一个关键特征是从包括图像在内的大量复杂数据中提取有用模式的能力。由于医疗保健的许多分支都围绕图像的生成、处理和分析展开,这些技术变得越来越普遍。放射治疗尤其如此,它依赖于使用来自一系列成像方式的解剖和功能图像,例如计算机断层扫描(CT)。这篇综述的目的是提供对深度学习方法的理解,包括神经网络的类型和结构,以及将这些一般概念与放射治疗的医学CT图像处理联系起来。具体来说,它侧重于增强和分析的阶段,包括图像去噪、超分辨率、生成、配准和分割,并以最近的文献为例提供支持。
{"title":"A review of image processing and analysis of computed tomography images using deep learning methods.","authors":"Darcie Anderson, Prabhakar Ramachandran, Jamie Trapp, Andrew Fielding","doi":"10.1007/s13246-025-01635-w","DOIUrl":"10.1007/s13246-025-01635-w","url":null,"abstract":"<p><p>The use of machine learning has seen extraordinary growth since the development of deep learning techniques, notably the deep artificial neural network. Deep learning methodology excels in addressing complicated problems such as image classification, object detection, and natural language processing. A key feature of these networks is the capability to extract useful patterns from vast quantities of complex data, including images. As many branches of healthcare revolves around the generation, processing, and analysis of images, these techniques have become increasingly commonplace. This is especially true for radiotherapy, which relies on the use of anatomical and functional images from a range of imaging modalities, such as Computed Tomography (CT). The aim of this review is to provide an understanding of deep learning methodologies, including neural network types and structure, as well as linking these general concepts to medical CT image processing for radiotherapy. Specifically, it focusses on the stages of enhancement and analysis, incorporating image denoising, super-resolution, generation, registration, and segmentation, supported by examples of recent literature.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1491-1523"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12738611/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144993950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable hierarchical machine-learning approaches for multimodal prediction of conversion from mild cognitive impairment to Alzheimer's disease. 从轻度认知障碍到阿尔茨海默病转换的多模态预测的可解释的分层机器学习方法。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-11 DOI: 10.1007/s13246-025-01618-x
Soheil Zarei, Mohsen Saffar, Reza Shalbaf, Peyman Hassani Abharian, Ahmad Shalbaf

Alzheimer's disease (AD) is a neurodegenerative disorder that challenges early diagnosis and intervention, yet the black-box nature of many predictive models limits clinical adoption. In this study, we developed an advanced machine learning (ML) framework that integrates hierarchical feature selection with multiple classifiers to predict progression from mild cognitive impairment (MCI) to AD. Using baseline data from 580 participants in the Alzheimer's Disease Neuroimaging Initiative (ADNI), categorized into stable MCI (sMCI) and progressive MCI (pMCI) subgroups, we analyzed features both individually and across seven key groups. The neuropsychological test group exhibited the highest predictive power, with several of the top individual predictors drawn from this domain. Hierarchical feature selection combining initial statistical filtering and machine learning based refinement, narrowed the feature set to the eight most informative variables. To demystify model decisions, we applied SHAP-based (SHapley Additive exPlanations) explainability analysis, quantifying each feature's contribution to conversion risk. The explainable random forest classifier, optimized on these selected features, achieved 83.79% accuracy (84.93% sensitivity, 83.32% specificity), outperforming other methods and revealing hippocampal volume, delayed memory recall (LDELTOTAL), and Functional Activities Questionnaire (FAQ) scores as the top drivers of conversion. These results underscore the effectiveness of combining diverse data sources with advanced ML models, and demonstrate that transparent, SHAP-driven insights align with known AD biomarkers, transforming our model from a predictive black box into a clinically actionable tool for early diagnosis and patient stratification.

阿尔茨海默病(AD)是一种神经退行性疾病,对早期诊断和干预具有挑战性,但许多预测模型的黑箱性质限制了临床应用。在这项研究中,我们开发了一个先进的机器学习(ML)框架,该框架将分层特征选择与多个分类器集成在一起,以预测从轻度认知障碍(MCI)到AD的进展。使用来自580名阿尔茨海默病神经影像学倡议(ADNI)参与者的基线数据,将其分为稳定型MCI (sMCI)和进行性MCI (pMCI)亚组,我们分析了个体和七个关键组的特征。神经心理测试组表现出最高的预测能力,有几个最重要的个体预测来自这个领域。分层特征选择结合初始统计过滤和基于机器学习的细化,将特征集缩小到8个信息量最大的变量。为了揭开模型决策的神秘面纱,我们应用了基于shap (SHapley可加解释)的可解释性分析,量化每个特征对转换风险的贡献。基于这些特征进行优化的可解释随机森林分类器准确率达到83.79%(灵敏度84.93%,特异性83.32%),优于其他方法,并显示海马体积、延迟记忆回忆(LDELTOTAL)和功能活动问卷(FAQ)得分是转换的主要驱动因素。这些结果强调了将不同数据源与先进的ML模型相结合的有效性,并证明了透明的、shap驱动的见解与已知的AD生物标志物相一致,将我们的模型从预测黑箱转变为临床可操作的工具,用于早期诊断和患者分层。
{"title":"Explainable hierarchical machine-learning approaches for multimodal prediction of conversion from mild cognitive impairment to Alzheimer's disease.","authors":"Soheil Zarei, Mohsen Saffar, Reza Shalbaf, Peyman Hassani Abharian, Ahmad Shalbaf","doi":"10.1007/s13246-025-01618-x","DOIUrl":"10.1007/s13246-025-01618-x","url":null,"abstract":"<p><p>Alzheimer's disease (AD) is a neurodegenerative disorder that challenges early diagnosis and intervention, yet the black-box nature of many predictive models limits clinical adoption. In this study, we developed an advanced machine learning (ML) framework that integrates hierarchical feature selection with multiple classifiers to predict progression from mild cognitive impairment (MCI) to AD. Using baseline data from 580 participants in the Alzheimer's Disease Neuroimaging Initiative (ADNI), categorized into stable MCI (sMCI) and progressive MCI (pMCI) subgroups, we analyzed features both individually and across seven key groups. The neuropsychological test group exhibited the highest predictive power, with several of the top individual predictors drawn from this domain. Hierarchical feature selection combining initial statistical filtering and machine learning based refinement, narrowed the feature set to the eight most informative variables. To demystify model decisions, we applied SHAP-based (SHapley Additive exPlanations) explainability analysis, quantifying each feature's contribution to conversion risk. The explainable random forest classifier, optimized on these selected features, achieved 83.79% accuracy (84.93% sensitivity, 83.32% specificity), outperforming other methods and revealing hippocampal volume, delayed memory recall (LDELTOTAL), and Functional Activities Questionnaire (FAQ) scores as the top drivers of conversion. These results underscore the effectiveness of combining diverse data sources with advanced ML models, and demonstrate that transparent, SHAP-driven insights align with known AD biomarkers, transforming our model from a predictive black box into a clinically actionable tool for early diagnosis and patient stratification.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1741-1759"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144817988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prop scan versus roll scan: selection for cranial three-dimensional rotational angiography using in-house phantom and Figure of Merit as parameter. 支柱扫描与滚动扫描:颅内三维旋转血管造影的选择,使用内部幻影和优点图作为参数。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-09-10 DOI: 10.1007/s13246-025-01632-z
Ika Hariyati, Ani Sulistyani, Matthew Gregorius, Harimulti Aribowo, Ungguh Prawoto, Defri Dwi Yana, Thariqah Salamah, Lukmanda Evan Lubis, Djarwani Soeharso Soejoko

This study introduces a novel optimization framework for cranial three-dimensional rotational angiography (3DRA), combining the development of a brain equivalent in-house phantom with Figure of Merit (FOM) a quantitative evaluation method. The technical contribution involves the development of an in-house phantom constructed using iodine-infused epoxy and lycal resins, validated against clinical Hounsfield Units (HU). A customized head phantom was developed to simulate brain tissue and cranial vasculature for 3DRA optimization. The phantom was constructed using epoxy resin with 0.15-0.2% iodine to replicate brain tissue and lycal resin with iodine concentrations ranging from 0.65 to 0.7% to simulate blood vessels of varying diameters. The phantom materials validation was performed by comparing their HU values to clinical reference HU values from brain tissue and cranial vessels, ensuring accurate tissue simulation. The validated phantom was used to acquire images using cranial 3DRA protocols, specifically Prop-Scan and Roll-Scan. Image quality was assessed using Signal-Difference-to-Noise Ratio (SDNR), Dose-Area Product (DAP), and Modulation Transfer Function (MTF). Imaging efficiency was quantified using the Figure of Merit (FOM), calculated as SDNR2/DAP, to objectively compare the performance of two cranial 3DRA protocols. The task-based optimization showed that Roll-Scan consistently outperformed Prop-Scan across all vessel sizes and regions. Roll-Scan yields FOM values ranging from 183 to 337, while Prop-Scan FOM values ranged from 96 to 189. Additionally, Roll-Scan (0.27 lp/pixel) delivered better spatial resolution, as indicated by higher MTF 10% value than Prop-Scan (0.23 lp/pixel). Most notably, Roll-Scan consistently detecting 2 mm vessel structures among all regions of the phantom. This capability is clinically important in cerebral angiography, which is accurate visualization of small vessels, i.e. the Anterior Cerebral Artery (ACA), Posterior Cerebral Artery (PCA), and Middle Cerebral Artery (MCA). These findings highlight Roll-Scan as the superior protocol for brain interventional imaging, underscoring the significance of FOM as a comprehensive parameter for optimizing imaging protocols in clinical practice. The experimental results support the use of the Roll-Scan protocol as the preferred acquisition method for cerebral angiography in clinical practice. The analysis using FOM provides substantial and quantifiable evidence in determining the acquisition methods. Furthermore, the customized in-house phantom is recommended as a candidate to optimization tools for clinical medical physicists.

本研究介绍了一种新的颅三维旋转血管造影(3DRA)优化框架,将脑等效内部幻像的开发与优点图(FOM)的定量评估方法相结合。技术贡献包括使用碘注入环氧树脂和local树脂构建内部模体,并通过临床Hounsfield单位(HU)进行验证。开发了一个定制的头部幻影来模拟脑组织和颅血管系统,以进行3DRA优化。用含碘量为0.15-0.2%的环氧树脂来模拟脑组织,用含碘量为0.65 - 0.7%的局部树脂来模拟不同直径的血管。通过将虚拟材料的HU值与临床参考脑组织和颅血管的HU值进行比较,以确保准确的组织模拟。通过颅3DRA协议,特别是Prop-Scan和Roll-Scan,使用验证过的假体获取图像。使用信噪比(SDNR)、剂量面积积(DAP)和调制传递函数(MTF)评估图像质量。成像效率采用优点图(FOM)量化,计算为SDNR2/DAP,客观比较两种颅3DRA方案的性能。基于任务的优化表明,在所有船舶尺寸和区域,Roll-Scan的性能始终优于Prop-Scan。Roll-Scan的FOM值范围从183到337,而Prop-Scan的FOM值范围从96到189。此外,Roll-Scan (0.27 lp/像素)提供了更好的空间分辨率,MTF值比Prop-Scan (0.23 lp/像素)高10%。最值得注意的是,Roll-Scan在幻体的所有区域中都能持续检测到2mm的血管结构。这种能力在脑血管造影中具有重要的临床意义,它可以准确地显示小血管,即大脑前动脉(ACA)、大脑后动脉(PCA)和大脑中动脉(MCA)。这些发现强调了Roll-Scan作为脑介入成像的优越方案,强调了FOM作为优化临床实践中成像方案的综合参数的重要性。实验结果支持在临床实践中使用Roll-Scan协议作为脑血管造影的首选采集方法。使用FOM的分析为确定获取方法提供了大量和可量化的证据。此外,定制的内部幻影被推荐为临床医学物理学家优化工具的候选。
{"title":"Prop scan versus roll scan: selection for cranial three-dimensional rotational angiography using in-house phantom and Figure of Merit as parameter.","authors":"Ika Hariyati, Ani Sulistyani, Matthew Gregorius, Harimulti Aribowo, Ungguh Prawoto, Defri Dwi Yana, Thariqah Salamah, Lukmanda Evan Lubis, Djarwani Soeharso Soejoko","doi":"10.1007/s13246-025-01632-z","DOIUrl":"10.1007/s13246-025-01632-z","url":null,"abstract":"<p><p>This study introduces a novel optimization framework for cranial three-dimensional rotational angiography (3DRA), combining the development of a brain equivalent in-house phantom with Figure of Merit (FOM) a quantitative evaluation method. The technical contribution involves the development of an in-house phantom constructed using iodine-infused epoxy and lycal resins, validated against clinical Hounsfield Units (HU). A customized head phantom was developed to simulate brain tissue and cranial vasculature for 3DRA optimization. The phantom was constructed using epoxy resin with 0.15-0.2% iodine to replicate brain tissue and lycal resin with iodine concentrations ranging from 0.65 to 0.7% to simulate blood vessels of varying diameters. The phantom materials validation was performed by comparing their HU values to clinical reference HU values from brain tissue and cranial vessels, ensuring accurate tissue simulation. The validated phantom was used to acquire images using cranial 3DRA protocols, specifically Prop-Scan and Roll-Scan. Image quality was assessed using Signal-Difference-to-Noise Ratio (SDNR), Dose-Area Product (DAP), and Modulation Transfer Function (MTF). Imaging efficiency was quantified using the Figure of Merit (FOM), calculated as SDNR<sup>2</sup>/DAP, to objectively compare the performance of two cranial 3DRA protocols. The task-based optimization showed that Roll-Scan consistently outperformed Prop-Scan across all vessel sizes and regions. Roll-Scan yields FOM values ranging from 183 to 337, while Prop-Scan FOM values ranged from 96 to 189. Additionally, Roll-Scan (0.27 lp/pixel) delivered better spatial resolution, as indicated by higher MTF 10% value than Prop-Scan (0.23 lp/pixel). Most notably, Roll-Scan consistently detecting 2 mm vessel structures among all regions of the phantom. This capability is clinically important in cerebral angiography, which is accurate visualization of small vessels, i.e. the Anterior Cerebral Artery (ACA), Posterior Cerebral Artery (PCA), and Middle Cerebral Artery (MCA). These findings highlight Roll-Scan as the superior protocol for brain interventional imaging, underscoring the significance of FOM as a comprehensive parameter for optimizing imaging protocols in clinical practice. The experimental results support the use of the Roll-Scan protocol as the preferred acquisition method for cerebral angiography in clinical practice. The analysis using FOM provides substantial and quantifiable evidence in determining the acquisition methods. Furthermore, the customized in-house phantom is recommended as a candidate to optimization tools for clinical medical physicists.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"1935-1947"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145030952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced detection of ovarian cancer using AI-optimized 3D CNNs for PET/CT scan analysis. 利用ai优化的3D cnn增强卵巢癌的PET/CT扫描分析。
IF 2 4区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-04 DOI: 10.1007/s13246-025-01615-0
Mohammad Hossein Sadeghi, Sedigheh Sina, Reza Faghihi, Mehrosadat Alavi, Francesco Giammarile, Hamid Omidi

This study investigates how deep learning (DL) can enhance ovarian cancer diagnosis and staging using large imaging datasets. Specifically, we compare six conventional convolutional neural network (CNN) architectures-ResNet, DenseNet, GoogLeNet, U-Net, VGG, and AlexNet-with OCDA-Net, an enhanced model designed for [18F]FDG PET image analysis. The OCDA-Net, an advancement on the ResNet architecture, was thoroughly compared using randomly split datasets of training (80%), validation (10%), and test (10%) images. Trained over 100 epochs, OCDA-Net achieved superior diagnostic classification with an accuracy of 92%, and staging results of 94%, supported by robust precision, recall, and F-measure metrics. Grad-CAM ++ heat-maps confirmed that the network attends to hyper-metabolic lesions, supporting clinical interpretability. Our findings show that OCDA-Net outperforms existing CNN models and has strong potential to transform ovarian cancer diagnosis and staging. The study suggests that implementing these DL models in clinical practice could ultimately improve patient prognoses. Future research should expand datasets, enhance model interpretability, and validate these models in clinical settings.

本研究探讨了深度学习(DL)如何利用大型成像数据集增强卵巢癌的诊断和分期。具体来说,我们将六种传统的卷积神经网络(CNN)架构——resnet、DenseNet、GoogLeNet、U-Net、VGG和alexnet与OCDA-Net进行了比较,OCDA-Net是一种为[18F]FDG PET图像分析设计的增强模型。OCDA-Net是ResNet架构的一个进步,使用随机分割的训练(80%)、验证(10%)和测试(10%)图像数据集进行了彻底的比较。OCDA-Net训练了超过100个epoch,在强大的精度、召回率和F-measure指标的支持下,OCDA-Net的诊断分类准确率达到92%,分期结果达到94%。Grad-CAM ++热图证实该网络关注高代谢病变,支持临床可解释性。我们的研究结果表明,OCDA-Net优于现有的CNN模型,具有很大的潜力来改变卵巢癌的诊断和分期。该研究表明,在临床实践中实施这些DL模型最终可以改善患者预后。未来的研究应该扩展数据集,增强模型的可解释性,并在临床环境中验证这些模型。
{"title":"Enhanced detection of ovarian cancer using AI-optimized 3D CNNs for PET/CT scan analysis.","authors":"Mohammad Hossein Sadeghi, Sedigheh Sina, Reza Faghihi, Mehrosadat Alavi, Francesco Giammarile, Hamid Omidi","doi":"10.1007/s13246-025-01615-0","DOIUrl":"10.1007/s13246-025-01615-0","url":null,"abstract":"<p><p>This study investigates how deep learning (DL) can enhance ovarian cancer diagnosis and staging using large imaging datasets. Specifically, we compare six conventional convolutional neural network (CNN) architectures-ResNet, DenseNet, GoogLeNet, U-Net, VGG, and AlexNet-with OCDA-Net, an enhanced model designed for [<sup>18</sup>F]FDG PET image analysis. The OCDA-Net, an advancement on the ResNet architecture, was thoroughly compared using randomly split datasets of training (80%), validation (10%), and test (10%) images. Trained over 100 epochs, OCDA-Net achieved superior diagnostic classification with an accuracy of 92%, and staging results of 94%, supported by robust precision, recall, and F-measure metrics. Grad-CAM ++ heat-maps confirmed that the network attends to hyper-metabolic lesions, supporting clinical interpretability. Our findings show that OCDA-Net outperforms existing CNN models and has strong potential to transform ovarian cancer diagnosis and staging. The study suggests that implementing these DL models in clinical practice could ultimately improve patient prognoses. Future research should expand datasets, enhance model interpretability, and validate these models in clinical settings.</p>","PeriodicalId":48490,"journal":{"name":"Physical and Engineering Sciences in Medicine","volume":" ","pages":"2087-2102"},"PeriodicalIF":2.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144785686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Physical and Engineering Sciences in Medicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1