首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
Detecting thyroid nodules along with surrounding tissues and tracking nodules using motion prior in ultrasound videos 检测甲状腺结节和周围组织,并利用超声视频中的运动先验追踪结节。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-10-01 DOI: 10.1016/j.compmedimag.2024.102439
Song Gao , Yueyang Li , Haichi Luo
Ultrasound examination plays a crucial role in the clinical diagnosis of thyroid nodules. Although deep learning technology has been applied to thyroid nodule examinations, the existing methods all overlook the prior knowledge of nodules moving along a straight line in the video. We propose a new detection model, DiffusionVID-Line, and design a novel tracking algorithm, ByteTrack-Line, both of which fully leverage the prior knowledge of linear motion of nodules in thyroid ultrasound videos. Among them, ByteTrack-Line groups detected nodules, further reducing the workload of doctors and significantly improving their diagnostic speed and accuracy. In DiffusionVID-Line, we propose two new modules: Freq-FPN and Attn-Line. Freq-FPN module is used to extract frequency features, taking advantage of these features to reduce the impact of image blur in ultrasound videos. Based on the standard practice of segmented scanning by doctors, Attn-Line module enhances the attention on targets moving along a straight line, thus improving the accuracy of detection. In ByteTrack-Line, considering the characteristic of linear motion of nodules, we propose the Match-Line association module, which reduces the number of nodule ID switches. In the testing of the detection and tracking datasets, DiffusionVID-Line achieved a mean Average Precision (mAP50) of 74.2 for multiple tissues and 85.6 for nodules, while ByteTrack-Line achieved a Multiple Object Tracking Accuracy (MOTA) of 83.4. Both nodule detection and tracking have achieved state-of-the-art performance.
超声波检查在甲状腺结节的临床诊断中起着至关重要的作用。虽然深度学习技术已被应用于甲状腺结节检查,但现有方法都忽略了结节在视频中沿直线运动的先验知识。我们提出了一种新的检测模型--DiffusionVID-Line,并设计了一种新的跟踪算法--ByteTrack-Line,这两种方法都充分利用了甲状腺超声视频中结节直线运动的先验知识。其中,ByteTrack-Line 对检测到的结节进行分组,进一步减轻了医生的工作量,显著提高了诊断速度和准确性。在 DiffusionVID-Line 中,我们提出了两个新模块:Freq-FPN 和 Attn-Line。Freq-FPN 模块用于提取频率特性,利用这些特性降低超声视频中图像模糊的影响。Attn-Line 模块基于医生分段扫描的标准做法,加强了对沿直线运动目标的关注,从而提高了检测的准确性。在 ByteTrack-Line 中,考虑到结节直线运动的特点,我们提出了 Match-Line 关联模块,减少了结节 ID 的切换次数。在检测和跟踪数据集测试中,DiffusionVID-Line 的多组织平均精度 (mAP50) 为 74.2,结节平均精度 (mAP50) 为 85.6,而 ByteTrack-Line 的多目标跟踪精度 (MOTA) 为 83.4。结节检测和跟踪都达到了最先进的性能。
{"title":"Detecting thyroid nodules along with surrounding tissues and tracking nodules using motion prior in ultrasound videos","authors":"Song Gao ,&nbsp;Yueyang Li ,&nbsp;Haichi Luo","doi":"10.1016/j.compmedimag.2024.102439","DOIUrl":"10.1016/j.compmedimag.2024.102439","url":null,"abstract":"<div><div>Ultrasound examination plays a crucial role in the clinical diagnosis of thyroid nodules. Although deep learning technology has been applied to thyroid nodule examinations, the existing methods all overlook the prior knowledge of nodules moving along a straight line in the video. We propose a new detection model, DiffusionVID-Line, and design a novel tracking algorithm, ByteTrack-Line, both of which fully leverage the prior knowledge of linear motion of nodules in thyroid ultrasound videos. Among them, ByteTrack-Line groups detected nodules, further reducing the workload of doctors and significantly improving their diagnostic speed and accuracy. In DiffusionVID-Line, we propose two new modules: Freq-FPN and Attn-Line. Freq-FPN module is used to extract frequency features, taking advantage of these features to reduce the impact of image blur in ultrasound videos. Based on the standard practice of segmented scanning by doctors, Attn-Line module enhances the attention on targets moving along a straight line, thus improving the accuracy of detection. In ByteTrack-Line, considering the characteristic of linear motion of nodules, we propose the Match-Line association module, which reduces the number of nodule ID switches. In the testing of the detection and tracking datasets, DiffusionVID-Line achieved a mean Average Precision (mAP50) of 74.2 for multiple tissues and 85.6 for nodules, while ByteTrack-Line achieved a Multiple Object Tracking Accuracy (MOTA) of 83.4. Both nodule detection and tracking have achieved state-of-the-art performance.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102439"},"PeriodicalIF":5.4,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142367305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RibFractureSys: A gem in the face of acute rib fracture diagnoses RibFractureSys:诊断急性肋骨骨折的瑰宝。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-10-01 DOI: 10.1016/j.compmedimag.2024.102429
Riel Castro-Zunti , Kaike Li , Aleti Vardhan , Younhee Choi , Gong Yong Jin , Seok-bum Ko
Rib fracture patients, common in trauma wards, have different mortality rates and comorbidities depending on how many and which ribs are fractured. This knowledge is therefore paramount to make accurate prognoses and prioritize patient care. However, tracking 24 ribs over upwards 200+ frames in a patient’s scan is time-consuming and error-prone for radiologists, especially depending on their experience.
We propose an automated, modular, three-stage solution to assist radiologists. Using 9 fully annotated patient scans, we trained a multi-class U-Net to segment rib lesions and common anatomical clutter. To recognize rib fractures and mitigate false positives, we fine-tuned a ResNet-based model using 5698 false positives, 2037 acute fractures, 4786 healed fractures, and 14,904 unfractured rib lesions. Using almost 200 patient cases, we developed a highly task-customized multi-object rib lesion tracker to determine which lesions in a frame belong to which of the 12 ribs on either side; bounding box intersection over union- and centroid-based tracking, a line-crossing methodology, and various heuristics were utilized. Our system accepts an axial CT scan and processes, labels, and color-codes the scan.
Over an internal validation dataset of 1000 acute rib fracture and 1000 control patients, our system, assessed by a 3-year radiologist resident, achieved 96.1% and 97.3% correct fracture classification accuracy for rib fracture and control patients, respectively. However, 18.0% and 20.8% of these patients, respectively, had incorrect rib labeling. Percentages remained consistent across sex and age demographics. Labeling issues include anatomical clutter being mislabeled as ribs and ribs going unlabeled.
肋骨骨折患者是创伤病房的常见病,其死亡率和合并症因肋骨骨折的数量和部位而异。因此,这些知识对于做出准确预后和优先护理病人至关重要。然而,对放射科医生来说,在病人扫描的 200 多帧图像中追踪 24 根肋骨既费时又容易出错,尤其是根据他们的经验。我们提出了一种自动化、模块化、三阶段的解决方案来帮助放射科医生。我们使用 9 张完整标注的患者扫描图像,训练了一个多类 U-Net 来分割肋骨病变和常见的解剖杂波。为了识别肋骨骨折并减少误报,我们利用 5698 例误报、2037 例急性骨折、4786 例愈合骨折和 14904 例未骨折肋骨病变对基于 ResNet 的模型进行了微调。利用近 200 例患者,我们开发了高度任务定制化的多对象肋骨病变追踪器,以确定一帧图像中的哪些病变属于两侧 12 根肋骨中的哪一根;我们采用了基于联合和中心点追踪的边界框交叉、线交叉方法和各种启发式方法。我们的系统接受轴向 CT 扫描,并对扫描结果进行处理、标记和颜色编码。在一个包含 1000 名急性肋骨骨折患者和 1000 名对照组患者的内部验证数据集上,我们的系统在一名 3 年放射科住院医生的评估下,对肋骨骨折患者和对照组患者的骨折分类正确率分别达到了 96.1% 和 97.3%。但是,在这些患者中,分别有 18.0% 和 20.8% 的肋骨标记不正确。不同性别和年龄人口的百分比保持一致。标注问题包括解剖杂乱的肋骨被误标为肋骨和肋骨未被标注。
{"title":"RibFractureSys: A gem in the face of acute rib fracture diagnoses","authors":"Riel Castro-Zunti ,&nbsp;Kaike Li ,&nbsp;Aleti Vardhan ,&nbsp;Younhee Choi ,&nbsp;Gong Yong Jin ,&nbsp;Seok-bum Ko","doi":"10.1016/j.compmedimag.2024.102429","DOIUrl":"10.1016/j.compmedimag.2024.102429","url":null,"abstract":"<div><div>Rib fracture patients, common in trauma wards, have different mortality rates and comorbidities depending on how many and which ribs are fractured. This knowledge is therefore paramount to make accurate prognoses and prioritize patient care. However, tracking 24 ribs over upwards 200+ frames in a patient’s scan is time-consuming and error-prone for radiologists, especially depending on their experience.</div><div>We propose an automated, modular, three-stage solution to assist radiologists. Using 9 fully annotated patient scans, we trained a multi-class U-Net to segment rib lesions and common anatomical clutter. To recognize rib fractures and mitigate false positives, we fine-tuned a ResNet-based model using 5698 false positives, 2037 acute fractures, 4786 healed fractures, and 14,904 unfractured rib lesions. Using almost 200 patient cases, we developed a highly task-customized multi-object rib lesion tracker to determine which lesions in a frame belong to which of the 12 ribs on either side; bounding box intersection over union- and centroid-based tracking, a line-crossing methodology, and various heuristics were utilized. Our system accepts an axial CT scan and processes, labels, and color-codes the scan.</div><div>Over an internal validation dataset of 1000 acute rib fracture and 1000 control patients, our system, assessed by a 3-year radiologist resident, achieved 96.1% and 97.3% correct fracture classification accuracy for rib fracture and control patients, respectively. However, 18.0% and 20.8% of these patients, respectively, had incorrect rib labeling. Percentages remained consistent across sex and age demographics. Labeling issues include anatomical clutter being mislabeled as ribs and ribs going unlabeled.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102429"},"PeriodicalIF":5.4,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142367306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning-based diagnostics of capsular invasion in thyroid nodules with wide-field second harmonic generation microscopy 基于机器学习的宽视场二次谐波发生显微镜甲状腺结节囊性侵袭诊断。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-10-01 DOI: 10.1016/j.compmedimag.2024.102440
Yaraslau Padrez , Lena Golubewa , Igor Timoshchenko , Adrian Enache , Lucian G. Eftimie , Radu Hristu , Danielis Rutkauskas
Papillary thyroid carcinoma (PTC) is one of the most common, well-differentiated carcinomas of the thyroid gland. PTC nodules are often surrounded by a collagen capsule that prevents the spread of cancer cells. However, as the malignant tumor progresses, the integrity of this protective barrier is compromised, and cancer cells invade the surroundings. The detection of capsular invasion is, therefore, crucial for the diagnosis and the choice of treatment and the development of new approaches aimed at the increase of diagnostic performance are of great importance. In the present study, we exploited the wide-field second harmonic generation (SHG) microscopy in combination with texture analysis and unsupervised machine learning (ML) to explore the possibility of quantitative characterization of collagen structure in the capsule and designation of different capsule areas as either intact, disrupted by invasion, or apt to invasion. Two-step k-means clustering showed that the collagen capsules in all analyzed tissue sections were highly heterogeneous and exhibited distinct segments described by characteristic ML parameter sets. The latter allowed a structural interpretation of the collagen fibers at the sites of overt invasion as fragmented and curled fibers with rarely formed distributed networks. Clustering analysis also distinguished areas in the PTC capsule that were not categorized as invasion sites by the initial histopathological analysis but could be recognized as prospective micro-invasions after additional inspection. The characteristic features of suspicious and invasive sites identified by the proposed unsupervised ML approach can become a reliable complement to existing methods for diagnosing encapsulated PTC, increase the reliability of diagnosis, simplify decision making, and prevent human-related diagnostic errors. In addition, the proposed automated ML-based selection of collagen capsule images and exclusion of non-informative regions can greatly accelerate and simplify the development of reliable methods for fully automated ML diagnosis that can be integrated into clinical practice.
甲状腺乳头状癌(PTC)是甲状腺中最常见的一种分化良好的癌症。PTC结节周围通常有一层胶原蛋白囊,可以防止癌细胞扩散。然而,随着恶性肿瘤的发展,这一保护屏障的完整性受到破坏,癌细胞就会侵入周围环境。因此,囊肿侵犯的检测对于诊断和选择治疗方法至关重要,而开发新方法以提高诊断性能则具有重要意义。在本研究中,我们利用宽视场二次谐波发生(SHG)显微镜,结合纹理分析和无监督机器学习(ML),探索了胶囊中胶原结构的定量表征以及将不同胶囊区域指定为完整、受侵袭破坏或易受侵袭区域的可能性。两步 K-均值聚类显示,所有分析组织切片中的胶原囊高度异质,并呈现出由特征性 ML 参数集描述的不同区段。后者允许对明显入侵部位的胶原纤维进行结构性解释,将其视为碎裂和卷曲的纤维,很少形成分布式网络。聚类分析还能区分 PTC 胶囊中初步组织病理学分析未将其归类为侵袭部位,但经进一步检查后可确认为前瞻性微侵袭的区域。所提出的无监督 ML 方法所识别的可疑部位和浸润部位的特征可以成为现有诊断包裹型 PTC 方法的可靠补充,提高诊断的可靠性,简化决策,并防止人为诊断错误。此外,所提出的基于 ML 的胶原囊图像自动选择和排除非信息区域的方法可以大大加快和简化可靠的全自动 ML 诊断方法的开发,并将其融入临床实践中。
{"title":"Machine learning-based diagnostics of capsular invasion in thyroid nodules with wide-field second harmonic generation microscopy","authors":"Yaraslau Padrez ,&nbsp;Lena Golubewa ,&nbsp;Igor Timoshchenko ,&nbsp;Adrian Enache ,&nbsp;Lucian G. Eftimie ,&nbsp;Radu Hristu ,&nbsp;Danielis Rutkauskas","doi":"10.1016/j.compmedimag.2024.102440","DOIUrl":"10.1016/j.compmedimag.2024.102440","url":null,"abstract":"<div><div>Papillary thyroid carcinoma (PTC) is one of the most common, well-differentiated carcinomas of the thyroid gland. PTC nodules are often surrounded by a collagen capsule that prevents the spread of cancer cells. However, as the malignant tumor progresses, the integrity of this protective barrier is compromised, and cancer cells invade the surroundings. The detection of capsular invasion is, therefore, crucial for the diagnosis and the choice of treatment and the development of new approaches aimed at the increase of diagnostic performance are of great importance. In the present study, we exploited the wide-field second harmonic generation (SHG) microscopy in combination with texture analysis and unsupervised machine learning (ML) to explore the possibility of quantitative characterization of collagen structure in the capsule and designation of different capsule areas as either intact, disrupted by invasion, or apt to invasion. Two-step <em>k</em>-means clustering showed that the collagen capsules in all analyzed tissue sections were highly heterogeneous and exhibited distinct segments described by characteristic ML parameter sets. The latter allowed a structural interpretation of the collagen fibers at the sites of overt invasion as fragmented and curled fibers with rarely formed distributed networks. Clustering analysis also distinguished areas in the PTC capsule that were not categorized as invasion sites by the initial histopathological analysis but could be recognized as prospective micro-invasions after additional inspection. The characteristic features of suspicious and invasive sites identified by the proposed unsupervised ML approach can become a reliable complement to existing methods for diagnosing encapsulated PTC, increase the reliability of diagnosis, simplify decision making, and prevent human-related diagnostic errors. In addition, the proposed automated ML-based selection of collagen capsule images and exclusion of non-informative regions can greatly accelerate and simplify the development of reliable methods for fully automated ML diagnosis that can be integrated into clinical practice.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102440"},"PeriodicalIF":5.4,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142394946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic MRI interpolation in temporal direction using an unsupervised generative model 利用无监督生成模型进行动态磁共振成像时间方向插值
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-22 DOI: 10.1016/j.compmedimag.2024.102435
Corbin Maciel , Qing Zou

Purpose

Cardiac cine magnetic resonance imaging (MRI) is an important tool in assessing dynamic heart function. However, this technique requires long acquisition time and long breath holds, which presents difficulties. The aim of this study is to propose an unsupervised neural network framework that can perform cardiac cine interpolation in time, so that we can increase the temporal resolution of cardiac cine without increasing acquisition time.

Methods

In this study, a subject-specific unsupervised generative neural network is designed to perform temporal interpolation for cardiac cine MRI. The network takes in a 2D latent vector in which each element corresponds to one cardiac phase in the cardiac cycle and then the network outputs the cardiac cine images which are acquired on the scanner. After the training of the generative network, we can interpolate the 2D latent vector and input the interpolated latent vector into the network and the network will output the frame-interpolated cine images. The results of the proposed cine interpolation neural network (CINN) framework are compared quantitatively and qualitatively with other state-of-the-art methods, the ground truth training cine frames, and the ground truth frames removed from the original acquisition. Signal-to-noise ratio (SNR), structural similarity index measures (SSIM), peak signal-to-noise ratio (PSNR), strain analysis, as well as the sharpness calculated using the Tenengrad algorithm were used for image quality assessment.

Results

As shown quantitatively and qualitatively, the proposed framework learns the generative task well and hence performs the temporal interpolation task well. Furthermore, both quantitative and qualitative comparison studies show the effectiveness of the proposed framework in cardiac cine interpolation in time.

Conclusion

The proposed generative model can effectively learn the generative task and perform high quality cardiac cine interpolation in time.
目的 心脏电影磁共振成像(MRI)是评估动态心脏功能的重要工具。然而,这项技术需要较长的采集时间和较长的屏气时间,这给研究带来了困难。本研究的目的是提出一种无监督神经网络框架,该框架可对心脏彩超进行时间插值,从而在不增加采集时间的情况下提高心脏彩超的时间分辨率。方法在本研究中,我们设计了一种针对特定对象的无监督生成神经网络,用于对心脏彩超进行时间插值。该网络接收二维潜向量,其中每个元素对应心脏周期中的一个心脏相位,然后该网络输出在扫描仪上获取的心脏显像图像。生成式网络训练完成后,我们可以对二维潜向量进行插值,然后将插值后的潜向量输入网络,网络将输出帧插值后的电影图像。我们将拟议的电影插值神经网络(CINN)框架的结果与其他最先进的方法、地面实况训练电影帧以及从原始采集中移除的地面实况帧进行了定量和定性比较。信噪比(SNR)、结构相似性指数(SSIM)、峰值信噪比(PSNR)、应变分析以及使用 Tenengrad 算法计算的清晰度都被用于图像质量评估。此外,定量和定性比较研究表明,所提出的框架在心脏实时插值中非常有效。结论所提出的生成模型可以有效地学习生成任务,并执行高质量的心脏实时插值。
{"title":"Dynamic MRI interpolation in temporal direction using an unsupervised generative model","authors":"Corbin Maciel ,&nbsp;Qing Zou","doi":"10.1016/j.compmedimag.2024.102435","DOIUrl":"10.1016/j.compmedimag.2024.102435","url":null,"abstract":"<div><h3>Purpose</h3><div>Cardiac cine magnetic resonance imaging (MRI) is an important tool in assessing dynamic heart function. However, this technique requires long acquisition time and long breath holds, which presents difficulties. The aim of this study is to propose an unsupervised neural network framework that can perform cardiac cine interpolation in time, so that we can increase the temporal resolution of cardiac cine without increasing acquisition time.</div></div><div><h3>Methods</h3><div>In this study, a subject-specific unsupervised generative neural network is designed to perform temporal interpolation for cardiac cine MRI. The network takes in a 2D latent vector in which each element corresponds to one cardiac phase in the cardiac cycle and then the network outputs the cardiac cine images which are acquired on the scanner. After the training of the generative network, we can interpolate the 2D latent vector and input the interpolated latent vector into the network and the network will output the frame-interpolated cine images. The results of the proposed cine interpolation neural network (CINN) framework are compared quantitatively and qualitatively with other state-of-the-art methods, the ground truth training cine frames, and the ground truth frames removed from the original acquisition. Signal-to-noise ratio (SNR), structural similarity index measures (SSIM), peak signal-to-noise ratio (PSNR), strain analysis, as well as the sharpness calculated using the Tenengrad algorithm were used for image quality assessment.</div></div><div><h3>Results</h3><div>As shown quantitatively and qualitatively, the proposed framework learns the generative task well and hence performs the temporal interpolation task well. Furthermore, both quantitative and qualitative comparison studies show the effectiveness of the proposed framework in cardiac cine interpolation in time.</div></div><div><h3>Conclusion</h3><div>The proposed generative model can effectively learn the generative task and perform high quality cardiac cine interpolation in time.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102435"},"PeriodicalIF":5.4,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142318453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BreasTDLUSeg: A coarse-to-fine framework for segmentation of breast terminal duct lobular units on histopathological whole-slide images BreasTDLUSeg:在组织病理学全切片图像上分割乳腺末端导管小叶单元的粗到细框架。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-19 DOI: 10.1016/j.compmedimag.2024.102432
Zixiao Lu , Kai Tang , Yi Wu , Xiaoxuan Zhang , Ziqi An , Xiongfeng Zhu , Qianjin Feng , Yinghua Zhao
Automatic segmentation of breast terminal duct lobular units (TDLUs) on histopathological whole-slide images (WSIs) is crucial for the quantitative evaluation of TDLUs in the diagnostic and prognostic analysis of breast cancer. However, TDLU segmentation remains a great challenge due to its highly heterogeneous sizes, structures, and morphologies as well as the small areas on WSIs. In this study, we propose BreasTDLUSeg, an efficient coarse-to-fine two-stage framework based on multi-scale attention to achieve localization and precise segmentation of TDLUs on hematoxylin and eosin (H&E)-stained WSIs. BreasTDLUSeg consists of two networks: a superpatch-based patch-level classification network (SPPC-Net) and a patch-based pixel-level segmentation network (PPS-Net). SPPC-Net takes a superpatch as input and adopts a sub-region classification head to classify each patch within the superpatch as TDLU positive or negative. PPS-Net takes the TDLU positive patches derived from SPPC-Net as input. PPS-Net deploys a multi-scale CNN-Transformer as an encoder to learn enhanced multi-scale morphological representations and an upsampler to generate pixel-wise segmentation masks for the TDLU positive patches. We also constructed two breast cancer TDLU datasets containing a total of 530 superpatch images with patch-level annotations and 2322 patch images with pixel-level annotations to enable the development of TDLU segmentation methods. Experiments on the two datasets demonstrate that BreasTDLUSeg outperforms other state-of-the-art methods with the highest Dice similarity coefficients of 79.97% and 92.93%, respectively. The proposed method shows great potential to assist pathologists in the pathological analysis of breast cancer. An open-source implementation of our approach can be found at https://github.com/Dian-kai/BreasTDLUSeg.
在组织病理全切片图像(WSI)上自动分割乳腺末端导管小叶单位(TDLU)对于乳腺癌诊断和预后分析中定量评估 TDLU 至关重要。然而,由于 TDLU 的大小、结构和形态差异很大,而且在 WSIs 上的面积很小,因此 TDLU 的分割仍然是一项巨大的挑战。在本研究中,我们提出了基于多尺度关注的高效粗到细两阶段框架 BreasTDLUSeg,以实现苏木精和伊红(H&E)染色 WSI 上 TDLU 的定位和精确分割。BreasTDLUSeg 由两个网络组成:基于超斑的斑块级分类网络(SPPC-Net)和基于斑块的像素级分割网络(PPS-Net)。SPPC-Net 以超级斑块为输入,采用子区域分类头将超级斑块中的每个斑块划分为 TDLU 正片或负片。PPS-Net 将 SPPC-Net 得出的 TDLU 正片作为输入。PPS-Net 部署了一个多尺度 CNN 变换器作为编码器,以学习增强的多尺度形态学表示,并部署了一个上采样器,为 TDLU 阳性斑块生成像素分割掩码。我们还构建了两个乳腺癌 TDLU 数据集,共包含 530 幅带有斑块级注释的超级斑块图像和 2322 幅带有像素级注释的斑块图像,以便开发 TDLU 分割方法。在这两个数据集上的实验表明,BreasTDLUSeg 优于其他最先进的方法,其最高 Dice 相似系数分别为 79.97% 和 92.93%。所提出的方法在协助病理学家进行乳腺癌病理分析方面显示出巨大的潜力。我们的方法的开源实现可以在 https://github.com/Dian-kai/BreasTDLUSeg 上找到。
{"title":"BreasTDLUSeg: A coarse-to-fine framework for segmentation of breast terminal duct lobular units on histopathological whole-slide images","authors":"Zixiao Lu ,&nbsp;Kai Tang ,&nbsp;Yi Wu ,&nbsp;Xiaoxuan Zhang ,&nbsp;Ziqi An ,&nbsp;Xiongfeng Zhu ,&nbsp;Qianjin Feng ,&nbsp;Yinghua Zhao","doi":"10.1016/j.compmedimag.2024.102432","DOIUrl":"10.1016/j.compmedimag.2024.102432","url":null,"abstract":"<div><div>Automatic segmentation of breast terminal duct lobular units (TDLUs) on histopathological whole-slide images (WSIs) is crucial for the quantitative evaluation of TDLUs in the diagnostic and prognostic analysis of breast cancer. However, TDLU segmentation remains a great challenge due to its highly heterogeneous sizes, structures, and morphologies as well as the small areas on WSIs. In this study, we propose BreasTDLUSeg, an efficient coarse-to-fine two-stage framework based on multi-scale attention to achieve localization and precise segmentation of TDLUs on hematoxylin and eosin (H&amp;E)-stained WSIs. BreasTDLUSeg consists of two networks: a superpatch-based patch-level classification network (SPPC-Net) and a patch-based pixel-level segmentation network (PPS-Net). SPPC-Net takes a superpatch as input and adopts a sub-region classification head to classify each patch within the superpatch as TDLU positive or negative. PPS-Net takes the TDLU positive patches derived from SPPC-Net as input. PPS-Net deploys a multi-scale CNN-Transformer as an encoder to learn enhanced multi-scale morphological representations and an upsampler to generate pixel-wise segmentation masks for the TDLU positive patches. We also constructed two breast cancer TDLU datasets containing a total of 530 superpatch images with patch-level annotations and 2322 patch images with pixel-level annotations to enable the development of TDLU segmentation methods. Experiments on the two datasets demonstrate that BreasTDLUSeg outperforms other state-of-the-art methods with the highest Dice similarity coefficients of 79.97% and 92.93%, respectively. The proposed method shows great potential to assist pathologists in the pathological analysis of breast cancer. An open-source implementation of our approach can be found at <span><span>https://github.com/Dian-kai/BreasTDLUSeg</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"118 ","pages":"Article 102432"},"PeriodicalIF":5.4,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142512335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Main challenges on the curation of large scale datasets for pancreas segmentation using deep learning in multi-phase CT scans: Focus on cardinality, manual refinement, and annotation quality 利用深度学习在多期 CT 扫描中进行胰腺分割的大型数据集整理工作面临的主要挑战:重点关注万有引力、人工完善和注释质量
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-13 DOI: 10.1016/j.compmedimag.2024.102434
Matteo Cavicchioli , Andrea Moglia , Ludovica Pierelli , Giacomo Pugliese , Pietro Cerveri

Accurate segmentation of the pancreas in computed tomography (CT) holds paramount importance in diagnostics, surgical planning, and interventions. Recent studies have proposed supervised deep-learning models for segmentation, but their efficacy relies on the quality and quantity of the training data. Most of such works employed small-scale public datasets, without proving the efficacy of generalization to external datasets. This study explored the optimization of pancreas segmentation accuracy by pinpointing the ideal dataset size, understanding resource implications, examining manual refinement impact, and assessing the influence of anatomical subregions. We present the AIMS-1300 dataset encompassing 1,300 CT scans. Its manual annotation by medical experts required 938 h. A 2.5D UNet was implemented to assess the impact of training sample size on segmentation accuracy by partitioning the original AIMS-1300 dataset into 11 smaller subsets of progressively increasing numerosity. The findings revealed that training sets exceeding 440 CTs did not lead to better segmentation performance. In contrast, nnU-Net and UNet with Attention Gate reached a plateau for 585 CTs. Tests on generalization on the publicly available AMOS-CT dataset confirmed this outcome. As the size of the partition of the AIMS-1300 training set increases, the number of error slices decreases, reaching a minimum with 730 and 440 CTs, for AIMS-1300 and AMOS-CT datasets, respectively. Segmentation metrics on the AIMS-1300 and AMOS-CT datasets improved more on the head than the body and tail of the pancreas as the dataset size increased. By carefully considering the task and the characteristics of the available data, researchers can develop deep learning models without sacrificing performance even with limited data. This could accelerate developing and deploying artificial intelligence tools for pancreas surgery and other surgical data science applications.

计算机断层扫描(CT)中胰腺的精确分割在诊断、手术规划和干预中至关重要。最近的研究提出了用于分割的有监督深度学习模型,但其有效性取决于训练数据的质量和数量。大多数此类研究都采用了小规模的公共数据集,但并未证明其对外部数据集的泛化效果。本研究通过确定理想的数据集大小、了解资源影响、检查人工细化的影响以及评估解剖亚区的影响,来探索胰腺分割准确性的优化。我们展示的 AIMS-1300 数据集包含 1,300 张 CT 扫描图像。通过将原始 AIMS-1300 数据集划分为 11 个数量逐渐增加的较小子集,我们采用了 2.5D UNet 来评估训练样本数量对分割准确性的影响。结果显示,训练集超过 440 个 CT 并不能带来更好的分割性能。相比之下,nnU-Net 和 UNet with Attention Gate 在 585 CTs 时达到了最高点。在公开的 AMOS-CT 数据集上进行的泛化测试证实了这一结果。随着 AIMS-1300 训练集分区大小的增加,错误片段的数量也在减少,AIMS-1300 和 AMOS-CT 数据集分别在 730 和 440 CT 时达到最小值。随着数据集大小的增加,AIMS-1300 和 AMOS-CT 数据集上胰腺头部的分割指标比胰腺体部和尾部的改进更大。通过仔细考虑任务和可用数据的特征,研究人员可以开发出深度学习模型,即使数据有限,也不会牺牲性能。这将加速开发和部署用于胰腺手术和其他手术数据科学应用的人工智能工具。
{"title":"Main challenges on the curation of large scale datasets for pancreas segmentation using deep learning in multi-phase CT scans: Focus on cardinality, manual refinement, and annotation quality","authors":"Matteo Cavicchioli ,&nbsp;Andrea Moglia ,&nbsp;Ludovica Pierelli ,&nbsp;Giacomo Pugliese ,&nbsp;Pietro Cerveri","doi":"10.1016/j.compmedimag.2024.102434","DOIUrl":"10.1016/j.compmedimag.2024.102434","url":null,"abstract":"<div><p>Accurate segmentation of the pancreas in computed tomography (CT) holds paramount importance in diagnostics, surgical planning, and interventions. Recent studies have proposed supervised deep-learning models for segmentation, but their efficacy relies on the quality and quantity of the training data. Most of such works employed small-scale public datasets, without proving the efficacy of generalization to external datasets. This study explored the optimization of pancreas segmentation accuracy by pinpointing the ideal dataset size, understanding resource implications, examining manual refinement impact, and assessing the influence of anatomical subregions. We present the AIMS-1300 dataset encompassing 1,300 CT scans. Its manual annotation by medical experts required 938 h. A 2.5D UNet was implemented to assess the impact of training sample size on segmentation accuracy by partitioning the original AIMS-1300 dataset into 11 smaller subsets of progressively increasing numerosity. The findings revealed that training sets exceeding 440 CTs did not lead to better segmentation performance. In contrast, nnU-Net and UNet with Attention Gate reached a plateau for 585 CTs. Tests on generalization on the publicly available AMOS-CT dataset confirmed this outcome. As the size of the partition of the AIMS-1300 training set increases, the number of error slices decreases, reaching a minimum with 730 and 440 CTs, for AIMS-1300 and AMOS-CT datasets, respectively. Segmentation metrics on the AIMS-1300 and AMOS-CT datasets improved more on the head than the body and tail of the pancreas as the dataset size increased. By carefully considering the task and the characteristics of the available data, researchers can develop deep learning models without sacrificing performance even with limited data. This could accelerate developing and deploying artificial intelligence tools for pancreas surgery and other surgical data science applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102434"},"PeriodicalIF":5.4,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001113/pdfft?md5=45c57efca48ce73af49c585837730bf7&pid=1-s2.0-S0895611124001113-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards explainable oral cancer recognition: Screening on imperfect images via Informed Deep Learning and Case-Based Reasoning 实现可解释的口腔癌识别:通过知情深度学习和基于案例的推理对不完美图像进行筛查
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-11 DOI: 10.1016/j.compmedimag.2024.102433
Marco Parola , Federico A. Galatolo , Gaetano La Mantia , Mario G.C.A. Cimino , Giuseppina Campisi , Olga Di Fede

Oral squamous cell carcinoma recognition presents a challenge due to late diagnosis and costly data acquisition. A cost-efficient, computerized screening system is crucial for early disease detection, minimizing the need for expert intervention and expensive analysis. Besides, transparency is essential to align these systems with critical sector applications. Explainable Artificial Intelligence (XAI) provides techniques for understanding models. However, current XAI is mostly data-driven and focused on addressing developers’ requirements of improving models rather than clinical users’ demands for expressing relevant insights. Among different XAI strategies, we propose a solution composed of Case-Based Reasoning paradigm to provide visual output explanations and Informed Deep Learning (IDL) to integrate medical knowledge within the system. A key aspect of our solution lies in its capability to handle data imperfections, including labeling inaccuracies and artifacts, thanks to an ensemble architecture on top of the deep learning (DL) workflow. We conducted several experimental benchmarks on a dataset collected in collaboration with medical centers. Our findings reveal that employing the IDL approach yields an accuracy of 85%, surpassing the 77% accuracy achieved by DL alone. Furthermore, we measured the human-centered explainability of the two approaches and IDL generates explanations more congruent with the clinical user demands.

口腔鳞状细胞癌的识别因诊断较晚和数据采集成本高昂而面临挑战。具有成本效益的计算机化筛查系统对于早期疾病检测至关重要,可最大限度地减少对专家干预和昂贵分析的需求。此外,要使这些系统与关键领域的应用保持一致,透明度也至关重要。可解释人工智能(XAI)提供了理解模型的技术。然而,目前的 XAI 大多是数据驱动的,侧重于满足开发人员改进模型的要求,而不是满足临床用户表达相关见解的需求。在不同的 XAI 策略中,我们提出了一种由基于案例的推理范式和知情深度学习(IDL)组成的解决方案,前者用于提供可视化的输出解释,后者用于在系统中整合医学知识。我们的解决方案的一个关键方面在于它能够处理数据不完善的问题,包括标记不准确和伪造,这要归功于深度学习(DL)工作流程之上的集合架构。我们在与医疗中心合作收集的数据集上进行了多项实验基准测试。我们的研究结果表明,采用 IDL 方法可获得 85% 的准确率,超过了单独使用 DL 所获得的 77% 的准确率。此外,我们还测量了这两种方法以人为本的可解释性,IDL 生成的解释更符合临床用户的需求。
{"title":"Towards explainable oral cancer recognition: Screening on imperfect images via Informed Deep Learning and Case-Based Reasoning","authors":"Marco Parola ,&nbsp;Federico A. Galatolo ,&nbsp;Gaetano La Mantia ,&nbsp;Mario G.C.A. Cimino ,&nbsp;Giuseppina Campisi ,&nbsp;Olga Di Fede","doi":"10.1016/j.compmedimag.2024.102433","DOIUrl":"10.1016/j.compmedimag.2024.102433","url":null,"abstract":"<div><p>Oral squamous cell carcinoma recognition presents a challenge due to late diagnosis and costly data acquisition. A cost-efficient, computerized screening system is crucial for early disease detection, minimizing the need for expert intervention and expensive analysis. Besides, transparency is essential to align these systems with critical sector applications. Explainable Artificial Intelligence (XAI) provides techniques for understanding models. However, current XAI is mostly data-driven and focused on addressing developers’ requirements of improving models rather than clinical users’ demands for expressing relevant insights. Among different XAI strategies, we propose a solution composed of Case-Based Reasoning paradigm to provide visual output explanations and Informed Deep Learning (IDL) to integrate medical knowledge within the system. A key aspect of our solution lies in its capability to handle data imperfections, including labeling inaccuracies and artifacts, thanks to an ensemble architecture on top of the deep learning (DL) workflow. We conducted several experimental benchmarks on a dataset collected in collaboration with medical centers. Our findings reveal that employing the IDL approach yields an accuracy of 85%, surpassing the 77% accuracy achieved by DL alone. Furthermore, we measured the human-centered explainability of the two approaches and IDL generates explanations more congruent with the clinical user demands.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102433"},"PeriodicalIF":5.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001101/pdfft?md5=96dc2c2aa07fe4189f5653f58184b9b0&pid=1-s2.0-S0895611124001101-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142228887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hematoma expansion prediction in intracerebral hemorrhage patients by using synthesized CT images in an end-to-end deep learning framework 在端到端深度学习框架中使用合成 CT 图像预测脑内出血患者血肿扩大情况
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-05 DOI: 10.1016/j.compmedimag.2024.102430
Cansu Yalcin , Valeriia Abramova , Mikel Terceño , Arnau Oliver , Yolanda Silva , Xavier Lladó

Spontaneous intracerebral hemorrhage (ICH) is a type of stroke less prevalent than ischemic stroke but associated with high mortality rates. Hematoma expansion (HE) is an increase in the bleeding that affects 30%–38% of hemorrhagic stroke patients. It is observed within 24 h of onset and associated with patient worsening. Clinically it is relevant to detect the patients that will develop HE from their initial computed tomography (CT) scans which could improve patient management and treatment decisions. However, this is a significant challenge due to the predictive nature of the task and its low prevalence, which hinders the availability of large datasets with the required longitudinal information. In this work, we present an end-to-end deep learning framework capable of predicting which cases will exhibit HE using only the initial basal image. We introduce a deep learning framework based on the 2D EfficientNet B0 model to predict the occurrence of HE using initial non-contrasted CT scans and their corresponding lesion annotation as priors. We used an in-house acquired dataset of 122 ICH patients, including 35 HE cases, containing longitudinal CT scans with manual lesion annotations in both basal and follow-up (obtained within 24 h after the basal scan). Experiments were conducted using a 5-fold cross-validation strategy. We addressed the limited data problem by incorporating synthetic images into the training process. To the best of our knowledge, our approach is novel in the field of HE prediction, being the first to use image synthesis to enhance results. We studied different scenarios such as training only with the original scans, using standard image augmentation techniques, and using synthetic image generation. The best performance was achieved by adding five generated versions of each image, along with standard data augmentation, during the training process. This significantly improved (p=0.0003) the performance obtained with our baseline model using directly the original CT scans from an Accuracy of 0.56 to 0.84, F1-Score of 0.53 to 0.82, Sensitivity of 0.51 to 0.77, and Specificity of 0.60 to 0.91, respectively. The proposed approach shows promising results in predicting HE, especially with the inclusion of synthetically generated images. The obtained results highlight the significance of this research direction, which has the potential to improve the clinical management of patients with hemorrhagic stroke. The code is available at: https://github.com/NIC-VICOROB/HE-prediction-SynthCT.

自发性脑内出血(ICH)是一种发病率低于缺血性中风但死亡率很高的中风类型。血肿扩大(HE)是指出血量增加,影响 30%-38% 的出血性中风患者。在发病后 24 小时内即可观察到,并与患者病情恶化有关。在临床上,从最初的计算机断层扫描(CT)中检测出会出现 HE 的患者具有重要意义,可改善患者管理和治疗决策。然而,由于这项任务具有预测性,而且发病率较低,这阻碍了具有所需纵向信息的大型数据集的可用性,因此这是一项重大挑战。在这项工作中,我们提出了一种端到端的深度学习框架,能够仅利用初始基底图像预测哪些病例会表现出 HE。我们引入了一个基于二维 EfficientNet B0 模型的深度学习框架,利用初始非对比 CT 扫描及其相应的病变注释作为先验,预测 HE 的发生。我们使用了内部获得的 122 例 ICH 患者数据集,其中包括 35 例 HE 病例,该数据集包含纵向 CT 扫描,并在基底扫描和随访(基底扫描后 24 小时内获得)中进行了人工病灶注释。实验采用了 5 倍交叉验证策略。我们在训练过程中加入了合成图像,从而解决了数据有限的问题。据我们所知,我们的方法是 HE 预测领域的新方法,也是第一个使用图像合成来提高结果的方法。我们研究了不同的情况,如仅使用原始扫描图像进行训练、使用标准图像增强技术以及使用合成图像生成技术。在训练过程中,在标准数据增强的同时,每张图像添加五个生成版本,从而达到最佳效果。这大大提高了(p=0.0003)直接使用原始 CT 扫描图像的基线模型的性能,准确率从 0.56 提高到 0.84,F1 分数从 0.53 提高到 0.82,灵敏度从 0.51 提高到 0.77,特异性从 0.60 提高到 0.91。所提出的方法在预测 HE 方面取得了可喜的成果,尤其是在包含合成图像的情况下。所取得的结果凸显了这一研究方向的重要意义,有望改善出血性中风患者的临床管理。代码见:https://github.com/NIC-VICOROB/HE-prediction-SynthCT。
{"title":"Hematoma expansion prediction in intracerebral hemorrhage patients by using synthesized CT images in an end-to-end deep learning framework","authors":"Cansu Yalcin ,&nbsp;Valeriia Abramova ,&nbsp;Mikel Terceño ,&nbsp;Arnau Oliver ,&nbsp;Yolanda Silva ,&nbsp;Xavier Lladó","doi":"10.1016/j.compmedimag.2024.102430","DOIUrl":"10.1016/j.compmedimag.2024.102430","url":null,"abstract":"<div><p>Spontaneous intracerebral hemorrhage (ICH) is a type of stroke less prevalent than ischemic stroke but associated with high mortality rates. Hematoma expansion (HE) is an increase in the bleeding that affects 30%–38% of hemorrhagic stroke patients. It is observed within 24 h of onset and associated with patient worsening. Clinically it is relevant to detect the patients that will develop HE from their initial computed tomography (CT) scans which could improve patient management and treatment decisions. However, this is a significant challenge due to the predictive nature of the task and its low prevalence, which hinders the availability of large datasets with the required longitudinal information. In this work, we present an end-to-end deep learning framework capable of predicting which cases will exhibit HE using only the initial basal image. We introduce a deep learning framework based on the 2D EfficientNet B0 model to predict the occurrence of HE using initial non-contrasted CT scans and their corresponding lesion annotation as priors. We used an in-house acquired dataset of 122 ICH patients, including 35 HE cases, containing longitudinal CT scans with manual lesion annotations in both basal and follow-up (obtained within 24 h after the basal scan). Experiments were conducted using a 5-fold cross-validation strategy. We addressed the limited data problem by incorporating synthetic images into the training process. To the best of our knowledge, our approach is novel in the field of HE prediction, being the first to use image synthesis to enhance results. We studied different scenarios such as training only with the original scans, using standard image augmentation techniques, and using synthetic image generation. The best performance was achieved by adding five generated versions of each image, along with standard data augmentation, during the training process. This significantly improved (<span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>0003</mn></mrow></math></span>) the performance obtained with our baseline model using directly the original CT scans from an Accuracy of 0.56 to 0.84, F1-Score of 0.53 to 0.82, Sensitivity of 0.51 to 0.77, and Specificity of 0.60 to 0.91, respectively. The proposed approach shows promising results in predicting HE, especially with the inclusion of synthetically generated images. The obtained results highlight the significance of this research direction, which has the potential to improve the clinical management of patients with hemorrhagic stroke. The code is available at: <span><span>https://github.com/NIC-VICOROB/HE-prediction-SynthCT</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102430"},"PeriodicalIF":5.4,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001071/pdfft?md5=c274f51b14553bc438e5f61e76ba2a00&pid=1-s2.0-S0895611124001071-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CycleSGAN: A cycle-consistent and semantics-preserving generative adversarial network for unpaired MR-to-CT image synthesis CycleSGAN:用于无配对 MR-CT 图像合成的周期一致性和语义保留生成对抗网络。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-04 DOI: 10.1016/j.compmedimag.2024.102431
Runze Wang , Alexander F. Heimann , Moritz Tannast , Guoyan Zheng

CycleGAN has been leveraged to synthesize a CT image from an available MR image after trained on unpaired data. Due to the lack of direct constraints between the synthetic and the input images, CycleGAN cannot guarantee structural consistency and often generates inaccurate mappings that shift the anatomy, which is highly undesirable for downstream clinical applications such as MRI-guided radiotherapy treatment planning and PET/MRI attenuation correction. In this paper, we propose a cycle-consistent and semantics-preserving generative adversarial network, referred as CycleSGAN, for unpaired MR-to-CT image synthesis. Our design features a novel and generic way to incorporate semantic information into CycleGAN. This is done by designing a pair of three-player games within the CycleGAN framework where each three-player game consists of one generator and two discriminators to formulate two distinct types of adversarial learning: appearance adversarial learning and structure adversarial learning. These two types of adversarial learning are alternately trained to ensure both realistic image synthesis and semantic structure preservation. Results on unpaired hip MR-to-CT image synthesis show that our method produces better synthetic CT images in both accuracy and visual quality as compared to other state-of-the-art (SOTA) unpaired MR-to-CT image synthesis methods.

在对非配对数据进行训练后,CycleGAN 被用于根据可用的 MR 图像合成 CT 图像。由于合成图像与输入图像之间缺乏直接约束,CycleGAN 无法保证结构的一致性,经常会生成不准确的映射,从而使解剖结构发生偏移,这对于下游临床应用(如 MRI 引导的放射治疗规划和 PET/MRI 衰减校正)来说是非常不可取的。在本文中,我们提出了一种循环一致性和语义保护生成对抗网络(称为 CycleSGAN),用于非配对 MR-CT 图像合成。我们的设计采用了一种新颖而通用的方法,将语义信息纳入 CycleGAN。具体做法是在 CycleGAN 框架内设计一对三人博弈,每个三人博弈由一个生成器和两个判别器组成,从而形成两种不同类型的对抗学习:外观对抗学习和结构对抗学习。这两类对抗学习交替进行训练,以确保既能合成真实图像,又能保留语义结构。非配对髋关节 MR 到 CT 图像合成的结果表明,与其他最先进的(SOTA)非配对 MR 到 CT 图像合成方法相比,我们的方法在准确性和视觉质量方面都能生成更好的合成 CT 图像。
{"title":"CycleSGAN: A cycle-consistent and semantics-preserving generative adversarial network for unpaired MR-to-CT image synthesis","authors":"Runze Wang ,&nbsp;Alexander F. Heimann ,&nbsp;Moritz Tannast ,&nbsp;Guoyan Zheng","doi":"10.1016/j.compmedimag.2024.102431","DOIUrl":"10.1016/j.compmedimag.2024.102431","url":null,"abstract":"<div><p>CycleGAN has been leveraged to synthesize a CT image from an available MR image after trained on unpaired data. Due to the lack of direct constraints between the synthetic and the input images, CycleGAN cannot guarantee structural consistency and often generates inaccurate mappings that shift the anatomy, which is highly undesirable for downstream clinical applications such as MRI-guided radiotherapy treatment planning and PET/MRI attenuation correction. In this paper, we propose a cycle-consistent and semantics-preserving generative adversarial network, referred as CycleSGAN, for unpaired MR-to-CT image synthesis. Our design features a novel and generic way to incorporate semantic information into CycleGAN. This is done by designing a pair of three-player games within the CycleGAN framework where each three-player game consists of one generator and two discriminators to formulate two distinct types of adversarial learning: appearance adversarial learning and structure adversarial learning. These two types of adversarial learning are alternately trained to ensure both realistic image synthesis and semantic structure preservation. Results on unpaired hip MR-to-CT image synthesis show that our method produces better synthetic CT images in both accuracy and visual quality as compared to other state-of-the-art (SOTA) unpaired MR-to-CT image synthesis methods.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102431"},"PeriodicalIF":5.4,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142146765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A lung biopsy path planning algorithm based on the double spherical constraint Pareto and indicators’ importance-correlation degree 基于双球约束帕累托和指标重要性相关度的肺活检路径规划算法
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-08-31 DOI: 10.1016/j.compmedimag.2024.102426
Hui Yang , Yu Zhang , Yuhang Gong , Jing Zhang , Ling He , Jianquan Zhong , Ling Tang

Lung cancer has the highest mortality rate among cancers. The commonly used clinical method for diagnosing lung cancer is the CT-guided percutaneous transthoracic lung biopsy (CT-PTLB), but this method requires a high level of clinical experience from doctors. In this work, an automatic path planning method for CT-PTLB is proposed to provide doctors with auxiliary advice on puncture paths. The proposed method comprises three steps: preprocessing, initial path selection, and path evaluation. During preprocessing, the chest organs required for subsequent path planning are segmented. During the initial path selection, a target point selection method for selecting biopsy samples according to biopsy sampling requirements is proposed, which includes a down-sampling algorithm suitable for different nodule shapes. Entry points are selected according to the selected target points and clinical constraints. During the path evaluation, the clinical needs of lung biopsy surgery are first quantified as path evaluation indicators and then divided according to their evaluation perspective into risk and execution indicators. Then, considering the impact of the correlation between indicators, a path scoring system based on the double spherical constraint Pareto and the importance-correlation degree of the indicators is proposed to evaluate the comprehensive performance of the planned paths. The proposed method is retrospectively tested on 6 CT images and prospectively tested on 25 CT images. The experimental results indicate that the method proposed in this work can be used to plan feasible puncture paths for different cases and can serve as an auxiliary tool for lung biopsy surgery.

肺癌是死亡率最高的癌症。临床上常用的肺癌诊断方法是CT引导下经皮经胸肺穿刺活检术(CT-PTLB),但这种方法对医生的临床经验要求很高。本研究提出了一种 CT-PTLB 自动路径规划方法,为医生提供穿刺路径的辅助建议。该方法包括三个步骤:预处理、初始路径选择和路径评估。在预处理过程中,对后续路径规划所需的胸部器官进行分割。在初始路径选择过程中,提出了根据活检取样要求选择活检样本的目标点选择方法,其中包括适合不同结节形状的向下取样算法。根据选定的目标点和临床限制条件选择入口点。在路径评价过程中,首先将肺活检手术的临床需求量化为路径评价指标,然后根据其评价角度分为风险指标和执行指标。然后,考虑到指标间相关性的影响,提出了基于双球面约束帕累托和指标重要性-相关度的路径评分体系,对规划路径的综合性能进行评价。提出的方法在 6 幅 CT 图像上进行了回顾性测试,并在 25 幅 CT 图像上进行了前瞻性测试。实验结果表明,本文提出的方法可用于规划不同病例的可行穿刺路径,可作为肺活检手术的辅助工具。
{"title":"A lung biopsy path planning algorithm based on the double spherical constraint Pareto and indicators’ importance-correlation degree","authors":"Hui Yang ,&nbsp;Yu Zhang ,&nbsp;Yuhang Gong ,&nbsp;Jing Zhang ,&nbsp;Ling He ,&nbsp;Jianquan Zhong ,&nbsp;Ling Tang","doi":"10.1016/j.compmedimag.2024.102426","DOIUrl":"10.1016/j.compmedimag.2024.102426","url":null,"abstract":"<div><p>Lung cancer has the highest mortality rate among cancers. The commonly used clinical method for diagnosing lung cancer is the CT-guided percutaneous transthoracic lung biopsy (CT-PTLB), but this method requires a high level of clinical experience from doctors. In this work, an automatic path planning method for CT-PTLB is proposed to provide doctors with auxiliary advice on puncture paths. The proposed method comprises three steps: preprocessing, initial path selection, and path evaluation. During preprocessing, the chest organs required for subsequent path planning are segmented. During the initial path selection, a target point selection method for selecting biopsy samples according to biopsy sampling requirements is proposed, which includes a down-sampling algorithm suitable for different nodule shapes. Entry points are selected according to the selected target points and clinical constraints. During the path evaluation, the clinical needs of lung biopsy surgery are first quantified as path evaluation indicators and then divided according to their evaluation perspective into risk and execution indicators. Then, considering the impact of the correlation between indicators, a path scoring system based on the double spherical constraint Pareto and the importance-correlation degree of the indicators is proposed to evaluate the comprehensive performance of the planned paths. The proposed method is retrospectively tested on 6 CT images and prospectively tested on 25 CT images. The experimental results indicate that the method proposed in this work can be used to plan feasible puncture paths for different cases and can serve as an auxiliary tool for lung biopsy surgery.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102426"},"PeriodicalIF":5.4,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142243550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1