首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
Deep local-to-global feature learning for medical image super-resolution 用于医学图像超分辨率的局部到全局深度特征学习
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-26 DOI: 10.1016/j.compmedimag.2024.102374
Wenfeng Huang , Xiangyun Liao , Hao Chen , Ying Hu , Wenjing Jia , Qiong Wang

Medical images play a vital role in medical analysis by providing crucial information about patients’ pathological conditions. However, the quality of these images can be compromised by many factors, such as limited resolution of the instruments, artifacts caused by movements, and the complexity of the scanned areas. As a result, low-resolution (LR) images cannot provide sufficient information for diagnosis. To address this issue, researchers have attempted to apply image super-resolution (SR) techniques to restore the high-resolution (HR) images from their LR counterparts. However, these techniques are designed for generic images, and thus suffer from many challenges unique to medical images. An obvious one is the diversity of the scanned objects; for example, the organs, tissues, and vessels typically appear in different sizes and shapes, and are thus hard to restore with standard convolution neural networks (CNNs). In this paper, we develop a dynamic-local learning framework to capture the details of these diverse areas, consisting of deformable convolutions with adjustable kernel shapes. Moreover, the global information between the tissues and organs is vital for medical diagnosis. To preserve global information, we propose pixel–pixel and patch–patch global learning using a non-local mechanism and a vision transformer (ViT), respectively. The result is a novel CNN-ViT neural network with Local-to-Global feature learning for medical image SR, referred to as LGSR, which can accurately restore both local details and global information. We evaluate our method on six public datasets and one large-scale private dataset, which include five different types of medical images (i.e., Ultrasound, OCT, Endoscope, CT, and MRI images). Experiments show that the proposed method achieves superior PSNR/SSIM and visual performance than the state of the arts with competitive computational costs, measured in network parameters, runtime, and FLOPs. What is more, the experiment conducted on OCT image segmentation for the downstream task demonstrates a significantly positive performance effect of LGSR.

医学影像在医学分析中起着至关重要的作用,能提供有关病人病理状况的关键信息。然而,这些图像的质量可能会受到许多因素的影响,例如仪器的分辨率有限、移动造成的伪影以及扫描区域的复杂性。因此,低分辨率(LR)图像无法为诊断提供足够的信息。为了解决这个问题,研究人员尝试应用图像超分辨率(SR)技术来还原低分辨率图像中的高分辨率(HR)图像。然而,这些技术都是针对普通图像设计的,因此存在许多医学图像特有的难题。其中一个明显的挑战是扫描对象的多样性,例如,器官、组织和血管通常具有不同的尺寸和形状,因此很难通过标准卷积神经网络(CNN)进行还原。在本文中,我们开发了一种动态局部学习框架来捕捉这些不同区域的细节,该框架由具有可调内核形状的可变形卷积组成。此外,组织和器官之间的全局信息对于医学诊断至关重要。为了保留全局信息,我们分别使用非局部机制和视觉变换器(ViT)提出了像素-像素和斑块-斑块全局学习。其结果是一种新颖的 CNN-ViT 神经网络,具有用于医学影像 SR 的从局部到全局的特征学习功能,简称为 LGSR,它能准确还原局部细节和全局信息。我们在六个公共数据集和一个大规模私有数据集上评估了我们的方法,这些数据集包括五种不同类型的医学图像(即超声波、OCT、内窥镜、CT 和 MRI 图像)。实验表明,所提方法的 PSNR/SSIM 和视觉效果优于目前的技术水平,而计算成本(以网络参数、运行时间和 FLOPs 计)却很有竞争力。更重要的是,针对下游任务的 OCT 图像分割实验表明,LGSR 的性能效果显著。
{"title":"Deep local-to-global feature learning for medical image super-resolution","authors":"Wenfeng Huang ,&nbsp;Xiangyun Liao ,&nbsp;Hao Chen ,&nbsp;Ying Hu ,&nbsp;Wenjing Jia ,&nbsp;Qiong Wang","doi":"10.1016/j.compmedimag.2024.102374","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102374","url":null,"abstract":"<div><p>Medical images play a vital role in medical analysis by providing crucial information about patients’ pathological conditions. However, the quality of these images can be compromised by many factors, such as limited resolution of the instruments, artifacts caused by movements, and the complexity of the scanned areas. As a result, low-resolution (LR) images cannot provide sufficient information for diagnosis. To address this issue, researchers have attempted to apply image super-resolution (SR) techniques to restore the high-resolution (HR) images from their LR counterparts. However, these techniques are designed for generic images, and thus suffer from many challenges unique to medical images. An obvious one is the diversity of the scanned objects; for example, the organs, tissues, and vessels typically appear in different sizes and shapes, and are thus hard to restore with standard convolution neural networks (CNNs). In this paper, we develop a dynamic-local learning framework to capture the details of these diverse areas, consisting of deformable convolutions with adjustable kernel shapes. Moreover, the global information between the tissues and organs is vital for medical diagnosis. To preserve global information, we propose pixel–pixel and patch–patch global learning using a non-local mechanism and a vision transformer (ViT), respectively. The result is a novel CNN-ViT neural network with Local-to-Global feature learning for medical image SR, referred to as LGSR, which can accurately restore both local details and global information. We evaluate our method on six public datasets and one large-scale private dataset, which include five different types of medical images (<em>i.e.</em>, Ultrasound, OCT, Endoscope, CT, and MRI images). Experiments show that the proposed method achieves superior PSNR/SSIM and visual performance than the state of the arts with competitive computational costs, measured in network parameters, runtime, and FLOPs. What is more, the experiment conducted on OCT image segmentation for the downstream task demonstrates a significantly positive performance effect of LGSR.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140332733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised classification of multi-contrast magnetic resonance histology of peripheral arterial disease lesions using a convolutional variational autoencoder with a Gaussian mixture model in latent space: A technical feasibility study 利用卷积变异自动编码器和潜空间高斯混合模型对外周动脉疾病病变的多对比磁共振组织学进行无监督分类:技术可行性研究
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-26 DOI: 10.1016/j.compmedimag.2024.102372
Judit Csore , Trisha L. Roy , Graham Wright , Christof Karmonik

Purpose

To investigate the feasibility of a deep learning algorithm combining variational autoencoder (VAE) and two-dimensional (2D) convolutional neural networks (CNN) for automatically quantifying hard tissue presence and morphology in multi-contrast magnetic resonance (MR) images of peripheral arterial disease (PAD) occlusive lesions.

Methods

Multi-contrast MR images (T2-weighted and ultrashort echo time) were acquired from lesions harvested from six amputated legs with high isotropic spatial resolution (0.078 mm and 0.156 mm, respectively) at 9.4 T. A total of 4014 pseudo-color combined images were generated, with 75% used to train a VAE employing custom 2D CNN layers. A Gaussian mixture model (GMM) was employed to classify the latent space data into four tissue classes: I) concentric calcified (c), II) eccentric calcified (e), III) occluded with hard tissue (h) and IV) occluded with soft tissue (s). Test image probabilities, encoded by the trained VAE were used to evaluate model performance.

Results

GMM component classification probabilities ranged from 0.92 to 0.97 for class (c), 1.00 for class (e), 0.82–0.95 for class (h) and 0.56–0.93 for the remaining class (s). Due to the complexity of soft-tissue lesions reflected in the heterogeneity of the pseudo-color images, more GMM components (n=17) were attributed to class (s), compared to the other three (c, e and h) (n=6).

Conclusion

Combination of 2D CNN VAE and GMM achieves high classification probabilities for hard tissue-containing lesions. Automatic recognition of these classes may aid therapeutic decision-making and identifying uncrossable lesions prior to endovascular intervention.

目的研究结合变异自动编码器(VAE)和二维卷积神经网络(CNN)的深度学习算法自动量化外周动脉疾病(PAD)闭塞病变的多对比磁共振(MR)图像中硬质组织的存在和形态的可行性。方法在 9.4 T 下以高各向同性空间分辨率(分别为 0.078 毫米和 0.156 毫米)从六条截肢腿上采集病变部位的多对比 MR 图像(T2 加权和超短回波时间)。共生成了 4014 张伪彩色合成图像,其中 75% 用于训练采用自定义二维 CNN 层的 VAE。采用高斯混合模型(GMM)将潜空间数据分为四类组织:I)同心钙化(c);II)偏心钙化(e);III)硬组织闭塞(h);IV)软组织闭塞(s)。由训练有素的 VAE 编码的测试图像概率用于评估模型性能。结果GMM 组件分类概率范围为:类(c)0.92-0.97,类(e)1.00,类(h)0.82-0.95,其余类(s)0.56-0.93。由于伪彩色图像的异质性反映了软组织病变的复杂性,与其他三个类别(c、e 和 h)(n=6)相比,更多的 GMM 成分(n=17)被归入类别(s)。这些类别的自动识别可帮助治疗决策,并在血管内介入治疗前识别不可穿越的病变。
{"title":"Unsupervised classification of multi-contrast magnetic resonance histology of peripheral arterial disease lesions using a convolutional variational autoencoder with a Gaussian mixture model in latent space: A technical feasibility study","authors":"Judit Csore ,&nbsp;Trisha L. Roy ,&nbsp;Graham Wright ,&nbsp;Christof Karmonik","doi":"10.1016/j.compmedimag.2024.102372","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102372","url":null,"abstract":"<div><h3>Purpose</h3><p>To investigate the feasibility of a deep learning algorithm combining variational autoencoder (VAE) and two-dimensional (2D) convolutional neural networks (CNN) for automatically quantifying hard tissue presence and morphology in multi-contrast magnetic resonance (MR) images of peripheral arterial disease (PAD) occlusive lesions.</p></div><div><h3>Methods</h3><p>Multi-contrast MR images (T2-weighted and ultrashort echo time) were acquired from lesions harvested from six amputated legs with high isotropic spatial resolution (0.078 mm and 0.156 mm, respectively) at 9.4 T. A total of 4014 pseudo-color combined images were generated, with 75% used to train a VAE employing custom 2D CNN layers. A Gaussian mixture model (GMM) was employed to classify the latent space data into four tissue classes: I) concentric calcified (c), II) eccentric calcified (e), III) occluded with hard tissue (h) and IV) occluded with soft tissue (s). Test image probabilities, encoded by the trained VAE were used to evaluate model performance.</p></div><div><h3>Results</h3><p>GMM component classification probabilities ranged from 0.92 to 0.97 for class (c), 1.00 for class (e), 0.82–0.95 for class (h) and 0.56–0.93 for the remaining class (s). Due to the complexity of soft-tissue lesions reflected in the heterogeneity of the pseudo-color images, more GMM components (n=17) were attributed to class (s), compared to the other three (c, e and h) (n=6).</p></div><div><h3>Conclusion</h3><p>Combination of 2D CNN VAE and GMM achieves high classification probabilities for hard tissue-containing lesions. Automatic recognition of these classes may aid therapeutic decision-making and identifying uncrossable lesions prior to endovascular intervention.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140347178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Annotation-free prediction of treatment-specific tissue outcome from 4D CT perfusion imaging in acute ischemic stroke 从急性缺血性脑卒中四维 CT 灌注成像无注释预测特异性组织治疗结果
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-23 DOI: 10.1016/j.compmedimag.2024.102376
Alejandro Gutierrez , Kimberly Amador , Anthony Winder , Matthias Wilms , Jens Fiehler , Nils D. Forkert

Acute ischemic stroke is a critical health condition that requires timely intervention. Following admission, clinicians typically use perfusion imaging to facilitate treatment decision-making. While deep learning models leveraging perfusion data have demonstrated the ability to predict post-treatment tissue infarction for individual patients, predictions are often represented as binary or probabilistic masks that are not straightforward to interpret or easy to obtain. Moreover, these models typically rely on large amounts of subjectively segmented data and non-standard perfusion analysis techniques. To address these challenges, we propose a novel deep learning approach that directly predicts follow-up computed tomography images from full spatio-temporal 4D perfusion scans through a temporal compression. The results show that this method leads to realistic follow-up image predictions containing the infarcted tissue outcomes. The proposed compression method achieves comparable prediction results to using perfusion maps as inputs but without the need for perfusion analysis or arterial input function selection. Additionally, separate models trained on 45 patients treated with thrombolysis and 102 treated with thrombectomy showed that each model correctly captured the different patient-specific treatment effects as shown by image difference maps. The findings of this work clearly highlight the potential of our method to provide interpretable stroke treatment decision support without requiring manual annotations.

急性缺血性脑卒中是一种需要及时干预的危重症。入院后,临床医生通常会使用灌注成像来帮助做出治疗决策。虽然利用灌注数据的深度学习模型已证明有能力预测个别患者治疗后的组织梗死,但预测结果通常表现为二元或概率掩码,并不能直接解释,也不容易获得。此外,这些模型通常依赖于大量主观分割的数据和非标准的灌注分析技术。为了应对这些挑战,我们提出了一种新颖的深度学习方法,通过时间压缩,直接预测来自全时空 4D 灌注扫描的后续计算机断层扫描图像。结果表明,这种方法能预测出包含梗死组织结果的真实随访图像。所提出的压缩方法与使用灌注图作为输入的预测结果相当,但无需进行灌注分析或动脉输入函数选择。此外,对 45 名接受溶栓治疗的患者和 102 名接受血栓切除术治疗的患者分别训练的模型显示,每个模型都能正确捕捉图像差异图所显示的不同患者的治疗效果。这项工作的研究结果清楚地凸显了我们的方法在提供可解释的中风治疗决策支持方面的潜力,而无需人工注释。
{"title":"Annotation-free prediction of treatment-specific tissue outcome from 4D CT perfusion imaging in acute ischemic stroke","authors":"Alejandro Gutierrez ,&nbsp;Kimberly Amador ,&nbsp;Anthony Winder ,&nbsp;Matthias Wilms ,&nbsp;Jens Fiehler ,&nbsp;Nils D. Forkert","doi":"10.1016/j.compmedimag.2024.102376","DOIUrl":"10.1016/j.compmedimag.2024.102376","url":null,"abstract":"<div><p>Acute ischemic stroke is a critical health condition that requires timely intervention. Following admission, clinicians typically use perfusion imaging to facilitate treatment decision-making. While deep learning models leveraging perfusion data have demonstrated the ability to predict post-treatment tissue infarction for individual patients, predictions are often represented as binary or probabilistic masks that are not straightforward to interpret or easy to obtain. Moreover, these models typically rely on large amounts of subjectively segmented data and non-standard perfusion analysis techniques. To address these challenges, we propose a novel deep learning approach that directly predicts follow-up computed tomography images from full spatio-temporal 4D perfusion scans through a temporal compression. The results show that this method leads to realistic follow-up image predictions containing the infarcted tissue outcomes. The proposed compression method achieves comparable prediction results to using perfusion maps as inputs but without the need for perfusion analysis or arterial input function selection. Additionally, separate models trained on 45 patients treated with thrombolysis and 102 treated with thrombectomy showed that each model correctly captured the different patient-specific treatment effects as shown by image difference maps. The findings of this work clearly highlight the potential of our method to provide interpretable stroke treatment decision support without requiring manual annotations.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140273444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel center-based deep contrastive metric learning method for the detection of polymicrogyria in pediatric brain MRI 用于检测小儿脑部核磁共振成像多微结构的基于中心的新型深度对比度学习方法
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-21 DOI: 10.1016/j.compmedimag.2024.102373
Lingfeng Zhang , Nishard Abdeen , Jochen Lang

Polymicrogyria (PMG) is a disorder of cortical organization mainly seen in children, which can be associated with seizures, developmental delay and motor weakness. PMG is typically diagnosed on magnetic resonance imaging (MRI) but some cases can be challenging to detect even for experienced radiologists. In this study, we create an open pediatric MRI dataset (PPMR) containing both PMG and control cases from the Children’s Hospital of Eastern Ontario (CHEO), Ottawa, Canada. The differences between PMG and control MRIs are subtle and the true distribution of the features of the disease is unknown. This makes automatic detection of potential PMG cases in MRI difficult. To enable the automatic detection of potential PMG cases, we propose an anomaly detection method based on a novel center-based deep contrastive metric learning loss function (cDCM). Despite working with a small and imbalanced dataset our method achieves 88.07% recall at 71.86% precision. This will facilitate a computer-aided tool for radiologists to select potential PMG MRIs. To the best of our knowledge, our research is the first to apply machine learning techniques to identify PMG solely from MRI.

Our code is available at: https://github.com/RichardChangCA/Deep-Contrastive-Metric-Learning-Method-to-Detect-Polymicrogyria-in-Pediatric-Brain-MRI.

Our pediatric MRI dataset is available at: https://www.kaggle.com/datasets/lingfengzhang/pediatric-polymicrogyria-mri-dataset.

多小脑症(PMG)是一种大脑皮层组织障碍,主要见于儿童,可伴有癫痫发作、发育迟缓和运动无力。多小脑症通常通过磁共振成像(MRI)诊断,但有些病例即使是经验丰富的放射科医生也很难发现。在本研究中,我们创建了一个开放式儿科磁共振成像数据集(PPMR),其中包含来自加拿大渥太华东安大略省儿童医院(CHEO)的 PMG 和对照病例。原发性骨髓增生异常综合征与对照组核磁共振成像之间的差异很微妙,而且该疾病特征的真实分布情况尚不清楚。这使得在磁共振成像中自动检测潜在的 PMG 病例变得困难。为了能够自动检测潜在的 PMG 病例,我们提出了一种异常检测方法,该方法基于一种新颖的基于中心的深度对比度度量学习损失函数(cDCM)。尽管使用的数据集较小且不平衡,但我们的方法仍实现了 88.07% 的召回率和 71.86% 的精确率。这将为放射科医生选择潜在的 PMG MRI 提供计算机辅助工具。据我们所知,我们的研究是首次应用机器学习技术仅从核磁共振成像中识别PMG。我们的代码可在以下网址获取:https://github.com/RichardChangCA/Deep-Contrastive-Metric-Learning-Method-to-Detect-Polymicrogyria-in-Pediatric-Brain-MRI.Our 儿科核磁共振成像数据集可在以下网址获取:https://www.kaggle.com/datasets/lingfengzhang/pediatric-polymicrogyria-mri-dataset。
{"title":"A novel center-based deep contrastive metric learning method for the detection of polymicrogyria in pediatric brain MRI","authors":"Lingfeng Zhang ,&nbsp;Nishard Abdeen ,&nbsp;Jochen Lang","doi":"10.1016/j.compmedimag.2024.102373","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102373","url":null,"abstract":"<div><p>Polymicrogyria (PMG) is a disorder of cortical organization mainly seen in children, which can be associated with seizures, developmental delay and motor weakness. PMG is typically diagnosed on magnetic resonance imaging (MRI) but some cases can be challenging to detect even for experienced radiologists. In this study, we create an open pediatric MRI dataset (PPMR) containing both PMG and control cases from the Children’s Hospital of Eastern Ontario (CHEO), Ottawa, Canada. The differences between PMG and control MRIs are subtle and the true distribution of the features of the disease is unknown. This makes automatic detection of potential PMG cases in MRI difficult. To enable the automatic detection of potential PMG cases, we propose an anomaly detection method based on a novel center-based deep contrastive metric learning loss function (cDCM). Despite working with a small and imbalanced dataset our method achieves 88.07% recall at 71.86% precision. This will facilitate a computer-aided tool for radiologists to select potential PMG MRIs. To the best of our knowledge, our research is the first to apply machine learning techniques to identify PMG solely from MRI.</p><p>Our code is available at: <span>https://github.com/RichardChangCA/Deep-Contrastive-Metric-Learning-Method-to-Detect-Polymicrogyria-in-Pediatric-Brain-MRI</span><svg><path></path></svg>.</p><p>Our pediatric MRI dataset is available at: <span>https://www.kaggle.com/datasets/lingfengzhang/pediatric-polymicrogyria-mri-dataset</span><svg><path></path></svg>.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140190733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi modality fusion transformer with spatio-temporal feature aggregation module for psychiatric disorder diagnosis 多模态融合转换器与时空特征聚合模块用于精神障碍诊断
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-19 DOI: 10.1016/j.compmedimag.2024.102368
Guoxin Wang , Fengmei Fan , Sheng Shi , Shan An , Xuyang Cao , Wenshu Ge , Feng Yu , Qi Wang , Xiaole Han , Shuping Tan , Yunlong Tan , Zhiren Wang

Bipolar disorder (BD) is characterized by recurrent episodes of depression and mild mania. In this paper, to address the common issue of insufficient accuracy in existing methods and meet the requirements of clinical diagnosis, we propose a framework called Spatio-temporal Feature Fusion Transformer (STF2Former). It improves on our previous work — MFFormer by introducing a Spatio-temporal Feature Aggregation Module (STFAM) to learn the temporal and spatial features of rs-fMRI data. It promotes intra-modality attention and information fusion across different modalities. Specifically, this method decouples the temporal and spatial dimensions and designs two feature extraction modules for extracting temporal and spatial information separately. Extensive experiments demonstrate the effectiveness of our proposed STFAM in extracting features from rs-fMRI, and prove that our STF2Former can significantly outperform MFFormer and achieve much better results among other state-of-the-art methods.

双相情感障碍(BD)以反复发作的抑郁和轻度躁狂为特征。为了解决现有方法准确性不足的普遍问题并满足临床诊断的要求,我们在本文中提出了一种名为时空特征融合转换器(STF2Former)的框架。它改进了我们之前的工作--MFFormer,引入了时空特征聚合模块(STFAM)来学习 rs-fMRI 数据的时空特征。它促进了不同模态之间的模态内关注和信息融合。具体来说,该方法将时间和空间维度解耦,并设计了两个特征提取模块,分别提取时间和空间信息。广泛的实验证明了我们提出的 STFAM 在从 rs-fMRI 提取特征方面的有效性,并证明我们的 STF2Former 可以显著超越 MFFormer,在其他最先进的方法中取得更好的结果。
{"title":"Multi modality fusion transformer with spatio-temporal feature aggregation module for psychiatric disorder diagnosis","authors":"Guoxin Wang ,&nbsp;Fengmei Fan ,&nbsp;Sheng Shi ,&nbsp;Shan An ,&nbsp;Xuyang Cao ,&nbsp;Wenshu Ge ,&nbsp;Feng Yu ,&nbsp;Qi Wang ,&nbsp;Xiaole Han ,&nbsp;Shuping Tan ,&nbsp;Yunlong Tan ,&nbsp;Zhiren Wang","doi":"10.1016/j.compmedimag.2024.102368","DOIUrl":"10.1016/j.compmedimag.2024.102368","url":null,"abstract":"<div><p>Bipolar disorder (BD) is characterized by recurrent episodes of depression and mild mania. In this paper, to address the common issue of insufficient accuracy in existing methods and meet the requirements of clinical diagnosis, we propose a framework called Spatio-temporal Feature Fusion Transformer (STF2Former). It improves on our previous work — MFFormer by introducing a Spatio-temporal Feature Aggregation Module (STFAM) to learn the temporal and spatial features of rs-fMRI data. It promotes intra-modality attention and information fusion across different modalities. Specifically, this method decouples the temporal and spatial dimensions and designs two feature extraction modules for extracting temporal and spatial information separately. Extensive experiments demonstrate the effectiveness of our proposed STFAM in extracting features from rs-fMRI, and prove that our STF2Former can significantly outperform MFFormer and achieve much better results among other state-of-the-art methods.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140182275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SSTU: Swin-Spectral Transformer U-Net for hyperspectral whole slide image reconstruction SSTU:用于高光谱全切片图像重建的斯温-光谱变换器 U-Net
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-16 DOI: 10.1016/j.compmedimag.2024.102367
Yukun Wang , Yanfeng Gu , Abiyasi Nanding

Whole Slide Imaging and Hyperspectral Microscopic Imaging provide great quality data with high spatial and spectral resolution for histopathology. Existing Hyperspectral Whole Slide Imaging systems combine the advantages of the techniques above, thus providing rich information for pathological diagnosis. However, it cannot avoid the problems of slow acquisition speed and mass data storage demand. Inspired by the spectral reconstruction task in computer vision and remote sensing, the Swin-Spectral Transformer U-Net (SSTU) has been developed to reconstruct Hyperspectral Whole Slide images (HWSis) from multiple Hyperspectral Microscopic images (HMis) of small Field of View and Whole Slide images (WSis). The Swin-Spectral Transformer (SST) module in SSTU takes full advantage of Transformer in extracting global attention. Firstly, Swin Transformer is exploited in space domain, which overcomes the high computation cost in Vision Transformer structures, while it maintains the spatial features extracted from WSis. Furthermore, Spectral Transformer is exploited to collect the long-range spectral features in HMis. Combined with the multi-scale encoder-bottleneck-decoder structure of U-Net, SSTU network is formed by sequential and symmetric residual connections of SSTs, which reconstructs a selected area of HWSi from coarse to fine. Qualitative and quantitative experiments prove the performance of SSTU in HWSi reconstruction task superior to other state-of-the-art spectral reconstruction methods.

整片成像和高光谱显微成像可为组织病理学提供高质量、高空间分辨率和高光谱分辨率的数据。现有的高光谱整片成像系统结合了上述技术的优点,从而为病理诊断提供了丰富的信息。然而,它无法避免采集速度慢和大量数据存储需求的问题。受计算机视觉和遥感中光谱重建任务的启发,Swin-Spectral Transformer U-Net(SSTU)被开发出来,用于从多个小视场高光谱显微图像(HMis)和全切片图像(WSis)重建高光谱全切片图像(HWSis)。SSTU 中的斯温-光谱变换器(SST)模块充分利用了变换器在提取全局注意力方面的优势。首先,在空间域利用斯温变换器,克服了视觉变换器结构的高计算成本,同时保留了从 WSis 提取的空间特征。此外,还利用光谱变换器收集 HMis 中的长距离光谱特征。结合 U-Net 的多尺度编码器-瓶颈-解码器结构,通过 SST 的顺序和对称残差连接形成 SSTU 网络,从而从粗到细地重建 HWSi 的选定区域。定性和定量实验证明,SSTU 在 HWSi 重建任务中的性能优于其他最先进的频谱重建方法。
{"title":"SSTU: Swin-Spectral Transformer U-Net for hyperspectral whole slide image reconstruction","authors":"Yukun Wang ,&nbsp;Yanfeng Gu ,&nbsp;Abiyasi Nanding","doi":"10.1016/j.compmedimag.2024.102367","DOIUrl":"10.1016/j.compmedimag.2024.102367","url":null,"abstract":"<div><p>Whole Slide Imaging and Hyperspectral Microscopic Imaging provide great quality data with high spatial and spectral resolution for histopathology. Existing Hyperspectral Whole Slide Imaging systems combine the advantages of the techniques above, thus providing rich information for pathological diagnosis. However, it cannot avoid the problems of slow acquisition speed and mass data storage demand. Inspired by the spectral reconstruction task in computer vision and remote sensing, the Swin-Spectral Transformer U-Net (SSTU) has been developed to reconstruct Hyperspectral Whole Slide images (HWSis) from multiple Hyperspectral Microscopic images (HMis) of small Field of View and Whole Slide images (WSis). The Swin-Spectral Transformer (SST) module in SSTU takes full advantage of Transformer in extracting global attention. Firstly, Swin Transformer is exploited in space domain, which overcomes the high computation cost in Vision Transformer structures, while it maintains the spatial features extracted from WSis. Furthermore, Spectral Transformer is exploited to collect the long-range spectral features in HMis. Combined with the multi-scale encoder-bottleneck-decoder structure of U-Net, SSTU network is formed by sequential and symmetric residual connections of SSTs, which reconstructs a selected area of HWSi from coarse to fine. Qualitative and quantitative experiments prove the performance of SSTU in HWSi reconstruction task superior to other state-of-the-art spectral reconstruction methods.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140182277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MEF-UNet: An end-to-end ultrasound image segmentation algorithm based on multi-scale feature extraction and fusion MEF-UNet:基于多尺度特征提取和融合的端到端超声图像分割算法
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-16 DOI: 10.1016/j.compmedimag.2024.102370
Mengqi Xu , Qianting Ma , Huajie Zhang , Dexing Kong , Tieyong Zeng

Ultrasound image segmentation is a challenging task due to the complexity of lesion types, fuzzy boundaries, and low-contrast images along with the presence of noises and artifacts. To address these issues, we propose an end-to-end multi-scale feature extraction and fusion network (MEF-UNet) for the automatic segmentation of ultrasound images. Specifically, we first design a selective feature extraction encoder, including detail extraction stage and structure extraction stage, to precisely capture the edge details and overall shape features of the lesions. In order to enhance the representation capacity of contextual information, we develop a context information storage module in the skip-connection section, responsible for integrating information from adjacent two-layer feature maps. In addition, we design a multi-scale feature fusion module in the decoder section to merge feature maps with different scales. Experimental results indicate that our MEF-UNet can significantly improve the segmentation results in both quantitative analysis and visual effects.

由于病变类型复杂、边界模糊、图像对比度低以及存在噪声和伪影,超声图像分割是一项具有挑战性的任务。为解决这些问题,我们提出了一种端到端的多尺度特征提取和融合网络(MEF-UNet),用于超声图像的自动分割。具体来说,我们首先设计了一个选择性特征提取编码器,包括细节提取阶段和结构提取阶段,以精确捕捉病变的边缘细节和整体形状特征。为了增强上下文信息的表示能力,我们在跳接部分开发了上下文信息存储模块,负责整合相邻两层特征图的信息。此外,我们还在解码器部分设计了一个多尺度特征融合模块,用于合并不同尺度的特征图。实验结果表明,我们的 MEF-UNet 在定量分析和视觉效果方面都能显著改善分割结果。
{"title":"MEF-UNet: An end-to-end ultrasound image segmentation algorithm based on multi-scale feature extraction and fusion","authors":"Mengqi Xu ,&nbsp;Qianting Ma ,&nbsp;Huajie Zhang ,&nbsp;Dexing Kong ,&nbsp;Tieyong Zeng","doi":"10.1016/j.compmedimag.2024.102370","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102370","url":null,"abstract":"<div><p>Ultrasound image segmentation is a challenging task due to the complexity of lesion types, fuzzy boundaries, and low-contrast images along with the presence of noises and artifacts. To address these issues, we propose an end-to-end multi-scale feature extraction and fusion network (MEF-UNet) for the automatic segmentation of ultrasound images. Specifically, we first design a selective feature extraction encoder, including detail extraction stage and structure extraction stage, to precisely capture the edge details and overall shape features of the lesions. In order to enhance the representation capacity of contextual information, we develop a context information storage module in the skip-connection section, responsible for integrating information from adjacent two-layer feature maps. In addition, we design a multi-scale feature fusion module in the decoder section to merge feature maps with different scales. Experimental results indicate that our MEF-UNet can significantly improve the segmentation results in both quantitative analysis and visual effects.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140180654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A semi-supervised multiview-MRI network for the detection of Knee Osteoarthritis 用于检测膝骨关节炎的半监督多视图-MRI 网络
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-16 DOI: 10.1016/j.compmedimag.2024.102371
Mohamed Berrimi , Didier Hans , Rachid Jennane

Knee OsteoArthritis (OA) is a prevalent chronic condition, affecting a significant proportion of the global population. Detecting knee OA is crucial as the degeneration of the knee joint is irreversible. In this paper, we introduce a semi-supervised multi-view framework and a 3D CNN model for detecting knee OA using 3D Magnetic Resonance Imaging (MRI) scans. We introduce a semi-supervised learning approach combining labeled and unlabeled data to improve the performance and generalizability of the proposed model. Experimental results show the efficacy of our proposed approach in detecting knee OA from 3D MRI scans using a large cohort of 4297 subjects. An ablation study was conducted to investigate the contributions of various components of the proposed model, providing insights into the optimal design of the model. Our results indicate the potential of the proposed approach to improve the accuracy and efficiency of OA diagnosis. The proposed framework reported an AUC of 93.20% for the detection of knee OA.

膝关节骨性关节炎(OA)是一种普遍存在的慢性疾病,影响着全球很大一部分人口。由于膝关节的退化是不可逆的,因此检测膝关节骨性关节炎至关重要。本文介绍了一种半监督多视角框架和三维 CNN 模型,用于利用三维核磁共振成像(MRI)扫描检测膝关节 OA。我们引入了一种结合标记和非标记数据的半监督学习方法,以提高所提模型的性能和通用性。实验结果表明,我们提出的方法在使用 4297 名受试者组成的大型群组从三维核磁共振成像扫描中检测膝关节 OA 方面效果显著。我们还进行了一项消融研究,以调查拟议模型各组成部分的贡献,为模型的优化设计提供启示。我们的研究结果表明,所提出的方法具有提高 OA 诊断准确性和效率的潜力。在膝关节 OA 的检测中,拟议框架的 AUC 为 93.20%。
{"title":"A semi-supervised multiview-MRI network for the detection of Knee Osteoarthritis","authors":"Mohamed Berrimi ,&nbsp;Didier Hans ,&nbsp;Rachid Jennane","doi":"10.1016/j.compmedimag.2024.102371","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102371","url":null,"abstract":"<div><p>Knee OsteoArthritis (OA) is a prevalent chronic condition, affecting a significant proportion of the global population. Detecting knee OA is crucial as the degeneration of the knee joint is irreversible. In this paper, we introduce a semi-supervised multi-view framework and a 3D CNN model for detecting knee OA using 3D Magnetic Resonance Imaging (MRI) scans. We introduce a semi-supervised learning approach combining labeled and unlabeled data to improve the performance and generalizability of the proposed model. Experimental results show the efficacy of our proposed approach in detecting knee OA from 3D MRI scans using a large cohort of 4297 subjects. An ablation study was conducted to investigate the contributions of various components of the proposed model, providing insights into the optimal design of the model. Our results indicate the potential of the proposed approach to improve the accuracy and efficiency of OA diagnosis. The proposed framework reported an AUC of 93.20% for the detection of knee OA.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140163720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Vessel Segmentation with Multi-Task Learning and Auxiliary Data Available Only During Model Training 利用多任务学习和仅在模型训练期间可用的辅助数据改进血管分割
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-16 DOI: 10.1016/j.compmedimag.2024.102369
Daniel Sobotka , Alexander Herold , Matthias Perkonigg , Lucian Beer , Nina Bastati , Alina Sablatnig , Ahmed Ba-Ssalamah , Georg Langs

Liver vessel segmentation in magnetic resonance imaging data is important for the computational analysis of vascular remodeling, associated with a wide spectrum of diffuse liver diseases. Existing approaches rely on contrast enhanced imaging data, but the necessary dedicated imaging sequences are not uniformly acquired. Images without contrast enhancement are acquired more frequently, but vessel segmentation is challenging, and requires large-scale annotated data. We propose a multi-task learning framework to segment vessels in liver MRI without contrast. It exploits auxiliary contrast enhanced MRI data available only during training to reduce the need for annotated training examples. Our approach draws on paired native and contrast enhanced data with and without vessel annotations for model training. Results show that auxiliary data improves the accuracy of vessel segmentation, even if they are not available during inference. The advantage is most pronounced if only few annotations are available for training, since the feature representation benefits from the shared task structure. A validation of this approach to augment a model for brain tumor segmentation confirms its benefits across different domains. An auxiliary informative imaging modality can augment expert annotations even if it is only available during training.

磁共振成像数据中的肝脏血管分割对于计算分析与多种弥漫性肝病相关的血管重塑非常重要。现有方法依赖于对比度增强成像数据,但必要的专用成像序列并不是统一获取的。无对比度增强的图像获取频率更高,但血管分割具有挑战性,需要大规模的注释数据。我们提出了一种多任务学习框架,用于分割无对比度的肝脏磁共振成像中的血管。它利用仅在训练期间可用的辅助对比增强 MRI 数据,减少了对注释训练示例的需求。我们的方法利用有血管注释和无血管注释的成对原始数据和对比度增强数据进行模型训练。结果表明,即使在推理过程中没有辅助数据,辅助数据也能提高血管分割的准确性。如果只有很少的注释数据可用于训练,那么辅助数据的优势会更加明显,因为特征表示可以从共享的任务结构中获益。对这种用于增强脑肿瘤分割模型的方法进行的验证证实了它在不同领域的优势。辅助信息成像模式可以增强专家注释,即使它只能在训练过程中使用。
{"title":"Improving Vessel Segmentation with Multi-Task Learning and Auxiliary Data Available Only During Model Training","authors":"Daniel Sobotka ,&nbsp;Alexander Herold ,&nbsp;Matthias Perkonigg ,&nbsp;Lucian Beer ,&nbsp;Nina Bastati ,&nbsp;Alina Sablatnig ,&nbsp;Ahmed Ba-Ssalamah ,&nbsp;Georg Langs","doi":"10.1016/j.compmedimag.2024.102369","DOIUrl":"10.1016/j.compmedimag.2024.102369","url":null,"abstract":"<div><p>Liver vessel segmentation in magnetic resonance imaging data is important for the computational analysis of vascular remodeling, associated with a wide spectrum of diffuse liver diseases. Existing approaches rely on contrast enhanced imaging data, but the necessary dedicated imaging sequences are not uniformly acquired. Images without contrast enhancement are acquired more frequently, but vessel segmentation is challenging, and requires large-scale annotated data. We propose a multi-task learning framework to segment vessels in liver MRI without contrast. It exploits auxiliary contrast enhanced MRI data available only during training to reduce the need for annotated training examples. Our approach draws on paired native and contrast enhanced data with and without vessel annotations for model training. Results show that auxiliary data improves the accuracy of vessel segmentation, even if they are not available during inference. The advantage is most pronounced if only few annotations are available for training, since the feature representation benefits from the shared task structure. A validation of this approach to augment a model for brain tumor segmentation confirms its benefits across different domains. An auxiliary informative imaging modality can augment expert annotations even if it is only available during training.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124000466/pdfft?md5=1b424839be9b9c32109a0d0f6a4f10e1&pid=1-s2.0-S0895611124000466-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140182296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deformable registration of preoperative MR and intraoperative long-length tomosynthesis images for guidance of spine surgery via image synthesis 通过图像合成对术前磁共振和术中长断层扫描图像进行可变形配准,为脊柱手术提供指导
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-06 DOI: 10.1016/j.compmedimag.2024.102365
Yixuan Huang , Xiaoxuan Zhang , Yicheng Hu , Ashley R. Johnston , Craig K. Jones , Wojciech B. Zbijewski , Jeffrey H. Siewerdsen , Patrick A. Helm , Timothy F. Witham , Ali Uneri

Purpose

Improved integration and use of preoperative imaging during surgery hold significant potential for enhancing treatment planning and instrument guidance through surgical navigation. Despite its prevalent use in diagnostic settings, MR imaging is rarely used for navigation in spine surgery. This study aims to leverage MR imaging for intraoperative visualization of spine anatomy, particularly in cases where CT imaging is unavailable or when minimizing radiation exposure is essential, such as in pediatric surgery.

Methods

This work presents a method for deformable 3D-2D registration of preoperative MR images with a novel intraoperative long-length tomosynthesis imaging modality (viz., Long-Film [LF]). A conditional generative adversarial network is used to translate MR images to an intermediate bone image suitable for registration, followed by a model-based 3D-2D registration algorithm to deformably map the synthesized images to LF images. The algorithm’s performance was evaluated on cadaveric specimens with implanted markers and controlled deformation, and in clinical images of patients undergoing spine surgery as part of a large-scale clinical study on LF imaging.

Results

The proposed method yielded a median 2D projection distance error of 2.0 mm (interquartile range [IQR]: 1.1–3.3 mm) and a 3D target registration error of 1.5 mm (IQR: 0.8–2.1 mm) in cadaver studies. Notably, the multi-scale approach exhibited significantly higher accuracy compared to rigid solutions and effectively managed the challenges posed by piecewise rigid spine deformation. The robustness and consistency of the method were evaluated on clinical images, yielding no outliers on vertebrae without surgical instrumentation and 3% outliers on vertebrae with instrumentation.

Conclusions

This work constitutes the first reported approach for deformable MR to LF registration based on deep image synthesis. The proposed framework provides access to the preoperative annotations and planning information during surgery and enables surgical navigation within the context of MR images and/or dual-plane LF images.

目的在手术过程中改进术前成像的整合和使用,为通过手术导航加强治疗规划和器械引导带来巨大潜力。尽管核磁共振成像在诊断中应用广泛,但很少用于脊柱手术导航。本研究旨在利用磁共振成像对脊柱解剖结构进行术中可视化,尤其是在无法使用 CT 成像或必须尽量减少辐射暴露的情况下,例如在儿科手术中。方法本研究提出了一种将术前磁共振图像与新型术中长片断层成像模式(即长片[LF])进行可变形 3D-2D 配准的方法。利用条件生成对抗网络将核磁共振图像转换为适合配准的中间骨图像,然后利用基于模型的三维-二维配准算法将合成图像变形映射到长片图像。作为低频成像大规模临床研究的一部分,该算法的性能在植入标记和可控变形的尸体标本以及接受脊柱手术患者的临床图像上进行了评估。结果在尸体研究中,所提出的方法产生的中位二维投影距离误差为 2.0 毫米(四分位间距 [IQR]:1.1-3.3 毫米),三维目标配准误差为 1.5 毫米(IQR:0.8-2.1 毫米)。值得注意的是,与刚性解决方案相比,多尺度方法表现出更高的准确性,并有效地应对了片状刚性脊柱变形带来的挑战。在临床图像上对该方法的稳健性和一致性进行了评估,结果显示在没有手术器械的椎体上没有异常值,在有器械的椎体上异常值为 3%。所提出的框架可在手术期间访问术前注释和规划信息,并可在 MR 图像和/或双平面 LF 图像的背景下进行手术导航。
{"title":"Deformable registration of preoperative MR and intraoperative long-length tomosynthesis images for guidance of spine surgery via image synthesis","authors":"Yixuan Huang ,&nbsp;Xiaoxuan Zhang ,&nbsp;Yicheng Hu ,&nbsp;Ashley R. Johnston ,&nbsp;Craig K. Jones ,&nbsp;Wojciech B. Zbijewski ,&nbsp;Jeffrey H. Siewerdsen ,&nbsp;Patrick A. Helm ,&nbsp;Timothy F. Witham ,&nbsp;Ali Uneri","doi":"10.1016/j.compmedimag.2024.102365","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102365","url":null,"abstract":"<div><h3>Purpose</h3><p>Improved integration and use of preoperative imaging during surgery hold significant potential for enhancing treatment planning and instrument guidance through surgical navigation. Despite its prevalent use in diagnostic settings, MR imaging is rarely used for navigation in spine surgery. This study aims to leverage MR imaging for intraoperative visualization of spine anatomy, particularly in cases where CT imaging is unavailable or when minimizing radiation exposure is essential, such as in pediatric surgery.</p></div><div><h3>Methods</h3><p>This work presents a method for deformable 3D-2D registration of preoperative MR images with a novel intraoperative long-length tomosynthesis imaging modality (viz., Long-Film [LF]). A conditional generative adversarial network is used to translate MR images to an intermediate bone image suitable for registration, followed by a model-based 3D-2D registration algorithm to deformably map the synthesized images to LF images. The algorithm’s performance was evaluated on cadaveric specimens with implanted markers and controlled deformation, and in clinical images of patients undergoing spine surgery as part of a large-scale clinical study on LF imaging.</p></div><div><h3>Results</h3><p>The proposed method yielded a median 2D projection distance error of 2.0 mm (interquartile range [IQR]: 1.1–3.3 mm) and a 3D target registration error of 1.5 mm (IQR: 0.8–2.1 mm) in cadaver studies. Notably, the multi-scale approach exhibited significantly higher accuracy compared to rigid solutions and effectively managed the challenges posed by piecewise rigid spine deformation. The robustness and consistency of the method were evaluated on clinical images, yielding no outliers on vertebrae without surgical instrumentation and 3% outliers on vertebrae with instrumentation.</p></div><div><h3>Conclusions</h3><p>This work constitutes the first reported approach for deformable MR to LF registration based on deep image synthesis. The proposed framework provides access to the preoperative annotations and planning information during surgery and enables surgical navigation within the context of MR images and/or dual-plane LF images.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140095918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1