首页 > 最新文献

Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention最新文献

英文 中文
MUTUAL: Towards Holistic Sensing and Inference in the Operating Room. 互助:在手术室中实现整体感知和推理。
Julien Quarez, Yang Li, Hassna Irzan, Matthew Elliot, Oscar MacCormac, James Knigth, Martin Huber, Toktam Mahmoodi, Prokar Dasgupta, Sebastien Ourselin, Nicholas Raison, Jonathan Shapey, Alejandro Granados

Embodied AI (E-AI) in the form of intelligent surgical robotics and other agents is calling for data platforms to facilitate its development and deployment. In this work, we present a cross-platform multimodal data recording and streaming software, MUTUAL, successfully deployed on two clinical studies, along with its ROS 2 distributed adaptation, MUTUAL-ROS 2. We describe and compare the two implementations of MUTUAL through their recording performance under different settings. MUTUAL offers robust recording performance at target configurations for multiple modalities, including video, audio, and live expert commentary. While this recording performance is not matched by MUTUAL-ROS 2, we demonstrate its advantages related to real-time streaming capabilities for AI inference and more horizontal scalability, key aspects for E-AI systems in the operating room. Our findings demonstrate that the baseline MUTUAL is well-suited for data curation and offline analysis, whereas MUTUAL-ROS 2, should match the recording reliability of the baseline system under a fully distributed manner where modalities are handled independently by edge computing devices. These insights are critical for advancing the integration of E-AI in surgical practice, ensuring that data infrastructure can support both robust recording and real-time processing needs.

智能手术机器人和其他代理形式的嵌入式人工智能(E-AI)需要数据平台来促进其发展和部署。在这项工作中,我们介绍了一款跨平台多模态数据记录和流软件 MUTUAL,该软件已在两项临床研究中成功部署,同时还介绍了其 ROS 2 分布式适配软件 MUTUAL-ROS2。 我们通过不同设置下的记录性能来描述和比较 MUTUAL 的两种实现方式。MUTUAL 在多种模式的目标配置下提供了强大的记录性能,包括视频、音频和现场专家评论。虽然 MUTUAL-ROS 2 的录制性能无法与 MUTUAL-ROS 2 相提并论,但我们展示了它在人工智能推理的实时流功能和更多横向扩展性方面的优势,这些都是手术室中电子人工智能系统的关键所在。我们的研究结果表明,基线 MUTUAL 非常适合数据整理和离线分析,而 MUTUAL-ROS 2 在完全分布式的情况下应与基线系统的记录可靠性相匹配,在这种情况下,各种模式均由边缘计算设备独立处理。这些见解对于推进电子人工智能在外科实践中的整合至关重要,可确保数据基础设施同时支持强大的记录和实时处理需求。
{"title":"MUTUAL: Towards Holistic Sensing and Inference in the Operating Room.","authors":"Julien Quarez, Yang Li, Hassna Irzan, Matthew Elliot, Oscar MacCormac, James Knigth, Martin Huber, Toktam Mahmoodi, Prokar Dasgupta, Sebastien Ourselin, Nicholas Raison, Jonathan Shapey, Alejandro Granados","doi":"10.1007/978-3-031-77610-6_17","DOIUrl":"10.1007/978-3-031-77610-6_17","url":null,"abstract":"<p><p>Embodied AI (E-AI) in the form of intelligent surgical robotics and other agents is calling for data platforms to facilitate its development and deployment. In this work, we present a cross-platform multimodal data recording and streaming software, MUTUAL, successfully deployed on two clinical studies, along with its ROS 2 distributed adaptation, MUTUAL-ROS 2. We describe and compare the two implementations of MUTUAL through their recording performance under different settings. MUTUAL offers robust recording performance at target configurations for multiple modalities, including video, audio, and live expert commentary. While this recording performance is not matched by MUTUAL-ROS 2, we demonstrate its advantages related to real-time streaming capabilities for AI inference and more horizontal scalability, key aspects for E-AI systems in the operating room. Our findings demonstrate that the baseline MUTUAL is well-suited for data curation and offline analysis, whereas MUTUAL-ROS 2, should match the recording reliability of the baseline system under a fully distributed manner where modalities are handled independently by edge computing devices. These insights are critical for advancing the integration of E-AI in surgical practice, ensuring that data infrastructure can support both robust recording and real-time processing needs.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":" ","pages":"178-188"},"PeriodicalIF":0.0,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7617325/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143049392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Zoom Pattern Signatures for Fetal Ultrasound Structures. 胎儿超声结构的缩放模式特征。
Mohammad Alsharid, Robail Yasrab, Lior Drukker, Aris T Papageorghiou, J Alison Noble

During a fetal ultrasound scan, a sonographer will zoom in and zoom out as they attempt to get clearer images of the anatomical structures of interest. This paper explores how to use this zoom information which is an under-utilised piece of information that is extractable from fetal ultrasound images. We explore associating zooming patterns to specific structures. The presence of such patterns would indicate that each individual anatomical structure has a unique signature associated with it, thereby allowing for classification of fetal ultrasound clips without directly reading the actual fetal ultrasound images in a convolutional neural network.

在胎儿超声波扫描过程中,超声波技师会放大和缩小图像,试图获得更清晰的相关解剖结构图像。本文探讨了如何利用这种缩放信息,因为这种信息可从胎儿超声图像中提取,但未得到充分利用。我们探索将缩放模式与特定结构联系起来。这种模式的存在将表明每个单独的解剖结构都有与之相关的独特特征,从而可以在卷积神经网络中对胎儿超声片段进行分类,而无需直接读取实际的胎儿超声图像。
{"title":"Zoom Pattern Signatures for Fetal Ultrasound Structures.","authors":"Mohammad Alsharid, Robail Yasrab, Lior Drukker, Aris T Papageorghiou, J Alison Noble","doi":"10.1007/978-3-031-72083-3_73","DOIUrl":"10.1007/978-3-031-72083-3_73","url":null,"abstract":"<p><p>During a fetal ultrasound scan, a sonographer will zoom in and zoom out as they attempt to get clearer images of the anatomical structures of interest. This paper explores how to use this zoom information which is an under-utilised piece of information that is extractable from fetal ultrasound images. We explore associating zooming patterns to specific structures. The presence of such patterns would indicate that each individual anatomical structure has a unique signature associated with it, thereby allowing for classification of fetal ultrasound clips without directly reading the actual fetal ultrasound images in a convolutional neural network.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15004 ","pages":"786-795"},"PeriodicalIF":0.0,"publicationDate":"2024-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7616787/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142635510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimation and Analysis of Slice Propagation Uncertainty in 3D Anatomy Segmentation. 三维解剖分割中切片传播不确定性的估计与分析。
Rachaell Nihalaani, Tushar Kataria, Jadie Adams, Shireen Y Elhabian

Supervised methods for 3D anatomy segmentation demonstrate superior performance but are often limited by the availability of annotated data. This limitation has led to a growing interest in self-supervised approaches in tandem with the abundance of available unannotated data. Slice propagation has emerged as a self-supervised approach that leverages slice registration as a self-supervised task to achieve full anatomy segmentation with minimal supervision. This approach significantly reduces the need for domain expertise, time, and the cost associated with building fully annotated datasets required for training segmentation networks. However, this shift toward reduced supervision via deterministic networks raises concerns about the trustworthiness and reliability of predictions, especially when compared with more accurate supervised approaches. To address this concern, we propose integrating calibrated uncertainty quantification (UQ) into slice propagation methods, which would provide insights into the model's predictive reliability and confidence levels. Incorporating uncertainty measures enhances user confidence in self-supervised approaches, thereby improving their practical applicability. We conducted experiments on three datasets for 3D abdominal segmentation using five UQ methods. The results illustrate that incorporating UQ improves not only model trustworthiness but also segmentation accuracy. Furthermore, our analysis reveals various failure modes of slice propagation methods that might not be immediately apparent to end-users. This study opens up new research avenues to improve the accuracy and trustworthiness of slice propagation methods.

用于三维解剖分割的监督方法表现出卓越的性能,但往往受到注释数据可用性的限制。这种局限性导致人们对自监督方法以及大量可用的未注释数据越来越感兴趣。切片传播是一种自我监督方法,它利用切片配准作为一项自我监督任务,以最少的监督实现全面解剖分割。这种方法大大减少了对领域专业知识的需求、时间,以及与建立训练分割网络所需的完全注释数据集相关的成本。然而,这种通过确定性网络减少监督的转变引发了人们对预测可信度和可靠性的担忧,尤其是与更精确的监督方法相比。为了解决这个问题,我们建议将校准的不确定性量化(UQ)整合到切片传播方法中,从而深入了解模型的预测可靠性和置信度。纳入不确定性度量可增强用户对自我监督方法的信心,从而提高其实际应用性。我们在三个数据集上使用五种 UQ 方法进行了三维腹部分割实验。结果表明,纳入 UQ 不仅能提高模型的可信度,还能提高分割的准确性。此外,我们的分析还揭示了切片传播方法的各种失效模式,而这些失效模式对于最终用户来说可能并不是显而易见的。这项研究为提高切片传播方法的准确性和可信度开辟了新的研究途径。
{"title":"Estimation and Analysis of Slice Propagation Uncertainty in 3D Anatomy Segmentation.","authors":"Rachaell Nihalaani, Tushar Kataria, Jadie Adams, Shireen Y Elhabian","doi":"10.1007/978-3-031-72117-5_26","DOIUrl":"10.1007/978-3-031-72117-5_26","url":null,"abstract":"<p><p>Supervised methods for 3D anatomy segmentation demonstrate superior performance but are often limited by the availability of annotated data. This limitation has led to a growing interest in self-supervised approaches in tandem with the abundance of available unannotated data. Slice propagation has emerged as a self-supervised approach that leverages slice registration as a self-supervised task to achieve full anatomy segmentation with minimal supervision. This approach significantly reduces the need for domain expertise, time, and the cost associated with building fully annotated datasets required for training segmentation networks. However, this shift toward reduced supervision via deterministic networks raises concerns about the trustworthiness and reliability of predictions, especially when compared with more accurate supervised approaches. To address this concern, we propose integrating calibrated uncertainty quantification (UQ) into slice propagation methods, which would provide insights into the model's predictive reliability and confidence levels. Incorporating uncertainty measures enhances user confidence in self-supervised approaches, thereby improving their practical applicability. We conducted experiments on three datasets for 3D abdominal segmentation using five UQ methods. The results illustrate that incorporating UQ improves not only model trustworthiness but also segmentation accuracy. Furthermore, our analysis reveals various failure modes of slice propagation methods that might not be immediately apparent to end-users. This study opens up new research avenues to improve the accuracy and trustworthiness of slice propagation methods.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15010 ","pages":"273-285"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11520486/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142549934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Subtype and Stage Inference for Alzheimer's Disease. 阿尔茨海默病的适应性亚型和分期推断。
Xinkai Wang, Yonggang Shi

Subtype and Stage Inference (SuStaIn) is a useful Event-based Model for capturing both the temporal and the phenotypical patterns for any progressive disorders, which is essential for understanding the heterogeneous nature of such diseases. However, this model cannot capture subtypes with different progression rates with respect to predefined biomarkers with fixed events prior to inference. Therefore, we propose an adaptive algorithm for learning subtype-specific events while making subtype and stage inference. We use simulation to demonstrate the improvement with respect to various performance metrics. Finally, we provide snapshots of different levels of biomarker abnormality within different subtypes on Alzheimer's Disease (AD) data to demonstrate the effectiveness of our algorithm.

亚型和阶段推断(SuStaIn)是一种有用的基于事件的模型,用于捕获任何进行性疾病的时间和表型模式,这对于理解此类疾病的异质性至关重要。然而,该模型不能捕获具有不同进展率的亚型,相对于预定义的生物标志物,在推理之前具有固定的事件。因此,我们提出了一种自适应算法,用于在进行子类型和阶段推理的同时学习特定于子类型的事件。我们使用模拟来演示有关各种性能指标的改进。最后,我们提供了阿尔茨海默病(AD)数据中不同亚型中不同水平的生物标志物异常的快照,以证明我们的算法的有效性。
{"title":"Adaptive Subtype and Stage Inference for Alzheimer's Disease.","authors":"Xinkai Wang, Yonggang Shi","doi":"10.1007/978-3-031-72384-1_5","DOIUrl":"10.1007/978-3-031-72384-1_5","url":null,"abstract":"<p><p>Subtype and Stage Inference (SuStaIn) is a useful Event-based Model for capturing both the temporal and the phenotypical patterns for any progressive disorders, which is essential for understanding the heterogeneous nature of such diseases. However, this model cannot capture subtypes with different progression rates with respect to predefined biomarkers with fixed events prior to inference. Therefore, we propose an adaptive algorithm for learning subtype-specific events while making subtype and stage inference. We use simulation to demonstrate the improvement with respect to various performance metrics. Finally, we provide snapshots of different levels of biomarker abnormality within different subtypes on Alzheimer's Disease (AD) data to demonstrate the effectiveness of our algorithm.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15003 ","pages":"46-55"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11632966/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142815445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tagged-to-Cine MRI Sequence Synthesis via Light Spatial-Temporal Transformer. 通过光时空变换器实现标记-正片磁共振成像序列合成
Xiaofeng Liu, Fangxu Xing, Zhangxing Bian, Tomas Arias-Vergara, Paula Andrea Pérez-Toro, Andreas Maier, Maureen Stone, Jiachen Zhuo, Jerry L Prince, Jonghye Woo

Tagged magnetic resonance imaging (MRI) has been successfully used to track the motion of internal tissue points within moving organs. Typically, to analyze motion using tagged MRI, cine MRI data in the same coordinate system are acquired, incurring additional time and costs. Consequently, tagged-to-cine MR synthesis holds the potential to reduce the extra acquisition time and costs associated with cine MRI, without disrupting downstream motion analysis tasks. Previous approaches have processed each frame independently, thereby overlooking the fact that complementary information from occluded regions of the tag patterns could be present in neighboring frames exhibiting motion. Furthermore, the inconsistent visual appearance, e.g., tag fading, across frames can reduce synthesis performance. To address this, we propose an efficient framework for tagged-to-cine MR sequence synthesis, leveraging both spatial and temporal information with relatively limited data. Specifically, we follow a split-and-integral protocol to balance spatialtemporal modeling efficiency and consistency. The light spatial-temporal transformer (LiST2) is designed to exploit the local and global attention in motion sequence with relatively lightweight training parameters. The directional product relative position-time bias is adapted to make the model aware of the spatial-temporal correlation, while the shifted window is used for motion alignment. Then, a recurrent sliding fine-tuning (ReST) scheme is applied to further enhance the temporal consistency. Our framework is evaluated on paired tagged and cine MRI sequences, demonstrating superior performance over comparison methods.

标记磁共振成像(MRI)已成功用于跟踪移动器官内部组织点的运动。通常情况下,要使用标记磁共振成像分析运动,需要获取同一坐标系的 cine MRI 数据,这就需要额外的时间和成本。因此,从标记到线性磁共振合成有望减少与线性磁共振成像相关的额外采集时间和成本,同时又不会影响下游的运动分析任务。以往的方法对每一帧图像进行独立处理,从而忽略了标签图案闭塞区域的补充信息可能存在于显示运动的相邻帧图像中这一事实。此外,各帧之间不一致的视觉外观(如标签褪色)也会降低合成性能。为了解决这个问题,我们提出了一个高效的框架,利用空间和时间信息,在数据相对有限的情况下进行标记到线性 MR 序列合成。具体来说,我们采用分割-积分协议来平衡时空建模效率和一致性。轻型时空变换器(LiST2)旨在利用运动序列中的局部和全局注意力,训练参数相对较轻。通过调整方向积相对位置-时间偏置,使模型意识到时空相关性,同时使用移动窗口进行运动对齐。然后,采用循环滑动微调(ReST)方案进一步增强时间一致性。我们的框架在成对标记和电影核磁共振成像序列上进行了评估,证明其性能优于比较方法。
{"title":"Tagged-to-Cine MRI Sequence Synthesis via Light Spatial-Temporal Transformer.","authors":"Xiaofeng Liu, Fangxu Xing, Zhangxing Bian, Tomas Arias-Vergara, Paula Andrea Pérez-Toro, Andreas Maier, Maureen Stone, Jiachen Zhuo, Jerry L Prince, Jonghye Woo","doi":"10.1007/978-3-031-72104-5_67","DOIUrl":"10.1007/978-3-031-72104-5_67","url":null,"abstract":"<p><p>Tagged magnetic resonance imaging (MRI) has been successfully used to track the motion of internal tissue points within moving organs. Typically, to analyze motion using tagged MRI, cine MRI data in the same coordinate system are acquired, incurring additional time and costs. Consequently, tagged-to-cine MR synthesis holds the potential to reduce the extra acquisition time and costs associated with cine MRI, without disrupting downstream motion analysis tasks. Previous approaches have processed each frame independently, thereby overlooking the fact that complementary information from occluded regions of the tag patterns could be present in neighboring frames exhibiting motion. Furthermore, the inconsistent visual appearance, e.g., tag fading, across frames can reduce synthesis performance. To address this, we propose an efficient framework for tagged-to-cine MR sequence synthesis, leveraging both spatial and temporal information with relatively limited data. Specifically, we follow a split-and-integral protocol to balance spatialtemporal modeling efficiency and consistency. The light spatial-temporal transformer (LiST<sup>2</sup>) is designed to exploit the local and global attention in motion sequence with relatively lightweight training parameters. The directional product relative position-time bias is adapted to make the model aware of the spatial-temporal correlation, while the shifted window is used for motion alignment. Then, a recurrent sliding fine-tuning (ReST) scheme is applied to further enhance the temporal consistency. Our framework is evaluated on paired tagged and cine MRI sequences, demonstrating superior performance over comparison methods.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15007 ","pages":"701-711"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11517403/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142524019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CLEFT: Language-Image Contrastive Learning with Efficient Large Language Model and Prompt Fine-Tuning. 基于高效大语言模型和快速微调的语言-图像对比学习。
Yuexi Du, Brian Chang, Nicha C Dvornek

Recent advancements in Contrastive Language-Image Pre-training (CLIP) [21] have demonstrated notable success in self-supervised representation learning across various tasks. However, the existing CLIP-like approaches often demand extensive GPU resources and prolonged training times due to the considerable size of the model and dataset, making them poor for medical applications, in which large datasets are not always common. Meanwhile, the language model prompts are mainly manually derived from labels tied to images, potentially overlooking the richness of information within training samples. We introduce a novel language-image Contrastive Learning method with an Efficient large language model and prompt Fine-Tuning (CLEFT) that harnesses the strengths of the extensive pre-trained language and visual models. Furthermore, we present an efficient strategy for learning context-based prompts that mitigates the gap between informative clinical diagnostic data and simple class labels. Our method demonstrates state-of-the-art performance on multiple chest X-ray and mammography datasets compared with various baselines. The proposed parameter efficient framework can reduce the total trainable model size by 39% and reduce the trainable language model to only 4% compared with the current BERT encoder.

对比语言图像预训练(CLIP)的最新进展已经在各种任务的自监督表示学习中取得了显著的成功。然而,由于模型和数据集的相当大的规模,现有的类似clip的方法通常需要大量的GPU资源和较长的训练时间,这使得它们不适合医疗应用,在医疗应用中,大型数据集并不总是常见的。同时,语言模型提示主要是手动从与图像绑定的标签中获得的,可能忽略了训练样本中信息的丰富性。我们介绍了一种新的语言-图像对比学习方法,该方法利用了广泛的预训练语言和视觉模型的优势,采用高效的大语言模型和快速微调(CLEFT)。此外,我们提出了一种有效的策略来学习基于上下文的提示,以减轻信息丰富的临床诊断数据和简单的类标签之间的差距。与各种基线相比,我们的方法在多个胸部x线和乳房x线摄影数据集上展示了最先进的性能。与现有的BERT编码器相比,所提出的参数高效框架可以将可训练模型的总大小减少39%,将可训练语言模型的大小减少到4%。
{"title":"CLEFT: Language-Image Contrastive Learning with Efficient Large Language Model and Prompt Fine-Tuning.","authors":"Yuexi Du, Brian Chang, Nicha C Dvornek","doi":"10.1007/978-3-031-72390-2_44","DOIUrl":"10.1007/978-3-031-72390-2_44","url":null,"abstract":"<p><p>Recent advancements in Contrastive Language-Image Pre-training (CLIP) [21] have demonstrated notable success in self-supervised representation learning across various tasks. However, the existing CLIP-like approaches often demand extensive GPU resources and prolonged training times due to the considerable size of the model and dataset, making them poor for medical applications, in which large datasets are not always common. Meanwhile, the language model prompts are mainly manually derived from labels tied to images, potentially overlooking the richness of information within training samples. We introduce a novel language-image Contrastive Learning method with an Efficient large language model and prompt Fine-Tuning (CLEFT) that harnesses the strengths of the extensive pre-trained language and visual models. Furthermore, we present an efficient strategy for learning context-based prompts that mitigates the gap between informative clinical diagnostic data and simple class labels. Our method demonstrates state-of-the-art performance on multiple chest X-ray and mammography datasets compared with various baselines. The proposed parameter efficient framework can reduce the total trainable model size by 39% and reduce the trainable language model to only 4% compared with the current BERT encoder.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15012 ","pages":"465-475"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11709740/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142960994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-Enhanced Fusion of Structural and Functional MRI for Analyzing HIV-Associated Asymptomatic Neurocognitive Impairment. 结构性和功能性核磁共振成像的注意力增强融合,用于分析艾滋病毒相关的无症状神经认知障碍。
Yuqi Fang, Wei Wang, Qianqian Wang, Hong-Jun Li, Mingxia Liu

Asymptomatic neurocognitive impairment (ANI) is a predominant form of cognitive impairment among individuals infected with human immunodeficiency virus (HIV). The current diagnostic criteria for ANI primarily rely on subjective clinical assessments, possibly leading to different interpretations among clinicians. Some recent studies leverage structural or functional MRI containing objective biomarkers for ANI analysis, offering clinicians companion diagnostic tools. However, they mainly utilize a single imaging modality, neglecting complementary information provided by structural and functional MRI. To this end, we propose an attention-enhanced structural and functional MRI fusion (ASFF) framework for HIV-associated ANI analysis. Specifically, the ASFF first extracts data-driven and human-engineered features from structural MRI, and also captures functional MRI features via a graph isomorphism network and Transformer. A mutual cross-attention fusion module is then designed to model the underlying relationship between structural and functional MRI. Additionally, a semantic inter-modality constraint is introduced to encourage consistency of multimodal features, facilitating effective feature fusion. Experimental results on 137 subjects from an HIV-associated ANI dataset with T1-weighted MRI and resting-state functional MRI show the effectiveness of our ASFF in ANI identification. Furthermore, our method can identify both modality-shared and modality-specific brain regions, which may advance our understanding of the structural and functional pathology underlying ANI.

无症状神经认知功能障碍(ANI)是人类免疫缺陷病毒(HIV)感染者认知功能障碍的主要表现形式。目前 ANI 的诊断标准主要依赖于主观临床评估,这可能会导致临床医生之间产生不同的解释。最近的一些研究利用含有客观生物标志物的结构性或功能性磁共振成像进行 ANI 分析,为临床医生提供了辅助诊断工具。然而,这些研究主要利用单一成像模式,忽略了结构性和功能性 MRI 提供的互补信息。为此,我们提出了一种用于艾滋病相关 ANI 分析的注意力增强结构和功能 MRI 融合(ASFF)框架。具体来说,ASFF 首先从结构磁共振成像中提取数据驱动和人为设计的特征,然后通过图同构网络和 Transformer 捕捉功能磁共振成像特征。然后设计一个相互交叉关注融合模块,以模拟结构性和功能性 MRI 之间的潜在关系。此外,还引入了语义跨模态约束,以鼓励多模态特征的一致性,从而促进有效的特征融合。实验结果显示,我们的 ASFF 在 ANI 识别方面非常有效。此外,我们的方法还能识别模式共享和模式特异的脑区,这可能会促进我们对 ANI 的结构和功能病理的理解。
{"title":"Attention-Enhanced Fusion of Structural and Functional MRI for Analyzing HIV-Associated Asymptomatic Neurocognitive Impairment.","authors":"Yuqi Fang, Wei Wang, Qianqian Wang, Hong-Jun Li, Mingxia Liu","doi":"10.1007/978-3-031-72120-5_11","DOIUrl":"10.1007/978-3-031-72120-5_11","url":null,"abstract":"<p><p>Asymptomatic neurocognitive impairment (ANI) is a predominant form of cognitive impairment among individuals infected with human immunodeficiency virus (HIV). The current diagnostic criteria for ANI primarily rely on subjective clinical assessments, possibly leading to different interpretations among clinicians. Some recent studies leverage structural or functional MRI containing objective biomarkers for ANI analysis, offering clinicians companion diagnostic tools. However, they mainly utilize a single imaging modality, neglecting complementary information provided by structural and functional MRI. To this end, we propose an attention-enhanced structural and functional MRI fusion (ASFF) framework for HIV-associated ANI analysis. Specifically, the ASFF first extracts data-driven and human-engineered features from structural MRI, and also captures functional MRI features via a graph isomorphism network and Transformer. A <i>mutual cross-attention fusion module</i> is then designed to model the underlying relationship between structural and functional MRI. Additionally, a <i>semantic inter-modality constraint</i> is introduced to encourage consistency of multimodal features, facilitating effective feature fusion. Experimental results on 137 subjects from an HIV-associated ANI dataset with T1-weighted MRI and resting-state functional MRI show the effectiveness of our ASFF in ANI identification. Furthermore, our method can identify both modality-shared and modality-specific brain regions, which may advance our understanding of the structural and functional pathology underlying ANI.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15011 ","pages":"113-123"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11512738/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142516842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Physics Informed Neural Networks for Estimation of Tissue Properties from Multi-echo Configuration State MRI. 基于物理信息的神经网络在多回波构型状态MRI中估计组织特性。
Samuel I Adams-Tew, Henrik Odéen, Dennis L Parker, Cheng-Chieh Cheng, Bruno Madore, Allison Payne, Sarang Joshi

This work investigates the use of configuration state imaging together with deep neural networks to develop quantitative MRI techniques for deployment in an interventional setting. A physics modeling technique for inhomogeneous fields and heterogeneous tissues is presented and used to evaluate the theoretical capability of neural networks to estimate parameter maps from configuration state signal data. All tested normalization strategies achieved similar performance in estimating T 2 and T 2 * . Varying network architecture and data normalization had substantial impacts on estimated flip angle and T 1 , highlighting their importance in developing neural networks to solve these inverse problems. The developed signal modeling technique provides an environment that will enable the development and evaluation of physics-informed machine learning techniques for MR parameter mapping and facilitate the development of quantitative MRI techniques to inform clinical decisions during MR-guided treatments.

这项工作研究了配置状态成像与深度神经网络的使用,以开发用于介入性设置的定量MRI技术。提出了一种非均匀场和非均匀组织的物理建模技术,并用于评估神经网络从组态信号数据估计参数映射的理论能力。所有测试的归一化策略在估计t2和t2 *方面都取得了相似的性能。不同的网络结构和数据归一化对估计的翻转角和t1有实质性的影响,突出了它们在开发神经网络来解决这些逆问题中的重要性。开发的信号建模技术提供了一个环境,可以开发和评估用于MR参数映射的物理信息机器学习技术,并促进定量MRI技术的开发,以便在MR引导治疗期间为临床决策提供信息。
{"title":"Physics Informed Neural Networks for Estimation of Tissue Properties from Multi-echo Configuration State MRI.","authors":"Samuel I Adams-Tew, Henrik Odéen, Dennis L Parker, Cheng-Chieh Cheng, Bruno Madore, Allison Payne, Sarang Joshi","doi":"10.1007/978-3-031-72120-5_47","DOIUrl":"10.1007/978-3-031-72120-5_47","url":null,"abstract":"<p><p>This work investigates the use of configuration state imaging together with deep neural networks to develop quantitative MRI techniques for deployment in an interventional setting. A physics modeling technique for inhomogeneous fields and heterogeneous tissues is presented and used to evaluate the theoretical capability of neural networks to estimate parameter maps from configuration state signal data. All tested normalization strategies achieved similar performance in estimating <math> <msub><mrow><mi>T</mi></mrow> <mrow><mn>2</mn></mrow> </msub> </math> and <math> <msubsup><mrow><mi>T</mi></mrow> <mrow><mn>2</mn></mrow> <mrow><mi>*</mi></mrow> </msubsup> </math> . Varying network architecture and data normalization had substantial impacts on estimated flip angle and <math> <msub><mrow><mi>T</mi></mrow> <mrow><mn>1</mn></mrow> </msub> </math> , highlighting their importance in developing neural networks to solve these inverse problems. The developed signal modeling technique provides an environment that will enable the development and evaluation of physics-informed machine learning techniques for MR parameter mapping and facilitate the development of quantitative MRI techniques to inform clinical decisions during MR-guided treatments.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15011 ","pages":"502-511"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11653200/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142857422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-guided Knowledge-Injected Graph Neural Network for Alzheimer's Diseases. 针对阿尔茨海默病的自导式知识注入图神经网络。
Zhepeng Wang, Runxue Bao, Yawen Wu, Guodong Liu, Lei Yang, Liang Zhan, Feng Zheng, Weiwen Jiang, Yanfu Zhang

Graph neural networks (GNNs) are proficient machine learning models in handling irregularly structured data. Nevertheless, their generic formulation falls short when applied to the analysis of brain connectomes in Alzheimer's Disease (AD), necessitating the incorporation of domain-specific knowledge to achieve optimal model performance. The integration of AD-related expertise into GNNs presents a significant challenge. Current methodologies reliant on manual design often demand substantial expertise from external domain specialists to guide the development of novel models, thereby consuming considerable time and resources. To mitigate the need for manual curation, this paper introduces a novel self-guided knowledge-infused multimodal GNN to autonomously integrate domain knowledge into the model development process. We propose to conceptualize existing domain knowledge as natural language, and devise a specialized multimodal GNN framework tailored to leverage this uncurated knowledge to direct the learning of the GNN submodule, thereby enhancing its efficacy and improving prediction interpretability. To assess the effectiveness of our framework, we compile a comprehensive literature dataset comprising recent peer-reviewed publications on AD. By integrating this literature dataset with several real-world AD datasets, our experimental results illustrate the effectiveness of the proposed method in extracting curated knowledge and offering explanations on graphs for domain-specific applications. Furthermore, our approach successfully utilizes the extracted information to enhance the performance of the GNN.

图神经网络(GNN)是处理不规则结构数据的熟练机器学习模型。然而,在应用于分析阿尔茨海默病(AD)的大脑连接组时,它们的通用表述并不完善,需要结合特定领域的知识才能实现最佳模型性能。将老年痴呆症相关专业知识整合到 GNN 中是一项重大挑战。目前依赖人工设计的方法往往需要外部领域专家提供大量专业知识,以指导新型模型的开发,从而耗费大量时间和资源。为了减少对人工策划的需求,本文介绍了一种新型的自引导知识注入多模态 GNN,可自主地将领域知识整合到模型开发过程中。我们建议将现有的领域知识概念化为自然语言,并设计一个专门的多模态 GNN 框架,利用这些未经整理的知识来指导 GNN 子模块的学习,从而增强其功效并提高预测的可解释性。为了评估我们的框架的有效性,我们汇编了一个全面的文献数据集,其中包括最近发表的有关注意力缺失症的同行评议出版物。通过将该文献数据集与几个真实世界的注意力缺失症数据集进行整合,我们的实验结果表明了所提出的方法在为特定领域应用提取策划知识和提供图解方面的有效性。此外,我们的方法还成功地利用了提取的信息来提高 GNN 的性能。
{"title":"Self-guided Knowledge-Injected Graph Neural Network for Alzheimer's Diseases.","authors":"Zhepeng Wang, Runxue Bao, Yawen Wu, Guodong Liu, Lei Yang, Liang Zhan, Feng Zheng, Weiwen Jiang, Yanfu Zhang","doi":"10.1007/978-3-031-72069-7_36","DOIUrl":"10.1007/978-3-031-72069-7_36","url":null,"abstract":"<p><p>Graph neural networks (GNNs) are proficient machine learning models in handling irregularly structured data. Nevertheless, their generic formulation falls short when applied to the analysis of brain connectomes in Alzheimer's Disease (AD), necessitating the incorporation of domain-specific knowledge to achieve optimal model performance. The integration of AD-related expertise into GNNs presents a significant challenge. Current methodologies reliant on manual design often demand substantial expertise from external domain specialists to guide the development of novel models, thereby consuming considerable time and resources. To mitigate the need for manual curation, this paper introduces a novel self-guided knowledge-infused multimodal GNN to autonomously integrate domain knowledge into the model development process. We propose to conceptualize existing domain knowledge as natural language, and devise a specialized multimodal GNN framework tailored to leverage this uncurated knowledge to direct the learning of the GNN submodule, thereby enhancing its efficacy and improving prediction interpretability. To assess the effectiveness of our framework, we compile a comprehensive literature dataset comprising recent peer-reviewed publications on AD. By integrating this literature dataset with several real-world AD datasets, our experimental results illustrate the effectiveness of the proposed method in extracting curated knowledge and offering explanations on graphs for domain-specific applications. Furthermore, our approach successfully utilizes the extracted information to enhance the performance of the GNN.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15002 ","pages":"378-388"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11488260/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142484836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of Effective Connectome from Infancy to Adolescence. 从婴儿期到青春期有效连接组的发展。
Guoshi Li, Kim-Han Thung, Hoyt Taylor, Zhengwang Wu, Gang Li, Li Wang, Weili Lin, Sahar Ahmad, Pew-Thian Yap

Delineating the normative developmental profile of functional connectome is important for both standardized assessment of individual growth and early detection of diseases. However, functional connectome has been mostly studied using functional connectivity (FC), where undirected connectivity strengths are estimated from statistical correlation of resting-state functional MRI (rs-fMRI) signals. To address this limitation, we applied regression dynamic causal modeling (rDCM) to delineate the developmental trajectories of effective connectivity (EC), the directed causal influence among neuronal populations, in whole-brain networks from infancy to adolescence (0-22 years old) based on high-quality rs-fMRI data from Baby Connectome Project (BCP) and Human Connectome Project Development (HCP-D). Analysis with linear mixed model demonstrates significant age effect on the mean nodal EC which is best fit by a "U" shaped quadratic curve with minimal EC at around 2 years old. Further analysis indicates that five brain regions including the left and right cuneus, left precuneus, left supramarginal gyrus and right inferior temporal gyrus have the most significant age effect on nodal EC (p < 0.05, FDR corrected). Moreover, the frontoparietal control (FPC) network shows the fastest increase from early childhood to adolescence followed by the visual and salience networks. Our findings suggest complex nonlinear developmental profile of EC from infancy to adolescence, which may reflect dynamic structural and functional maturation during this critical growth period.

{"title":"Development of Effective Connectome from Infancy to Adolescence.","authors":"Guoshi Li, Kim-Han Thung, Hoyt Taylor, Zhengwang Wu, Gang Li, Li Wang, Weili Lin, Sahar Ahmad, Pew-Thian Yap","doi":"10.1007/978-3-031-72384-1_13","DOIUrl":"10.1007/978-3-031-72384-1_13","url":null,"abstract":"<p><p>Delineating the normative developmental profile of functional connectome is important for both standardized assessment of individual growth and early detection of diseases. However, functional connectome has been mostly studied using functional connectivity (FC), where undirected connectivity strengths are estimated from statistical correlation of resting-state functional MRI (rs-fMRI) signals. To address this limitation, we applied regression dynamic causal modeling (rDCM) to delineate the developmental trajectories of effective connectivity (EC), the directed causal influence among neuronal populations, in whole-brain networks from infancy to adolescence (0-22 years old) based on high-quality rs-fMRI data from Baby Connectome Project (BCP) and Human Connectome Project Development (HCP-D). Analysis with linear mixed model demonstrates significant age effect on the mean nodal EC which is best fit by a \"U\" shaped quadratic curve with minimal EC at around 2 years old. Further analysis indicates that five brain regions including the left and right cuneus, left precuneus, left supramarginal gyrus and right inferior temporal gyrus have the most significant age effect on nodal EC (<i>p</i> < 0.05, FDR corrected). Moreover, the frontoparietal control (FPC) network shows the fastest increase from early childhood to adolescence followed by the visual and salience networks. Our findings suggest complex nonlinear developmental profile of EC from infancy to adolescence, which may reflect dynamic structural and functional maturation during this critical growth period.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"15003 ","pages":"131-140"},"PeriodicalIF":0.0,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11758277/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143049390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1