Pub Date : 2025-07-01Epub Date: 2025-08-05DOI: 10.1117/1.JMI.12.4.044002
Juming Xiong, Ruining Deng, Jialin Yue, Siqi Lu, Junlin Guo, Marilyn Lionts, Tianyuan Yao, Can Cui, Junchao Zhu, Chongyu Qu, Yuechen Yang, Mengmeng Yin, Haichun Yang, Yuankai Huo
Purpose: Histological analysis plays a crucial role in understanding tissue structure and pathology. Although recent advancements in registration methods have improved 2D histological analysis, they often struggle to preserve critical 3D spatial relationships, limiting their utility in both clinical and research applications. Specifically, constructing accurate 3D models from 2D slices remains challenging due to tissue deformation, sectioning artifacts, variability in imaging techniques, and inconsistent illumination. Deep learning-based registration methods have demonstrated improved performance but suffer from limited generalizability and require large-scale training data. In contrast, non-deep-learning approaches offer better generalizability but often compromise on accuracy.
Approach: We introduce ZeroReg3D, a zero-shot registration pipeline that integrates zero-shot deep learning-based keypoint matching and non-deep-learning registration techniques to effectively mitigate deformation and sectioning artifacts without requiring extensive training data.
Results: Comprehensive evaluations demonstrate that our pairwise 2D image registration method improves registration accuracy by over baseline methods, outperforming existing strategies in both accuracy and robustness. High-fidelity 3D reconstructions further validate the effectiveness of our approach, establishing ZeroReg3D as a reliable framework for precise 3D reconstruction from consecutive 2D histological images.
Conclusions: We introduced ZeroReg3D, a zero-shot registration pipeline tailored for accurate 3D reconstruction from serial histological sections. By combining zero-shot deep learning-based keypoint matching with optimization-based affine and non-rigid registration techniques, ZeroReg3D effectively addresses critical challenges such as tissue deformation, sectioning artifacts, staining variability, and inconsistent illumination without requiring retraining or fine-tuning.
{"title":"ZeroReg3D: a zero-shot registration pipeline for 3D consecutive histopathology image reconstruction.","authors":"Juming Xiong, Ruining Deng, Jialin Yue, Siqi Lu, Junlin Guo, Marilyn Lionts, Tianyuan Yao, Can Cui, Junchao Zhu, Chongyu Qu, Yuechen Yang, Mengmeng Yin, Haichun Yang, Yuankai Huo","doi":"10.1117/1.JMI.12.4.044002","DOIUrl":"10.1117/1.JMI.12.4.044002","url":null,"abstract":"<p><strong>Purpose: </strong>Histological analysis plays a crucial role in understanding tissue structure and pathology. Although recent advancements in registration methods have improved 2D histological analysis, they often struggle to preserve critical 3D spatial relationships, limiting their utility in both clinical and research applications. Specifically, constructing accurate 3D models from 2D slices remains challenging due to tissue deformation, sectioning artifacts, variability in imaging techniques, and inconsistent illumination. Deep learning-based registration methods have demonstrated improved performance but suffer from limited generalizability and require large-scale training data. In contrast, non-deep-learning approaches offer better generalizability but often compromise on accuracy.</p><p><strong>Approach: </strong>We introduce ZeroReg3D, a zero-shot registration pipeline that integrates zero-shot deep learning-based keypoint matching and non-deep-learning registration techniques to effectively mitigate deformation and sectioning artifacts without requiring extensive training data.</p><p><strong>Results: </strong>Comprehensive evaluations demonstrate that our pairwise 2D image registration method improves registration accuracy by <math><mrow><mo>∼</mo> <mn>10</mn> <mo>%</mo></mrow> </math> over baseline methods, outperforming existing strategies in both accuracy and robustness. High-fidelity 3D reconstructions further validate the effectiveness of our approach, establishing ZeroReg3D as a reliable framework for precise 3D reconstruction from consecutive 2D histological images.</p><p><strong>Conclusions: </strong>We introduced ZeroReg3D, a zero-shot registration pipeline tailored for accurate 3D reconstruction from serial histological sections. By combining zero-shot deep learning-based keypoint matching with optimization-based affine and non-rigid registration techniques, ZeroReg3D effectively addresses critical challenges such as tissue deformation, sectioning artifacts, staining variability, and inconsistent illumination without requiring retraining or fine-tuning.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044002"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12322837/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144790347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-08-29DOI: 10.1117/1.JMI.12.4.040101
Bennett A Landman
The editorial discusses current JMI special sections/issues and calls for papers.
该社论讨论了当前JMI的特殊部分/问题和论文征集。
{"title":"JMI's Special Issues and Shared Journeys.","authors":"Bennett A Landman","doi":"10.1117/1.JMI.12.4.040101","DOIUrl":"10.1117/1.JMI.12.4.040101","url":null,"abstract":"<p><p>The editorial discusses current JMI special sections/issues and calls for papers.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"040101"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12395497/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144974286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-07-23DOI: 10.1117/1.JMI.12.4.044503
Hridoy Biswas, Rui Tang, Shamim Mollah, Mikhail Y Berezin
Purpose: Hyperspectral imaging (HSI) collects detailed spectral information across hundreds of narrow bands, providing valuable datasets for applications such as medical diagnostics. However, the large size of HSI datasets, often exceeding several gigabytes, creates significant challenges in data transmission, storage, and processing. We aim to develop a wavelet-based compression method that addresses these challenges while preserving the integrity and quality of spectral information.
Approach: The proposed method applies wavelet transforms to the spectral dimension of hyperspectral data in three steps: (1) wavelet transformation for dimensionality reduction, (2) spectral cropping to eliminate low-intensity bands, and (3) scale matching to maintain accurate wavelength mapping. Daubechies wavelets were used to achieve up to 32× compression while ensuring spectral fidelity and spatial feature retention.
Results: The wavelet-based method achieved up to 32× compression, corresponding to a 96.88% reduction in data size without significant loss of important data. Unlike principal component analysis and independent component analysis, the method preserved the original wavelength scale, enabling straightforward spectral interpretation. In addition, the compressed data exhibited minimal loss in spatial features, with improvements in contrast and noise reduction compared with spectral binning.
Conclusions: We demonstrate that wavelet-based compression is an effective solution for managing large HSI datasets in medical imaging. The method preserves critical spectral and spatial information and therefore facilitates efficient data storage and processing, providing a way for the practical integration of HSI technology in clinical applications.
{"title":"Wavelet-based compression method for scale-preserving in VNIR and SWIR hyperspectral data.","authors":"Hridoy Biswas, Rui Tang, Shamim Mollah, Mikhail Y Berezin","doi":"10.1117/1.JMI.12.4.044503","DOIUrl":"10.1117/1.JMI.12.4.044503","url":null,"abstract":"<p><strong>Purpose: </strong>Hyperspectral imaging (HSI) collects detailed spectral information across hundreds of narrow bands, providing valuable datasets for applications such as medical diagnostics. However, the large size of HSI datasets, often exceeding several gigabytes, creates significant challenges in data transmission, storage, and processing. We aim to develop a wavelet-based compression method that addresses these challenges while preserving the integrity and quality of spectral information.</p><p><strong>Approach: </strong>The proposed method applies wavelet transforms to the spectral dimension of hyperspectral data in three steps: (1) wavelet transformation for dimensionality reduction, (2) spectral cropping to eliminate low-intensity bands, and (3) scale matching to maintain accurate wavelength mapping. Daubechies wavelets were used to achieve up to 32× compression while ensuring spectral fidelity and spatial feature retention.</p><p><strong>Results: </strong>The wavelet-based method achieved up to 32× compression, corresponding to a 96.88% reduction in data size without significant loss of important data. Unlike principal component analysis and independent component analysis, the method preserved the original wavelength scale, enabling straightforward spectral interpretation. In addition, the compressed data exhibited minimal loss in spatial features, with improvements in contrast and noise reduction compared with spectral binning.</p><p><strong>Conclusions: </strong>We demonstrate that wavelet-based compression is an effective solution for managing large HSI datasets in medical imaging. The method preserves critical spectral and spatial information and therefore facilitates efficient data storage and processing, providing a way for the practical integration of HSI technology in clinical applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044503"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12285520/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144700099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-08-13DOI: 10.1117/1.JMI.12.4.046501
Nirmal Choradia, Nathan Lay, Alex Chen, James Latanski, Meredith McAdams, Shannon Swift, Christine Feierabend, Testi Sherif, Susan Sansone, Laercio DaSilva, James L Gulley, Arlene Sirajuddin, Stephanie Harmon, Arun Rajan, Baris Turkbey, Chen Zhao
Purpose: The Response Evaluation Criteria in Solid Tumors (RECIST) relies solely on one-dimensional measurements to evaluate tumor response to treatments. However, thymic epithelial tumors (TETs), which frequently metastasize to the pleural cavity, exhibit a curvilinear morphology that complicates accurate measurement. To address this, we developed a physician-guided deep learning model and performed a retrospective study based on a patient cohort derived from clinical trials, aiming at efficient and reproducible volumetric assessments of TETs.
Approach: We used 231 computed tomography scans comprising 572 TETs from 81 patients. Tumors within the scans were identified and manually outlined to develop a ground truth that was used to measure model performance. TETs were characterized by their general location within the chest cavity: lung parenchyma, pleura, or mediastinum. Model performance was quantified on an unseen test set of 61 scans by mask Dice similarity coefficient (DSC), tumor DSC, absolute volume difference, and relative volume difference.
Results: We included 81 patients: 47 (58.0%) had thymic carcinoma; the remaining patients had thymoma B1, B2, B2/B3, or B3. The artificial intelligence (AI) model achieved an overall DSC of 0.77 per scan when provided with boxes surrounding the tumors as identified by physicians, corresponding to a mean absolute volume difference between the AI measurement and the ground truth of and a mean relative volume difference of 22%.
Conclusion: We have successfully developed a robust annotation workflow and AI segmentation model for analyzing advanced TETs. The model has been integrated into the Picture Archiving and Communication System alongside RECIST measurements to enhance outcome assessments for patients with metastatic TETs.
目的:实体肿瘤反应评价标准(RECIST)仅依赖于一维测量来评估肿瘤对治疗的反应。然而,胸腺上皮肿瘤(TETs)经常转移到胸膜腔,表现出曲线形态,使精确测量复杂化。为了解决这个问题,我们开发了一个医生指导的深度学习模型,并基于来自临床试验的患者队列进行了一项回顾性研究,旨在对TETs进行有效和可重复的体积评估。方法:我们使用了231次计算机断层扫描,包括来自81名患者的572次tet。扫描中的肿瘤被识别并手动勾画出来,以建立一个用于测量模型性能的基本事实。tet的特征在于其在胸腔内的一般位置:肺实质、胸膜或纵隔。通过掩模骰子相似系数(DSC)、肿瘤DSC、绝对体积差和相对体积差对61次扫描的未见测试集的模型性能进行量化。结果:我们纳入81例患者:47例(58.0%)患有胸腺癌;其余患者为胸腺瘤B1、B2、B2/B3或B3。当提供医生识别的肿瘤周围的盒子时,人工智能(AI)模型每次扫描的总体DSC为0.77,对应于AI测量值与地面真实值之间的平均绝对体积差为16.1 cm 3,平均相对体积差为22%。结论:我们成功开发了一个鲁棒的注释工作流和AI分割模型,用于分析高级考试。该模型已与RECIST测量一起集成到图像存档和通信系统中,以增强对转移性tet患者的结果评估。
{"title":"Physician-guided deep learning model for assessing thymic epithelial tumor volume.","authors":"Nirmal Choradia, Nathan Lay, Alex Chen, James Latanski, Meredith McAdams, Shannon Swift, Christine Feierabend, Testi Sherif, Susan Sansone, Laercio DaSilva, James L Gulley, Arlene Sirajuddin, Stephanie Harmon, Arun Rajan, Baris Turkbey, Chen Zhao","doi":"10.1117/1.JMI.12.4.046501","DOIUrl":"10.1117/1.JMI.12.4.046501","url":null,"abstract":"<p><strong>Purpose: </strong>The Response Evaluation Criteria in Solid Tumors (RECIST) relies solely on one-dimensional measurements to evaluate tumor response to treatments. However, thymic epithelial tumors (TETs), which frequently metastasize to the pleural cavity, exhibit a curvilinear morphology that complicates accurate measurement. To address this, we developed a physician-guided deep learning model and performed a retrospective study based on a patient cohort derived from clinical trials, aiming at efficient and reproducible volumetric assessments of TETs.</p><p><strong>Approach: </strong>We used 231 computed tomography scans comprising 572 TETs from 81 patients. Tumors within the scans were identified and manually outlined to develop a ground truth that was used to measure model performance. TETs were characterized by their general location within the chest cavity: lung parenchyma, pleura, or mediastinum. Model performance was quantified on an unseen test set of 61 scans by mask Dice similarity coefficient (DSC), tumor DSC, absolute volume difference, and relative volume difference.</p><p><strong>Results: </strong>We included 81 patients: 47 (58.0%) had thymic carcinoma; the remaining patients had thymoma B1, B2, B2/B3, or B3. The artificial intelligence (AI) model achieved an overall DSC of 0.77 per scan when provided with boxes surrounding the tumors as identified by physicians, corresponding to a mean absolute volume difference between the AI measurement and the ground truth of <math><mrow><mn>16.1</mn> <mtext> </mtext> <msup><mrow><mi>cm</mi></mrow> <mrow><mn>3</mn></mrow> </msup> </mrow> </math> and a mean relative volume difference of 22%.</p><p><strong>Conclusion: </strong>We have successfully developed a robust annotation workflow and AI segmentation model for analyzing advanced TETs. The model has been integrated into the Picture Archiving and Communication System alongside RECIST measurements to enhance outcome assessments for patients with metastatic TETs.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"046501"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12344731/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144849395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-07-16DOI: 10.1117/1.JMI.12.4.044501
Junmei Sun, Xin Zhang, Xiumei Li, Lei Xiao, Huang Bai, Meixi Wang, Maoqun Yao
Purpose: Medical image segmentation based on deep learning has played a crucial role in computer-aided medical diagnosis. However, they are still vulnerable to imperceptible adversarial attacks, which lead to potential misdiagnosis in clinical practice. Research on adversarial attack methods is beneficial for improving the robustness design of medical image segmentation models. Currently, there is a lack of research on adversarial attack methods toward deep learning-based medical image segmentation models. Existing attack methods often yield poor results in terms of both attack effects and image quality of adversarial examples and primarily focus on nontargeted attacks. To address these limitations and further investigate adversarial attacks on segmentation models, we propose an adversarial attack approach.
Approach: We propose an approach called momentum-driven adaptive feature-cosine-similarity with low-frequency constraint attack (MAFL-Attack). The proposed feature-cosine-similarity loss uses high-level abstract semantic information to interfere with the understanding of models about adversarial examples. The low-frequency component constraint ensures the imperceptibility of adversarial examples by constraining the low-frequency components. In addition, the momentum and dynamic step-size calculator are used to enhance the attack process.
Results: Experimental results demonstrate that MAFL-Attack generates adversarial examples with superior targeted attack effects compared with the existing Adaptive Segmentation Mask Attack method, in terms of the evaluation metrics of Intersection over Union, accuracy, , , Peak Signal to Noise Ratio, and Structure Similarity Index Measure.
Conclusions: The design idea of the MAFL-Attack inspires researchers to take corresponding defensive measures to strengthen the robustness of segmentation models.
{"title":"MAFL-Attack: a targeted attack method against deep learning-based medical image segmentation models.","authors":"Junmei Sun, Xin Zhang, Xiumei Li, Lei Xiao, Huang Bai, Meixi Wang, Maoqun Yao","doi":"10.1117/1.JMI.12.4.044501","DOIUrl":"10.1117/1.JMI.12.4.044501","url":null,"abstract":"<p><strong>Purpose: </strong>Medical image segmentation based on deep learning has played a crucial role in computer-aided medical diagnosis. However, they are still vulnerable to imperceptible adversarial attacks, which lead to potential misdiagnosis in clinical practice. Research on adversarial attack methods is beneficial for improving the robustness design of medical image segmentation models. Currently, there is a lack of research on adversarial attack methods toward deep learning-based medical image segmentation models. Existing attack methods often yield poor results in terms of both attack effects and image quality of adversarial examples and primarily focus on nontargeted attacks. To address these limitations and further investigate adversarial attacks on segmentation models, we propose an adversarial attack approach.</p><p><strong>Approach: </strong>We propose an approach called momentum-driven adaptive feature-cosine-similarity with low-frequency constraint attack (MAFL-Attack). The proposed feature-cosine-similarity loss uses high-level abstract semantic information to interfere with the understanding of models about adversarial examples. The low-frequency component constraint ensures the imperceptibility of adversarial examples by constraining the low-frequency components. In addition, the momentum and dynamic step-size calculator are used to enhance the attack process.</p><p><strong>Results: </strong>Experimental results demonstrate that MAFL-Attack generates adversarial examples with superior targeted attack effects compared with the existing Adaptive Segmentation Mask Attack method, in terms of the evaluation metrics of Intersection over Union, accuracy, <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mn>2</mn></mrow> </msub> </mrow> </math> , <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mo>∞</mo></mrow> </msub> </mrow> </math> , Peak Signal to Noise Ratio, and Structure Similarity Index Measure.</p><p><strong>Conclusions: </strong>The design idea of the MAFL-Attack inspires researchers to take corresponding defensive measures to strengthen the robustness of segmentation models.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044501"},"PeriodicalIF":1.9,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12266980/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144676110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-12DOI: 10.1117/1.JMI.12.3.035002
Naeeme Modir, Maysam Shahedi, James Dormer, Ling Ma, Baowei Fei
Purpose: This study demonstrates the feasibility of using an LED array for hyperspectral imaging (HSI). The prototype validates the concept and provides insights into the design of future HSI applications. Our goal is to design, develop, and test a real-time, LED-based HSI prototype as a proof-of-principle device for in situ hyperspectral imaging using LEDs.
Approach: A prototype was designed based on a multiwavelength LED array and a monochrome camera and was tested to investigate the properties of the LED-based HSI. The LED array consisted of 18 LEDs in 18 different wavelengths from 405 nm to 910 nm. The performance of the imaging system was evaluated on different normal and cancerous ex vivo tissues. The impact of imaging conditions on the HSI quality was investigated. The LED-based HSI device was compared with a reference hyperspectral camera system.
Results: The hyperspectral signatures of different imaging targets were acquired using our prototype HSI device, which are comparable to the data obtained using the reference HSI system.
Conclusions: The feasibility of employing a spectral LED array as the illumination source for high-speed and high-quality HSI has been demonstrated. The use of LEDs for HSI can open the door to numerous applications in endoscopic, laparoscopic, and handheld HSI devices.
{"title":"LED-based, real-time, hyperspectral imaging device.","authors":"Naeeme Modir, Maysam Shahedi, James Dormer, Ling Ma, Baowei Fei","doi":"10.1117/1.JMI.12.3.035002","DOIUrl":"10.1117/1.JMI.12.3.035002","url":null,"abstract":"<p><strong>Purpose: </strong>This study demonstrates the feasibility of using an LED array for hyperspectral imaging (HSI). The prototype validates the concept and provides insights into the design of future HSI applications. Our goal is to design, develop, and test a real-time, LED-based HSI prototype as a proof-of-principle device for <i>in situ</i> hyperspectral imaging using LEDs.</p><p><strong>Approach: </strong>A prototype was designed based on a multiwavelength LED array and a monochrome camera and was tested to investigate the properties of the LED-based HSI. The LED array consisted of 18 LEDs in 18 different wavelengths from 405 nm to 910 nm. The performance of the imaging system was evaluated on different normal and cancerous <i>ex vivo</i> tissues. The impact of imaging conditions on the HSI quality was investigated. The LED-based HSI device was compared with a reference hyperspectral camera system.</p><p><strong>Results: </strong>The hyperspectral signatures of different imaging targets were acquired using our prototype HSI device, which are comparable to the data obtained using the reference HSI system.</p><p><strong>Conclusions: </strong>The feasibility of employing a spectral LED array as the illumination source for high-speed and high-quality HSI has been demonstrated. The use of LEDs for HSI can open the door to numerous applications in endoscopic, laparoscopic, and handheld HSI devices.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"035002"},"PeriodicalIF":1.7,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12162177/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144303315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-28DOI: 10.1117/1.JMI.12.3.030101
Bennett A Landman
The editorial celebrates emerging breakthroughs and the foundational work that continues to shape the field.
这篇社论赞扬了新兴的突破和继续塑造该领域的基础工作。
{"title":"Summer of Ideas, Community, and Recognition.","authors":"Bennett A Landman","doi":"10.1117/1.JMI.12.3.030101","DOIUrl":"https://doi.org/10.1117/1.JMI.12.3.030101","url":null,"abstract":"<p><p>The editorial celebrates emerging breakthroughs and the foundational work that continues to shape the field.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"030101"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12205331/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144530365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-19DOI: 10.1117/1.JMI.12.3.034506
Bohan Jiang, Andrew J McNeil, Yihao Liu, David W House, Placide Mbala-Kingebeni, Olivier Tshiani Mbaya, Tyra Silaphet, Lori E Dodd, Edward W Cowen, Veronique Nussenblatt, Tyler Bonnett, Ziche Chen, Inga Saknite, Benoit M Dawant, Eric R Tkaczyk
Purpose: Mpox is a viral illness with symptoms similar to smallpox. A key clinical metric to monitor disease progression is the number of skin lesions. Manually counting mpox skin lesions is labor-intensive and susceptible to human error.
Approach: We previously developed an mpox lesion counting method based on the UNet segmentation model using 66 photographs from 18 patients. We have compared four additional methods: the instance segmentation methods Mask R-CNN, YOLOv8, and E2EC, in addition to a UNet++ model. We designed a patient-level leave-one-out experiment, assessing their performance using score and lesion count metrics. Finally, we tested whether an ensemble of the networks outperformed any single model.
Results: Mask R-CNN model achieved an score of 0.75, YOLOv8 a score of 0.75, E2EC a score of 0.70, UNet++ a score of 0.81, and baseline UNet a score of 0.79. Bland-Altman analysis of lesion count performance showed a limit of agreement (LoA) width of 62.2 for Mask R-CNN, 91.3 for YOLOv8, 94.2 for E2EC, and 62.1 for UNet++, with the baseline UNet model achieving 69.1. The ensemble showed an score performance of 0.78 and LoA width of 67.4.
Conclusions: Instance segmentation methods and UNet-based semantic segmentation methods performed equally well in lesion counting. Furthermore, the ensemble of the trained models showed no performance increase over the best-performing model UNet, likely because errors are frequently shared across models. Performance is likely limited by the availability of high-quality photographs for this complex problem, rather than the methodologies used.
{"title":"Mpox lesion counting with semantic and instance segmentation methods.","authors":"Bohan Jiang, Andrew J McNeil, Yihao Liu, David W House, Placide Mbala-Kingebeni, Olivier Tshiani Mbaya, Tyra Silaphet, Lori E Dodd, Edward W Cowen, Veronique Nussenblatt, Tyler Bonnett, Ziche Chen, Inga Saknite, Benoit M Dawant, Eric R Tkaczyk","doi":"10.1117/1.JMI.12.3.034506","DOIUrl":"10.1117/1.JMI.12.3.034506","url":null,"abstract":"<p><strong>Purpose: </strong>Mpox is a viral illness with symptoms similar to smallpox. A key clinical metric to monitor disease progression is the number of skin lesions. Manually counting mpox skin lesions is labor-intensive and susceptible to human error.</p><p><strong>Approach: </strong>We previously developed an mpox lesion counting method based on the UNet segmentation model using 66 photographs from 18 patients. We have compared four additional methods: the instance segmentation methods Mask R-CNN, YOLOv8, and E2EC, in addition to a UNet++ model. We designed a patient-level leave-one-out experiment, assessing their performance using <math><mrow><mi>F</mi> <mn>1</mn></mrow> </math> score and lesion count metrics. Finally, we tested whether an ensemble of the networks outperformed any single model.</p><p><strong>Results: </strong>Mask R-CNN model achieved an <math><mrow><mi>F</mi> <mn>1</mn></mrow> </math> score of 0.75, YOLOv8 a score of 0.75, E2EC a score of 0.70, UNet++ a score of 0.81, and baseline UNet a score of 0.79. Bland-Altman analysis of lesion count performance showed a limit of agreement (LoA) width of 62.2 for Mask R-CNN, 91.3 for YOLOv8, 94.2 for E2EC, and 62.1 for UNet++, with the baseline UNet model achieving 69.1. The ensemble showed an <math><mrow><mi>F</mi> <mn>1</mn></mrow> </math> score performance of 0.78 and LoA width of 67.4.</p><p><strong>Conclusions: </strong>Instance segmentation methods and UNet-based semantic segmentation methods performed equally well in lesion counting. Furthermore, the ensemble of the trained models showed no performance increase over the best-performing model UNet, likely because errors are frequently shared across models. Performance is likely limited by the availability of high-quality photographs for this complex problem, rather than the methodologies used.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"034506"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12177574/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144369413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-03DOI: 10.1117/1.JMI.12.3.035001
Constant R Noordman, Steffan J W Borgers, Martijn F Boomsma, Thomas C Kwee, Marloes M G van der Lees, Christiaan G Overduin, Maarten de Rooij, Derya Yakar, Jurgen J Fütterer, Henkjan J Huisman
Purpose: Interventional MR imaging struggles with speed and efficiency. We aim to accelerate transrectal in-bore MR-guided biopsies for prostate cancer through undersampled image reconstruction and instrument localization by image segmentation.
Approach: In this single-center retrospective study, we used 8464 MR 2D multislice scans from 1289 patients undergoing a prostate biopsy to train and test a deep learning-based spatiotemporal MR image reconstruction model and a nnU-Net segmentation model. The dataset was synthetically undersampled using various undersampling rates ( , 16, 25, 32). An annotated, unseen subset of these data was used to compare our model with a nontemporal model and readers in a reader study involving seven radiologists from three centers based in the Netherlands. We assessed a maximum noninferior undersampling rate using instrument prediction success rate and instrument tip position (ITP) error.
Results: The maximum noninferior undersampling rate is 16-times for the temporal model (ITP error: 2.28 mm, 95% CI: 1.68 to 3.31, mean difference from reference standard: 0.63 mm, ), whereas a nontemporal model could not produce noninferior image reconstructions comparable to our reference standard. Furthermore, the nontemporal model (ITP error: 6.27 mm, 95% CI: 3.90 to 9.07) and readers (ITP error: 6.87 mm, 95% CI: 6.38 to 7.40) had low instrument prediction success rates (46% and 60%, respectively) compared with the temporal model's 95%.
Conclusion: Deep learning-based spatiotemporal MR image reconstruction can improve time-critical intervention tasks such as instrument tracking. We found 16 times undersampling as the maximum noninferior acceleration where image quality is preserved, ITP error is minimized, and the instrument prediction success rate is maximized.
目的:介入磁共振成像的速度和效率。我们的目标是通过采样不足的图像重建和图像分割的仪器定位来加速经直肠前列腺癌的磁共振引导活检。方法:在这项单中心回顾性研究中,我们使用了1289例前列腺活检患者的8464张磁共振二维多层扫描图来训练和测试基于深度学习的时空磁共振图像重建模型和nnU-Net分割模型。使用不同的欠采样率(R = 8,16,25,32)对数据集进行综合欠采样。在一项涉及来自荷兰三个中心的七名放射科医生的读者研究中,使用这些数据的一个注释的、未见过的子集将我们的模型与非时间模型和读者进行比较。我们使用仪器预测成功率和仪器尖端位置(ITP)误差来评估最大非劣欠采样率。结果:时间模型的最大非劣欠采样率为16次(ITP误差:2.28 mm, 95% CI: 1.68 ~ 3.31,与参考标准的平均差值:0.63 mm, P =。09),而非时间模型无法产生与我们的参考标准相当的非劣质图像重建。此外,与时间模型的95%相比,非时间模型(ITP误差:6.27 mm, 95% CI: 3.90至9.07)和读取器(ITP误差:6.87 mm, 95% CI: 6.38至7.40)的仪器预测成功率较低(分别为46%和60%)。结论:基于深度学习的时空磁共振图像重建可以改善仪器跟踪等时间关键型干预任务。我们发现16次欠采样作为最大非劣等加速,在此条件下,图像质量得以保留,ITP误差最小化,仪器预测成功率最大化。
{"title":"Deep learning-based temporal MR image reconstruction for accelerated interventional imaging during in-bore biopsies.","authors":"Constant R Noordman, Steffan J W Borgers, Martijn F Boomsma, Thomas C Kwee, Marloes M G van der Lees, Christiaan G Overduin, Maarten de Rooij, Derya Yakar, Jurgen J Fütterer, Henkjan J Huisman","doi":"10.1117/1.JMI.12.3.035001","DOIUrl":"10.1117/1.JMI.12.3.035001","url":null,"abstract":"<p><strong>Purpose: </strong>Interventional MR imaging struggles with speed and efficiency. We aim to accelerate transrectal in-bore MR-guided biopsies for prostate cancer through undersampled image reconstruction and instrument localization by image segmentation.</p><p><strong>Approach: </strong>In this single-center retrospective study, we used 8464 MR 2D multislice scans from 1289 patients undergoing a prostate biopsy to train and test a deep learning-based spatiotemporal MR image reconstruction model and a nnU-Net segmentation model. The dataset was synthetically undersampled using various undersampling rates ( <math><mrow><mi>R</mi> <mo>=</mo> <mn>8</mn></mrow> </math> , 16, 25, 32). An annotated, unseen subset of these data was used to compare our model with a nontemporal model and readers in a reader study involving seven radiologists from three centers based in the Netherlands. We assessed a maximum noninferior undersampling rate using instrument prediction success rate and instrument tip position (ITP) error.</p><p><strong>Results: </strong>The maximum noninferior undersampling rate is 16-times for the temporal model (ITP error: 2.28 mm, 95% CI: 1.68 to 3.31, mean difference from reference standard: 0.63 mm, <math><mrow><mi>P</mi> <mo>=</mo> <mo>.</mo> <mn>09</mn></mrow> </math> ), whereas a nontemporal model could not produce noninferior image reconstructions comparable to our reference standard. Furthermore, the nontemporal model (ITP error: 6.27 mm, 95% CI: 3.90 to 9.07) and readers (ITP error: 6.87 mm, 95% CI: 6.38 to 7.40) had low instrument prediction success rates (46% and 60%, respectively) compared with the temporal model's 95%.</p><p><strong>Conclusion: </strong>Deep learning-based spatiotemporal MR image reconstruction can improve time-critical intervention tasks such as instrument tracking. We found 16 times undersampling as the maximum noninferior acceleration where image quality is preserved, ITP error is minimized, and the instrument prediction success rate is maximized.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"035001"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12131189/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144227256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-05-21DOI: 10.1117/1.JMI.12.3.035501
Vaibhav Sharma, Alina Jade Barnett, Julia Yang, Sangwook Cheon, Giyoung Kim, Fides Regina Schwartz, Avivah Wang, Neal Hall, Lars Grimm, Chaofan Chen, Joseph Y Lo, Cynthia Rudin
Purpose: Breast cancer remains a leading cause of death for women. Screening programs are deployed to detect cancer at early stages. One current barrier identified by breast imaging researchers is a shortage of labeled image datasets. Addressing this problem is crucial to improve early detection models. We present an active learning (AL) framework for segmenting breast masses from 2D digital mammography, and we publish labeled data. Our method aims to reduce the input needed from expert annotators to reach a fully labeled dataset.
Approach: We create a dataset of 1136 mammographic masses with pixel-wise binary segmentation labels, with the test subset labeled independently by two different teams. With this dataset, we simulate a human annotator within an AL framework to develop and compare AI-assisted labeling methods, using a discriminator model and a simulated oracle to collect acceptable segmentation labels. A UNet model is retrained on these labels, generating new segmentations. We evaluate various oracle heuristics using the percentage of segmentations that the oracle relabels and measure the quality of the proposed labels by evaluating the intersection over union over a validation dataset.
Results: Our method reduces expert annotator input by 44%. We present a dataset of 1136 binary segmentation labels approved by board-certified radiologists and make the 143-image validation set public for comparison with other researchers' methods.
Conclusions: We demonstrate that AL can significantly improve the efficiency and time-effectiveness of creating labeled mammogram datasets. Our framework facilitates the development of high-quality datasets while minimizing manual effort in the domain of digital mammography.
{"title":"Improving annotation efficiency for fully labeling a breast mass segmentation dataset.","authors":"Vaibhav Sharma, Alina Jade Barnett, Julia Yang, Sangwook Cheon, Giyoung Kim, Fides Regina Schwartz, Avivah Wang, Neal Hall, Lars Grimm, Chaofan Chen, Joseph Y Lo, Cynthia Rudin","doi":"10.1117/1.JMI.12.3.035501","DOIUrl":"10.1117/1.JMI.12.3.035501","url":null,"abstract":"<p><strong>Purpose: </strong>Breast cancer remains a leading cause of death for women. Screening programs are deployed to detect cancer at early stages. One current barrier identified by breast imaging researchers is a shortage of labeled image datasets. Addressing this problem is crucial to improve early detection models. We present an active learning (AL) framework for segmenting breast masses from 2D digital mammography, and we publish labeled data. Our method aims to reduce the input needed from expert annotators to reach a fully labeled dataset.</p><p><strong>Approach: </strong>We create a dataset of 1136 mammographic masses with pixel-wise binary segmentation labels, with the test subset labeled independently by two different teams. With this dataset, we simulate a human annotator within an AL framework to develop and compare AI-assisted labeling methods, using a discriminator model and a simulated oracle to collect acceptable segmentation labels. A UNet model is retrained on these labels, generating new segmentations. We evaluate various oracle heuristics using the percentage of segmentations that the oracle relabels and measure the quality of the proposed labels by evaluating the intersection over union over a validation dataset.</p><p><strong>Results: </strong>Our method reduces expert annotator input by 44%. We present a dataset of 1136 binary segmentation labels approved by board-certified radiologists and make the 143-image validation set public for comparison with other researchers' methods.</p><p><strong>Conclusions: </strong>We demonstrate that AL can significantly improve the efficiency and time-effectiveness of creating labeled mammogram datasets. Our framework facilitates the development of high-quality datasets while minimizing manual effort in the domain of digital mammography.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"035501"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12094908/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144144147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}