Pub Date : 2025-07-01Epub Date: 2025-07-23DOI: 10.1117/1.JMI.12.4.044503
Hridoy Biswas, Rui Tang, Shamim Mollah, Mikhail Y Berezin
Purpose: Hyperspectral imaging (HSI) collects detailed spectral information across hundreds of narrow bands, providing valuable datasets for applications such as medical diagnostics. However, the large size of HSI datasets, often exceeding several gigabytes, creates significant challenges in data transmission, storage, and processing. We aim to develop a wavelet-based compression method that addresses these challenges while preserving the integrity and quality of spectral information.
Approach: The proposed method applies wavelet transforms to the spectral dimension of hyperspectral data in three steps: (1) wavelet transformation for dimensionality reduction, (2) spectral cropping to eliminate low-intensity bands, and (3) scale matching to maintain accurate wavelength mapping. Daubechies wavelets were used to achieve up to 32× compression while ensuring spectral fidelity and spatial feature retention.
Results: The wavelet-based method achieved up to 32× compression, corresponding to a 96.88% reduction in data size without significant loss of important data. Unlike principal component analysis and independent component analysis, the method preserved the original wavelength scale, enabling straightforward spectral interpretation. In addition, the compressed data exhibited minimal loss in spatial features, with improvements in contrast and noise reduction compared with spectral binning.
Conclusions: We demonstrate that wavelet-based compression is an effective solution for managing large HSI datasets in medical imaging. The method preserves critical spectral and spatial information and therefore facilitates efficient data storage and processing, providing a way for the practical integration of HSI technology in clinical applications.
{"title":"Wavelet-based compression method for scale-preserving in VNIR and SWIR hyperspectral data.","authors":"Hridoy Biswas, Rui Tang, Shamim Mollah, Mikhail Y Berezin","doi":"10.1117/1.JMI.12.4.044503","DOIUrl":"10.1117/1.JMI.12.4.044503","url":null,"abstract":"<p><strong>Purpose: </strong>Hyperspectral imaging (HSI) collects detailed spectral information across hundreds of narrow bands, providing valuable datasets for applications such as medical diagnostics. However, the large size of HSI datasets, often exceeding several gigabytes, creates significant challenges in data transmission, storage, and processing. We aim to develop a wavelet-based compression method that addresses these challenges while preserving the integrity and quality of spectral information.</p><p><strong>Approach: </strong>The proposed method applies wavelet transforms to the spectral dimension of hyperspectral data in three steps: (1) wavelet transformation for dimensionality reduction, (2) spectral cropping to eliminate low-intensity bands, and (3) scale matching to maintain accurate wavelength mapping. Daubechies wavelets were used to achieve up to 32× compression while ensuring spectral fidelity and spatial feature retention.</p><p><strong>Results: </strong>The wavelet-based method achieved up to 32× compression, corresponding to a 96.88% reduction in data size without significant loss of important data. Unlike principal component analysis and independent component analysis, the method preserved the original wavelength scale, enabling straightforward spectral interpretation. In addition, the compressed data exhibited minimal loss in spatial features, with improvements in contrast and noise reduction compared with spectral binning.</p><p><strong>Conclusions: </strong>We demonstrate that wavelet-based compression is an effective solution for managing large HSI datasets in medical imaging. The method preserves critical spectral and spatial information and therefore facilitates efficient data storage and processing, providing a way for the practical integration of HSI technology in clinical applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044503"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12285520/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144700099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-08-13DOI: 10.1117/1.JMI.12.4.046501
Nirmal Choradia, Nathan Lay, Alex Chen, James Latanski, Meredith McAdams, Shannon Swift, Christine Feierabend, Testi Sherif, Susan Sansone, Laercio DaSilva, James L Gulley, Arlene Sirajuddin, Stephanie Harmon, Arun Rajan, Baris Turkbey, Chen Zhao
Purpose: The Response Evaluation Criteria in Solid Tumors (RECIST) relies solely on one-dimensional measurements to evaluate tumor response to treatments. However, thymic epithelial tumors (TETs), which frequently metastasize to the pleural cavity, exhibit a curvilinear morphology that complicates accurate measurement. To address this, we developed a physician-guided deep learning model and performed a retrospective study based on a patient cohort derived from clinical trials, aiming at efficient and reproducible volumetric assessments of TETs.
Approach: We used 231 computed tomography scans comprising 572 TETs from 81 patients. Tumors within the scans were identified and manually outlined to develop a ground truth that was used to measure model performance. TETs were characterized by their general location within the chest cavity: lung parenchyma, pleura, or mediastinum. Model performance was quantified on an unseen test set of 61 scans by mask Dice similarity coefficient (DSC), tumor DSC, absolute volume difference, and relative volume difference.
Results: We included 81 patients: 47 (58.0%) had thymic carcinoma; the remaining patients had thymoma B1, B2, B2/B3, or B3. The artificial intelligence (AI) model achieved an overall DSC of 0.77 per scan when provided with boxes surrounding the tumors as identified by physicians, corresponding to a mean absolute volume difference between the AI measurement and the ground truth of and a mean relative volume difference of 22%.
Conclusion: We have successfully developed a robust annotation workflow and AI segmentation model for analyzing advanced TETs. The model has been integrated into the Picture Archiving and Communication System alongside RECIST measurements to enhance outcome assessments for patients with metastatic TETs.
目的:实体肿瘤反应评价标准(RECIST)仅依赖于一维测量来评估肿瘤对治疗的反应。然而,胸腺上皮肿瘤(TETs)经常转移到胸膜腔,表现出曲线形态,使精确测量复杂化。为了解决这个问题,我们开发了一个医生指导的深度学习模型,并基于来自临床试验的患者队列进行了一项回顾性研究,旨在对TETs进行有效和可重复的体积评估。方法:我们使用了231次计算机断层扫描,包括来自81名患者的572次tet。扫描中的肿瘤被识别并手动勾画出来,以建立一个用于测量模型性能的基本事实。tet的特征在于其在胸腔内的一般位置:肺实质、胸膜或纵隔。通过掩模骰子相似系数(DSC)、肿瘤DSC、绝对体积差和相对体积差对61次扫描的未见测试集的模型性能进行量化。结果:我们纳入81例患者:47例(58.0%)患有胸腺癌;其余患者为胸腺瘤B1、B2、B2/B3或B3。当提供医生识别的肿瘤周围的盒子时,人工智能(AI)模型每次扫描的总体DSC为0.77,对应于AI测量值与地面真实值之间的平均绝对体积差为16.1 cm 3,平均相对体积差为22%。结论:我们成功开发了一个鲁棒的注释工作流和AI分割模型,用于分析高级考试。该模型已与RECIST测量一起集成到图像存档和通信系统中,以增强对转移性tet患者的结果评估。
{"title":"Physician-guided deep learning model for assessing thymic epithelial tumor volume.","authors":"Nirmal Choradia, Nathan Lay, Alex Chen, James Latanski, Meredith McAdams, Shannon Swift, Christine Feierabend, Testi Sherif, Susan Sansone, Laercio DaSilva, James L Gulley, Arlene Sirajuddin, Stephanie Harmon, Arun Rajan, Baris Turkbey, Chen Zhao","doi":"10.1117/1.JMI.12.4.046501","DOIUrl":"10.1117/1.JMI.12.4.046501","url":null,"abstract":"<p><strong>Purpose: </strong>The Response Evaluation Criteria in Solid Tumors (RECIST) relies solely on one-dimensional measurements to evaluate tumor response to treatments. However, thymic epithelial tumors (TETs), which frequently metastasize to the pleural cavity, exhibit a curvilinear morphology that complicates accurate measurement. To address this, we developed a physician-guided deep learning model and performed a retrospective study based on a patient cohort derived from clinical trials, aiming at efficient and reproducible volumetric assessments of TETs.</p><p><strong>Approach: </strong>We used 231 computed tomography scans comprising 572 TETs from 81 patients. Tumors within the scans were identified and manually outlined to develop a ground truth that was used to measure model performance. TETs were characterized by their general location within the chest cavity: lung parenchyma, pleura, or mediastinum. Model performance was quantified on an unseen test set of 61 scans by mask Dice similarity coefficient (DSC), tumor DSC, absolute volume difference, and relative volume difference.</p><p><strong>Results: </strong>We included 81 patients: 47 (58.0%) had thymic carcinoma; the remaining patients had thymoma B1, B2, B2/B3, or B3. The artificial intelligence (AI) model achieved an overall DSC of 0.77 per scan when provided with boxes surrounding the tumors as identified by physicians, corresponding to a mean absolute volume difference between the AI measurement and the ground truth of <math><mrow><mn>16.1</mn> <mtext> </mtext> <msup><mrow><mi>cm</mi></mrow> <mrow><mn>3</mn></mrow> </msup> </mrow> </math> and a mean relative volume difference of 22%.</p><p><strong>Conclusion: </strong>We have successfully developed a robust annotation workflow and AI segmentation model for analyzing advanced TETs. The model has been integrated into the Picture Archiving and Communication System alongside RECIST measurements to enhance outcome assessments for patients with metastatic TETs.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"046501"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12344731/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144849395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-07-16DOI: 10.1117/1.JMI.12.4.044501
Junmei Sun, Xin Zhang, Xiumei Li, Lei Xiao, Huang Bai, Meixi Wang, Maoqun Yao
Purpose: Medical image segmentation based on deep learning has played a crucial role in computer-aided medical diagnosis. However, they are still vulnerable to imperceptible adversarial attacks, which lead to potential misdiagnosis in clinical practice. Research on adversarial attack methods is beneficial for improving the robustness design of medical image segmentation models. Currently, there is a lack of research on adversarial attack methods toward deep learning-based medical image segmentation models. Existing attack methods often yield poor results in terms of both attack effects and image quality of adversarial examples and primarily focus on nontargeted attacks. To address these limitations and further investigate adversarial attacks on segmentation models, we propose an adversarial attack approach.
Approach: We propose an approach called momentum-driven adaptive feature-cosine-similarity with low-frequency constraint attack (MAFL-Attack). The proposed feature-cosine-similarity loss uses high-level abstract semantic information to interfere with the understanding of models about adversarial examples. The low-frequency component constraint ensures the imperceptibility of adversarial examples by constraining the low-frequency components. In addition, the momentum and dynamic step-size calculator are used to enhance the attack process.
Results: Experimental results demonstrate that MAFL-Attack generates adversarial examples with superior targeted attack effects compared with the existing Adaptive Segmentation Mask Attack method, in terms of the evaluation metrics of Intersection over Union, accuracy, , , Peak Signal to Noise Ratio, and Structure Similarity Index Measure.
Conclusions: The design idea of the MAFL-Attack inspires researchers to take corresponding defensive measures to strengthen the robustness of segmentation models.
{"title":"MAFL-Attack: a targeted attack method against deep learning-based medical image segmentation models.","authors":"Junmei Sun, Xin Zhang, Xiumei Li, Lei Xiao, Huang Bai, Meixi Wang, Maoqun Yao","doi":"10.1117/1.JMI.12.4.044501","DOIUrl":"10.1117/1.JMI.12.4.044501","url":null,"abstract":"<p><strong>Purpose: </strong>Medical image segmentation based on deep learning has played a crucial role in computer-aided medical diagnosis. However, they are still vulnerable to imperceptible adversarial attacks, which lead to potential misdiagnosis in clinical practice. Research on adversarial attack methods is beneficial for improving the robustness design of medical image segmentation models. Currently, there is a lack of research on adversarial attack methods toward deep learning-based medical image segmentation models. Existing attack methods often yield poor results in terms of both attack effects and image quality of adversarial examples and primarily focus on nontargeted attacks. To address these limitations and further investigate adversarial attacks on segmentation models, we propose an adversarial attack approach.</p><p><strong>Approach: </strong>We propose an approach called momentum-driven adaptive feature-cosine-similarity with low-frequency constraint attack (MAFL-Attack). The proposed feature-cosine-similarity loss uses high-level abstract semantic information to interfere with the understanding of models about adversarial examples. The low-frequency component constraint ensures the imperceptibility of adversarial examples by constraining the low-frequency components. In addition, the momentum and dynamic step-size calculator are used to enhance the attack process.</p><p><strong>Results: </strong>Experimental results demonstrate that MAFL-Attack generates adversarial examples with superior targeted attack effects compared with the existing Adaptive Segmentation Mask Attack method, in terms of the evaluation metrics of Intersection over Union, accuracy, <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mn>2</mn></mrow> </msub> </mrow> </math> , <math> <mrow> <msub><mrow><mi>L</mi></mrow> <mrow><mo>∞</mo></mrow> </msub> </mrow> </math> , Peak Signal to Noise Ratio, and Structure Similarity Index Measure.</p><p><strong>Conclusions: </strong>The design idea of the MAFL-Attack inspires researchers to take corresponding defensive measures to strengthen the robustness of segmentation models.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044501"},"PeriodicalIF":1.9,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12266980/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144676110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-12DOI: 10.1117/1.JMI.12.3.035002
Naeeme Modir, Maysam Shahedi, James Dormer, Ling Ma, Baowei Fei
Purpose: This study demonstrates the feasibility of using an LED array for hyperspectral imaging (HSI). The prototype validates the concept and provides insights into the design of future HSI applications. Our goal is to design, develop, and test a real-time, LED-based HSI prototype as a proof-of-principle device for in situ hyperspectral imaging using LEDs.
Approach: A prototype was designed based on a multiwavelength LED array and a monochrome camera and was tested to investigate the properties of the LED-based HSI. The LED array consisted of 18 LEDs in 18 different wavelengths from 405 nm to 910 nm. The performance of the imaging system was evaluated on different normal and cancerous ex vivo tissues. The impact of imaging conditions on the HSI quality was investigated. The LED-based HSI device was compared with a reference hyperspectral camera system.
Results: The hyperspectral signatures of different imaging targets were acquired using our prototype HSI device, which are comparable to the data obtained using the reference HSI system.
Conclusions: The feasibility of employing a spectral LED array as the illumination source for high-speed and high-quality HSI has been demonstrated. The use of LEDs for HSI can open the door to numerous applications in endoscopic, laparoscopic, and handheld HSI devices.
{"title":"LED-based, real-time, hyperspectral imaging device.","authors":"Naeeme Modir, Maysam Shahedi, James Dormer, Ling Ma, Baowei Fei","doi":"10.1117/1.JMI.12.3.035002","DOIUrl":"10.1117/1.JMI.12.3.035002","url":null,"abstract":"<p><strong>Purpose: </strong>This study demonstrates the feasibility of using an LED array for hyperspectral imaging (HSI). The prototype validates the concept and provides insights into the design of future HSI applications. Our goal is to design, develop, and test a real-time, LED-based HSI prototype as a proof-of-principle device for <i>in situ</i> hyperspectral imaging using LEDs.</p><p><strong>Approach: </strong>A prototype was designed based on a multiwavelength LED array and a monochrome camera and was tested to investigate the properties of the LED-based HSI. The LED array consisted of 18 LEDs in 18 different wavelengths from 405 nm to 910 nm. The performance of the imaging system was evaluated on different normal and cancerous <i>ex vivo</i> tissues. The impact of imaging conditions on the HSI quality was investigated. The LED-based HSI device was compared with a reference hyperspectral camera system.</p><p><strong>Results: </strong>The hyperspectral signatures of different imaging targets were acquired using our prototype HSI device, which are comparable to the data obtained using the reference HSI system.</p><p><strong>Conclusions: </strong>The feasibility of employing a spectral LED array as the illumination source for high-speed and high-quality HSI has been demonstrated. The use of LEDs for HSI can open the door to numerous applications in endoscopic, laparoscopic, and handheld HSI devices.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"035002"},"PeriodicalIF":1.7,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12162177/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144303315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-28DOI: 10.1117/1.JMI.12.3.030101
Bennett A Landman
The editorial celebrates emerging breakthroughs and the foundational work that continues to shape the field.
这篇社论赞扬了新兴的突破和继续塑造该领域的基础工作。
{"title":"Summer of Ideas, Community, and Recognition.","authors":"Bennett A Landman","doi":"10.1117/1.JMI.12.3.030101","DOIUrl":"https://doi.org/10.1117/1.JMI.12.3.030101","url":null,"abstract":"<p><p>The editorial celebrates emerging breakthroughs and the foundational work that continues to shape the field.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"030101"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12205331/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144530365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-19DOI: 10.1117/1.JMI.12.3.034506
Bohan Jiang, Andrew J McNeil, Yihao Liu, David W House, Placide Mbala-Kingebeni, Olivier Tshiani Mbaya, Tyra Silaphet, Lori E Dodd, Edward W Cowen, Veronique Nussenblatt, Tyler Bonnett, Ziche Chen, Inga Saknite, Benoit M Dawant, Eric R Tkaczyk
Purpose: Mpox is a viral illness with symptoms similar to smallpox. A key clinical metric to monitor disease progression is the number of skin lesions. Manually counting mpox skin lesions is labor-intensive and susceptible to human error.
Approach: We previously developed an mpox lesion counting method based on the UNet segmentation model using 66 photographs from 18 patients. We have compared four additional methods: the instance segmentation methods Mask R-CNN, YOLOv8, and E2EC, in addition to a UNet++ model. We designed a patient-level leave-one-out experiment, assessing their performance using score and lesion count metrics. Finally, we tested whether an ensemble of the networks outperformed any single model.
Results: Mask R-CNN model achieved an score of 0.75, YOLOv8 a score of 0.75, E2EC a score of 0.70, UNet++ a score of 0.81, and baseline UNet a score of 0.79. Bland-Altman analysis of lesion count performance showed a limit of agreement (LoA) width of 62.2 for Mask R-CNN, 91.3 for YOLOv8, 94.2 for E2EC, and 62.1 for UNet++, with the baseline UNet model achieving 69.1. The ensemble showed an score performance of 0.78 and LoA width of 67.4.
Conclusions: Instance segmentation methods and UNet-based semantic segmentation methods performed equally well in lesion counting. Furthermore, the ensemble of the trained models showed no performance increase over the best-performing model UNet, likely because errors are frequently shared across models. Performance is likely limited by the availability of high-quality photographs for this complex problem, rather than the methodologies used.
{"title":"Mpox lesion counting with semantic and instance segmentation methods.","authors":"Bohan Jiang, Andrew J McNeil, Yihao Liu, David W House, Placide Mbala-Kingebeni, Olivier Tshiani Mbaya, Tyra Silaphet, Lori E Dodd, Edward W Cowen, Veronique Nussenblatt, Tyler Bonnett, Ziche Chen, Inga Saknite, Benoit M Dawant, Eric R Tkaczyk","doi":"10.1117/1.JMI.12.3.034506","DOIUrl":"10.1117/1.JMI.12.3.034506","url":null,"abstract":"<p><strong>Purpose: </strong>Mpox is a viral illness with symptoms similar to smallpox. A key clinical metric to monitor disease progression is the number of skin lesions. Manually counting mpox skin lesions is labor-intensive and susceptible to human error.</p><p><strong>Approach: </strong>We previously developed an mpox lesion counting method based on the UNet segmentation model using 66 photographs from 18 patients. We have compared four additional methods: the instance segmentation methods Mask R-CNN, YOLOv8, and E2EC, in addition to a UNet++ model. We designed a patient-level leave-one-out experiment, assessing their performance using <math><mrow><mi>F</mi> <mn>1</mn></mrow> </math> score and lesion count metrics. Finally, we tested whether an ensemble of the networks outperformed any single model.</p><p><strong>Results: </strong>Mask R-CNN model achieved an <math><mrow><mi>F</mi> <mn>1</mn></mrow> </math> score of 0.75, YOLOv8 a score of 0.75, E2EC a score of 0.70, UNet++ a score of 0.81, and baseline UNet a score of 0.79. Bland-Altman analysis of lesion count performance showed a limit of agreement (LoA) width of 62.2 for Mask R-CNN, 91.3 for YOLOv8, 94.2 for E2EC, and 62.1 for UNet++, with the baseline UNet model achieving 69.1. The ensemble showed an <math><mrow><mi>F</mi> <mn>1</mn></mrow> </math> score performance of 0.78 and LoA width of 67.4.</p><p><strong>Conclusions: </strong>Instance segmentation methods and UNet-based semantic segmentation methods performed equally well in lesion counting. Furthermore, the ensemble of the trained models showed no performance increase over the best-performing model UNet, likely because errors are frequently shared across models. Performance is likely limited by the availability of high-quality photographs for this complex problem, rather than the methodologies used.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"034506"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12177574/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144369413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-06-03DOI: 10.1117/1.JMI.12.3.035001
Constant R Noordman, Steffan J W Borgers, Martijn F Boomsma, Thomas C Kwee, Marloes M G van der Lees, Christiaan G Overduin, Maarten de Rooij, Derya Yakar, Jurgen J Fütterer, Henkjan J Huisman
Purpose: Interventional MR imaging struggles with speed and efficiency. We aim to accelerate transrectal in-bore MR-guided biopsies for prostate cancer through undersampled image reconstruction and instrument localization by image segmentation.
Approach: In this single-center retrospective study, we used 8464 MR 2D multislice scans from 1289 patients undergoing a prostate biopsy to train and test a deep learning-based spatiotemporal MR image reconstruction model and a nnU-Net segmentation model. The dataset was synthetically undersampled using various undersampling rates ( , 16, 25, 32). An annotated, unseen subset of these data was used to compare our model with a nontemporal model and readers in a reader study involving seven radiologists from three centers based in the Netherlands. We assessed a maximum noninferior undersampling rate using instrument prediction success rate and instrument tip position (ITP) error.
Results: The maximum noninferior undersampling rate is 16-times for the temporal model (ITP error: 2.28 mm, 95% CI: 1.68 to 3.31, mean difference from reference standard: 0.63 mm, ), whereas a nontemporal model could not produce noninferior image reconstructions comparable to our reference standard. Furthermore, the nontemporal model (ITP error: 6.27 mm, 95% CI: 3.90 to 9.07) and readers (ITP error: 6.87 mm, 95% CI: 6.38 to 7.40) had low instrument prediction success rates (46% and 60%, respectively) compared with the temporal model's 95%.
Conclusion: Deep learning-based spatiotemporal MR image reconstruction can improve time-critical intervention tasks such as instrument tracking. We found 16 times undersampling as the maximum noninferior acceleration where image quality is preserved, ITP error is minimized, and the instrument prediction success rate is maximized.
目的:介入磁共振成像的速度和效率。我们的目标是通过采样不足的图像重建和图像分割的仪器定位来加速经直肠前列腺癌的磁共振引导活检。方法:在这项单中心回顾性研究中,我们使用了1289例前列腺活检患者的8464张磁共振二维多层扫描图来训练和测试基于深度学习的时空磁共振图像重建模型和nnU-Net分割模型。使用不同的欠采样率(R = 8,16,25,32)对数据集进行综合欠采样。在一项涉及来自荷兰三个中心的七名放射科医生的读者研究中,使用这些数据的一个注释的、未见过的子集将我们的模型与非时间模型和读者进行比较。我们使用仪器预测成功率和仪器尖端位置(ITP)误差来评估最大非劣欠采样率。结果:时间模型的最大非劣欠采样率为16次(ITP误差:2.28 mm, 95% CI: 1.68 ~ 3.31,与参考标准的平均差值:0.63 mm, P =。09),而非时间模型无法产生与我们的参考标准相当的非劣质图像重建。此外,与时间模型的95%相比,非时间模型(ITP误差:6.27 mm, 95% CI: 3.90至9.07)和读取器(ITP误差:6.87 mm, 95% CI: 6.38至7.40)的仪器预测成功率较低(分别为46%和60%)。结论:基于深度学习的时空磁共振图像重建可以改善仪器跟踪等时间关键型干预任务。我们发现16次欠采样作为最大非劣等加速,在此条件下,图像质量得以保留,ITP误差最小化,仪器预测成功率最大化。
{"title":"Deep learning-based temporal MR image reconstruction for accelerated interventional imaging during in-bore biopsies.","authors":"Constant R Noordman, Steffan J W Borgers, Martijn F Boomsma, Thomas C Kwee, Marloes M G van der Lees, Christiaan G Overduin, Maarten de Rooij, Derya Yakar, Jurgen J Fütterer, Henkjan J Huisman","doi":"10.1117/1.JMI.12.3.035001","DOIUrl":"10.1117/1.JMI.12.3.035001","url":null,"abstract":"<p><strong>Purpose: </strong>Interventional MR imaging struggles with speed and efficiency. We aim to accelerate transrectal in-bore MR-guided biopsies for prostate cancer through undersampled image reconstruction and instrument localization by image segmentation.</p><p><strong>Approach: </strong>In this single-center retrospective study, we used 8464 MR 2D multislice scans from 1289 patients undergoing a prostate biopsy to train and test a deep learning-based spatiotemporal MR image reconstruction model and a nnU-Net segmentation model. The dataset was synthetically undersampled using various undersampling rates ( <math><mrow><mi>R</mi> <mo>=</mo> <mn>8</mn></mrow> </math> , 16, 25, 32). An annotated, unseen subset of these data was used to compare our model with a nontemporal model and readers in a reader study involving seven radiologists from three centers based in the Netherlands. We assessed a maximum noninferior undersampling rate using instrument prediction success rate and instrument tip position (ITP) error.</p><p><strong>Results: </strong>The maximum noninferior undersampling rate is 16-times for the temporal model (ITP error: 2.28 mm, 95% CI: 1.68 to 3.31, mean difference from reference standard: 0.63 mm, <math><mrow><mi>P</mi> <mo>=</mo> <mo>.</mo> <mn>09</mn></mrow> </math> ), whereas a nontemporal model could not produce noninferior image reconstructions comparable to our reference standard. Furthermore, the nontemporal model (ITP error: 6.27 mm, 95% CI: 3.90 to 9.07) and readers (ITP error: 6.87 mm, 95% CI: 6.38 to 7.40) had low instrument prediction success rates (46% and 60%, respectively) compared with the temporal model's 95%.</p><p><strong>Conclusion: </strong>Deep learning-based spatiotemporal MR image reconstruction can improve time-critical intervention tasks such as instrument tracking. We found 16 times undersampling as the maximum noninferior acceleration where image quality is preserved, ITP error is minimized, and the instrument prediction success rate is maximized.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"035001"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12131189/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144227256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-05-21DOI: 10.1117/1.JMI.12.3.035501
Vaibhav Sharma, Alina Jade Barnett, Julia Yang, Sangwook Cheon, Giyoung Kim, Fides Regina Schwartz, Avivah Wang, Neal Hall, Lars Grimm, Chaofan Chen, Joseph Y Lo, Cynthia Rudin
Purpose: Breast cancer remains a leading cause of death for women. Screening programs are deployed to detect cancer at early stages. One current barrier identified by breast imaging researchers is a shortage of labeled image datasets. Addressing this problem is crucial to improve early detection models. We present an active learning (AL) framework for segmenting breast masses from 2D digital mammography, and we publish labeled data. Our method aims to reduce the input needed from expert annotators to reach a fully labeled dataset.
Approach: We create a dataset of 1136 mammographic masses with pixel-wise binary segmentation labels, with the test subset labeled independently by two different teams. With this dataset, we simulate a human annotator within an AL framework to develop and compare AI-assisted labeling methods, using a discriminator model and a simulated oracle to collect acceptable segmentation labels. A UNet model is retrained on these labels, generating new segmentations. We evaluate various oracle heuristics using the percentage of segmentations that the oracle relabels and measure the quality of the proposed labels by evaluating the intersection over union over a validation dataset.
Results: Our method reduces expert annotator input by 44%. We present a dataset of 1136 binary segmentation labels approved by board-certified radiologists and make the 143-image validation set public for comparison with other researchers' methods.
Conclusions: We demonstrate that AL can significantly improve the efficiency and time-effectiveness of creating labeled mammogram datasets. Our framework facilitates the development of high-quality datasets while minimizing manual effort in the domain of digital mammography.
{"title":"Improving annotation efficiency for fully labeling a breast mass segmentation dataset.","authors":"Vaibhav Sharma, Alina Jade Barnett, Julia Yang, Sangwook Cheon, Giyoung Kim, Fides Regina Schwartz, Avivah Wang, Neal Hall, Lars Grimm, Chaofan Chen, Joseph Y Lo, Cynthia Rudin","doi":"10.1117/1.JMI.12.3.035501","DOIUrl":"10.1117/1.JMI.12.3.035501","url":null,"abstract":"<p><strong>Purpose: </strong>Breast cancer remains a leading cause of death for women. Screening programs are deployed to detect cancer at early stages. One current barrier identified by breast imaging researchers is a shortage of labeled image datasets. Addressing this problem is crucial to improve early detection models. We present an active learning (AL) framework for segmenting breast masses from 2D digital mammography, and we publish labeled data. Our method aims to reduce the input needed from expert annotators to reach a fully labeled dataset.</p><p><strong>Approach: </strong>We create a dataset of 1136 mammographic masses with pixel-wise binary segmentation labels, with the test subset labeled independently by two different teams. With this dataset, we simulate a human annotator within an AL framework to develop and compare AI-assisted labeling methods, using a discriminator model and a simulated oracle to collect acceptable segmentation labels. A UNet model is retrained on these labels, generating new segmentations. We evaluate various oracle heuristics using the percentage of segmentations that the oracle relabels and measure the quality of the proposed labels by evaluating the intersection over union over a validation dataset.</p><p><strong>Results: </strong>Our method reduces expert annotator input by 44%. We present a dataset of 1136 binary segmentation labels approved by board-certified radiologists and make the 143-image validation set public for comparison with other researchers' methods.</p><p><strong>Conclusions: </strong>We demonstrate that AL can significantly improve the efficiency and time-effectiveness of creating labeled mammogram datasets. Our framework facilitates the development of high-quality datasets while minimizing manual effort in the domain of digital mammography.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"035501"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12094908/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144144147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-05-23DOI: 10.1117/1.JMI.12.3.034503
Tianyang Wang, Xiumei Li, Ruyu Liu, Meixi Wang, Junmei Sun
Purpose: Early-stage pneumonia is not easily detected, leading to many patients missing the optimal treatment window. This is because segmenting lesion areas from CT images presents several challenges, including low-intensity contrast between the lesion and normal areas, as well as variations in the shape and size of lesion areas. To overcome these challenges, we propose a segmentation network called DECE-Net to segment the pneumonia lesions from CT images automatically.
Approach: The DECE-Net adds an extra encoder path to the U-Net, where one encoder path extracts the features of the original CT image with the attention multi-scale feature fusion module, and the other encoder path extracts the contour features in the CT contour image with the contour feature extraction module to compensate and enhance the boundary information that is lost in the downsampling process. The network further fuses the low-level features from both encoder paths through the feature fusion attention connection module and connects them to the upsampled high-level features to replace the skip connections in the U-Net. Finally, multi-point deep supervision is applied to the segmentation results at each scale to improve segmentation accuracy.
Results: We evaluate the DECE-Net using four public COVID-19 segmentation datasets. The mIoU results for the four datasets are 80.76%, 84.59%, 84.41%, and 78.55%, respectively.
Conclusions: The experimental results indicate that the proposed DECE-Net achieves state-of-the-art performance, especially in the precise segmentation of small lesion areas.
{"title":"DECE-Net: a dual-path encoder network with contour enhancement for pneumonia lesion segmentation.","authors":"Tianyang Wang, Xiumei Li, Ruyu Liu, Meixi Wang, Junmei Sun","doi":"10.1117/1.JMI.12.3.034503","DOIUrl":"10.1117/1.JMI.12.3.034503","url":null,"abstract":"<p><strong>Purpose: </strong>Early-stage pneumonia is not easily detected, leading to many patients missing the optimal treatment window. This is because segmenting lesion areas from CT images presents several challenges, including low-intensity contrast between the lesion and normal areas, as well as variations in the shape and size of lesion areas. To overcome these challenges, we propose a segmentation network called DECE-Net to segment the pneumonia lesions from CT images automatically.</p><p><strong>Approach: </strong>The DECE-Net adds an extra encoder path to the U-Net, where one encoder path extracts the features of the original CT image with the attention multi-scale feature fusion module, and the other encoder path extracts the contour features in the CT contour image with the contour feature extraction module to compensate and enhance the boundary information that is lost in the downsampling process. The network further fuses the low-level features from both encoder paths through the feature fusion attention connection module and connects them to the upsampled high-level features to replace the skip connections in the U-Net. Finally, multi-point deep supervision is applied to the segmentation results at each scale to improve segmentation accuracy.</p><p><strong>Results: </strong>We evaluate the DECE-Net using four public COVID-19 segmentation datasets. The mIoU results for the four datasets are 80.76%, 84.59%, 84.41%, and 78.55%, respectively.</p><p><strong>Conclusions: </strong>The experimental results indicate that the proposed DECE-Net achieves state-of-the-art performance, especially in the precise segmentation of small lesion areas.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"034503"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12101900/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144144146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-05-01Epub Date: 2025-05-21DOI: 10.1117/1.JMI.12.3.034501
Harshani Fonseka, Soheil Varastehpour, Masoud Shakiba, Ehsan Golkar, David Tien
Purpose: Alzheimer's disease (AD) is becoming more prevalent among the elderly, with projections indicating that it will affect a significantly large population in the future. Regardless of substantial research efforts and investments focused on exploring the underlying biological factors, a definitive cure has yet to be discovered. The currently available treatments are only effective in slowing disease progression if it is identified in the early stages of the disease. Therefore, early diagnosis has become critical in treating AD.
Approach: Recently, the use of deep learning techniques has demonstrated remarkable improvement in enhancing the precision and speed of automatic AD diagnosis through medical image analysis. We propose a hybrid model that integrates a convolutional variational auto-encoder (CVAE) with a vision transformer (ViT). During the encoding phase, the CVAE captures key features from the MRI scans, whereas the decoding phase reduces irrelevant details in MRIs. These refined inputs enhance the ViT's ability to analyze complex patterns through its multihead attention mechanism.
Results: The model was trained and evaluated using 14,000 structural MRI samples from the ADNI and SCAN databases. Compared with three benchmark methods and previous studies with Alzheimer's classification techniques, our approach achieved a significant improvement, with a test accuracy of 93.3%.
Conclusions: Through this research, we identified the potential of the CVAE-ViT hybrid approach in detecting minor structural abnormalities related to AD. Integrating unsupervised feature extraction via CVAE can significantly enhance transformer-based models in distinguishing between stages of cognitive impairment, thereby identifying early indicators of AD.
{"title":"Convolutional variational auto-encoder and vision transformer hybrid approach for enhanced early Alzheimer's detection.","authors":"Harshani Fonseka, Soheil Varastehpour, Masoud Shakiba, Ehsan Golkar, David Tien","doi":"10.1117/1.JMI.12.3.034501","DOIUrl":"10.1117/1.JMI.12.3.034501","url":null,"abstract":"<p><strong>Purpose: </strong>Alzheimer's disease (AD) is becoming more prevalent among the elderly, with projections indicating that it will affect a significantly large population in the future. Regardless of substantial research efforts and investments focused on exploring the underlying biological factors, a definitive cure has yet to be discovered. The currently available treatments are only effective in slowing disease progression if it is identified in the early stages of the disease. Therefore, early diagnosis has become critical in treating AD.</p><p><strong>Approach: </strong>Recently, the use of deep learning techniques has demonstrated remarkable improvement in enhancing the precision and speed of automatic AD diagnosis through medical image analysis. We propose a hybrid model that integrates a convolutional variational auto-encoder (CVAE) with a vision transformer (ViT). During the encoding phase, the CVAE captures key features from the MRI scans, whereas the decoding phase reduces irrelevant details in MRIs. These refined inputs enhance the ViT's ability to analyze complex patterns through its multihead attention mechanism.</p><p><strong>Results: </strong>The model was trained and evaluated using 14,000 structural MRI samples from the ADNI and SCAN databases. Compared with three benchmark methods and previous studies with Alzheimer's classification techniques, our approach achieved a significant improvement, with a test accuracy of 93.3%.</p><p><strong>Conclusions: </strong>Through this research, we identified the potential of the CVAE-ViT hybrid approach in detecting minor structural abnormalities related to AD. Integrating unsupervised feature extraction via CVAE can significantly enhance transformer-based models in distinguishing between stages of cognitive impairment, thereby identifying early indicators of AD.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 3","pages":"034501"},"PeriodicalIF":1.9,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12094909/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144144145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}