首页 > 最新文献

International Journal of Computer Assisted Radiology and Surgery最新文献

英文 中文
Multi-modal dataset creation for federated learning with DICOM-structured reports.
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-02-03 DOI: 10.1007/s11548-025-03327-y
Malte Tölle, Lukas Burger, Halvar Kelm, Florian André, Peter Bannas, Gerhard Diller, Norbert Frey, Philipp Garthe, Stefan Groß, Anja Hennemuth, Lars Kaderali, Nina Krüger, Andreas Leha, Simon Martin, Alexander Meyer, Eike Nagel, Stefan Orwat, Clemens Scherer, Moritz Seiffert, Jan Moritz Seliger, Stefan Simm, Tim Friede, Tim Seidler, Sandy Engelhardt

Purpose Federated training is often challenging on heterogeneous datasets due to divergent data storage options, inconsistent naming schemes, varied annotation procedures, and disparities in label quality. This is particularly evident in the emerging multi-modal learning paradigms, where dataset harmonization including a uniform data representation and filtering options are of paramount importance.Methods DICOM-structured reports enable the standardized linkage of arbitrary information beyond the imaging domain and can be used within Python deep learning pipelines with highdicom. Building on this, we developed an open platform for data integration with interactive filtering capabilities, thereby simplifying the process of creation of patient cohorts over several sites with consistent multi-modal data.Results In this study, we extend our prior work by showing its applicability to more and divergent data types, as well as streamlining datasets for federated training within an established consortium of eight university hospitals in Germany. We prove its concurrent filtering ability by creating harmonized multi-modal datasets across all locations for predicting the outcome after minimally invasive heart valve replacement. The data include imaging and waveform data (i.e., computed tomography images, electrocardiography scans) as well as annotations (i.e., calcification segmentations, and pointsets), and metadata (i.e., prostheses and pacemaker dependency).Conclusion Structured reports bridge the traditional gap between imaging systems and information systems. Utilizing the inherent DICOM reference system arbitrary data types can be queried concurrently to create meaningful cohorts for multi-centric data analysis. The graphical interface as well as example structured report templates are available at https://github.com/Cardio-AI/fl-multi-modal-dataset-creation .

{"title":"Multi-modal dataset creation for federated learning with DICOM-structured reports.","authors":"Malte Tölle, Lukas Burger, Halvar Kelm, Florian André, Peter Bannas, Gerhard Diller, Norbert Frey, Philipp Garthe, Stefan Groß, Anja Hennemuth, Lars Kaderali, Nina Krüger, Andreas Leha, Simon Martin, Alexander Meyer, Eike Nagel, Stefan Orwat, Clemens Scherer, Moritz Seiffert, Jan Moritz Seliger, Stefan Simm, Tim Friede, Tim Seidler, Sandy Engelhardt","doi":"10.1007/s11548-025-03327-y","DOIUrl":"https://doi.org/10.1007/s11548-025-03327-y","url":null,"abstract":"<p><p>Purpose Federated training is often challenging on heterogeneous datasets due to divergent data storage options, inconsistent naming schemes, varied annotation procedures, and disparities in label quality. This is particularly evident in the emerging multi-modal learning paradigms, where dataset harmonization including a uniform data representation and filtering options are of paramount importance.Methods DICOM-structured reports enable the standardized linkage of arbitrary information beyond the imaging domain and can be used within Python deep learning pipelines with highdicom. Building on this, we developed an open platform for data integration with interactive filtering capabilities, thereby simplifying the process of creation of patient cohorts over several sites with consistent multi-modal data.Results In this study, we extend our prior work by showing its applicability to more and divergent data types, as well as streamlining datasets for federated training within an established consortium of eight university hospitals in Germany. We prove its concurrent filtering ability by creating harmonized multi-modal datasets across all locations for predicting the outcome after minimally invasive heart valve replacement. The data include imaging and waveform data (i.e., computed tomography images, electrocardiography scans) as well as annotations (i.e., calcification segmentations, and pointsets), and metadata (i.e., prostheses and pacemaker dependency).Conclusion Structured reports bridge the traditional gap between imaging systems and information systems. Utilizing the inherent DICOM reference system arbitrary data types can be queried concurrently to create meaningful cohorts for multi-centric data analysis. The graphical interface as well as example structured report templates are available at https://github.com/Cardio-AI/fl-multi-modal-dataset-creation .</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143081992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DenseSeg: joint learning for semantic segmentation and landmark detection using dense image-to-shape representation.
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-23 DOI: 10.1007/s11548-024-03315-8
Ron Keuth, Lasse Hansen, Maren Balks, Ronja Jäger, Anne-Nele Schröder, Ludger Tüshaus, Mattias Heinrich

Purpose: Semantic segmentation and landmark detection are fundamental tasks of medical image processing, facilitating further analysis of anatomical objects. Although deep learning-based pixel-wise classification has set a new-state-of-the-art for segmentation, it falls short in landmark detection, a strength of shape-based approaches.

Methods: In this work, we propose a dense image-to-shape representation that enables the joint learning of landmarks and semantic segmentation by employing a fully convolutional architecture. Our method intuitively allows the extraction of arbitrary landmarks due to its representation of anatomical correspondences. We benchmark our method against the state-of-the-art for semantic segmentation (nnUNet), a shape-based approach employing geometric deep learning and a convolutional neural network-based method for landmark detection.

Results: We evaluate our method on two medical datasets: one common benchmark featuring the lungs, heart, and clavicle from thorax X-rays, and another with 17 different bones in the paediatric wrist. While our method is on par with the landmark detection baseline in the thorax setting (error in mm of 2.6 ± 0.9 vs. 2.7 ± 0.9 ), it substantially surpassed it in the more complex wrist setting ( 1.1 ± 0.6 vs. 1.9 ± 0.5 ).

Conclusion: We demonstrate that dense geometric shape representation is beneficial for challenging landmark detection tasks and outperforms previous state-of-the-art using heatmap regression. While it does not require explicit training on the landmarks themselves, allowing for the addition of new landmarks without necessitating retraining.

{"title":"DenseSeg: joint learning for semantic segmentation and landmark detection using dense image-to-shape representation.","authors":"Ron Keuth, Lasse Hansen, Maren Balks, Ronja Jäger, Anne-Nele Schröder, Ludger Tüshaus, Mattias Heinrich","doi":"10.1007/s11548-024-03315-8","DOIUrl":"https://doi.org/10.1007/s11548-024-03315-8","url":null,"abstract":"<p><strong>Purpose: </strong>Semantic segmentation and landmark detection are fundamental tasks of medical image processing, facilitating further analysis of anatomical objects. Although deep learning-based pixel-wise classification has set a new-state-of-the-art for segmentation, it falls short in landmark detection, a strength of shape-based approaches.</p><p><strong>Methods: </strong>In this work, we propose a dense image-to-shape representation that enables the joint learning of landmarks and semantic segmentation by employing a fully convolutional architecture. Our method intuitively allows the extraction of arbitrary landmarks due to its representation of anatomical correspondences. We benchmark our method against the state-of-the-art for semantic segmentation (nnUNet), a shape-based approach employing geometric deep learning and a convolutional neural network-based method for landmark detection.</p><p><strong>Results: </strong>We evaluate our method on two medical datasets: one common benchmark featuring the lungs, heart, and clavicle from thorax X-rays, and another with 17 different bones in the paediatric wrist. While our method is on par with the landmark detection baseline in the thorax setting (error in mm of <math><mrow><mn>2.6</mn> <mo>±</mo> <mn>0.9</mn></mrow> </math> vs. <math><mrow><mn>2.7</mn> <mo>±</mo> <mn>0.9</mn></mrow> </math> ), it substantially surpassed it in the more complex wrist setting ( <math><mrow><mn>1.1</mn> <mo>±</mo> <mn>0.6</mn></mrow> </math> vs. <math><mrow><mn>1.9</mn> <mo>±</mo> <mn>0.5</mn></mrow> </math> ).</p><p><strong>Conclusion: </strong>We demonstrate that dense geometric shape representation is beneficial for challenging landmark detection tasks and outperforms previous state-of-the-art using heatmap regression. While it does not require explicit training on the landmarks themselves, allowing for the addition of new landmarks without necessitating retraining.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143030205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Volume and quality of the gluteal muscles are associated with early physical function after total hip arthroplasty. 臀肌的体积和质量与全髋关节置换术后早期的身体功能有关。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-21 DOI: 10.1007/s11548-025-03321-4
Makoto Iwasa, Keisuke Uemura, Mazen Soufi, Yoshito Otake, Tomofumi Kinoshita, Tatsuhiko Kutsuna, Kazuma Takashima, Hidetoshi Hamada, Yoshinobu Sato, Nobuhiko Sugano, Seiji Okada, Masaki Takao

Purpose: Identifying muscles linked to postoperative physical function can guide protocols to enhance early recovery following total hip arthroplasty (THA). This study aimed to evaluate the association of preoperative pelvic and thigh muscle volume and quality with early physical function after THA in patients with unilateral hip osteoarthritis (HOA).

Methods: Preoperative Computed tomography (CT) images of 61 patients (eight males and 53 females) with HOA were analyzed. Six muscle groups were segmented from CT images, and muscle volume and quality were calculated on the healthy and affected sides. Muscle quality was quantified using the mean CT values (Hounsfield units [HU]). Early postoperative physical function was evaluated using the Timed Up & Go test (TUG) at three weeks after THA. The effect of preoperative muscle volume and quality of both sides on early postoperative physical function was assessed.

Results: On the healthy and affected sides, mean muscle mass was 9.7 cm3/kg and 8.1 cm3/kg, and mean muscle HU values were 46.0 HU and 39.1 HU, respectively. Significant differences in muscle volume and quality were observed between the affected and healthy sides. On analyzing the function of various muscle groups, the TUG score showed a significant association with the gluteus maximum volume and the gluteus medius/minimus quality on the affected side.

Conclusion: Patients with HOA showed significant muscle atrophy and fatty degeneration in the affected pelvic and thigh regions. The gluteus maximum volume and gluteus medius/minimus quality were associated with early postoperative physical function. Preoperative rehabilitation targeting the gluteal muscles on the affected side could potentially enhance recovery of physical function in the early postoperative period.

目的:识别与术后身体功能相关的肌肉可以指导方案,以增强全髋关节置换术(THA)后的早期恢复。本研究旨在评估单侧髋关节骨关节炎(HOA)患者术前骨盆和大腿肌肉体积和质量与THA术后早期身体功能的关系。方法:对61例HOA患者(男8例,女53例)术前CT图像进行分析。从CT图像中分割出6组肌肉,计算健康侧和病变侧的肌肉体积和质量。肌肉质量采用CT平均值(Hounsfield单位[HU])进行量化。THA术后三周采用定时Up & Go测试(TUG)评估早期术后身体功能。评估术前两侧肌肉体积和质量对术后早期身体功能的影响。结果:健康侧和患侧平均肌肉质量分别为9.7 cm3/kg和8.1 cm3/kg,平均肌肉HU值分别为46.0 HU和39.1 HU。在患病侧和健康侧之间观察到肌肉体积和质量的显著差异。在分析各肌肉群的功能时,TUG评分显示与患侧臀肌最大体积和臀中/臀小肌质量有显著关联。结论:HOA患者在骨盆和大腿部位表现出明显的肌肉萎缩和脂肪变性。臀肌最大体积和臀中/臀小肌质量与术后早期身体功能有关。术前以患侧臀肌为目标的康复治疗可促进术后早期身体功能的恢复。
{"title":"Volume and quality of the gluteal muscles are associated with early physical function after total hip arthroplasty.","authors":"Makoto Iwasa, Keisuke Uemura, Mazen Soufi, Yoshito Otake, Tomofumi Kinoshita, Tatsuhiko Kutsuna, Kazuma Takashima, Hidetoshi Hamada, Yoshinobu Sato, Nobuhiko Sugano, Seiji Okada, Masaki Takao","doi":"10.1007/s11548-025-03321-4","DOIUrl":"https://doi.org/10.1007/s11548-025-03321-4","url":null,"abstract":"<p><strong>Purpose: </strong>Identifying muscles linked to postoperative physical function can guide protocols to enhance early recovery following total hip arthroplasty (THA). This study aimed to evaluate the association of preoperative pelvic and thigh muscle volume and quality with early physical function after THA in patients with unilateral hip osteoarthritis (HOA).</p><p><strong>Methods: </strong>Preoperative Computed tomography (CT) images of 61 patients (eight males and 53 females) with HOA were analyzed. Six muscle groups were segmented from CT images, and muscle volume and quality were calculated on the healthy and affected sides. Muscle quality was quantified using the mean CT values (Hounsfield units [HU]). Early postoperative physical function was evaluated using the Timed Up & Go test (TUG) at three weeks after THA. The effect of preoperative muscle volume and quality of both sides on early postoperative physical function was assessed.</p><p><strong>Results: </strong>On the healthy and affected sides, mean muscle mass was 9.7 cm<sup>3</sup>/kg and 8.1 cm<sup>3</sup>/kg, and mean muscle HU values were 46.0 HU and 39.1 HU, respectively. Significant differences in muscle volume and quality were observed between the affected and healthy sides. On analyzing the function of various muscle groups, the TUG score showed a significant association with the gluteus maximum volume and the gluteus medius/minimus quality on the affected side.</p><p><strong>Conclusion: </strong>Patients with HOA showed significant muscle atrophy and fatty degeneration in the affected pelvic and thigh regions. The gluteus maximum volume and gluteus medius/minimus quality were associated with early postoperative physical function. Preoperative rehabilitation targeting the gluteal muscles on the affected side could potentially enhance recovery of physical function in the early postoperative period.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143015925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perfusion estimation from dynamic non-contrast computed tomography using self-supervised learning and a physics-inspired U-net transformer architecture. 使用自监督学习和物理启发的U-net变压器架构的动态非对比计算机断层扫描灌注估计。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-20 DOI: 10.1007/s11548-025-03323-2
Yi-Kuan Liu, Jorge Cisneros, Girish Nair, Craig Stevens, Richard Castillo, Yevgeniy Vinogradskiy, Edward Castillo

Purpose: Pulmonary perfusion imaging is a key lung health indicator with clinical utility as a diagnostic and treatment planning tool. However, current nuclear medicine modalities face challenges like low spatial resolution and long acquisition times which limit clinical utility to non-emergency settings and often placing extra financial burden on the patient. This study introduces a novel deep learning approach to predict perfusion imaging from non-contrast inhale and exhale computed tomography scans (IE-CT).

Methods: We developed a U-Net Transformer architecture modified for Siamese IE-CT inputs, integrating insights from physical models and utilizing a self-supervised learning strategy tailored for lung function prediction. We aggregated 523 IE-CT images from nine different 4DCT imaging datasets for self-supervised training, aiming to learn a low-dimensional IE-CT feature space by reconstructing image volumes from random data augmentations. Supervised training for perfusion prediction used this feature space and transfer learning on a cohort of 44 patients who had both IE-CT and single-photon emission CT (SPECT/CT) perfusion scans.

Results: Testing with random bootstrapping, we estimated the mean and standard deviation of the spatial Spearman correlation between our predictions and the ground truth (SPECT perfusion) to be 0.742 ± 0.037, with a mean median correlation of 0.792 ± 0.036. These results represent a new state-of-the-art accuracy for predicting perfusion imaging from non-contrast CT.

Conclusion: Our approach combines low-dimensional feature representations of both inhale and exhale images into a deep learning model, aligning with previous physical modeling methods for characterizing perfusion from IE-CT. This likely contributes to the high spatial correlation with ground truth. With further development, our method could provide faster and more accurate lung function imaging, potentially expanding its clinical applications beyond what is currently possible with nuclear medicine.

目的:肺灌注显像是一项重要的肺健康指标,可作为临床诊断和治疗计划的工具。然而,目前的核医学模式面临着诸如低空间分辨率和长采集时间等挑战,这些挑战限制了非紧急情况下的临床应用,并往往给患者带来额外的经济负担。本研究引入了一种新的深度学习方法来预测非对比吸气和呼气计算机断层扫描(IE-CT)的灌注成像。方法:我们开发了一种针对Siamese IE-CT输入进行修改的U-Net Transformer架构,整合了来自物理模型的见解,并利用了为肺功能预测量身定制的自监督学习策略。我们收集了来自9个不同的4DCT成像数据集的523张IE-CT图像进行自监督训练,旨在通过随机数据增强重建图像体来学习低维IE-CT特征空间。对44名同时进行IE-CT和单光子发射CT (SPECT/CT)灌注扫描的患者进行灌注预测的监督训练,使用该特征空间和迁移学习。结果:随机自适应检验,我们估计我们的预测与地面真相(SPECT灌注)之间的空间Spearman相关性的平均值和标准差为0.742±0.037,平均中位数相关性为0.792±0.036。这些结果代表了一个新的国家的最先进的准确性预测灌注成像从非对比CT。结论:我们的方法将吸气和呼气图像的低维特征表示结合到一个深度学习模型中,与先前用于表征IE-CT灌注的物理建模方法保持一致。这可能有助于与地面真值的高空间相关性。随着进一步的发展,我们的方法可以提供更快、更准确的肺功能成像,有可能扩大其临床应用范围,超越目前核医学的可能性。
{"title":"Perfusion estimation from dynamic non-contrast computed tomography using self-supervised learning and a physics-inspired U-net transformer architecture.","authors":"Yi-Kuan Liu, Jorge Cisneros, Girish Nair, Craig Stevens, Richard Castillo, Yevgeniy Vinogradskiy, Edward Castillo","doi":"10.1007/s11548-025-03323-2","DOIUrl":"https://doi.org/10.1007/s11548-025-03323-2","url":null,"abstract":"<p><strong>Purpose: </strong>Pulmonary perfusion imaging is a key lung health indicator with clinical utility as a diagnostic and treatment planning tool. However, current nuclear medicine modalities face challenges like low spatial resolution and long acquisition times which limit clinical utility to non-emergency settings and often placing extra financial burden on the patient. This study introduces a novel deep learning approach to predict perfusion imaging from non-contrast inhale and exhale computed tomography scans (IE-CT).</p><p><strong>Methods: </strong>We developed a U-Net Transformer architecture modified for Siamese IE-CT inputs, integrating insights from physical models and utilizing a self-supervised learning strategy tailored for lung function prediction. We aggregated 523 IE-CT images from nine different 4DCT imaging datasets for self-supervised training, aiming to learn a low-dimensional IE-CT feature space by reconstructing image volumes from random data augmentations. Supervised training for perfusion prediction used this feature space and transfer learning on a cohort of 44 patients who had both IE-CT and single-photon emission CT (SPECT/CT) perfusion scans.</p><p><strong>Results: </strong>Testing with random bootstrapping, we estimated the mean and standard deviation of the spatial Spearman correlation between our predictions and the ground truth (SPECT perfusion) to be 0.742 ± 0.037, with a mean median correlation of 0.792 ± 0.036. These results represent a new state-of-the-art accuracy for predicting perfusion imaging from non-contrast CT.</p><p><strong>Conclusion: </strong>Our approach combines low-dimensional feature representations of both inhale and exhale images into a deep learning model, aligning with previous physical modeling methods for characterizing perfusion from IE-CT. This likely contributes to the high spatial correlation with ground truth. With further development, our method could provide faster and more accurate lung function imaging, potentially expanding its clinical applications beyond what is currently possible with nuclear medicine.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143015924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-guided erasing for enhanced transfer learning in breast abnormality classification. 注意引导擦除增强乳房异常分类中的迁移学习。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-15 DOI: 10.1007/s11548-024-03317-6
Adarsh Bhandary Panambur, Sheethal Bhat, Hui Yu, Prathmesh Madhu, Siming Bayer, Andreas Maier

Purpose: Breast cancer remains one of the most prevalent cancers globally, necessitating effective early screening and diagnosis. This study investigates the effectiveness and generalizability of our recently proposed data augmentation technique, attention-guided erasing (AGE), across various transfer learning classification tasks for breast abnormality classification in mammography.

Methods: AGE utilizes attention head visualizations from DINO self-supervised pretraining to weakly localize regions of interest (ROI) in images. These localizations are then used to stochastically erase non-essential background information from training images during transfer learning. Our research evaluates AGE across two image-level and three patch-level classification tasks. The image-level tasks involve breast density categorization in digital mammography (DM) and malignancy classification in contrast-enhanced mammography (CEM). Patch-level tasks include classifying calcifications and masses in scanned film mammography (SFM), as well as malignancy classification of ROIs in CEM.

Results: AGE significantly boosts classification performance with statistically significant improvements in mean F1-scores across four tasks compared to baselines. Specifically, for image-level classification of breast density in DM and malignancy in CEM, we achieve gains of 2% and 1.5%, respectively. Additionally, for patch-level classification of calcifications in SFM and CEM ROIs, gains of 0.4% and 0.6% are observed, respectively. However, marginal improvement is noted in the mass classification task, indicating the necessity for further optimization in tasks where critical features may be obscured by erasing techniques.

Conclusion: Our findings underscore the potential of AGE, a dataset- and task-specific augmentation strategy powered by self-supervised learning, to enhance the downstream classification performance of DL models, particularly involving ViTs, in medical imaging.

目的:乳腺癌仍然是全球最常见的癌症之一,需要有效的早期筛查和诊断。本研究探讨了我们最近提出的数据增强技术,即注意力引导擦除(AGE),在各种迁移学习分类任务中用于乳房x光检查中乳房异常分类的有效性和普遍性。方法:AGE利用DINO自监督预训练的注意头可视化来弱定位图像中的兴趣区域(ROI)。然后使用这些定位在迁移学习过程中随机清除训练图像中的非必要背景信息。我们的研究通过两个图像级和三个补丁级分类任务来评估AGE。图像级任务包括数字乳房x线摄影(DM)中的乳腺密度分类和对比增强乳房x线摄影(CEM)中的恶性肿瘤分类。斑块级任务包括扫描乳房x线摄影(SFM)中的钙化和肿块分类,以及CEM中roi的恶性分类。结果:与基线相比,AGE显著提高了分类性能,在四个任务中的平均f1得分有统计学显著提高。具体来说,对于DM和CEM的乳腺密度图像级别分类,我们分别获得了2%和1.5%的增益。此外,对于SFM和CEM roi的斑块级钙化分类,分别观察到0.4%和0.6%的增益。然而,在大规模分类任务中注意到边际改进,这表明在关键特征可能被擦除技术模糊的任务中需要进一步优化。结论:我们的研究结果强调了AGE的潜力,AGE是一种由自我监督学习驱动的数据集和任务特定增强策略,可以增强DL模型的下游分类性能,特别是涉及vit的医学成像。
{"title":"Attention-guided erasing for enhanced transfer learning in breast abnormality classification.","authors":"Adarsh Bhandary Panambur, Sheethal Bhat, Hui Yu, Prathmesh Madhu, Siming Bayer, Andreas Maier","doi":"10.1007/s11548-024-03317-6","DOIUrl":"https://doi.org/10.1007/s11548-024-03317-6","url":null,"abstract":"<p><strong>Purpose: </strong>Breast cancer remains one of the most prevalent cancers globally, necessitating effective early screening and diagnosis. This study investigates the effectiveness and generalizability of our recently proposed data augmentation technique, attention-guided erasing (AGE), across various transfer learning classification tasks for breast abnormality classification in mammography.</p><p><strong>Methods: </strong>AGE utilizes attention head visualizations from DINO self-supervised pretraining to weakly localize regions of interest (ROI) in images. These localizations are then used to stochastically erase non-essential background information from training images during transfer learning. Our research evaluates AGE across two image-level and three patch-level classification tasks. The image-level tasks involve breast density categorization in digital mammography (DM) and malignancy classification in contrast-enhanced mammography (CEM). Patch-level tasks include classifying calcifications and masses in scanned film mammography (SFM), as well as malignancy classification of ROIs in CEM.</p><p><strong>Results: </strong>AGE significantly boosts classification performance with statistically significant improvements in mean F1-scores across four tasks compared to baselines. Specifically, for image-level classification of breast density in DM and malignancy in CEM, we achieve gains of 2% and 1.5%, respectively. Additionally, for patch-level classification of calcifications in SFM and CEM ROIs, gains of 0.4% and 0.6% are observed, respectively. However, marginal improvement is noted in the mass classification task, indicating the necessity for further optimization in tasks where critical features may be obscured by erasing techniques.</p><p><strong>Conclusion: </strong>Our findings underscore the potential of AGE, a dataset- and task-specific augmentation strategy powered by self-supervised learning, to enhance the downstream classification performance of DL models, particularly involving ViTs, in medical imaging.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142985295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shape-matching-based fracture reduction aid concept exemplified on the proximal humerus-a pilot study. 以肱骨近端为例的基于形状匹配的骨折复位辅助概念-一项试点研究。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-14 DOI: 10.1007/s11548-024-03318-5
Karen Mys, Luke Visscher, Sara Lindenmann, Torsten Pastor, Paolo Antonacci, Matthias Knobe, Martin Jaeger, Simon Lambert, Peter Varga

Purpose: Optimizing fracture reduction quality is key to achieve successful osteosynthesis, especially for epimetaphyseal regions such as the proximal humerus (PH), but can be challenging, partly due to the lack of a clear endpoint. We aimed to develop the prototype for a novel intraoperative C-arm-based aid to facilitate true anatomical reduction of fractures of the PH.

Methods: The proposed method designates the reduced endpoint position of fragments by superimposing the outer boundary of the premorbid bone shape on intraoperative C-arm images, taking the mirrored intact contralateral PH from the preoperative CT scan as a surrogate. The accuracy of the algorithm was tested on 60 synthetic C-arm images created from the preoperative CT images of 20 complex PH fracture cases (Dataset A) and on 12 real C-arm images of a prefractured human anatomical specimen (Dataset B). The predicted outer boundary shape was compared with the known exact solution by (1) a calculated matching error and (2) two experienced shoulder trauma surgeons.

Results: A prediction accuracy of 88% (with 73% 'good') was achieved according to the calculation method and an 87% accuracy (68% 'good') by surgeon assessment in Dataset A. Accuracy was 100% by both assessments for Dataset B.

Conclusion: By seamlessly integrating into the standard perioperative workflow and imaging, the intuitive shape-matching-based aid, once developed as a medical device, has the potential to optimize the accuracy of the reduction of PH fractures while reducing the number of X-rays and surgery time. Further studies are required to demonstrate the applicability and efficacy of this method in optimizing fracture reduction quality.

目的:优化骨折复位质量是成功进行骨合成的关键,尤其是对于肱骨近端(PH)等骺区,但这可能具有挑战性,部分原因是缺乏明确的终点。我们旨在开发一种基于 C 臂的新型术中辅助工具原型,以促进 PH 骨折的真正解剖复位:方法:所提出的方法通过在术中 C 臂图像上叠加原骨形状的外边界来指定骨折片的缩小端点位置,并以术前 CT 扫描中的完整对侧 PH 镜像作为替代物。该算法的准确性在 20 个复杂 PH 骨折病例(数据集 A)的术前 CT 图像制作的 60 张合成 C 臂图像和 12 张骨折前人体解剖标本的真实 C 臂图像(数据集 B)上进行了测试。通过(1)计算匹配误差和(2)两名经验丰富的肩部创伤外科医生,将预测的外部边界形状与已知的精确解进行比较:结果:根据计算方法,数据集 A 的预测准确率为 88%(73% 为 "良好"),根据外科医生的评估,准确率为 87%(68% 为 "良好"):通过无缝集成到标准围手术期工作流程和成像中,这种基于形状匹配的直观辅助工具一旦作为医疗设备开发出来,就有可能优化 PH 骨折复位的准确性,同时减少 X 射线的数量和手术时间。要证明这种方法在优化骨折复位质量方面的适用性和有效性,还需要进一步的研究。
{"title":"Shape-matching-based fracture reduction aid concept exemplified on the proximal humerus-a pilot study.","authors":"Karen Mys, Luke Visscher, Sara Lindenmann, Torsten Pastor, Paolo Antonacci, Matthias Knobe, Martin Jaeger, Simon Lambert, Peter Varga","doi":"10.1007/s11548-024-03318-5","DOIUrl":"https://doi.org/10.1007/s11548-024-03318-5","url":null,"abstract":"<p><strong>Purpose: </strong>Optimizing fracture reduction quality is key to achieve successful osteosynthesis, especially for epimetaphyseal regions such as the proximal humerus (PH), but can be challenging, partly due to the lack of a clear endpoint. We aimed to develop the prototype for a novel intraoperative C-arm-based aid to facilitate true anatomical reduction of fractures of the PH.</p><p><strong>Methods: </strong>The proposed method designates the reduced endpoint position of fragments by superimposing the outer boundary of the premorbid bone shape on intraoperative C-arm images, taking the mirrored intact contralateral PH from the preoperative CT scan as a surrogate. The accuracy of the algorithm was tested on 60 synthetic C-arm images created from the preoperative CT images of 20 complex PH fracture cases (Dataset A) and on 12 real C-arm images of a prefractured human anatomical specimen (Dataset B). The predicted outer boundary shape was compared with the known exact solution by (1) a calculated matching error and (2) two experienced shoulder trauma surgeons.</p><p><strong>Results: </strong>A prediction accuracy of 88% (with 73% 'good') was achieved according to the calculation method and an 87% accuracy (68% 'good') by surgeon assessment in Dataset A. Accuracy was 100% by both assessments for Dataset B.</p><p><strong>Conclusion: </strong>By seamlessly integrating into the standard perioperative workflow and imaging, the intuitive shape-matching-based aid, once developed as a medical device, has the potential to optimize the accuracy of the reduction of PH fractures while reducing the number of X-rays and surgery time. Further studies are required to demonstrate the applicability and efficacy of this method in optimizing fracture reduction quality.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142980620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A real-time approach for surgical activity recognition and prediction based on transformer models in robot-assisted surgery. 机器人辅助手术中基于变压器模型的手术活动实时识别与预测方法。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-12 DOI: 10.1007/s11548-024-03306-9
Ketai Chen, D S V Bandara, Jumpei Arata

Purpose: This paper presents a deep learning approach to recognize and predict surgical activity in robot-assisted minimally invasive surgery (RAMIS). Our primary objective is to deploy the developed model for implementing a real-time surgical risk monitoring system within the realm of RAMIS.

Methods: We propose a modified Transformer model with the architecture comprising no positional encoding, 5 fully connected layers, 1 encoder, and 3 decoders. This model is specifically designed to address 3 primary tasks in surgical robotics: gesture recognition, prediction, and end-effector trajectory prediction. Notably, it operates solely on kinematic data obtained from the joints of robotic arm.

Results: The model's performance was evaluated on JHU-ISI Gesture and Skill Assessment Working Set dataset, achieving highest accuracy of 94.4% for gesture recognition, 84.82% for gesture prediction, and significantly low distance error of 1.34 mm with a prediction of 1 s in advance. Notably, the computational time per iteration was minimal recorded at only 4.2 ms.

Conclusion: The results demonstrated the excellence of our proposed model compared to previous studies highlighting its potential for integration in real-time systems. We firmly believe that our model could significantly elevate realms of surgical activity recognition and prediction within RAS and make a substantial and meaningful contribution to the healthcare sector.

目的:本文提出了一种用于机器人辅助微创手术(RAMIS)手术活动识别和预测的深度学习方法。我们的主要目标是部署开发的模型,在RAMIS领域内实施实时手术风险监测系统。方法:我们提出了一个改进的Transformer模型,其架构包括无位置编码,5个完全连接层,1个编码器和3个解码器。该模型专门用于解决手术机器人中的3个主要任务:手势识别、预测和末端执行器轨迹预测。值得注意的是,它仅对从机械臂关节获得的运动学数据进行操作。结果:在JHU-ISI手势和技能评估工作集数据集上对该模型的性能进行了评估,手势识别准确率最高,为94.4%,手势预测准确率为84.82%,距离误差显著降低,为1.34 mm,预测时间提前1 s。值得注意的是,每次迭代的计算时间是最小的,只有4.2毫秒。结论:与之前的研究相比,结果证明了我们提出的模型的卓越性,突出了它在实时系统集成中的潜力。我们坚信,我们的模型可以显著提升RAS内手术活动识别和预测领域,并为医疗保健部门做出实质性和有意义的贡献。
{"title":"A real-time approach for surgical activity recognition and prediction based on transformer models in robot-assisted surgery.","authors":"Ketai Chen, D S V Bandara, Jumpei Arata","doi":"10.1007/s11548-024-03306-9","DOIUrl":"https://doi.org/10.1007/s11548-024-03306-9","url":null,"abstract":"<p><strong>Purpose: </strong>This paper presents a deep learning approach to recognize and predict surgical activity in robot-assisted minimally invasive surgery (RAMIS). Our primary objective is to deploy the developed model for implementing a real-time surgical risk monitoring system within the realm of RAMIS.</p><p><strong>Methods: </strong>We propose a modified Transformer model with the architecture comprising no positional encoding, 5 fully connected layers, 1 encoder, and 3 decoders. This model is specifically designed to address 3 primary tasks in surgical robotics: gesture recognition, prediction, and end-effector trajectory prediction. Notably, it operates solely on kinematic data obtained from the joints of robotic arm.</p><p><strong>Results: </strong>The model's performance was evaluated on JHU-ISI Gesture and Skill Assessment Working Set dataset, achieving highest accuracy of 94.4% for gesture recognition, 84.82% for gesture prediction, and significantly low distance error of 1.34 mm with a prediction of 1 s in advance. Notably, the computational time per iteration was minimal recorded at only 4.2 ms.</p><p><strong>Conclusion: </strong>The results demonstrated the excellence of our proposed model compared to previous studies highlighting its potential for integration in real-time systems. We firmly believe that our model could significantly elevate realms of surgical activity recognition and prediction within RAS and make a substantial and meaningful contribution to the healthcare sector.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142973200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Acknowledgement to reviewers. 感谢审稿人。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-11 DOI: 10.1007/s11548-024-03320-x
{"title":"Acknowledgement to reviewers.","authors":"","doi":"10.1007/s11548-024-03320-x","DOIUrl":"https://doi.org/10.1007/s11548-024-03320-x","url":null,"abstract":"","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142967329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
G-SET-DCL: a guided sequential episodic training with dual contrastive learning approach for colon segmentation. G-SET-DCL:一种带有双重对比学习方法的引导序列情景训练,用于结肠分割。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-09 DOI: 10.1007/s11548-024-03319-4
Samir Farag Harb, Asem Ali, Mohamed Yousuf, Salwa Elshazly, Aly Farag

Purpose: This article introduces a novel deep learning approach to substantially improve the accuracy of colon segmentation even with limited data annotation, which enhances the overall effectiveness of the CT colonography pipeline in clinical settings.

Methods: The proposed approach integrates 3D contextual information via guided sequential episodic training in which a query CT slice is segmented by exploiting its previous labeled CT slice (i.e., support). Segmentation starts by detecting the rectum using a Markov Random Field-based algorithm. Then, supervised sequential episodic training is applied to the remaining slices, while contrastive learning is employed to enhance feature discriminability, thereby improving segmentation accuracy.

Results: The proposed method, evaluated on 98 abdominal scans of prepped patients, achieved a Dice coefficient of 97.3% and a polyp information preservation accuracy of 98.28%. Statistical analysis, including 95% confidence intervals, underscores the method's robustness and reliability. Clinically, this high level of accuracy is vital for ensuring the preservation of critical polyp details, which are essential for accurate automatic diagnostic evaluation. The proposed method performs reliably in scenarios with limited annotated data. This is demonstrated by achieving a Dice coefficient of 97.15% when the model was trained on a smaller number of annotated CT scans (e.g., 10 scans) than the testing dataset (e.g., 88 scans).

Conclusions: The proposed sequential segmentation approach achieves promising results in colon segmentation. A key strength of the method is its ability to generalize effectively, even with limited annotated datasets-a common challenge in medical imaging.

目的:本文介绍了一种新颖的深度学习方法,即使在有限的数据注释下也能大幅提高结肠分割的准确性,从而提高CT结肠镜管道在临床环境中的整体有效性。方法:该方法通过引导序列情景训练集成3D上下文信息,其中查询CT切片通过利用其先前标记的CT切片(即支持)进行分割。分割首先使用基于马尔科夫随机场的算法检测直肠。然后对剩余的切片进行有监督的序列情景训练,同时利用对比学习增强特征的可判别性,从而提高分割的准确率。结果:对98例术前准备患者的腹部扫描进行评估,该方法的Dice系数为97.3%,息肉信息保存准确率为98.28%。统计分析,包括95%的置信区间,强调了方法的稳健性和可靠性。临床上,这种高水平的准确性对于确保保留关键的息肉细节至关重要,这对于准确的自动诊断评估至关重要。该方法在标注数据有限的情况下运行可靠。当模型在较少数量的带注释的CT扫描(例如,10次扫描)上进行训练时,其Dice系数达到97.15%,这一点得到了证明(例如,88次扫描)。结论:提出的顺序分割方法在结肠分割中取得了良好的效果。该方法的一个关键优势是它能够有效地泛化,即使是在有限的注释数据集上——这是医学成像中的一个常见挑战。
{"title":"G-SET-DCL: a guided sequential episodic training with dual contrastive learning approach for colon segmentation.","authors":"Samir Farag Harb, Asem Ali, Mohamed Yousuf, Salwa Elshazly, Aly Farag","doi":"10.1007/s11548-024-03319-4","DOIUrl":"https://doi.org/10.1007/s11548-024-03319-4","url":null,"abstract":"<p><strong>Purpose: </strong>This article introduces a novel deep learning approach to substantially improve the accuracy of colon segmentation even with limited data annotation, which enhances the overall effectiveness of the CT colonography pipeline in clinical settings.</p><p><strong>Methods: </strong>The proposed approach integrates 3D contextual information via guided sequential episodic training in which a query CT slice is segmented by exploiting its previous labeled CT slice (i.e., support). Segmentation starts by detecting the rectum using a Markov Random Field-based algorithm. Then, supervised sequential episodic training is applied to the remaining slices, while contrastive learning is employed to enhance feature discriminability, thereby improving segmentation accuracy.</p><p><strong>Results: </strong>The proposed method, evaluated on 98 abdominal scans of prepped patients, achieved a Dice coefficient of 97.3% and a polyp information preservation accuracy of 98.28%. Statistical analysis, including 95% confidence intervals, underscores the method's robustness and reliability. Clinically, this high level of accuracy is vital for ensuring the preservation of critical polyp details, which are essential for accurate automatic diagnostic evaluation. The proposed method performs reliably in scenarios with limited annotated data. This is demonstrated by achieving a Dice coefficient of 97.15% when the model was trained on a smaller number of annotated CT scans (e.g., 10 scans) than the testing dataset (e.g., 88 scans).</p><p><strong>Conclusions: </strong>The proposed sequential segmentation approach achieves promising results in colon segmentation. A key strength of the method is its ability to generalize effectively, even with limited annotated datasets-a common challenge in medical imaging.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparison of active learning algorithms in classifying head computed tomography reports using bidirectional encoder representations from transformers. 使用变压器双向编码器表示分类头部计算机断层扫描报告的主动学习算法的比较。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-01-08 DOI: 10.1007/s11548-024-03316-7
Tomohiro Wataya, Azusa Miura, Takahisa Sakisuka, Masahiro Fujiwara, Hisashi Tanaka, Yu Hiraoka, Junya Sato, Miyuki Tomiyama, Daiki Nishigaki, Kosuke Kita, Yuki Suzuki, Shoji Kido, Noriyuki Tomiyama

Purpose: Systems equipped with natural language (NLP) processing can reduce missed radiological findings by physicians, but the annotation costs are burden in the development. This study aimed to compare the effects of active learning (AL) algorithms in NLP for estimating the significance of head computed tomography (CT) reports using bidirectional encoder representations from transformers (BERT).

Methods: A total of 3728 head CT reports annotated with five categories of importance were used and UTH-BERT was adopted as the pre-trained BERT model. We assumed that 64% (2385 reports) of the data were initially in the unlabeled data pool (UDP), while the labeled data set (LD) used to train the model was empty. Twenty-five reports were repeatedly selected from the UDP and added to the LD, based on seven metrices: random sampling (RS: control), four uncertainty sampling (US) methods (least confidence (LC), margin sampling (MS), ratio of confidence (RC), and entropy sampling (ES)), and two distance-based sampling (DS) methods (cosine distance (CD) and Euclidian distance (ED)). The transition of accuracy of the model was evaluated using the test dataset.

Results: The accuracy of the models with US was significantly higher than RS when reports in LD were < 1800, whereas DS methods were significantly lower than RS. Among the US methods, MS and RC were even better than the others. With the US methods, the required labeled data decreased by 15.4-40.5%, and most efficient in RC. In addition, in the US methods, data for minor categories tended to be added to LD earlier than RS and DS.

Conclusions: In the classification task for the importance of head CT reports, US methods, especially RC and MS can lead to the effective fine-tuning of BERT models and reduce the imbalance of categories. AL can contribute to other studies on larger datasets by providing effective annotation.

目的:配备自然语言(NLP)处理的系统可以减少医生遗漏的放射发现,但注释成本是开发中的负担。本研究旨在比较NLP中主动学习(AL)算法在估计使用变压器(BERT)双向编码器表示的头部计算机断层扫描(CT)报告的重要性方面的效果。方法:选取3728份头部CT报告,标注5类重要性,采用UTH-BERT作为预训练的BERT模型。我们假设64%(2385份报告)的数据最初在未标记的数据池(UDP)中,而用于训练模型的标记数据集(LD)是空的。根据随机抽样(RS: control)、四种不确定性抽样(US)方法(最小置信度(LC)、边际抽样(MS)、置信比(RC)和熵抽样(ES))和两种基于距离的抽样(DS)方法(余弦距离(CD)和欧氏距离(ED)),从UDP中重复选择25份报告并添加到LD中。利用测试数据集对模型的精度过渡进行了评价。结论:在头部CT报告重要性的分类任务中,US方法,尤其是RC和MS方法,可以对BERT模型进行有效的微调,减少类别的不平衡。通过提供有效的注释,人工智能可以为更大数据集的其他研究做出贡献。
{"title":"Comparison of active learning algorithms in classifying head computed tomography reports using bidirectional encoder representations from transformers.","authors":"Tomohiro Wataya, Azusa Miura, Takahisa Sakisuka, Masahiro Fujiwara, Hisashi Tanaka, Yu Hiraoka, Junya Sato, Miyuki Tomiyama, Daiki Nishigaki, Kosuke Kita, Yuki Suzuki, Shoji Kido, Noriyuki Tomiyama","doi":"10.1007/s11548-024-03316-7","DOIUrl":"https://doi.org/10.1007/s11548-024-03316-7","url":null,"abstract":"<p><strong>Purpose: </strong>Systems equipped with natural language (NLP) processing can reduce missed radiological findings by physicians, but the annotation costs are burden in the development. This study aimed to compare the effects of active learning (AL) algorithms in NLP for estimating the significance of head computed tomography (CT) reports using bidirectional encoder representations from transformers (BERT).</p><p><strong>Methods: </strong>A total of 3728 head CT reports annotated with five categories of importance were used and UTH-BERT was adopted as the pre-trained BERT model. We assumed that 64% (2385 reports) of the data were initially in the unlabeled data pool (UDP), while the labeled data set (LD) used to train the model was empty. Twenty-five reports were repeatedly selected from the UDP and added to the LD, based on seven metrices: random sampling (RS: control), four uncertainty sampling (US) methods (least confidence (LC), margin sampling (MS), ratio of confidence (RC), and entropy sampling (ES)), and two distance-based sampling (DS) methods (cosine distance (CD) and Euclidian distance (ED)). The transition of accuracy of the model was evaluated using the test dataset.</p><p><strong>Results: </strong>The accuracy of the models with US was significantly higher than RS when reports in LD were < 1800, whereas DS methods were significantly lower than RS. Among the US methods, MS and RC were even better than the others. With the US methods, the required labeled data decreased by 15.4-40.5%, and most efficient in RC. In addition, in the US methods, data for minor categories tended to be added to LD earlier than RS and DS.</p><p><strong>Conclusions: </strong>In the classification task for the importance of head CT reports, US methods, especially RC and MS can lead to the effective fine-tuning of BERT models and reduce the imbalance of categories. AL can contribute to other studies on larger datasets by providing effective annotation.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Computer Assisted Radiology and Surgery
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1