首页 > 最新文献

International Journal of Computer Assisted Radiology and Surgery最新文献

英文 中文
BrightVAE: luminosity enhancement in underexposed endoscopic images. BrightVAE:曝光不足的内窥镜图像的亮度增强。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-02-02 DOI: 10.1007/s11548-026-03573-8
Farzaneh Koohestani, Zahra Nabizadeh, Nader Karimi, Shahram Shirani, Shadrokh Samavi

Purpose: Low-light endoscopic images often lack contrast and clarity, obscuring anatomical details and reducing diagnostic accuracy. This study develops a method to enhance image brightness and visibility, enabling clearer visualization of critical structures to support precise medical diagnoses and improve patient outcomes.

Methods: To specifically address nonuniform illumination, we propose BrightVAE, a model that uses a dual-receptive-field architecture to decouple global brightness correction from local texture preservation. Integrated attention-based modules (Attencoder and Attenquant) explicitly target and amplify underexposed regions while preventing over-saturation, thereby recovering human-evaluable details in shadowed areas. The model was trained and tested on a public endoscopic dataset, and its performance was evaluated against other techniques using quality metrics.

Results: The model outperformed alternatives, improving PSNR by 3.252 units, structural detail by 0.045, and perceptual quality by 0.014 compared to the best model before us, achieving a PSNR of 30.576, SSIM of 0.879, and LPIPS of 0.133, ensuring superior visibility of shadowed regions.

Conclusion: This approach advances endoscopic imaging by delivering sharper, reliable images, enhancing diagnostic precision in clinical practice. Improved visualization supports better detection of abnormalities, potentially leading to more effective treatment decisions and enhanced patient care.

目的:低光内镜图像往往缺乏对比度和清晰度,模糊解剖细节,降低诊断准确性。本研究开发了一种增强图像亮度和可见度的方法,使关键结构的可视化更加清晰,从而支持精确的医疗诊断并改善患者的预后。方法:为了解决非均匀照明问题,我们提出了BrightVAE模型,该模型使用双接受场架构将全局亮度校正与局部纹理保存分离开来。集成的基于注意力的模块(Attencoder和Attenquant)明确瞄准和放大曝光不足的区域,同时防止过度饱和,从而在阴影区域恢复人类可评估的细节。该模型在一个公共内窥镜数据集上进行了训练和测试,并使用质量指标对其性能与其他技术进行了评估。结果:与之前的最佳模型相比,该模型的PSNR提高了3.252个单位,结构细节提高了0.045个单位,感知质量提高了0.014个单位,实现了PSNR为30.576,SSIM为0.879,LPIPS为0.133,确保了阴影区域的卓越可见性。结论:该方法通过提供更清晰、可靠的图像,提高了临床诊断的准确性,从而促进了内窥镜成像。改进的可视化支持更好地检测异常,可能导致更有效的治疗决策和增强患者护理。
{"title":"BrightVAE: luminosity enhancement in underexposed endoscopic images.","authors":"Farzaneh Koohestani, Zahra Nabizadeh, Nader Karimi, Shahram Shirani, Shadrokh Samavi","doi":"10.1007/s11548-026-03573-8","DOIUrl":"https://doi.org/10.1007/s11548-026-03573-8","url":null,"abstract":"<p><strong>Purpose: </strong>Low-light endoscopic images often lack contrast and clarity, obscuring anatomical details and reducing diagnostic accuracy. This study develops a method to enhance image brightness and visibility, enabling clearer visualization of critical structures to support precise medical diagnoses and improve patient outcomes.</p><p><strong>Methods: </strong>To specifically address nonuniform illumination, we propose BrightVAE, a model that uses a dual-receptive-field architecture to decouple global brightness correction from local texture preservation. Integrated attention-based modules (Attencoder and Attenquant) explicitly target and amplify underexposed regions while preventing over-saturation, thereby recovering human-evaluable details in shadowed areas. The model was trained and tested on a public endoscopic dataset, and its performance was evaluated against other techniques using quality metrics.</p><p><strong>Results: </strong>The model outperformed alternatives, improving PSNR by 3.252 units, structural detail by 0.045, and perceptual quality by 0.014 compared to the best model before us, achieving a PSNR of 30.576, SSIM of 0.879, and LPIPS of 0.133, ensuring superior visibility of shadowed regions.</p><p><strong>Conclusion: </strong>This approach advances endoscopic imaging by delivering sharper, reliable images, enhancing diagnostic precision in clinical practice. Improved visualization supports better detection of abnormalities, potentially leading to more effective treatment decisions and enhanced patient care.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146108457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Acknowledgement to reviewers. 感谢审稿人。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-31 DOI: 10.1007/s11548-026-03575-6
{"title":"Acknowledgement to reviewers.","authors":"","doi":"10.1007/s11548-026-03575-6","DOIUrl":"https://doi.org/10.1007/s11548-026-03575-6","url":null,"abstract":"","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146094903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transient numerical simulation of hemodynamics in bioprosthetic heart valves: insights into valve sizing and thrombosis risk. 生物人工心脏瓣膜血流动力学的瞬态数值模拟:对瓣膜尺寸和血栓形成风险的见解。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1007/s11548-026-03577-4
M Berger, J Golks, L Gleissner, T Senfter, Ch Mayerl, N Bonaros, C G Tepeköylü, L Stastny, M Grimm, M Pillei

Purpose: This study aims to evaluate the hemodynamic effect of tilting after implantation of a bioprosthetic heart valve in an enlarged aortic annulus using computational fluid dynamics (CFD). The objective is to define the cost of excessive enlargement in transprosthetical flow and identify the optimal prosthesis size and implantation angle to reduce thrombosis and early degeneration.

Methods: Virtual implantation of a 23-mm bioprosthetic model was conducted in a prespecified virtual aortic annulus of 23 mm. CFD simulations were conducted to analyze the hemodynamic parameters, including wall shear stress and turbulent kinetic energy, for bioprosthetic heart valves of the 23 mm (perpendicular implantation), 25 mm, and 27 mm (tilted implantation at different angles) after virtual annular enlargement. The simulations utilized transient flow models and mesh convergence studies, to ensure numerical accuracy and clinical relevance.

Results: The 23-mm valve implanted without annular enlargement and in an aligned fashion to the annulus exhibited higher wall shear stress (WSS) and shear stress (SS) values than the 25-mm valve implanted at 12° tilted position and better hemodynamics than the 27-mm valve at 25° tilted position. The 25-mm valve after annular enlargement and implantation at 12° tilted position achieved the best hemodynamic performance together with a 23% and 7% reduction of WSS and SS as compared to the 23-mm valve.

Conclusions: Slight oversizing and tilting after annular enlargement yields the best performance of an aortic bioprosthetic valve with lowest WSS and balanced turbulence.

目的:本研究旨在利用计算流体动力学(CFD)评估生物人工心脏瓣膜植入扩大主动脉环后倾斜对血流动力学的影响。目的是确定假体血流过度扩大的成本,确定最佳假体尺寸和植入角度,以减少血栓形成和早期退变。方法:在预先设定的23 mm虚拟主动脉环内虚拟植入23 mm生物假体模型。采用CFD模拟方法,分析了23 mm(垂直植入)、25 mm和27 mm(不同角度倾斜植入)生物人工心脏瓣膜在虚拟环形放大后的血流动力学参数,包括壁面剪切应力和湍流动能。模拟利用瞬态流动模型和网格收敛研究,以确保数值准确性和临床相关性。结果:23-mm瓣膜植入时未扩大环空,且与环空对齐,其壁剪切应力(WSS)和剪切应力(SS)值高于25-mm瓣膜在12°倾斜位置植入,血流动力学优于27-mm瓣膜在25°倾斜位置植入。经环形放大并在12°倾斜位置植入的25mm瓣膜获得了最佳的血流动力学性能,与23mm瓣膜相比,WSS和SS分别降低23%和7%。结论:主动脉环扩大后的轻微过大和倾斜可获得最佳性能,WSS最低,湍流平衡。
{"title":"Transient numerical simulation of hemodynamics in bioprosthetic heart valves: insights into valve sizing and thrombosis risk.","authors":"M Berger, J Golks, L Gleissner, T Senfter, Ch Mayerl, N Bonaros, C G Tepeköylü, L Stastny, M Grimm, M Pillei","doi":"10.1007/s11548-026-03577-4","DOIUrl":"https://doi.org/10.1007/s11548-026-03577-4","url":null,"abstract":"<p><strong>Purpose: </strong>This study aims to evaluate the hemodynamic effect of tilting after implantation of a bioprosthetic heart valve in an enlarged aortic annulus using computational fluid dynamics (CFD). The objective is to define the cost of excessive enlargement in transprosthetical flow and identify the optimal prosthesis size and implantation angle to reduce thrombosis and early degeneration.</p><p><strong>Methods: </strong>Virtual implantation of a 23-mm bioprosthetic model was conducted in a prespecified virtual aortic annulus of 23 mm. CFD simulations were conducted to analyze the hemodynamic parameters, including wall shear stress and turbulent kinetic energy, for bioprosthetic heart valves of the 23 mm (perpendicular implantation), 25 mm, and 27 mm (tilted implantation at different angles) after virtual annular enlargement. The simulations utilized transient flow models and mesh convergence studies, to ensure numerical accuracy and clinical relevance.</p><p><strong>Results: </strong>The 23-mm valve implanted without annular enlargement and in an aligned fashion to the annulus exhibited higher wall shear stress (WSS) and shear stress (SS) values than the 25-mm valve implanted at 12° tilted position and better hemodynamics than the 27-mm valve at 25° tilted position. The 25-mm valve after annular enlargement and implantation at 12° tilted position achieved the best hemodynamic performance together with a 23% and 7% reduction of WSS and SS as compared to the 23-mm valve.</p><p><strong>Conclusions: </strong>Slight oversizing and tilting after annular enlargement yields the best performance of an aortic bioprosthetic valve with lowest WSS and balanced turbulence.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146087889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Digs: diffusion-guided Gaussian Splatting for dynamic occlusion surgical scene reconstruction. Digs:用于动态闭塞手术场景重建的扩散引导高斯溅射。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-30 DOI: 10.1007/s11548-026-03571-w
Huoling Luo, Xiangling Nan, Jiahao Yang, Changmiao Wang, Tianqiao Zhang, Yingfang Fan, Fucang Jia, Qin Zhang

Purpose: Accurate 3D reconstruction from endoscopic videos is crucial for advancing computer-assisted minimally invasive surgery. However, existing approaches struggle with dynamic surgical scenes where instrument occlusions cause significant reconstruction artifacts. Although 3D Gaussian Splatting (3DGS) enables rapid reconstruction, it often suffers from incomplete surface recovery due to occlusion-induced missing regions and error propagation from suboptimal initial point clouds during radiance field optimization. This study aims to enhance reconstruction accuracy in dynamically occluded surgical environments.

Methods: We propose a diffusion-guided Gaussian Splatting (DiGS) framework comprising two key components: (1) a diffusion-guided surface completion network that incorporates surgical scene priors to restore high-fidelity textures in occluded regions, improving surface completeness; and (2) a lightweight annealed smoothing mechanism designed to mitigate endoscope motion estimation errors, ensuring temporal coherence during continuous frame interpolation and stabilizing radiance field optimization.

Results: Extensive experiments on the EndoNeRF and StereoMIS datasets demonstrate the superiority of DiGS over state-of-the-art baselines. On EndoNeRF, DiGS achieves a 61.75% improvement in LPIPS, indicating stronger perceptual alignment in dynamically occluded scenes. On StereoMIS, DiGS delivers an 7.03% PSNR gain and a 40.79% LPIPS improvement, along with consistently higher SSIM scores confirming superior preservation of structural details.

Conclusion: The proposed DiGS framework effectively addresses the challenges of dynamic occlusions and motion-induced errors in surgical scene reconstruction, producing more accurate and temporally coherent 3D models. The code is publicly available at https://github.com/IGSResearch/DiGS .

目的:从内窥镜视频中获得准确的三维重建对于推进计算机辅助微创手术至关重要。然而,现有的方法与动态手术场景斗争,其中器械闭塞导致显著的重建伪影。虽然3D高斯溅射(3DGS)能够快速重建,但由于遮挡引起的缺失区域和辐射场优化过程中次优初始点云的错误传播,它经常遭受表面恢复不完全的问题。本研究旨在提高动态闭塞手术环境下的重建精度。方法:我们提出了一个弥散引导高斯飞溅(DiGS)框架,该框架包括两个关键组件:(1)弥散引导表面补全网络,该网络包含手术场景先验,以恢复遮挡区域的高保真纹理,提高表面完整性;(2)设计了一种轻型退火平滑机制,旨在减轻内窥镜运动估计误差,确保连续帧插值期间的时间相干性和稳定辐射场优化。结果:在EndoNeRF和StereoMIS数据集上进行的大量实验表明,DiGS优于最先进的基线。在EndoNeRF上,DiGS在LPIPS上实现了61.75%的改进,表明在动态遮挡的场景中有更强的感知对齐。在StereoMIS上,DiGS提供了7.03%的PSNR增益和40.79%的LPIPS改进,以及持续较高的SSIM分数,证实了优越的结构细节保存。结论:提出的DiGS框架有效地解决了手术场景重建中动态咬合和运动引起的错误的挑战,产生了更准确和时间连贯的3D模型。该代码可在https://github.com/IGSResearch/DiGS上公开获得。
{"title":"Digs: diffusion-guided Gaussian Splatting for dynamic occlusion surgical scene reconstruction.","authors":"Huoling Luo, Xiangling Nan, Jiahao Yang, Changmiao Wang, Tianqiao Zhang, Yingfang Fan, Fucang Jia, Qin Zhang","doi":"10.1007/s11548-026-03571-w","DOIUrl":"https://doi.org/10.1007/s11548-026-03571-w","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate 3D reconstruction from endoscopic videos is crucial for advancing computer-assisted minimally invasive surgery. However, existing approaches struggle with dynamic surgical scenes where instrument occlusions cause significant reconstruction artifacts. Although 3D Gaussian Splatting (3DGS) enables rapid reconstruction, it often suffers from incomplete surface recovery due to occlusion-induced missing regions and error propagation from suboptimal initial point clouds during radiance field optimization. This study aims to enhance reconstruction accuracy in dynamically occluded surgical environments.</p><p><strong>Methods: </strong>We propose a diffusion-guided Gaussian Splatting (DiGS) framework comprising two key components: (1) a diffusion-guided surface completion network that incorporates surgical scene priors to restore high-fidelity textures in occluded regions, improving surface completeness; and (2) a lightweight annealed smoothing mechanism designed to mitigate endoscope motion estimation errors, ensuring temporal coherence during continuous frame interpolation and stabilizing radiance field optimization.</p><p><strong>Results: </strong>Extensive experiments on the EndoNeRF and StereoMIS datasets demonstrate the superiority of DiGS over state-of-the-art baselines. On EndoNeRF, DiGS achieves a 61.75% improvement in LPIPS, indicating stronger perceptual alignment in dynamically occluded scenes. On StereoMIS, DiGS delivers an 7.03% PSNR gain and a 40.79% LPIPS improvement, along with consistently higher SSIM scores confirming superior preservation of structural details.</p><p><strong>Conclusion: </strong>The proposed DiGS framework effectively addresses the challenges of dynamic occlusions and motion-induced errors in surgical scene reconstruction, producing more accurate and temporally coherent 3D models. The code is publicly available at https://github.com/IGSResearch/DiGS .</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146087903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Left-right relationship-aware 3D volume classification method. 感知左右关系的三维体分类方法。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-28 DOI: 10.1007/s11548-025-03567-y
Masahiro Oda, Yuichiro Hayashi, Yoshito Otake, Masahiro Hashimoto, Toshiaki Akashi, Shigeki Aoki, Kensaku Mori

Purpose: This paper proposes a left-right (LR) relationship-aware classification model for 3D volumetric images (3D volume). Bilateral symmetry (LR relationship) is an essential property of the human body that can be used to detect abnormalities and understand anatomical structures. Checking the difference and similarity between the left and right anatomical structures is very important in diagnosis. We propose an LR relationship-aware classification model of 3D volume.

Methods: The proposed model employs an image feature extraction process from LR symmetric positions of human anatomy from 3D volume. Due to variations in body position and individual anatomical structure, small positional gaps among LR corresponding anatomical structures can be observed in medical images. We developed a multi-shift symmetric feature extraction module to accommodate such positional gaps.

Results: The model was applied to 3D volume classification tasks of the lung and brain. From experimental results, the proposed model achieved superior performances both in the lung and brain classification tasks compared to the previous models. The result indicates that the proposed model has generalized performance in classifying anatomical structures with bilateral symmetric or semi-symmetric structures.

Conclusion: We proposed the LR relationship-aware classification model of 3D volume. The proposed model effectively extracts image features from LR symmetric positions. The multi-shift symmetric feature extraction module was employed to accommodate small positional gaps among LR corresponding positions. The experimental results of 3D volume classification tasks of the lung and brain showed that the proposed method achieved superior performances compared to the previous models. Our code is available at https://github.com/modafone/lr3dvolumeclassification .

目的:本文提出了一种3D体积图像(3D volume)的左右(LR)关系感知分类模型。双侧对称(LR关系)是人体的基本属性,可以用来检测异常和了解解剖结构。检查左右解剖结构的异同在诊断中是非常重要的。提出了一种LR关系感知的三维体分类模型。方法:该模型采用从人体解剖体的LR对称位置提取图像特征的方法。由于体位和个体解剖结构的差异,在医学图像中可以观察到LR对应解剖结构之间的小位置间隙。我们开发了一个多移位对称特征提取模块来适应这种位置差距。结果:该模型可用于肺和脑的三维体积分类任务。实验结果表明,该模型在肺和脑的分类任务上都取得了较好的效果。结果表明,该模型对双侧对称或半对称解剖结构具有较好的分类效果。结论:提出了三维体LR关系感知分类模型。该模型能有效地从LR对称位置提取图像特征。采用多移位对称特征提取模块,以适应LR对应位置之间较小的位置间隙。对肺和脑三维体积分类任务的实验结果表明,该方法取得了较好的效果。我们的代码可在https://github.com/modafone/lr3dvolumeclassification上获得。
{"title":"Left-right relationship-aware 3D volume classification method.","authors":"Masahiro Oda, Yuichiro Hayashi, Yoshito Otake, Masahiro Hashimoto, Toshiaki Akashi, Shigeki Aoki, Kensaku Mori","doi":"10.1007/s11548-025-03567-y","DOIUrl":"https://doi.org/10.1007/s11548-025-03567-y","url":null,"abstract":"<p><strong>Purpose: </strong>This paper proposes a left-right (LR) relationship-aware classification model for 3D volumetric images (3D volume). Bilateral symmetry (LR relationship) is an essential property of the human body that can be used to detect abnormalities and understand anatomical structures. Checking the difference and similarity between the left and right anatomical structures is very important in diagnosis. We propose an LR relationship-aware classification model of 3D volume.</p><p><strong>Methods: </strong>The proposed model employs an image feature extraction process from LR symmetric positions of human anatomy from 3D volume. Due to variations in body position and individual anatomical structure, small positional gaps among LR corresponding anatomical structures can be observed in medical images. We developed a multi-shift symmetric feature extraction module to accommodate such positional gaps.</p><p><strong>Results: </strong>The model was applied to 3D volume classification tasks of the lung and brain. From experimental results, the proposed model achieved superior performances both in the lung and brain classification tasks compared to the previous models. The result indicates that the proposed model has generalized performance in classifying anatomical structures with bilateral symmetric or semi-symmetric structures.</p><p><strong>Conclusion: </strong>We proposed the LR relationship-aware classification model of 3D volume. The proposed model effectively extracts image features from LR symmetric positions. The multi-shift symmetric feature extraction module was employed to accommodate small positional gaps among LR corresponding positions. The experimental results of 3D volume classification tasks of the lung and brain showed that the proposed method achieved superior performances compared to the previous models. Our code is available at https://github.com/modafone/lr3dvolumeclassification .</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146068660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Algorithm for Automatic Osteotomy Plate Placement Planning in 3D: Application in Distal Radius Malunion. 一种三维自动截骨钢板放置规划算法:在桡骨远端畸形愈合中的应用。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-27 DOI: 10.1007/s11548-026-03576-5
Eva van de Nes, Camiel J Smees, Judith Olde Heuvel, Anne J H Vochteloo, Gabriëlle J M Tuijthof

Purpose: Positioning of an osteosynthesis plate is a key step in the preoperative 3D planning processes for the design of patient-specific guides. This step requires considerable time and expertise. To increase 3D planning efficiency, this study aims to develop an automated plate positioning algorithm.

Methods: A robust algorithm was developed to optimize osteosynthesis plate positioning on the distal radius, using STL properties and anatomical landmarks. The algorithm involved alignment, landmark detection, initial placement, and final optimization. Retrospective data of 34 planned radii and corresponding plate positions, including decimated and refined mesh versions, were used to compare algorithm output to manual placement based on runtime, Hausdorff distance, translation, and rotation (mean ± SD, 95% CI), thereby assessing robustness across different mesh resolutions.

Results: The average run time for the algorithm was 18.3 ± 16.8 s (95% CI 12.4-24.1 s) compared to a manual placement time of 12.45 ± 4.56 min (single expert, n = 10, 95% CI 9.22-16.28 min). The mean unpaired maximum Hausdorff distance between manual and algorithm placements was 5.5 ± 2.5 mm (95% CI 4.6-6.4 mm). The mean rotation and translation differences were 4.9 ± 3.2° (95% CI 3.8-6.0°) and 3.3 ± 1.7 mm (95% CI 2.8-3.9 mm), respectively.

Conclusion: In conclusion, while some manual adjustment remains necessary, the algorithm aids in reducing planning time and offers a modular, generalizable framework adaptable to other osteotomy-plate procedures, supporting clinical 3D planning.

目的:固定钢板的定位是术前3D规划过程中设计患者特异性导尿管的关键步骤。这一步需要大量的时间和专业知识。为了提高三维规划效率,本研究旨在开发一种自动板定位算法。方法:利用STL特性和解剖标志,开发了一种鲁棒算法来优化桡骨远端骨接骨板定位。该算法包括对齐、地标检测、初始放置和最终优化。34个计划半径和相应板位置的回顾性数据,包括抽取和精化网格版本,用于比较基于运行时间、Hausdorff距离、平移和旋转(mean±SD, 95% CI)的算法输出与手动放置的结果,从而评估不同网格分辨率下的鲁棒性。结果:该算法的平均运行时间为18.3±16.8 s (95% CI 12.4-24.1 s),而人工放置时间为12.45±4.56 min(单个专家,n = 10, 95% CI 9.22-16.28 min)。人工和算法放置之间的平均未配对最大Hausdorff距离为5.5±2.5 mm (95% CI 4.6-6.4 mm)。平均旋转和平移差异分别为4.9±3.2°(95% CI 3.8-6.0°)和3.3±1.7 mm (95% CI 2.8-3.9 mm)。结论:虽然仍然需要一些手动调整,但该算法有助于减少计划时间,并提供一个模块化的、可推广的框架,适用于其他截骨钢板手术,支持临床3D计划。
{"title":"An Algorithm for Automatic Osteotomy Plate Placement Planning in 3D: Application in Distal Radius Malunion.","authors":"Eva van de Nes, Camiel J Smees, Judith Olde Heuvel, Anne J H Vochteloo, Gabriëlle J M Tuijthof","doi":"10.1007/s11548-026-03576-5","DOIUrl":"https://doi.org/10.1007/s11548-026-03576-5","url":null,"abstract":"<p><strong>Purpose: </strong>Positioning of an osteosynthesis plate is a key step in the preoperative 3D planning processes for the design of patient-specific guides. This step requires considerable time and expertise. To increase 3D planning efficiency, this study aims to develop an automated plate positioning algorithm.</p><p><strong>Methods: </strong>A robust algorithm was developed to optimize osteosynthesis plate positioning on the distal radius, using STL properties and anatomical landmarks. The algorithm involved alignment, landmark detection, initial placement, and final optimization. Retrospective data of 34 planned radii and corresponding plate positions, including decimated and refined mesh versions, were used to compare algorithm output to manual placement based on runtime, Hausdorff distance, translation, and rotation (mean ± SD, 95% CI), thereby assessing robustness across different mesh resolutions.</p><p><strong>Results: </strong>The average run time for the algorithm was 18.3 ± 16.8 s (95% CI 12.4-24.1 s) compared to a manual placement time of 12.45 ± 4.56 min (single expert, n = 10, 95% CI 9.22-16.28 min). The mean unpaired maximum Hausdorff distance between manual and algorithm placements was 5.5 ± 2.5 mm (95% CI 4.6-6.4 mm). The mean rotation and translation differences were 4.9 ± 3.2° (95% CI 3.8-6.0°) and 3.3 ± 1.7 mm (95% CI 2.8-3.9 mm), respectively.</p><p><strong>Conclusion: </strong>In conclusion, while some manual adjustment remains necessary, the algorithm aids in reducing planning time and offers a modular, generalizable framework adaptable to other osteotomy-plate procedures, supporting clinical 3D planning.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward precision surgical education: quantitative evaluation of surgical performance using instantaneous screw axes. 走向精准外科教育:利用瞬时螺旋轴定量评价手术效果。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-26 DOI: 10.1007/s11548-025-03565-0
Heath Boyea, James Korndorffer, Ann Majewicz Fey

Purpose: In surgical skill development, a trainee's goal is to move through the learning curve and achieve expert performance. Goal directed training, with expert performance as the goal, can facilitate skill development; however, there are currently few methods available to encode expert-like simulation performance into learning strategies that can be practiced independently.

Methods: We propose a novel method of surgical simulation skill analysis through segmenting and evaluating kinematic data with instantaneous screw axes (ISA) theory and K-means clustering. In ISA, single degree of freedom (DOF) tasks can be represented as displacements about a single screw axis; however, we propose extending this method to more complex tasks defining them with clusters of similar ISAs in the surgical environment, decomposing them into a sequence of 1DOF movements. In this paper, we present an ISA algorithm and apply it to surgeon manipulator poses across fourteen suturing and knot-tying gestures obtained from the JIGSAWS surgical dataset. We also apply this method to entire simulated suturing demonstrations across a 6-month training period from the BGU-SKILLS dataset. We implemented K-means clustering to segment these movements into sub-gestures. We hypothesize that individuals with greater levels of expertise should exhibit more concise actions with minimal extraneous movement; therefore, fewer clusters should be required to decompose their simulation performance.

Results: Our ISA algorithm was applied to 1136 gestures from ten surgeons across three skill levels and 324 unsegmented demonstrations collected from 18 surgical residents over a training period of 6 months. We performed a Kruskal-Wallis analysis with a Dunn-Sidak post-hoc test on the number of ISA clusters required to decompose each gesture. We found that highly task-constrained gestures required significantly fewer numbers of clusters for expert and/or intermediate groups when compared to novices on suturing tasks only.

Conclusion: Our results suggest that this method can be used to identify task-constrained gestures within independently performed suturing surgical simulations and classify them into higher skill and lower skill sets. This analysis can also provide geometric feedback on performed gestures vs expert gestures, providing personalized automated performance analysis for surgical trainees leading to personalized educational training.

目的:在外科技能发展中,受训人员的目标是通过学习曲线,达到专家水平。目标导向训练,以专家表现为目标,促进技能发展;然而,目前很少有方法可以将类似专家的模拟性能编码为可以独立练习的学习策略。方法:利用瞬时螺旋轴(ISA)理论和K-means聚类方法对运动数据进行分割和评估,提出了一种新的手术模拟技能分析方法。在ISA中,单自由度(DOF)任务可以表示为围绕单个螺杆轴的位移;然而,我们建议将这种方法扩展到更复杂的任务中,将它们定义为手术环境中类似isa的集群,并将其分解为一系列1DOF运动。在本文中,我们提出了一种ISA算法,并将其应用于从JIGSAWS手术数据集中获得的14种缝合和打结手势的外科医生操纵器姿势。我们还将这种方法应用于BGU-SKILLS数据集中为期6个月的整个模拟缝合演示。我们实现了K-means聚类,将这些动作分割成子手势。我们假设,具有更高水平专业知识的个体应该表现出更简洁的行动和最小的外部运动;因此,需要更少的集群来分解它们的模拟性能。结果:在为期6个月的培训期间,我们的ISA算法应用于来自10名外科医生的1136个手势,涉及三个技能水平,以及来自18名外科住院医师的324个未分割的演示。我们对分解每个手势所需的ISA集群数量进行了Kruskal-Wallis分析和Dunn-Sidak事后测试。我们发现,与仅在缝合任务上的新手相比,高度任务约束手势对专家和/或中级组所需的集群数量明显更少。结论:我们的研究结果表明,该方法可用于识别独立进行缝合手术模拟中的任务受限手势,并将其分为高技能和低技能组。该分析还可以提供表演手势与专家手势的几何反馈,为外科培训生提供个性化的自动化性能分析,从而实现个性化的教育培训。
{"title":"Toward precision surgical education: quantitative evaluation of surgical performance using instantaneous screw axes.","authors":"Heath Boyea, James Korndorffer, Ann Majewicz Fey","doi":"10.1007/s11548-025-03565-0","DOIUrl":"https://doi.org/10.1007/s11548-025-03565-0","url":null,"abstract":"<p><strong>Purpose: </strong>In surgical skill development, a trainee's goal is to move through the learning curve and achieve expert performance. Goal directed training, with expert performance as the goal, can facilitate skill development; however, there are currently few methods available to encode expert-like simulation performance into learning strategies that can be practiced independently.</p><p><strong>Methods: </strong>We propose a novel method of surgical simulation skill analysis through segmenting and evaluating kinematic data with instantaneous screw axes (ISA) theory and K-means clustering. In ISA, single degree of freedom (DOF) tasks can be represented as displacements about a single screw axis; however, we propose extending this method to more complex tasks defining them with clusters of similar ISAs in the surgical environment, decomposing them into a sequence of 1DOF movements. In this paper, we present an ISA algorithm and apply it to surgeon manipulator poses across fourteen suturing and knot-tying gestures obtained from the JIGSAWS surgical dataset. We also apply this method to entire simulated suturing demonstrations across a 6-month training period from the BGU-SKILLS dataset. We implemented K-means clustering to segment these movements into sub-gestures. We hypothesize that individuals with greater levels of expertise should exhibit more concise actions with minimal extraneous movement; therefore, fewer clusters should be required to decompose their simulation performance.</p><p><strong>Results: </strong>Our ISA algorithm was applied to 1136 gestures from ten surgeons across three skill levels and 324 unsegmented demonstrations collected from 18 surgical residents over a training period of 6 months. We performed a Kruskal-Wallis analysis with a Dunn-Sidak post-hoc test on the number of ISA clusters required to decompose each gesture. We found that highly task-constrained gestures required significantly fewer numbers of clusters for expert and/or intermediate groups when compared to novices on suturing tasks only.</p><p><strong>Conclusion: </strong>Our results suggest that this method can be used to identify task-constrained gestures within independently performed suturing surgical simulations and classify them into higher skill and lower skill sets. This analysis can also provide geometric feedback on performed gestures vs expert gestures, providing personalized automated performance analysis for surgical trainees leading to personalized educational training.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
St-Swin TransNet: a spatiotemporal swin transformer-based network for self-supervised depth estimation in stereoscopic surgical videos. St-Swin TransNet:一个基于时空swin变压器的网络,用于立体手术视频的自监督深度估计。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-24 DOI: 10.1007/s11548-026-03569-4
Derong Yu, Wenyuan Sun, Junchen Wang, Guoyan Zheng

Purpose: Depth estimation from stereoscopic laparoscopic videos is of vital importance in computer-assisted intervention due to its potential for downstream tasks in laparoscopic surgical navigation. Previous works mostly focus on depth estimation from static frames, while temporal information in stereoscopic laparoscopic videos is largely ignored.

Methods: A spatiotemporal swin (ST-Swin) transformer-based network, referred to as ST-Swin TransNet, is proposed for depth estimation in stereoscopic surgical videos. Built upon a symmetric encoder-decoder architecture consisting of 12 ST-Swin blocks, ST-Swin TransNet extracts spatiotemporal features for efficient and accurate depth estimation, where the ST-Swin blocks are designed to capture spatiotemporal information from stereo video sequences via self-attention mechanism. Given binocular laparoscopic videos, ST-Swin TransNet exploits hierarchical spatiotemporal features to predict disparity maps.

Results: Comprehensive experiments are conducted on two typical yet challenging public datasets to evaluate the performance of the proposed method. We additionally demonstrate the feasibility of applying ST-Swin TransNet to video see-through augmented reality (VST-AR) navigation in laparoscopic surgery. Our method achieved a mean absolute depth error (mADE) of 3.33 mm in depth estimation and a mean absolute distance (mAD) of 1.07 mm in VST-AR navigation.

Conclusion: A spatiotemporal swin transformer-based network for self-supervised depth estimation in binocular laparoscopic surgical videos was developed. Results from the comprehensive experiments demonstrate the superior performance of the proposed method over the state-of-the-art methods.

目的:由于立体腹腔镜视频的深度估计在腹腔镜手术导航的下游任务中具有潜在的潜力,因此在计算机辅助干预中至关重要。以往的工作大多集中在静态帧的深度估计上,而立体腹腔镜视频中的时间信息在很大程度上被忽略了。方法:提出了一种基于时空swin (ST-Swin)变压器的网络,称为ST-Swin TransNet,用于立体手术视频的深度估计。ST-Swin TransNet基于由12个ST-Swin块组成的对称编码器-解码器架构,提取时空特征以实现高效准确的深度估计,其中ST-Swin块旨在通过自注意机制从立体视频序列中捕获时空信息。给定双目腹腔镜视频,ST-Swin TransNet利用分层时空特征来预测视差图。结果:在两个典型但具有挑战性的公共数据集上进行了综合实验,以评估所提出方法的性能。我们还演示了将ST-Swin TransNet应用于腹腔镜手术中视频透明增强现实(VST-AR)导航的可行性。该方法在VST-AR导航中实现了深度估计的平均绝对深度误差(mADE)为3.33 mm,平均绝对距离(mAD)为1.07 mm。结论:建立了一种用于双目腹腔镜手术视频自监督深度估计的基于时空漩涡变压器的网络。综合实验结果表明,该方法优于现有方法。
{"title":"St-Swin TransNet: a spatiotemporal swin transformer-based network for self-supervised depth estimation in stereoscopic surgical videos.","authors":"Derong Yu, Wenyuan Sun, Junchen Wang, Guoyan Zheng","doi":"10.1007/s11548-026-03569-4","DOIUrl":"https://doi.org/10.1007/s11548-026-03569-4","url":null,"abstract":"<p><strong>Purpose: </strong>Depth estimation from stereoscopic laparoscopic videos is of vital importance in computer-assisted intervention due to its potential for downstream tasks in laparoscopic surgical navigation. Previous works mostly focus on depth estimation from static frames, while temporal information in stereoscopic laparoscopic videos is largely ignored.</p><p><strong>Methods: </strong>A spatiotemporal swin (ST-Swin) transformer-based network, referred to as ST-Swin TransNet, is proposed for depth estimation in stereoscopic surgical videos. Built upon a symmetric encoder-decoder architecture consisting of 12 ST-Swin blocks, ST-Swin TransNet extracts spatiotemporal features for efficient and accurate depth estimation, where the ST-Swin blocks are designed to capture spatiotemporal information from stereo video sequences via self-attention mechanism. Given binocular laparoscopic videos, ST-Swin TransNet exploits hierarchical spatiotemporal features to predict disparity maps.</p><p><strong>Results: </strong>Comprehensive experiments are conducted on two typical yet challenging public datasets to evaluate the performance of the proposed method. We additionally demonstrate the feasibility of applying ST-Swin TransNet to video see-through augmented reality (VST-AR) navigation in laparoscopic surgery. Our method achieved a mean absolute depth error (mADE) of 3.33 mm in depth estimation and a mean absolute distance (mAD) of 1.07 mm in VST-AR navigation.</p><p><strong>Conclusion: </strong>A spatiotemporal swin transformer-based network for self-supervised depth estimation in binocular laparoscopic surgical videos was developed. Results from the comprehensive experiments demonstrate the superior performance of the proposed method over the state-of-the-art methods.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146042132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing open-surgery gesture recognition using 3D pose estimation. 利用三维姿态估计增强开放手术手势识别。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-14 DOI: 10.1007/s11548-025-03564-1
Ori Meiraz, Shlomi Laufer, Robert Spector, Itay Or, Gil Bolotin, Tom Friedman

Purpose Surgical gestures are fundamental components of surgical procedures, encompassing actions such as cutting, suturing, and knot-tying. Gesture recognition plays a pivotal role in the automatic analysis of surgical data. Although recent advancements have improved surgical gesture recognition, much of the existing research relies on simulations or minimally invasive surgery data, failing to capture the complexities of open surgery. In this study, we introduce and employ a new open surgery dataset focused on closing incisions after saphenous vein harvesting. Methods Our goal is to improve gesture recognition accuracy by incorporating tool pose estimation and 3D hand pose predictions of surgeons. We employ MS-TCN++  and LTContext  for gesture recognition, and further enhance performance through an ensemble of models using different modalities-video, tool pose, and hand pose data.Results The results reveal that using an ensemble model combining all three modalities provides a substantial improvement over video-only approaches, leading to statistically significant gains across multiple evaluation metrics. We further demonstrate that the model can rely solely on hand and tool poses, completely discarding the video input, while still achieving comparable performance. Additionally, we show that an ensemble model using only hand and tool poses produces results that are either: statistically significantly better than using video alone, or not statistically significantly different.Conclusion This study highlights the effectiveness of integrating multimodal data for surgical gesture recognition. By combining video, hand pose, and tool pose information, our approach achieves higher accuracy and robustness compared to video-only methods. Moreover, the comparable performance of pose-only models suggests a promising, privacy-preserving alternative for surgical data analysis.

手术手势是外科手术的基本组成部分,包括切割、缝合和打结等动作。手势识别在手术数据的自动分析中起着举足轻重的作用。尽管最近的进展改善了手术手势识别,但现有的许多研究依赖于模拟或微创手术数据,未能捕捉开放手术的复杂性。在这项研究中,我们介绍并采用了一个新的开放手术数据集,专注于隐静脉采集后闭合切口。方法通过结合工具姿态估计和三维手部姿态预测,提高外科医生的手势识别精度。我们使用MS-TCN++和LTContext进行手势识别,并通过使用不同模式(视频、工具姿势和手姿势数据)的模型集成进一步提高性能。结果表明,与仅使用视频的方法相比,使用结合所有三种方式的集成模型提供了实质性的改进,从而在多个评估指标上获得了统计上显著的收益。我们进一步证明,该模型可以完全依赖于手和工具的姿势,完全放弃视频输入,同时仍然达到相当的性能。此外,我们表明,仅使用手和工具姿势的集成模型产生的结果要么在统计上明显优于单独使用视频,要么在统计上没有显著差异。结论本研究强调了多模态数据集成在手术手势识别中的有效性。通过结合视频、手姿态和工具姿态信息,我们的方法比仅视频的方法具有更高的准确性和鲁棒性。此外,仅姿态模型的可比性能为外科数据分析提供了一个有前途的、保护隐私的替代方案。
{"title":"Enhancing open-surgery gesture recognition using 3D pose estimation.","authors":"Ori Meiraz, Shlomi Laufer, Robert Spector, Itay Or, Gil Bolotin, Tom Friedman","doi":"10.1007/s11548-025-03564-1","DOIUrl":"https://doi.org/10.1007/s11548-025-03564-1","url":null,"abstract":"<p><p>Purpose Surgical gestures are fundamental components of surgical procedures, encompassing actions such as cutting, suturing, and knot-tying. Gesture recognition plays a pivotal role in the automatic analysis of surgical data. Although recent advancements have improved surgical gesture recognition, much of the existing research relies on simulations or minimally invasive surgery data, failing to capture the complexities of open surgery. In this study, we introduce and employ a new open surgery dataset focused on closing incisions after saphenous vein harvesting. Methods Our goal is to improve gesture recognition accuracy by incorporating tool pose estimation and 3D hand pose predictions of surgeons. We employ MS-TCN++  and LTContext  for gesture recognition, and further enhance performance through an ensemble of models using different modalities-video, tool pose, and hand pose data.Results The results reveal that using an ensemble model combining all three modalities provides a substantial improvement over video-only approaches, leading to statistically significant gains across multiple evaluation metrics. We further demonstrate that the model can rely solely on hand and tool poses, completely discarding the video input, while still achieving comparable performance. Additionally, we show that an ensemble model using only hand and tool poses produces results that are either: statistically significantly better than using video alone, or not statistically significantly different.Conclusion This study highlights the effectiveness of integrating multimodal data for surgical gesture recognition. By combining video, hand pose, and tool pose information, our approach achieves higher accuracy and robustness compared to video-only methods. Moreover, the comparable performance of pose-only models suggests a promising, privacy-preserving alternative for surgical data analysis.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145967781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Environmental and economic costs behind LLMs. 法学硕士背后的环境和经济成本。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1007/s11548-026-03568-5
Pilar López-Úbeda, Teodoro Martín-Noguerol, Antonio Luna

Purpose: To discuss the economic and environmental implications of implementing large language models (LLMs) in radiology, highlighting both their transformative potential and the challenges they pose for equitable and sustainable adoption.

Methods: Current trends in AI investment, infrastructure requirements, operational costs, and environmental impact associated with LLMs are analyzed, highlighting the specific challenges of integrating LLMs into radiological workflows, including data privacy, regulatory compliance, and cost barriers for healthcare institutions. The analysis also considers the costs of model validation, maintenance, and updates, as well as investments in system integration, staff training, and cybersecurity for clinical implementation.

Results: LLMs have revolutionized natural language processing and offer promising applications in radiology, such as improved diagnostic support and workflow optimization. However, their deployment involves substantial financial and environmental costs. Training and operating these models require high-performance computing infrastructure, significant energy consumption, and large volumes of annotated data. Water usage and CO₂ emissions from data centers further raise sustainability concerns, while ongoing operational costs add to the financial burden. Subscription fees and per-query pricing may restrict access for smaller institutions, widening existing inequalities.

Conclusion: While LLMs offer significant benefits for radiology, their high economic and environmental costs present challenges to widespread and equitable adoption. Responsible use, sustainable practices, and policy frameworks are essential to ensure that AI-driven innovations do not exacerbate existing disparities in healthcare access and quality.

目的:讨论在放射学中实施大型语言模型(llm)的经济和环境影响,强调它们的变革潜力和它们为公平和可持续采用所带来的挑战。方法:分析了与llm相关的人工智能投资、基础设施需求、运营成本和环境影响的当前趋势,强调了将llm集成到放射工作流程中的具体挑战,包括数据隐私、法规遵从性和医疗机构的成本障碍。该分析还考虑了模型验证、维护和更新的成本,以及在系统集成、员工培训和临床实施的网络安全方面的投资。结果:法学硕士彻底改变了自然语言处理,并在放射学中提供了有前途的应用,例如改进的诊断支持和工作流程优化。然而,它们的部署涉及大量的财政和环境成本。训练和操作这些模型需要高性能的计算基础设施、大量的能源消耗和大量带注释的数据。数据中心的用水和二氧化碳排放进一步引发了可持续性问题,而持续的运营成本增加了财务负担。订阅费用和按查询收费可能会限制小型机构的访问权限,从而扩大现有的不平等。结论:虽然llm为放射学提供了显著的好处,但其高昂的经济和环境成本对广泛和公平的采用提出了挑战。负责任的使用、可持续的做法和政策框架对于确保人工智能驱动的创新不会加剧医疗保健可及性和质量方面的现有差距至关重要。
{"title":"Environmental and economic costs behind LLMs.","authors":"Pilar López-Úbeda, Teodoro Martín-Noguerol, Antonio Luna","doi":"10.1007/s11548-026-03568-5","DOIUrl":"https://doi.org/10.1007/s11548-026-03568-5","url":null,"abstract":"<p><strong>Purpose: </strong>To discuss the economic and environmental implications of implementing large language models (LLMs) in radiology, highlighting both their transformative potential and the challenges they pose for equitable and sustainable adoption.</p><p><strong>Methods: </strong>Current trends in AI investment, infrastructure requirements, operational costs, and environmental impact associated with LLMs are analyzed, highlighting the specific challenges of integrating LLMs into radiological workflows, including data privacy, regulatory compliance, and cost barriers for healthcare institutions. The analysis also considers the costs of model validation, maintenance, and updates, as well as investments in system integration, staff training, and cybersecurity for clinical implementation.</p><p><strong>Results: </strong>LLMs have revolutionized natural language processing and offer promising applications in radiology, such as improved diagnostic support and workflow optimization. However, their deployment involves substantial financial and environmental costs. Training and operating these models require high-performance computing infrastructure, significant energy consumption, and large volumes of annotated data. Water usage and CO₂ emissions from data centers further raise sustainability concerns, while ongoing operational costs add to the financial burden. Subscription fees and per-query pricing may restrict access for smaller institutions, widening existing inequalities.</p><p><strong>Conclusion: </strong>While LLMs offer significant benefits for radiology, their high economic and environmental costs present challenges to widespread and equitable adoption. Responsible use, sustainable practices, and policy frameworks are essential to ensure that AI-driven innovations do not exacerbate existing disparities in healthcare access and quality.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145960786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Computer Assisted Radiology and Surgery
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1