首页 > 最新文献

Machine Learning Science and Technology最新文献

英文 中文
Generative diffusion model surrogates for mechanistic agent-based biological models. 生成扩散模型替代了基于机械主体的生物模型。
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-30 Epub Date: 2025-10-28 DOI: 10.1088/2632-2153/ae11f8
Tien Comlekoglu, J Quetzalcoatl Toledo-Marín, Douglas W DeSimone, Shayn M Peirce, Geoffrey Fox, James A Glazier

Mechanistic, multicellular, agent-based models are commonly used to investigate tissue, organ, and organism-scale biology at single-cell resolution. The Cellular-Potts Model (CPM) is a powerful and popular framework for developing and interrogating these models. CPMs become computationally expensive at large space- and time- scales making application and investigation of developed models difficult. Surrogate models may allow for the accelerated evaluation of CPMs of complex biological systems. However, the stochastic nature of these models means each set of parameters may give rise to different model configurations, complicating surrogate model development. In this work, we leverage denoising diffusion probabilistic models (DDPMs) to train a generative AI surrogate of a CPM used to investigate in vitro vasculogenesis. We describe the use of an image classifier to learn the characteristics that define unique areas of a 2-dimensional parameter space. We then apply this classifier to aid in surrogate model selection and verification. Our CPM model surrogate generates model configurations 20,000 timesteps ahead of a reference configuration and demonstrates approximately a 22x reduction in computational time as compared to native code execution. Our work represents a step towards the implementation of DDPMs to develop digital twins of stochastic biological systems.

机械的、多细胞的、基于主体的模型通常用于在单细胞分辨率下研究组织、器官和有机体尺度的生物学。蜂窝-波茨模型(Cellular-Potts Model, CPM)是一个强大而流行的框架,用于开发和询问这些模型。cpm在大空间和时间尺度上的计算成本很高,这使得已开发模型的应用和研究变得困难。替代模型可以加速复杂生物系统cpm的评估。然而,这些模型的随机性意味着每组参数可能会产生不同的模型配置,从而使代理模型的开发复杂化。在这项工作中,我们利用去噪扩散概率模型(ddpm)来训练用于研究体外血管发生的CPM的生成人工智能替代品。我们描述了使用图像分类器来学习定义二维参数空间的唯一区域的特征。然后我们应用这个分类器来帮助代理模型的选择和验证。我们的CPM模型代理比参考配置提前20,000个时间步生成模型配置,并证明与本地代码执行相比,计算时间减少了大约22倍。我们的工作代表了实现ddpm以开发随机生物系统的数字双胞胎的一步。
{"title":"Generative diffusion model surrogates for mechanistic agent-based biological models.","authors":"Tien Comlekoglu, J Quetzalcoatl Toledo-Marín, Douglas W DeSimone, Shayn M Peirce, Geoffrey Fox, James A Glazier","doi":"10.1088/2632-2153/ae11f8","DOIUrl":"10.1088/2632-2153/ae11f8","url":null,"abstract":"<p><p>Mechanistic, multicellular, agent-based models are commonly used to investigate tissue, organ, and organism-scale biology at single-cell resolution. The Cellular-Potts Model (CPM) is a powerful and popular framework for developing and interrogating these models. CPMs become computationally expensive at large space- and time- scales making application and investigation of developed models difficult. Surrogate models may allow for the accelerated evaluation of CPMs of complex biological systems. However, the stochastic nature of these models means each set of parameters may give rise to different model configurations, complicating surrogate model development. In this work, we leverage denoising diffusion probabilistic models (DDPMs) to train a generative AI surrogate of a CPM used to investigate <i>in vitro</i> vasculogenesis. We describe the use of an image classifier to learn the characteristics that define unique areas of a 2-dimensional parameter space. We then apply this classifier to aid in surrogate model selection and verification. Our CPM model surrogate generates model configurations 20,000 timesteps ahead of a reference configuration and demonstrates approximately a 22x reduction in computational time as compared to native code execution. Our work represents a step towards the implementation of DDPMs to develop digital twins of stochastic biological systems.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 4","pages":"045024"},"PeriodicalIF":4.6,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12570967/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145408961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Depthwise-Dilated Convolutional Adapters for Medical Object Tracking and Segmentation Using the Segment Anything Model 2. 基于分段任意模型的深度扩展卷积医学对象跟踪和分割适配器
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-01 Epub Date: 2025-10-29 DOI: 10.1088/2632-2153/ae13d1
Guoping Xu, Christopher Kabat, You Zhang

Recent advances in medical image segmentation have been driven by deep learning; however, most existing methods remain limited by modality-specific designs and exhibit poor adaptability to dynamic medical imaging scenarios. The Segment Anything Model 2 (SAM2) and its related variants, which introduce a streaming memory mechanism for real-time video segmentation, present new opportunities for prompt-based, generalizable solutions. Nevertheless, adapting these models to medical video scenarios typically requires large-scale datasets for retraining or transfer learning, leading to high computational costs and the risk of catastrophic forgetting. To address these challenges, we propose DD-SAM2, an efficient adaptation framework for SAM2 that incorporates a Depthwise-Dilated Adapter (DD-Adapter) to enhance multi-scale feature extraction with minimal parameter overhead. This design enables effective fine-tuning of SAM2 on medical videos with limited training data. Unlike existing adapter-based methods focused solely on static images, DD-SAM2 fully exploits SAM2's streaming memory for medical video objects tracking and segmentation. Comprehensive evaluations on TrackRad2025 (tumor segmentation) and EchoNet-Dynamic (left ventricle tracking) datasets demonstrate superior performance, achieving Dice scores of 0.93±0.04 and 0.97±0.01, respectively. To the best of our knowledge, this work provides an initial attempt at systematically exploring adapter-based fine-tuning strategies for SAM2 applied medical video segmentation and tracking. Code, datasets, and models will be made publicly available at https://github.com/apple1986/DD-SAM2.

医学图像分割的最新进展是由深度学习推动的;然而,大多数现有方法仍然受到特定模式设计的限制,并且对动态医学成像场景的适应性较差。分段任意模型2 (SAM2)及其相关变体为实时视频分段引入了流存储器机制,为基于提示的通用解决方案提供了新的机会。然而,将这些模型应用于医疗视频场景通常需要大规模的数据集来进行再训练或迁移学习,这导致了高计算成本和灾难性遗忘的风险。为了解决这些挑战,我们提出了DD-SAM2,这是一个有效的SAM2自适应框架,它包含一个深度扩展适配器(DD-Adapter),以最小的参数开销增强多尺度特征提取。这种设计能够在训练数据有限的医学视频上有效地微调SAM2。与现有的仅关注静态图像的基于适配器的方法不同,DD-SAM2充分利用SAM2的流内存进行医疗视频对象跟踪和分割。在TrackRad2025(肿瘤分割)和EchoNet-Dynamic(左心室跟踪)数据集上的综合评价显示出优异的性能,Dice得分分别为0.93±0.04和0.97±0.01。据我们所知,这项工作为系统地探索基于适配器的SAM2应用医学视频分割和跟踪微调策略提供了初步尝试。代码、数据集和模型将在https://github.com/apple1986/DD-SAM2上公开提供。
{"title":"Depthwise-Dilated Convolutional Adapters for Medical Object Tracking and Segmentation Using the Segment Anything Model 2.","authors":"Guoping Xu, Christopher Kabat, You Zhang","doi":"10.1088/2632-2153/ae13d1","DOIUrl":"10.1088/2632-2153/ae13d1","url":null,"abstract":"<p><p>Recent advances in medical image segmentation have been driven by deep learning; however, most existing methods remain limited by modality-specific designs and exhibit poor adaptability to dynamic medical imaging scenarios. The Segment Anything Model 2 (SAM2) and its related variants, which introduce a streaming memory mechanism for real-time video segmentation, present new opportunities for prompt-based, generalizable solutions. Nevertheless, adapting these models to medical video scenarios typically requires large-scale datasets for retraining or transfer learning, leading to high computational costs and the risk of catastrophic forgetting. To address these challenges, we propose DD-SAM2, an efficient adaptation framework for SAM2 that incorporates a Depthwise-Dilated Adapter (DD-Adapter) to enhance multi-scale feature extraction with minimal parameter overhead. This design enables effective fine-tuning of SAM2 on medical videos with limited training data. Unlike existing adapter-based methods focused solely on static images, DD-SAM2 fully exploits SAM2's streaming memory for medical video objects tracking and segmentation. Comprehensive evaluations on TrackRad2025 (tumor segmentation) and EchoNet-Dynamic (left ventricle tracking) datasets demonstrate superior performance, achieving Dice scores of 0.93±0.04 and 0.97±0.01, respectively. To the best of our knowledge, this work provides an initial attempt at systematically exploring adapter-based fine-tuning strategies for SAM2 applied medical video segmentation and tracking. Code, datasets, and models will be made publicly available at https://github.com/apple1986/DD-SAM2.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 4","pages":""},"PeriodicalIF":4.6,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12806169/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145999235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mamba time series forecasting with uncertainty quantification. 不确定量化的曼巴时间序列预测。
IF 6.3 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-30 Epub Date: 2025-07-22 DOI: 10.1088/2632-2153/adec3b
Pedro Pessoa, Paul Campitelli, Douglas P Shepherd, S Banu Ozkan, Steve Pressé

State space models, such as Mamba, have recently garnered attention in time series forecasting (TSF) due to their ability to capture sequence patterns. However, in electricity consumption benchmarks, Mamba forecasts exhibit a mean error of approximately 8%. Similarly, in traffic occupancy benchmarks, the mean error reaches 18%. This discrepancy leaves us to wonder whether the prediction is simply inaccurate or falls within error given spread in historical data. To address this limitation, we propose a method to quantify the predictive uncertainty of Mamba forecasts. To achieve this, we propose a dual-network framework based on the Mamba architecture for probabilistic forecasting, where one network generates point forecasts while the other estimates predictive uncertainty by modeling variance. We abbreviate our tool, Mamba with probabilistic TSF, as Mamba-ProbTSF and the code for its implementation is available on GitHub https://github.com/PessoaP/Mamba-ProbTSF. Evaluating this approach on synthetic and real-world benchmark datasets, we find Kullback-Leibler divergence between the learned distributions and the data-which, in the limit of infinite data, should converge to zero if the model correctly captures the underlying probability distribution-reduced to the order of 10-3 for synthetic data and 10-1 for real-world benchmark. We find that in both the electricity consumption and traffic occupancy benchmark, the true trajectory stays within the predicted uncertainty interval at the two-sigma level about 95% of the time. We further compare Mamba-ProbTSF against leading probabilistic forecast methods, DeepAR and ARIMA, and show that our method consistently achieves lower forecast errors while offering more reliable uncertainty quantification. We end with a consideration of potential limitations, adjustments to improve performance, and considerations for applying this framework to processes for purely or largely stochastic dynamics where the stochastic changes accumulate as observed, for example, in pure Brownian motion or molecular dynamics trajectories.

状态空间模型,如Mamba,由于其捕获序列模式的能力,最近在时间序列预测(TSF)中引起了人们的关注。然而,在电力消耗基准中,曼巴预测的平均误差约为8%。同样,在交通占用率基准中,平均误差达到18%。这种差异让我们怀疑,这种预测是简单地不准确,还是在历史数据分布的情况下属于误差范围。为了解决这一限制,我们提出了一种量化曼巴预测的预测不确定性的方法。为了实现这一点,我们提出了一个基于Mamba架构的双网络框架,用于概率预测,其中一个网络生成点预测,而另一个网络通过建模方差来估计预测不确定性。我们将我们的工具Mamba with probabilistic TSF简称为Mamba- probtsf,其实现代码可在GitHub https://github.com/PessoaP/Mamba-ProbTSF上获得。在合成和真实世界的基准数据集上评估这种方法,我们发现学习分布和数据之间的Kullback-Leibler散度——在无限数据的限制下,如果模型正确捕获潜在的概率分布,应该收敛于零——对于合成数据减少到10-3的数量级,对于真实世界的基准降低到10-1。我们发现,在电力消耗和交通占用基准中,真实轨迹在大约95%的时间内保持在预测的2西格玛水平的不确定性区间内。我们进一步将Mamba-ProbTSF与领先的概率预测方法DeepAR和ARIMA进行了比较,结果表明,我们的方法在提供更可靠的不确定性量化的同时,始终实现更低的预测误差。我们最后考虑了潜在的局限性,改进性能的调整,以及将该框架应用于纯或大部分随机动力学过程的考虑,其中随机变化如观察到的那样累积,例如,在纯布朗运动或分子动力学轨迹中。
{"title":"Mamba time series forecasting with uncertainty quantification.","authors":"Pedro Pessoa, Paul Campitelli, Douglas P Shepherd, S Banu Ozkan, Steve Pressé","doi":"10.1088/2632-2153/adec3b","DOIUrl":"10.1088/2632-2153/adec3b","url":null,"abstract":"<p><p>State space models, such as Mamba, have recently garnered attention in time series forecasting (TSF) due to their ability to capture sequence patterns. However, in electricity consumption benchmarks, Mamba forecasts exhibit a mean error of approximately 8%. Similarly, in traffic occupancy benchmarks, the mean error reaches 18%. This discrepancy leaves us to wonder whether the prediction is simply inaccurate or falls within error given spread in historical data. To address this limitation, we propose a method to quantify the predictive uncertainty of Mamba forecasts. To achieve this, we propose a dual-network framework based on the Mamba architecture for probabilistic forecasting, where one network generates point forecasts while the other estimates predictive uncertainty by modeling variance. We abbreviate our tool, Mamba with probabilistic TSF, as Mamba-ProbTSF and the code for its implementation is available on GitHub https://github.com/PessoaP/Mamba-ProbTSF. Evaluating this approach on synthetic and real-world benchmark datasets, we find Kullback-Leibler divergence between the learned distributions and the data-which, in the limit of infinite data, should converge to zero if the model correctly captures the underlying probability distribution-reduced to the order of 10<sup>-3</sup> for synthetic data and 10<sup>-1</sup> for real-world benchmark. We find that in both the electricity consumption and traffic occupancy benchmark, the true trajectory stays within the predicted uncertainty interval at the two-sigma level about 95% of the time. We further compare Mamba-ProbTSF against leading probabilistic forecast methods, DeepAR and ARIMA, and show that our method consistently achieves lower forecast errors while offering more reliable uncertainty quantification. We end with a consideration of potential limitations, adjustments to improve performance, and considerations for applying this framework to processes for purely or largely stochastic dynamics where the stochastic changes accumulate as observed, for example, in pure Brownian motion or molecular dynamics trajectories.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 3","pages":"035012"},"PeriodicalIF":6.3,"publicationDate":"2025-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12281171/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144699735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
32 examples of LLM applications in materials science and chemistry: towards automation, assistants, agents, and accelerated scientific discovery. 32个LLM在材料科学和化学中的应用实例:走向自动化、助理、代理和加速科学发现。
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-30 Epub Date: 2025-09-29 DOI: 10.1088/2632-2153/ae011a
Yoel Zimmermann, Adib Bazgir, Alexander Al-Feghali, Mehrad Ansari, Joshua Bocarsly, L Catherine Brinson, Yuan Chiang, Defne Circi, Min-Hsueh Chiu, Nathan Daelman, Matthew L Evans, Abhijeet S Gangan, Janine George, Hassan Harb, Ghazal Khalighinejad, Sartaaj Takrim Khan, Sascha Klawohn, Magdalena Lederbauer, Soroush Mahjoubi, Bernadette Mohr, Seyed Mohamad Moosavi, Aakash Naik, Aleyna Beste Ozhan, Dieter Plessers, Aritra Roy, Fabian Schöppach, Philippe Schwaller, Carla Terboven, Katharina Ueltzen, Yue Wu, Shang Zhu, Jan Janssen, Calvin Li, Ian Foster, Ben Blaiszik

Large language models (LLMs) are reshaping many aspects of materials science and chemistry research, enabling advances in molecular property prediction, materials design, scientific automation, knowledge extraction, and more. Recent developments demonstrate that the latest class of models are able to integrate structured and unstructured data, assist in hypothesis generation, and streamline research workflows. To explore the frontier of LLM capabilities across the research lifecycle, we review applications of LLMs through 32 total projects developed during the second annual LLM hackathon for applications in materials science and chemistry, a global hybrid event. These projects spanned seven key research areas: (1) molecular and material property prediction, (2) molecular and material design, (3) automation and novel interfaces, (4) scientific communication and education, (5) research data management and automation, (6) hypothesis generation and evaluation, and (7) knowledge extraction and reasoning from the scientific literature. Collectively, these applications illustrate how LLMs serve as versatile predictive models, platforms for rapid prototyping of domain-specific tools, and much more. In particular, improvements in both open source and proprietary LLM performance through the addition of reasoning, additional training data, and new techniques have expanded effectiveness, particularly in low-data environments and interdisciplinary research. As LLMs continue to improve, their integration into scientific workflows presents both new opportunities and new challenges, requiring ongoing exploration, continued refinement, and further research to address reliability, interpretability, and reproducibility.

大型语言模型(llm)正在重塑材料科学和化学研究的许多方面,使分子性质预测、材料设计、科学自动化、知识提取等方面取得进展。最近的发展表明,最新的一类模型能够集成结构化和非结构化数据,协助假设生成,并简化研究工作流程。为了探索LLM能力在整个研究生命周期中的前沿,我们通过第二届年度LLM黑客马拉松期间开发的32个项目来审查LLM的应用,这些项目用于材料科学和化学的应用,这是一项全球性的混合活动。这些项目涵盖七个重点研究领域:(1)分子和材料性质预测;(2)分子和材料设计;(3)自动化和新型界面;(4)科学交流与教育;(5)研究数据管理与自动化;(6)假设生成与评估;(7)科学文献的知识提取与推理。总的来说,这些应用程序说明了llm如何作为通用的预测模型、领域特定工具的快速原型设计平台等等。特别是,通过添加推理、额外的训练数据和新技术,开源和专有LLM性能的改进扩展了有效性,特别是在低数据环境和跨学科研究中。随着法学硕士的不断进步,它们与科学工作流程的整合带来了新的机遇和新的挑战,需要持续的探索、不断的改进和进一步的研究来解决可靠性、可解释性和可重复性。
{"title":"32 examples of LLM applications in materials science and chemistry: towards automation, assistants, agents, and accelerated scientific discovery.","authors":"Yoel Zimmermann, Adib Bazgir, Alexander Al-Feghali, Mehrad Ansari, Joshua Bocarsly, L Catherine Brinson, Yuan Chiang, Defne Circi, Min-Hsueh Chiu, Nathan Daelman, Matthew L Evans, Abhijeet S Gangan, Janine George, Hassan Harb, Ghazal Khalighinejad, Sartaaj Takrim Khan, Sascha Klawohn, Magdalena Lederbauer, Soroush Mahjoubi, Bernadette Mohr, Seyed Mohamad Moosavi, Aakash Naik, Aleyna Beste Ozhan, Dieter Plessers, Aritra Roy, Fabian Schöppach, Philippe Schwaller, Carla Terboven, Katharina Ueltzen, Yue Wu, Shang Zhu, Jan Janssen, Calvin Li, Ian Foster, Ben Blaiszik","doi":"10.1088/2632-2153/ae011a","DOIUrl":"10.1088/2632-2153/ae011a","url":null,"abstract":"<p><p>Large language models (LLMs) are reshaping many aspects of materials science and chemistry research, enabling advances in molecular property prediction, materials design, scientific automation, knowledge extraction, and more. Recent developments demonstrate that the latest class of models are able to integrate structured and unstructured data, assist in hypothesis generation, and streamline research workflows. To explore the frontier of LLM capabilities across the research lifecycle, we review applications of LLMs through 32 total projects developed during the second annual LLM hackathon for applications in materials science and chemistry, a global hybrid event. These projects spanned seven key research areas: (1) molecular and material property prediction, (2) molecular and material design, (3) automation and novel interfaces, (4) scientific communication and education, (5) research data management and automation, (6) hypothesis generation and evaluation, and (7) knowledge extraction and reasoning from the scientific literature. Collectively, these applications illustrate how LLMs serve as versatile predictive models, platforms for rapid prototyping of domain-specific tools, and much more. In particular, improvements in both open source and proprietary LLM performance through the addition of reasoning, additional training data, and new techniques have expanded effectiveness, particularly in low-data environments and interdisciplinary research. As LLMs continue to improve, their integration into scientific workflows presents both new opportunities and new challenges, requiring ongoing exploration, continued refinement, and further research to address reliability, interpretability, and reproducibility.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 3","pages":"030701"},"PeriodicalIF":4.6,"publicationDate":"2025-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12492978/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145233532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond Euclid: an illustrated guide to modern machine learning with geometric, topological, and algebraic structures. 超越欧几里得:用几何,拓扑和代数结构的现代机器学习的图解指南。
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-30 Epub Date: 2025-08-01 DOI: 10.1088/2632-2153/adf375
Mathilde Papillon, Sophia Sanborn, Johan Mathe, Louisa Cornelis, Abby Bertics, Domas Buracas, Hansen J Lillemark, Christian Shewmake, Fatih Dinc, Xavier Pennec, Nina Miolane

The enduring legacy of Euclidean geometry underpins classical machine learning, which, for decades, has been primarily developed for data lying in Euclidean space. Yet, modern machine learning increasingly encounters richly structured data that is inherently non-Euclidean. This data can exhibit intricate geometric, topological and algebraic structure: from the geometry of the curvature of space-time, to topologically complex interactions between neurons in the brain, to the algebraic transformations describing symmetries of physical systems. Extracting knowledge from such non-Euclidean data necessitates a broader mathematical perspective. Echoing the 19th-century revolutions that gave rise to non-Euclidean geometry, an emerging line of research is redefining modern machine learning with non-Euclidean structures. Its goal: generalizing classical methods to unconventional data types with geometry, topology, and algebra. In this review, we provide an accessible gateway to this fast-growing field and propose a graphical taxonomy that integrates recent advances into an intuitive unified framework. We subsequently extract insights into current challenges and highlight exciting opportunities for future development in this field.

欧几里得几何的不朽遗产是经典机器学习的基础,几十年来,经典机器学习主要是为欧几里得空间中的数据而开发的。然而,现代机器学习越来越多地遇到本质上非欧几里得的丰富结构化数据。这些数据可以展示复杂的几何、拓扑和代数结构:从时空曲率的几何,到大脑神经元之间拓扑复杂的相互作用,再到描述物理系统对称性的代数变换。从这种非欧几里得数据中提取知识需要更广阔的数学视角。与19世纪催生非欧几里得几何的革命相呼应,一项新兴的研究正在用非欧几里得结构重新定义现代机器学习。它的目标是:将经典方法推广到具有几何、拓扑和代数的非常规数据类型。在这篇综述中,我们为这个快速发展的领域提供了一个可访问的门户,并提出了一个图形分类,将最新进展集成到一个直观的统一框架中。随后,我们深入分析当前的挑战,并强调该领域未来发展的激动人心的机遇。
{"title":"Beyond Euclid: an illustrated guide to modern machine learning with geometric, topological, and algebraic structures.","authors":"Mathilde Papillon, Sophia Sanborn, Johan Mathe, Louisa Cornelis, Abby Bertics, Domas Buracas, Hansen J Lillemark, Christian Shewmake, Fatih Dinc, Xavier Pennec, Nina Miolane","doi":"10.1088/2632-2153/adf375","DOIUrl":"10.1088/2632-2153/adf375","url":null,"abstract":"<p><p>The enduring legacy of Euclidean geometry underpins classical machine learning, which, for decades, has been primarily developed for data lying in Euclidean space. Yet, modern machine learning increasingly encounters richly structured data that is inherently non-Euclidean. This data can exhibit intricate geometric, topological and algebraic structure: from the geometry of the curvature of space-time, to topologically complex interactions between neurons in the brain, to the algebraic transformations describing symmetries of physical systems. Extracting knowledge from such non-Euclidean data necessitates a broader mathematical perspective. Echoing the 19th-century revolutions that gave rise to non-Euclidean geometry, an emerging line of research is redefining modern machine learning with non-Euclidean structures. Its goal: generalizing classical methods to unconventional data types with geometry, topology, and algebra. In this review, we provide an accessible gateway to this fast-growing field and propose a graphical taxonomy that integrates recent advances into an intuitive unified framework. We subsequently extract insights into current challenges and highlight exciting opportunities for future development in this field.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 3","pages":"031002"},"PeriodicalIF":4.6,"publicationDate":"2025-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12315666/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144776367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prior guided deep difference meta-learner for fast adaptation to stylized segmentation. 先验引导深度差异元学习器快速适应程式化分割。
IF 6.3 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-30 Epub Date: 2025-04-16 DOI: 10.1088/2632-2153/adc970
Dan Nguyen, Anjali Balagopal, Ti Bai, Michael Dohopolski, Mu-Han Lin, Steve Jiang

Radiotherapy treatment planning requires segmenting anatomical structures in various styles, influenced by guidelines, protocols, preferences, or dose planning needs. Deep learning-based auto-segmentation models, trained on anatomical definitions, may not match local clinicians' styles at new institutions. Adapting these models can be challenging without sufficient resources. We hypothesize that consistent differences between segmentation styles and anatomical definitions can be learned from initial patients and applied to pre-trained models for more precise segmentation. We propose a Prior-guided deep difference meta-learner (DDL) to learn and adapt these differences. We collected data from 440 patients for model development and 30 for testing. The dataset includes contours of the prostate clinical target volume (CTV), parotid, and rectum. We developed a deep learning framework that segments new images with a matching style using example styles as a prior, without model retraining. The pre-trained segmentation models were adapted to three different clinician styles for post-operative CTV for prostate, parotid gland, and rectum segmentation. We tested the model's ability to learn unseen styles and compared its performance with transfer learning, using varying amounts of prior patient style data (0-10 patients). Performance was quantitatively evaluated using dice similarity coefficient (DSC) and Hausdorff distance. With exposure to only three patients for the model, the average DSC (%) improved from 78.6, 71.9, 63.0, 69.6, 52.2 and 46.3-84.4, 77.8, 73.0, 77.8, 70.5, 68.1, for CTVstyle1, CTVstyle2, CTVstyle3, Parotidsuperficial, Rectumsuperior, and Rectumposterior, respectively. The proposed Prior-guided DDL is a fast and effortless network for adapting a structure to new styles. The improved segmentation accuracy may result in reduced contour editing time, providing a more efficient and streamlined clinical workflow.

放疗治疗计划需要分割不同风格的解剖结构,受指南、方案、偏好或剂量计划需求的影响。基于解剖定义的深度学习自动分割模型可能与新机构的当地临床医生的风格不匹配。在没有足够资源的情况下,调整这些模型可能具有挑战性。我们假设,分割风格和解剖定义之间的一致差异可以从初始患者身上学习到,并应用于预训练模型,以实现更精确的分割。我们提出了一个先验引导的深度差异元学习者(DDL)来学习和适应这些差异。我们收集了440例患者的数据用于模型开发,30例用于测试。该数据集包括前列腺临床靶体积(CTV)、腮腺和直肠的轮廓。我们开发了一个深度学习框架,该框架使用示例风格作为先验,将具有匹配风格的新图像分割,而无需模型再训练。预先训练的分割模型适用于三种不同的临床医生风格,用于前列腺、腮腺和直肠的术后CTV分割。我们测试了模型学习未知风格的能力,并使用不同数量的先前患者风格数据(0-10名患者)将其性能与迁移学习进行了比较。使用骰子相似系数(DSC)和豪斯多夫距离对性能进行定量评价。仅暴露于3例患者时,CTVstyle1、CTVstyle2、CTVstyle3、腮腺浅、上直肠和后直肠的平均DSC(%)分别从78.6、71.9、63.0、69.6、52.2和46.3提高到84.4、77.8、73.0、77.8、70.5、68.1。提出的先验引导DDL是一种快速且轻松的网络,可以使结构适应新的样式。改进的分割精度可以减少轮廓编辑时间,提供更有效和精简的临床工作流程。
{"title":"Prior guided deep difference meta-learner for fast adaptation to stylized segmentation.","authors":"Dan Nguyen, Anjali Balagopal, Ti Bai, Michael Dohopolski, Mu-Han Lin, Steve Jiang","doi":"10.1088/2632-2153/adc970","DOIUrl":"https://doi.org/10.1088/2632-2153/adc970","url":null,"abstract":"<p><p>Radiotherapy treatment planning requires segmenting anatomical structures in various styles, influenced by guidelines, protocols, preferences, or dose planning needs. Deep learning-based auto-segmentation models, trained on anatomical definitions, may not match local clinicians' styles at new institutions. Adapting these models can be challenging without sufficient resources. We hypothesize that consistent differences between segmentation styles and anatomical definitions can be learned from initial patients and applied to pre-trained models for more precise segmentation. We propose a Prior-guided deep difference meta-learner (DDL) to learn and adapt these differences. We collected data from 440 patients for model development and 30 for testing. The dataset includes contours of the prostate clinical target volume (CTV), parotid, and rectum. We developed a deep learning framework that segments new images with a matching style using example styles as a prior, without model retraining. The pre-trained segmentation models were adapted to three different clinician styles for post-operative CTV for prostate, parotid gland, and rectum segmentation. We tested the model's ability to learn unseen styles and compared its performance with transfer learning, using varying amounts of prior patient style data (0-10 patients). Performance was quantitatively evaluated using dice similarity coefficient (DSC) and Hausdorff distance. With exposure to only three patients for the model, the average DSC (%) improved from 78.6, 71.9, 63.0, 69.6, 52.2 and 46.3-84.4, 77.8, 73.0, 77.8, 70.5, 68.1, for CTV<sub>style1</sub>, CTV<sub>style2</sub>, CTV<sub>style3</sub>, Parotid<sub>superficial</sub>, Rectum<sub>superior</sub>, and Rectum<sub>posterior</sub>, respectively. The proposed Prior-guided DDL is a fast and effortless network for adapting a structure to new styles. The improved segmentation accuracy may result in reduced contour editing time, providing a more efficient and streamlined clinical workflow.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 2","pages":"025016"},"PeriodicalIF":6.3,"publicationDate":"2025-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12001319/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144002018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FDDM: Unsupervised Medical Image Translation with a Frequency-Decoupled Diffusion Model. 基于频率解耦扩散模型的无监督医学图像平移。
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-01 Epub Date: 2025-04-07 DOI: 10.1088/2632-2153/adc656
Yunxiang Li, Hua-Chieh Shao, Xiaoxue Qian, You Zhang

Diffusion models have demonstrated significant potential in producing high-quality images in medical image translation to aid disease diagnosis, localization, and treatment. Nevertheless, current diffusion models often fall short when it comes to faithfully translating medical images. They struggle to accurately preserve anatomical structures, especially when working with unpaired datasets. In this study, we introduce the Frequency Decoupled Diffusion Model (FDDM) for MR-to-CT conversion. The differences between MR and CT images lie in both anatomical structures (e.g., the outlines of organs or bones) and the data distribution (e.g., intensity values and contrast within). Therefore, FDDM first converts anatomical information using an initial conversion module. Then, the converted anatomical information guides a subsequent diffusion model to generate high-quality CT images. Our diffusion model uses a dual-path reverse diffusion process for low-frequency and high-frequency information, achieving a better balance between image quality and anatomical accuracy. We extensively evaluated FDDM using public datasets for brain MR-to-CT and pelvis MR-to-CT translations. The results show that FDDM outperforms other generative adversarial network (GAN)-based, variational autoencoder (VAE)-based, and diffusion-based models. The evaluation metrics included Fréchet Inception Distance (FID), mean absolute error (MAE), mean squared error (MSE), Structural Similarity Index Measure (SSIM), and Dice similarity coefficient (DICE). FDDM achieved the best scores on all metrics for both datasets, particularly excelling in FID, with scores of 25.9 for brain data and 29.2 for pelvis data, significantly outperforming other methods. These results demonstrate that FDDM can generate high-quality target domain images while maintaining the accuracy of translated anatomical structures, thereby facilitating more precise/accurate downstream tasks including anatomy segmentation and radiotherapy planning.

扩散模型在医学图像翻译中产生高质量图像以帮助疾病诊断、定位和治疗方面显示出巨大的潜力。然而,目前的扩散模型在忠实地翻译医学图像时往往存在不足。他们很难准确地保存解剖结构,特别是在处理不成对的数据集时。在这项研究中,我们引入了频率解耦扩散模型(FDDM)用于磁共振到ct的转换。MR和CT图像的区别在于解剖结构(如器官或骨骼的轮廓)和数据分布(如强度值和内部对比度)。因此,FDDM首先使用初始转换模块转换解剖信息。然后,转换后的解剖信息引导后续的扩散模型生成高质量的CT图像。我们的扩散模型对低频和高频信息采用双路径反向扩散过程,在图像质量和解剖精度之间实现了更好的平衡。我们使用脑MR-to-CT和骨盆MR-to-CT翻译的公共数据集广泛评估FDDM。结果表明,FDDM优于其他基于生成对抗网络(GAN)、基于变分自编码器(VAE)和基于扩散的模型。评价指标包括fr起始距离(FID)、平均绝对误差(MAE)、均方误差(MSE)、结构相似指数测量(SSIM)和Dice相似系数(Dice)。FDDM在两个数据集的所有指标上都取得了最好的分数,特别是在FID中表现出色,脑数据得分为25.9分,骨盆数据得分为29.2分,明显优于其他方法。这些结果表明,FDDM可以生成高质量的目标域图像,同时保持翻译的解剖结构的准确性,从而促进更精确/准确的下游任务,包括解剖分割和放疗计划。
{"title":"FDDM: Unsupervised Medical Image Translation with a Frequency-Decoupled Diffusion Model.","authors":"Yunxiang Li, Hua-Chieh Shao, Xiaoxue Qian, You Zhang","doi":"10.1088/2632-2153/adc656","DOIUrl":"10.1088/2632-2153/adc656","url":null,"abstract":"<p><p>Diffusion models have demonstrated significant potential in producing high-quality images in medical image translation to aid disease diagnosis, localization, and treatment. Nevertheless, current diffusion models often fall short when it comes to faithfully translating medical images. They struggle to accurately preserve anatomical structures, especially when working with unpaired datasets. In this study, we introduce the Frequency Decoupled Diffusion Model (FDDM) for MR-to-CT conversion. The differences between MR and CT images lie in both anatomical structures (e.g., the outlines of organs or bones) and the data distribution (e.g., intensity values and contrast within). Therefore, FDDM first converts anatomical information using an initial conversion module. Then, the converted anatomical information guides a subsequent diffusion model to generate high-quality CT images. Our diffusion model uses a dual-path reverse diffusion process for low-frequency and high-frequency information, achieving a better balance between image quality and anatomical accuracy. We extensively evaluated FDDM using public datasets for brain MR-to-CT and pelvis MR-to-CT translations. The results show that FDDM outperforms other generative adversarial network (GAN)-based, variational autoencoder (VAE)-based, and diffusion-based models. The evaluation metrics included Fréchet Inception Distance (FID), mean absolute error (MAE), mean squared error (MSE), Structural Similarity Index Measure (SSIM), and Dice similarity coefficient (DICE). FDDM achieved the best scores on all metrics for both datasets, particularly excelling in FID, with scores of 25.9 for brain data and 29.2 for pelvis data, significantly outperforming other methods. These results demonstrate that FDDM can generate high-quality target domain images while maintaining the accuracy of translated anatomical structures, thereby facilitating more precise/accurate downstream tasks including anatomy segmentation and radiotherapy planning.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 2","pages":""},"PeriodicalIF":4.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12826588/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146053995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Medical Image Segmentation Assisted with Clinical Inputs via Language Encoder in A Deep Learning Framework. 基于深度学习框架的语言编码器辅助临床输入医学图像分割。
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-01 Epub Date: 2025-02-14 DOI: 10.1088/2632-2153/adb371
Hengrui Zhao, Biling Wang, Deepkumar Mistry, Jing Wang, Michael Dohopolski, Daniel Yang, Weiguo Lu, Steve Jiang, Dan Nguyen

Introduction: Auto-segmentation of tumor volumes and organs at risk (OARs) is a critical step in cancer radiotherapy treatment planning, where rapid, precise adjustments to treatment plans are required to match the patient anatomy. Although auto-segmentation has been clinically accepted for most OARs, auto-segmentation of tumor volumes, particularly clinical target volumes (CTVs), remains a challenge. This difficulty arises because images alone are often insufficient to capture the necessary information for accurate delineation of microscopic tumor invasion invisible on the image itself.

Methods: We propose a deep learning-based medical image segmentation framework designed to mimic the clinical process of delineating CTVs and OARs. At its core, the model performs precise segmentation of medical images while enhancing accuracy by integrating clinical information in text format. A transformer-based text encoder converts textual clinical data into vectors, which are incorporated into the segmentation process with image features. This integration bridges the gap between traditional automated segmentation methods and clinician-guided, context-rich delineations. The framework's effectiveness is demonstrated through a prostate segmentation example in the context of radiation therapy for localized prostate cancer, where incorporating clinical context significantly impacts the delineation process.

Results: In our experiments, we included additional clinical information potentially influencing clinicians' prostate segmentation. The results show that our proposed method not only outperforms the baseline model, but also surpasses current state-of-the-art methods, with or without clinical contexts. Furthermore, our method demonstrates high performance even with limited data.

Conclusion: This proposed segmentation framework has shown to significantly improve auto-segmentation, particularly for CTVs, in cancer radiotherapy.

肿瘤体积和危险器官(OARs)的自动分割是癌症放疗治疗计划的关键步骤,需要快速,精确地调整治疗计划以匹配患者的解剖结构。尽管自动分割在临床上已被大多数OARs所接受,但肿瘤体积的自动分割,特别是临床靶体积(ctv)的自动分割仍然是一个挑战。这一困难的出现是因为单独的图像往往不足以捕获必要的信息,以准确描绘显微镜下的肿瘤侵袭,而在图像本身是看不见的。方法:我们提出了一个基于深度学习的医学图像分割框架,旨在模拟ctv和OARs的临床描绘过程。该模型的核心是对医学图像进行精确分割,同时通过将临床信息整合为文本格式来提高准确性。基于转换器的文本编码器将文本临床数据转换为矢量,并将其与图像特征结合到分割过程中。这种集成弥补了传统的自动分割方法和临床医生指导的、上下文丰富的描绘之间的差距。该框架的有效性通过局部前列腺癌放射治疗背景下的前列腺分割示例得到了证明,其中纳入临床背景显著影响了描绘过程。结果:在我们的实验中,我们纳入了可能影响临床医生前列腺分割的额外临床信息。结果表明,我们提出的方法不仅优于基线模型,而且超过了目前最先进的方法,无论是否有临床背景。此外,即使在有限的数据下,我们的方法也显示出高性能。结论:提出的分割框架已显示出显著改善自动分割,特别是对ctv,在癌症放疗。
{"title":"Medical Image Segmentation Assisted with Clinical Inputs via Language Encoder in A Deep Learning Framework.","authors":"Hengrui Zhao, Biling Wang, Deepkumar Mistry, Jing Wang, Michael Dohopolski, Daniel Yang, Weiguo Lu, Steve Jiang, Dan Nguyen","doi":"10.1088/2632-2153/adb371","DOIUrl":"10.1088/2632-2153/adb371","url":null,"abstract":"<p><strong>Introduction: </strong>Auto-segmentation of tumor volumes and organs at risk (OARs) is a critical step in cancer radiotherapy treatment planning, where rapid, precise adjustments to treatment plans are required to match the patient anatomy. Although auto-segmentation has been clinically accepted for most OARs, auto-segmentation of tumor volumes, particularly clinical target volumes (CTVs), remains a challenge. This difficulty arises because images alone are often insufficient to capture the necessary information for accurate delineation of microscopic tumor invasion invisible on the image itself.</p><p><strong>Methods: </strong>We propose a deep learning-based medical image segmentation framework designed to mimic the clinical process of delineating CTVs and OARs. At its core, the model performs precise segmentation of medical images while enhancing accuracy by integrating clinical information in text format. A transformer-based text encoder converts textual clinical data into vectors, which are incorporated into the segmentation process with image features. This integration bridges the gap between traditional automated segmentation methods and clinician-guided, context-rich delineations. The framework's effectiveness is demonstrated through a prostate segmentation example in the context of radiation therapy for localized prostate cancer, where incorporating clinical context significantly impacts the delineation process.</p><p><strong>Results: </strong>In our experiments, we included additional clinical information potentially influencing clinicians' prostate segmentation. The results show that our proposed method not only outperforms the baseline model, but also surpasses current state-of-the-art methods, with or without clinical contexts. Furthermore, our method demonstrates high performance even with limited data.</p><p><strong>Conclusion: </strong>This proposed segmentation framework has shown to significantly improve auto-segmentation, particularly for CTVs, in cancer radiotherapy.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 1","pages":""},"PeriodicalIF":4.6,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12509794/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145281319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Unsupervised Clustering for Prostate Auto-segmentation With and Without Hydrogel Spacer. 基于深度无监督聚类的前列腺自动分割。
IF 4.6 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-01 Epub Date: 2025-01-23 DOI: 10.1088/2632-2153/ada8f3
Hengrui Zhao, Biling Wang, Michael Dohopolski, Ti Bai, Steve Jiang, Dan Nguyen

Introduction: Clinical datasets for training deep learning (DL) models often exhibit high levels of heterogeneity due to differences such as patient characteristics, new medical techniques, and physician preferences. In recent years, hydrogel spacers have been used in some prostate cancer patients receiving radiotherapy to separate the prostate and the rectum to better spare the rectum while achieving adequate dose coverage on the prostate. However, this substantially affects the CT image appearance, which downstream reduced the contouring accuracy of auto-segmentation algorithms. This leads to highly heterogeneous dataset.

Methods: To address this issue, we propose to identify underlying clusters within the dataset and use the cluster labels for segmentation. We collected a clinical dataset of 909 patients, including those with two types of hydrogel spacers and those without. First, we trained a DL model to locate the prostate and limit our field of view to the local area surrounding the prostate and rectum. We then used Uniform Manifold Approximation and Projection (UMAP) for dimensionality reduction and employed k-means clustering to assign each patient to a cluster. To leverage this clustered data, we propose a text-guided segmentation model, CLIP-UNet, which encodes the cluster information using a text encoder and combines the encoded text information with image features for segmentation.

Results: The UMAP results indicated up to three clusters within the dataset. CLIP-UNet with cluster information achieved a Dice score of 86.2% compared to 84.4% from the baseline UNet. Additionally, CLIP-UNet outperforms other state-of-the-art models with or without cluster information.

Conclusion: Automatic clustering assisted by deep learning can reveal hidden data clusters in clinical datasets, and CLIP-UNet effectively utilizes clustered labels and achieves higher performance.

导读:用于训练深度学习(DL)模型的临床数据集通常由于患者特征、新医疗技术和医生偏好等差异而表现出高度的异质性。近年来,在一些接受放疗的前列腺癌患者中,水凝胶间隔剂被用于分离前列腺和直肠,以更好地保留直肠,同时使前列腺获得足够的剂量覆盖。然而,这实质上影响了CT图像的外观,从而降低了自动分割算法的轮廓精度。这导致了高度异构的数据集。方法:为了解决这个问题,我们建议在数据集中识别底层聚类,并使用聚类标签进行分割。我们收集了909例患者的临床数据集,包括两种类型的水凝胶间隔剂和没有。首先,我们训练了一个深度学习模型来定位前列腺,并将视野限制在前列腺和直肠周围的局部区域。然后,我们使用统一流形逼近和投影(UMAP)进行降维,并使用k-means聚类将每个患者分配到一个聚类中。为了利用这些聚类数据,我们提出了一个文本引导的分割模型CLIP-UNet,该模型使用文本编码器对聚类信息进行编码,并将编码的文本信息与图像特征结合起来进行分割。结果:UMAP结果显示数据集中最多有三个集群。与基线UNet的84.4%相比,带有集群信息的CLIP-UNet的Dice得分为86.2%。此外,CLIP-UNet在有或没有集群信息的情况下都优于其他最先进的模型。结论:深度学习辅助下的自动聚类可以揭示临床数据集中隐藏的数据聚类,CLIP-UNet有效地利用了聚类标签,实现了更高的性能。
{"title":"Deep Unsupervised Clustering for Prostate Auto-segmentation With and Without Hydrogel Spacer.","authors":"Hengrui Zhao, Biling Wang, Michael Dohopolski, Ti Bai, Steve Jiang, Dan Nguyen","doi":"10.1088/2632-2153/ada8f3","DOIUrl":"10.1088/2632-2153/ada8f3","url":null,"abstract":"<p><strong>Introduction: </strong>Clinical datasets for training deep learning (DL) models often exhibit high levels of heterogeneity due to differences such as patient characteristics, new medical techniques, and physician preferences. In recent years, hydrogel spacers have been used in some prostate cancer patients receiving radiotherapy to separate the prostate and the rectum to better spare the rectum while achieving adequate dose coverage on the prostate. However, this substantially affects the CT image appearance, which downstream reduced the contouring accuracy of auto-segmentation algorithms. This leads to highly heterogeneous dataset.</p><p><strong>Methods: </strong>To address this issue, we propose to identify underlying clusters within the dataset and use the cluster labels for segmentation. We collected a clinical dataset of 909 patients, including those with two types of hydrogel spacers and those without. First, we trained a DL model to locate the prostate and limit our field of view to the local area surrounding the prostate and rectum. We then used Uniform Manifold Approximation and Projection (UMAP) for dimensionality reduction and employed k-means clustering to assign each patient to a cluster. To leverage this clustered data, we propose a text-guided segmentation model, CLIP-UNet, which encodes the cluster information using a text encoder and combines the encoded text information with image features for segmentation.</p><p><strong>Results: </strong>The UMAP results indicated up to three clusters within the dataset. CLIP-UNet with cluster information achieved a Dice score of 86.2% compared to 84.4% from the baseline UNet. Additionally, CLIP-UNet outperforms other state-of-the-art models with or without cluster information.</p><p><strong>Conclusion: </strong>Automatic clustering assisted by deep learning can reveal hidden data clusters in clinical datasets, and CLIP-UNet effectively utilizes clustered labels and achieves higher performance.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"6 1","pages":""},"PeriodicalIF":4.6,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12509332/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145281229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality assurance for online adaptive radiotherapy: a secondary dose verification model with geometry-encoded U-Net. 在线自适应放射治疗的质量保证:采用几何编码 U-Net 的二次剂量验证模型。
IF 6.3 2区 物理与天体物理 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-01 Epub Date: 2024-10-11 DOI: 10.1088/2632-2153/ad829e
Shunyu Yan, Austen Maniscalco, Biling Wang, Dan Nguyen, Steve Jiang, Chenyang Shen

In online adaptive radiotherapy (ART), quick computation-based secondary dose verification is crucial for ensuring the quality of ART plans while the patient is positioned on the treatment couch. However, traditional dose verification algorithms are generally time-consuming, reducing the efficiency of ART workflow. This study aims to develop an ultra-fast deep-learning (DL) based secondary dose verification algorithm to accurately estimate dose distributions using computed tomography (CT) and fluence maps (FMs). We integrated FMs into the CT image domain by explicitly resolving the geometry of treatment delivery. For each gantry angle, an FM was constructed based on the optimized multi-leaf collimator apertures and corresponding monitoring units. To effectively encode treatment beam configuration, the constructed FMs were back-projected to 30 cm away from the isocenter with respect to the exact geometry of the treatment machines. Then, a 3D U-Net was utilized to take the integrated CT and FM volume as input to estimate dose. Training and validation were performed on 381 prostate cancer cases, with an additional 40 testing cases for independent evaluation of model performance. The proposed model can estimate dose in ∼ 15 ms for each patient. The average γ passing rate ( 3 % / 2 mm , 10 % threshold) for the estimated dose was 99.9% ± 0.15% on testing patients. The mean dose differences for the planning target volume and organs at risk were 0.07 % ± 0.34 % and 0.48 % ± 0.72 % , respectively. We have developed a geometry-resolved DL framework for accurate dose estimation and demonstrated its potential in real-time online ART doses verification.

在在线自适应放射治疗(ART)中,当病人被安置在治疗床上时,基于快速计算的二次剂量验证对于确保 ART 计划的质量至关重要。然而,传统的剂量验证算法一般都很耗时,降低了 ART 工作流程的效率。本研究旨在开发一种基于深度学习(DL)的超快速二次剂量验证算法,利用计算机断层成像(CT)和通量图(FMs)准确估计剂量分布。我们通过明确解析治疗投放的几何形状,将通量图整合到 CT 图像域中。对于每个龙门架角度,我们都根据优化的多叶准直器孔径和相应的监测单元构建了一个 FM。为有效编码治疗光束配置,根据治疗机的精确几何形状,将构建的调频反向投影到距离等中心 30 厘米的位置。然后,利用三维 U-Net 将集成 CT 和调频体积作为输入来估算剂量。对 381 个前列腺癌病例进行了训练和验证,另外还对 40 个测试病例进行了独立的模型性能评估。建议的模型能在 15 毫秒内估算出每位患者的剂量。在测试患者中,估计剂量的平均γ通过率(3 % / 2 mm,10 %阈值)为 99.9% ± 0.15%。规划靶体积和危险器官的平均剂量差异分别为 0.07 % ± 0.34 % 和 0.48 % ± 0.72 %。我们开发出了一种用于精确剂量估算的几何分辨 DL 框架,并证明了其在实时在线 ART 剂量验证中的潜力。
{"title":"Quality assurance for online adaptive radiotherapy: a secondary dose verification model with geometry-encoded U-Net.","authors":"Shunyu Yan, Austen Maniscalco, Biling Wang, Dan Nguyen, Steve Jiang, Chenyang Shen","doi":"10.1088/2632-2153/ad829e","DOIUrl":"10.1088/2632-2153/ad829e","url":null,"abstract":"<p><p>In online adaptive radiotherapy (ART), quick computation-based secondary dose verification is crucial for ensuring the quality of ART plans while the patient is positioned on the treatment couch. However, traditional dose verification algorithms are generally time-consuming, reducing the efficiency of ART workflow. This study aims to develop an ultra-fast deep-learning (DL) based secondary dose verification algorithm to accurately estimate dose distributions using computed tomography (CT) and fluence maps (FMs). We integrated FMs into the CT image domain by explicitly resolving the geometry of treatment delivery. For each gantry angle, an FM was constructed based on the optimized multi-leaf collimator apertures and corresponding monitoring units. To effectively encode treatment beam configuration, the constructed FMs were back-projected to <math><mrow><mn>30</mn></mrow> </math> cm away from the isocenter with respect to the exact geometry of the treatment machines. Then, a 3D U-Net was utilized to take the integrated CT and FM volume as input to estimate dose. Training and validation were performed on <math><mrow><mn>381</mn></mrow> </math> prostate cancer cases, with an additional <math><mrow><mn>40</mn></mrow> </math> testing cases for independent evaluation of model performance. The proposed model can estimate dose in ∼ <math><mrow><mn>15</mn></mrow> </math> ms for each patient. The average <i>γ</i> passing rate ( <math><mrow><mn>3</mn> <mi>%</mi> <mrow><mo>/</mo></mrow> <mn>2</mn> <mstyle></mstyle> <mrow><mtext>mm</mtext></mrow> </mrow> </math> , <math><mrow><mn>10</mn> <mi>%</mi></mrow> </math> threshold) for the estimated dose was 99.9% ± 0.15% on testing patients. The mean dose differences for the planning target volume and organs at risk were <math><mrow><mn>0.07</mn> <mi>%</mi> <mo>±</mo> <mn>0.34</mn> <mi>%</mi></mrow> </math> and <math><mrow><mn>0.48</mn> <mi>%</mi> <mo>±</mo> <mn>0.72</mn> <mi>%</mi></mrow> </math> , respectively. We have developed a geometry-resolved DL framework for accurate dose estimation and demonstrated its potential in real-time online ART doses verification.</p>","PeriodicalId":33757,"journal":{"name":"Machine Learning Science and Technology","volume":"5 4","pages":"045013"},"PeriodicalIF":6.3,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11467776/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142476443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Machine Learning Science and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1