首页 > 最新文献

Nature Machine Intelligence最新文献

英文 中文
Error-controlled non-additive interaction discovery in machine learning models 机器学习模型中误差控制的非加性交互发现
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-22 DOI: 10.1038/s42256-025-01086-8
Winston Chen, Yifan Jiang, William Stafford Noble, Yang Young Lu
Machine learning (ML) models are powerful tools for detecting complex patterns, yet their ‘black-box’ nature limits their interpretability, hindering their use in critical domains like healthcare and finance. Interpretable ML methods aim to explain how features influence model predictions but often focus on univariate feature importance, overlooking complex feature interactions. Although recent efforts extend interpretability to feature interactions, existing approaches struggle with robustness and error control, especially under data perturbations. In this study, we introduce Diamond, a method for trustworthy feature interaction discovery. Diamond uniquely integrates the model-X knockoffs framework to control the false discovery rate, ensuring a low proportion of falsely detected interactions. Diamond includes a non-additivity distillation procedure that refines existing interaction importance measures to isolate non-additive interaction effects and preserve false discovery rate control. This approach addresses the limitations of off-the-shelf interaction measures, which, when used naively, can lead to inaccurate discoveries. Diamond’s applicability spans a broad class of ML models, including deep neural networks, transformers, tree-based models and factorization-based models. Empirical evaluations on both simulated and real datasets across various biomedical studies demonstrate its utility in enabling reliable data-driven scientific discoveries. Diamond represents a significant step forward in leveraging ML for scientific innovation and hypothesis generation. Diamond, a statistically rigorous method, is capable of finding meaningful feature interactions within machine learning models, making black-box models more interpretable for science and medicine.
机器学习(ML)模型是检测复杂模式的强大工具,但它们的“黑箱”性质限制了它们的可解释性,阻碍了它们在医疗保健和金融等关键领域的应用。可解释的机器学习方法旨在解释特征如何影响模型预测,但往往侧重于单变量特征的重要性,忽略了复杂的特征相互作用。虽然最近的努力将可解释性扩展到特征交互,但现有的方法在鲁棒性和误差控制方面存在困难,特别是在数据扰动下。在本研究中,我们引入了一种可信特征交互发现方法Diamond。Diamond独特地集成了model-X仿冒框架,以控制错误发现率,确保低比例的错误检测交互。Diamond包括一个非加性蒸馏程序,该程序改进了现有的相互作用重要性度量,以隔离非加性相互作用效应并保持错误发现率控制。这种方法解决了现成的交互度量的局限性,如果使用不当,可能会导致不准确的发现。Diamond的适用性涵盖了广泛的机器学习模型,包括深度神经网络、变压器、基于树的模型和基于分解的模型。对各种生物医学研究的模拟和真实数据集的经验评估表明,它在实现可靠的数据驱动的科学发现方面具有实用价值。戴蒙德代表了利用机器学习进行科学创新和假设生成的重要一步。Diamond是一种统计严谨的方法,能够在机器学习模型中找到有意义的特征交互,使黑盒模型更易于科学和医学解释。
{"title":"Error-controlled non-additive interaction discovery in machine learning models","authors":"Winston Chen, Yifan Jiang, William Stafford Noble, Yang Young Lu","doi":"10.1038/s42256-025-01086-8","DOIUrl":"10.1038/s42256-025-01086-8","url":null,"abstract":"Machine learning (ML) models are powerful tools for detecting complex patterns, yet their ‘black-box’ nature limits their interpretability, hindering their use in critical domains like healthcare and finance. Interpretable ML methods aim to explain how features influence model predictions but often focus on univariate feature importance, overlooking complex feature interactions. Although recent efforts extend interpretability to feature interactions, existing approaches struggle with robustness and error control, especially under data perturbations. In this study, we introduce Diamond, a method for trustworthy feature interaction discovery. Diamond uniquely integrates the model-X knockoffs framework to control the false discovery rate, ensuring a low proportion of falsely detected interactions. Diamond includes a non-additivity distillation procedure that refines existing interaction importance measures to isolate non-additive interaction effects and preserve false discovery rate control. This approach addresses the limitations of off-the-shelf interaction measures, which, when used naively, can lead to inaccurate discoveries. Diamond’s applicability spans a broad class of ML models, including deep neural networks, transformers, tree-based models and factorization-based models. Empirical evaluations on both simulated and real datasets across various biomedical studies demonstrate its utility in enabling reliable data-driven scientific discoveries. Diamond represents a significant step forward in leveraging ML for scientific innovation and hypothesis generation. Diamond, a statistically rigorous method, is capable of finding meaningful feature interactions within machine learning models, making black-box models more interpretable for science and medicine.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1541-1554"},"PeriodicalIF":23.9,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.comhttps://www.nature.com/articles/s42256-025-01086-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145129499","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modelling neural coding in the auditory midbrain with high resolution and accuracy 高分辨率、高准确度的中脑听觉神经编码建模
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-18 DOI: 10.1038/s42256-025-01104-9
Fotios Drakopoulos, Lloyd Pellatt, Shievanie Sabesan, Yiqing Xia, Andreas Fragner, Nicholas A. Lesica
Computational models of auditory processing can be valuable tools for research and technology development. Models of the cochlea are highly accurate and widely used, but models of the auditory brain lag far behind in both performance and penetration. Here we present ICNet, a convolutional encoder–decoder model of neural coding in the inferior colliculus. We developed ICNet using large-scale intracranial recordings from anaesthetized gerbils, addressing three key modelling challenges that are common across all sensory systems: capturing the full statistical structure of neuronal response patterns; accounting for physiological and experimental non-stationarity; and extracting features of sensory processing that are shared across different brains. ICNet provides highly accurate simulation of multi-unit neural responses to a wide range of complex sounds, including near-perfect responses to speech. It also reproduces key neurophysiological phenomena such as forward masking and dynamic range adaptation. ICNet can be used to simulate activity from thousands of neural units or to provide a compact representation of early central auditory processing through its latent dynamics, facilitating a wide range of hearing and audio applications. It can also serve as a foundation core, providing a baseline neural representation for models of active listening or higher-level auditory processing. Drakopoulos et al. present a model that captures the transformation from sound waves to neural activity patterns underlying early auditory processing. The model reproduces neural responses to a range of complex sounds and key neurophysiological phenomena.
听觉处理的计算模型可以成为研究和技术开发的宝贵工具。耳蜗模型精度高,应用广泛,但听觉脑模型在性能和穿透性方面都远远落后。在这里,我们提出了ICNet,一个卷积编码器-解码器模型的神经编码在下丘。我们使用麻醉沙鼠的大规模颅内记录开发了ICNet,解决了所有感觉系统中常见的三个关键建模挑战:捕获神经元反应模式的完整统计结构;考虑生理和实验的非平稳性;提取不同大脑共有的感觉处理特征。ICNet提供了对各种复杂声音的多单元神经反应的高度精确模拟,包括对语音的近乎完美的反应。它还再现了关键的神经生理现象,如前向掩蔽和动态范围适应。ICNet可用于模拟来自数千个神经单元的活动,或通过其潜在动态提供早期中枢听觉处理的紧凑表示,促进广泛的听力和音频应用。它也可以作为基础核心,为主动倾听模型或更高层次的听觉处理提供基线神经表征。Drakopoulos等人提出了一个模型,该模型捕捉了从声波到早期听觉处理背后的神经活动模式的转换。该模型再现了对一系列复杂声音和关键神经生理现象的神经反应。
{"title":"Modelling neural coding in the auditory midbrain with high resolution and accuracy","authors":"Fotios Drakopoulos, Lloyd Pellatt, Shievanie Sabesan, Yiqing Xia, Andreas Fragner, Nicholas A. Lesica","doi":"10.1038/s42256-025-01104-9","DOIUrl":"10.1038/s42256-025-01104-9","url":null,"abstract":"Computational models of auditory processing can be valuable tools for research and technology development. Models of the cochlea are highly accurate and widely used, but models of the auditory brain lag far behind in both performance and penetration. Here we present ICNet, a convolutional encoder–decoder model of neural coding in the inferior colliculus. We developed ICNet using large-scale intracranial recordings from anaesthetized gerbils, addressing three key modelling challenges that are common across all sensory systems: capturing the full statistical structure of neuronal response patterns; accounting for physiological and experimental non-stationarity; and extracting features of sensory processing that are shared across different brains. ICNet provides highly accurate simulation of multi-unit neural responses to a wide range of complex sounds, including near-perfect responses to speech. It also reproduces key neurophysiological phenomena such as forward masking and dynamic range adaptation. ICNet can be used to simulate activity from thousands of neural units or to provide a compact representation of early central auditory processing through its latent dynamics, facilitating a wide range of hearing and audio applications. It can also serve as a foundation core, providing a baseline neural representation for models of active listening or higher-level auditory processing. Drakopoulos et al. present a model that captures the transformation from sound waves to neural activity patterns underlying early auditory processing. The model reproduces neural responses to a range of complex sounds and key neurophysiological phenomena.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1478-1493"},"PeriodicalIF":23.9,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.comhttps://www.nature.com/articles/s42256-025-01104-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145129516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Author Correction: Deep learning-based prediction of the selection factors for quantifying selection in immune receptor repertoires 作者更正:基于深度学习的选择因子预测,用于量化免疫受体库的选择
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-17 DOI: 10.1038/s42256-025-01128-1
Yuepeng Jiang, Pingping Zhang, Miaozhe Huo, Shuai Cheng Li
{"title":"Author Correction: Deep learning-based prediction of the selection factors for quantifying selection in immune receptor repertoires","authors":"Yuepeng Jiang, Pingping Zhang, Miaozhe Huo, Shuai Cheng Li","doi":"10.1038/s42256-025-01128-1","DOIUrl":"10.1038/s42256-025-01128-1","url":null,"abstract":"","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1587-1587"},"PeriodicalIF":23.9,"publicationDate":"2025-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.comhttps://www.nature.com/articles/s42256-025-01128-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145129498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Aligning generalization between humans and machines 调整人类和机器之间的泛化
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-15 DOI: 10.1038/s42256-025-01109-4
Filip Ilievski, Barbara Hammer, Frank van Harmelen, Benjamin Paassen, Sascha Saralajew, Ute Schmid, Michael Biehl, Marianna Bolognesi, Xin Luna Dong, Kiril Gashteovski, Pascal Hitzler, Giuseppe Marra, Pasquale Minervini, Martin Mundt, Axel-Cyrille Ngonga Ngomo, Alessandro Oltramari, Gabriella Pasi, Zeynep G. Saribatur, Luciano Serafini, John Shawe-Taylor, Vered Shwartz, Gabriella Skitalinskaya, Clemens Stachl, Gido M. van de Ven, Thomas Villmann
Recent advances in artificial intelligence (AI)—including generative approaches—have resulted in technology that can support humans in scientific discovery and forming decisions, but may also disrupt democracies and target individuals. The responsible use of AI and its participation in human–AI teams increasingly shows the need for AI alignment, that is, to make AI systems act according to our preferences. A crucial yet often overlooked aspect of these interactions is the different ways in which humans and machines generalize. In cognitive science, human generalization commonly involves abstraction and concept learning. By contrast, AI generalization encompasses out-of-domain generalization in machine learning, rule-based reasoning in symbolic AI, and abstraction in neurosymbolic AI. Here we combine insights from AI and cognitive science to identify key commonalities and differences across three dimensions: notions of, methods for, and evaluation of generalization. We map the different conceptualizations of generalization in AI and cognitive science along these three dimensions and consider their role for alignment in human–AI teaming. This results in interdisciplinary challenges across AI and cognitive science that must be tackled to support effective and cognitively supported alignment in human–AI teaming scenarios. Ilievski et al. examine differences and similarities in the various ways human and AI systems generalize. The insights are important for effectively supporting alignment in human–AI teams.
人工智能(AI)的最新进展-包括生成方法-已经产生了可以支持人类进行科学发现和形成决策的技术,但也可能破坏民主并针对个人。负责任地使用人工智能及其参与人类-人工智能团队越来越多地显示出人工智能一致性的必要性,也就是说,让人工智能系统根据我们的偏好行事。在这些互动中,一个重要但经常被忽视的方面是人类和机器进行泛化的不同方式。在认知科学中,人类的泛化通常涉及抽象和概念学习。相比之下,人工智能泛化包括机器学习中的域外泛化、符号人工智能中的基于规则的推理和神经符号人工智能中的抽象。在这里,我们结合了人工智能和认知科学的见解,以确定三个方面的关键共性和差异:泛化的概念、方法和评估。我们沿着这三个维度绘制了人工智能和认知科学中泛化的不同概念,并考虑了它们在人类-人工智能团队中的作用。这导致了人工智能和认知科学之间的跨学科挑战,必须解决这些挑战,以支持人类-人工智能团队场景中有效和认知支持的一致性。
{"title":"Aligning generalization between humans and machines","authors":"Filip Ilievski, Barbara Hammer, Frank van Harmelen, Benjamin Paassen, Sascha Saralajew, Ute Schmid, Michael Biehl, Marianna Bolognesi, Xin Luna Dong, Kiril Gashteovski, Pascal Hitzler, Giuseppe Marra, Pasquale Minervini, Martin Mundt, Axel-Cyrille Ngonga Ngomo, Alessandro Oltramari, Gabriella Pasi, Zeynep G. Saribatur, Luciano Serafini, John Shawe-Taylor, Vered Shwartz, Gabriella Skitalinskaya, Clemens Stachl, Gido M. van de Ven, Thomas Villmann","doi":"10.1038/s42256-025-01109-4","DOIUrl":"10.1038/s42256-025-01109-4","url":null,"abstract":"Recent advances in artificial intelligence (AI)—including generative approaches—have resulted in technology that can support humans in scientific discovery and forming decisions, but may also disrupt democracies and target individuals. The responsible use of AI and its participation in human–AI teams increasingly shows the need for AI alignment, that is, to make AI systems act according to our preferences. A crucial yet often overlooked aspect of these interactions is the different ways in which humans and machines generalize. In cognitive science, human generalization commonly involves abstraction and concept learning. By contrast, AI generalization encompasses out-of-domain generalization in machine learning, rule-based reasoning in symbolic AI, and abstraction in neurosymbolic AI. Here we combine insights from AI and cognitive science to identify key commonalities and differences across three dimensions: notions of, methods for, and evaluation of generalization. We map the different conceptualizations of generalization in AI and cognitive science along these three dimensions and consider their role for alignment in human–AI teaming. This results in interdisciplinary challenges across AI and cognitive science that must be tackled to support effective and cognitively supported alignment in human–AI teaming scenarios. Ilievski et al. examine differences and similarities in the various ways human and AI systems generalize. The insights are important for effectively supporting alignment in human–AI teams.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1378-1389"},"PeriodicalIF":23.9,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145059570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Target-specific de novo design of drug candidate molecules with graph-transformer-based generative adversarial networks 基于图转换器的生成对抗网络的候选药物分子靶向从头设计
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-15 DOI: 10.1038/s42256-025-01082-y
Atabey Ünlü, Elif Çevrim, Melih Gökay Yiğit, Ahmet Sarıgün, Hayriye Çelikbilek, Osman Bayram, Deniz Cansen Kahraman, Abdurrahman Olğaç, Ahmet Sureyya Rifaioglu, Erden Banoğlu, Tunca Doğan
Discovering novel drug candidate molecules is a fundamental step in drug development. Generative deep learning models can sample new molecular structures from learned probability distributions; however, their practical use in drug discovery hinges on generating compounds tailored to a specific target molecule. Here we introduce DrugGEN, an end-to-end generative system for the de novo design of drug candidate molecules that interact with a selected protein. The proposed method represents molecules as graphs and processes them using a generative adversarial network that comprises graph transformer layers. Trained on large datasets of drug-like compounds and target-specific bioactive molecules, DrugGEN designed candidate inhibitors for AKT1, a kinase crucial in many cancers. Docking and molecular dynamics simulations suggest that the generated compounds effectively bind to AKT1, and attention maps provide insights into the model’s reasoning. Furthermore, selected de novo molecules were synthesized and shown to inhibit AKT1 at low micromolar concentrations in the context of in vitro enzymatic assays. These results demonstrate the potential of DrugGEN for designing target-specific molecules. Using the open-access DrugGEN codebase, researchers can retrain the model for other druggable proteins, provided a dataset of known bioactive molecules is available. Inhibiting AKT1 kinase can have potentially positive uses against many types of cancer. To find novel molecules targeting this protein, a graph adversarial network is trained as a generative model.
发现新的候选药物分子是药物开发的基本步骤。生成式深度学习模型可以从学习到的概率分布中采样新的分子结构;然而,它们在药物发现中的实际应用取决于生成针对特定目标分子的化合物。在这里,我们介绍DrugGEN,一个端到端生成系统,用于与选定蛋白质相互作用的候选药物分子的从头设计。该方法将分子表示为图形,并使用由图形转换层组成的生成对抗网络对其进行处理。通过对药物样化合物和靶向特异性生物活性分子的大量数据集的训练,DrugGEN设计了AKT1(一种对许多癌症至关重要的激酶)的候选抑制剂。对接和分子动力学模拟表明,生成的化合物有效地与AKT1结合,注意图为模型的推理提供了见解。此外,选择的新生分子被合成,并在体外酶分析的背景下显示出在低微摩尔浓度下抑制AKT1。这些结果证明了DrugGEN在设计靶向分子方面的潜力。使用开放获取的DrugGEN代码库,研究人员可以为其他可药物蛋白质重新训练模型,前提是提供已知生物活性分子的数据集。
{"title":"Target-specific de novo design of drug candidate molecules with graph-transformer-based generative adversarial networks","authors":"Atabey Ünlü, Elif Çevrim, Melih Gökay Yiğit, Ahmet Sarıgün, Hayriye Çelikbilek, Osman Bayram, Deniz Cansen Kahraman, Abdurrahman Olğaç, Ahmet Sureyya Rifaioglu, Erden Banoğlu, Tunca Doğan","doi":"10.1038/s42256-025-01082-y","DOIUrl":"10.1038/s42256-025-01082-y","url":null,"abstract":"Discovering novel drug candidate molecules is a fundamental step in drug development. Generative deep learning models can sample new molecular structures from learned probability distributions; however, their practical use in drug discovery hinges on generating compounds tailored to a specific target molecule. Here we introduce DrugGEN, an end-to-end generative system for the de novo design of drug candidate molecules that interact with a selected protein. The proposed method represents molecules as graphs and processes them using a generative adversarial network that comprises graph transformer layers. Trained on large datasets of drug-like compounds and target-specific bioactive molecules, DrugGEN designed candidate inhibitors for AKT1, a kinase crucial in many cancers. Docking and molecular dynamics simulations suggest that the generated compounds effectively bind to AKT1, and attention maps provide insights into the model’s reasoning. Furthermore, selected de novo molecules were synthesized and shown to inhibit AKT1 at low micromolar concentrations in the context of in vitro enzymatic assays. These results demonstrate the potential of DrugGEN for designing target-specific molecules. Using the open-access DrugGEN codebase, researchers can retrain the model for other druggable proteins, provided a dataset of known bioactive molecules is available. Inhibiting AKT1 kinase can have potentially positive uses against many types of cancer. To find novel molecules targeting this protein, a graph adversarial network is trained as a generative model.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1524-1540"},"PeriodicalIF":23.9,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145059571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sampling-enabled scalable manifold learning unveils the discriminative cluster structure of high-dimensional data 支持采样的可扩展流形学习揭示了高维数据的判别聚类结构
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-10 DOI: 10.1038/s42256-025-01112-9
Dehua Peng, Zhipeng Gui, Wenzhang Wei, Fa Li, Jie Gui, Huayi Wu, Jianya Gong
As a pivotal branch of machine learning, manifold learning uncovers the intrinsic low-dimensional structure within complex non-linear manifolds in high-dimensional space for visualization, classification, clustering and gaining key insights. Although existing techniques have achieved remarkable successes, they suffer from extensive distortions of cluster structure, which hinders the understanding of underlying patterns. Scalability issues also limit their applicability for handling large-scale data. Here we propose a sampling-based scalable manifold learning technique that enables uniform and discriminative embedding (SUDE) for large-scale and high-dimensional data. It starts by seeking a set of landmarks to construct the low-dimensional skeleton of the entire data and then incorporates the non-landmarks into the learned space by constrained locally linear embedding. We empirically validated the effectiveness of SUDE on synthetic datasets and real-world benchmarks and applied it to analyse single-cell data and detect anomalies in electrocardiogram signals. SUDE exhibits a distinct advantage in scalability with respect to data size and embedding dimension and shows promising performance in cluster separation, integrity and global structure preservation. The experiments also demonstrate notable robustness in embedding quality as the sampling rate decreases. A sampling-based manifold learning method is proposed to study the cluster structure of high-dimensional data. Its applicability and scalability have been verified in single-cell data analysis and anomaly detection in electrocardiogram signals.
流形学习是机器学习的一个关键分支,它揭示了高维空间中复杂非线性流形内在的低维结构,用于可视化、分类、聚类和获得关键见解。虽然现有的技术已经取得了显著的成功,但它们受到团簇结构的广泛扭曲的影响,这阻碍了对潜在模式的理解。可伸缩性问题也限制了它们处理大规模数据的适用性。在这里,我们提出了一种基于采样的可扩展流形学习技术,该技术可以实现大规模和高维数据的均匀和判别嵌入(SUDE)。它首先寻找一组地标来构建整个数据的低维骨架,然后通过约束局部线性嵌入将非地标纳入学习空间。我们在合成数据集和现实世界基准上验证了SUDE的有效性,并将其应用于分析单细胞数据和检测心电图信号中的异常。SUDE在数据大小和嵌入维度方面具有明显的可伸缩性优势,并且在簇分离、完整性和全局结构保存方面表现出良好的性能。实验还表明,随着采样率的降低,嵌入质量具有显著的鲁棒性。
{"title":"Sampling-enabled scalable manifold learning unveils the discriminative cluster structure of high-dimensional data","authors":"Dehua Peng, Zhipeng Gui, Wenzhang Wei, Fa Li, Jie Gui, Huayi Wu, Jianya Gong","doi":"10.1038/s42256-025-01112-9","DOIUrl":"10.1038/s42256-025-01112-9","url":null,"abstract":"As a pivotal branch of machine learning, manifold learning uncovers the intrinsic low-dimensional structure within complex non-linear manifolds in high-dimensional space for visualization, classification, clustering and gaining key insights. Although existing techniques have achieved remarkable successes, they suffer from extensive distortions of cluster structure, which hinders the understanding of underlying patterns. Scalability issues also limit their applicability for handling large-scale data. Here we propose a sampling-based scalable manifold learning technique that enables uniform and discriminative embedding (SUDE) for large-scale and high-dimensional data. It starts by seeking a set of landmarks to construct the low-dimensional skeleton of the entire data and then incorporates the non-landmarks into the learned space by constrained locally linear embedding. We empirically validated the effectiveness of SUDE on synthetic datasets and real-world benchmarks and applied it to analyse single-cell data and detect anomalies in electrocardiogram signals. SUDE exhibits a distinct advantage in scalability with respect to data size and embedding dimension and shows promising performance in cluster separation, integrity and global structure preservation. The experiments also demonstrate notable robustness in embedding quality as the sampling rate decreases. A sampling-based manifold learning method is proposed to study the cluster structure of high-dimensional data. Its applicability and scalability have been verified in single-cell data analysis and anomaly detection in electrocardiogram signals.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 10","pages":"1669-1684"},"PeriodicalIF":23.9,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145025291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards agentic science for advancing scientific discovery 走向代理科学,推进科学发现
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-10 DOI: 10.1038/s42256-025-01110-x
Hongliang Xin, John R. Kitchin, Heather J. Kulik
Artificial intelligence is transforming scientific discovery through (semi-)autonomous agents capable of reasoning, planning, and interacting with digital and physical environments. This Comment explores the foundations and frontiers of agentic science, outlining its emerging directions, current limitations, and the pathways for responsible integration into scientific practice.
人工智能正在通过能够推理、规划以及与数字和物理环境交互的(半)自主代理改变科学发现。本评论探讨了自主科学的基础和前沿,概述了其新兴方向、当前的局限性以及将其负责任地纳入科学实践的途径。
{"title":"Towards agentic science for advancing scientific discovery","authors":"Hongliang Xin, John R. Kitchin, Heather J. Kulik","doi":"10.1038/s42256-025-01110-x","DOIUrl":"10.1038/s42256-025-01110-x","url":null,"abstract":"Artificial intelligence is transforming scientific discovery through (semi-)autonomous agents capable of reasoning, planning, and interacting with digital and physical environments. This Comment explores the foundations and frontiers of agentic science, outlining its emerging directions, current limitations, and the pathways for responsible integration into scientific practice.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1373-1375"},"PeriodicalIF":23.9,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145025286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real-world validation of a structure-aware pipeline for molecular design 分子设计的结构感知管道的实际验证
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-08 DOI: 10.1038/s42256-025-01102-x
Ana Laura Dias, Tiago Rodrigues
The next major challenge for artificial intelligence in drug development lies in proving its value in real-world settings. A new technology not only supports the generation of novel chemical entities but also accelerates a range of real-world molecular design tasks.
人工智能在药物开发中的下一个主要挑战在于证明其在现实世界中的价值。一项新技术不仅支持新的化学实体的产生,而且还加速了一系列现实世界的分子设计任务。
{"title":"Real-world validation of a structure-aware pipeline for molecular design","authors":"Ana Laura Dias, Tiago Rodrigues","doi":"10.1038/s42256-025-01102-x","DOIUrl":"10.1038/s42256-025-01102-x","url":null,"abstract":"The next major challenge for artificial intelligence in drug development lies in proving its value in real-world settings. A new technology not only supports the generation of novel chemical entities but also accelerates a range of real-world molecular design tasks.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1376-1377"},"PeriodicalIF":23.9,"publicationDate":"2025-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145009022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Conditional generation of real antigen-specific T cell receptor sequences 条件生成真正的抗原特异性T细胞受体序列
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-08 DOI: 10.1038/s42256-025-01096-6
Dhuvarakesh Karthikeyan, Sarah N. Bennett, Amy G. Reynolds, Benjamin G. Vincent, Alex Rubinsteyn
Despite recent advances in T cell receptor (TCR) engineering, designing functional TCRs against arbitrary targets remains challenging due to complex rules governing cross-reactivity and limited paired data. Here we present TCR-TRANSLATE, a sequence-to-sequence framework that adapts low-resource machine translation techniques to generate antigen-specific TCR sequences against unseen epitopes. By evaluating 12 model variants of the BART and T5 model architectures, we identified key factors affecting performance and utility, revealing discordances between these objectives. Our flagship model, TCRT5, outperforms existing approaches on computational benchmarks, prioritizing functionally relevant sequences at higher ranks. Most significantly, we experimentally validated a computationally designed TCR against Wilms’ tumour antigen, a therapeutically relevant target in leukaemia, excluded from our training and validation sets. Although the identified TCR shows cross-reactivity with pathogen-derived peptides, highlighting limitations in specificity, our work represents the successful computational design of a functional TCR construct against a non-viral epitope from the target sequence alone. Our findings establish a foundation for computational TCR design and reveal current limitations in data availability and methodology, providing a framework for accelerating personalized immunotherapy by reducing the search space for novel targets. TCR-TRANSLATE, a deep learning framework adapting machine translation to immune design, demonstrates the successful generation of a functional T cell receptor sequence for a cancer epitope from the target sequence alone.
尽管最近在T细胞受体(TCR)工程方面取得了进展,但由于控制交叉反应性的复杂规则和有限的配对数据,设计针对任意靶标的功能性TCR仍然具有挑战性。在这里,我们提出了TCR- translate,这是一个序列到序列的框架,它适应低资源机器翻译技术来生成针对未见表位的抗原特异性TCR序列。通过评估BART和T5模型架构的12个模型变体,我们确定了影响性能和效用的关键因素,揭示了这些目标之间的不一致。我们的旗舰模型TCRT5在计算基准上优于现有方法,在更高的等级上优先考虑功能相关序列。最重要的是,我们通过实验验证了计算设计的针对Wilms肿瘤抗原的TCR, Wilms肿瘤抗原是白血病的治疗相关靶标,排除在我们的训练和验证集之外。虽然鉴定出的TCR显示出与病原体衍生肽的交叉反应性,突出了特异性的局限性,但我们的工作代表了仅针对靶序列的非病毒表位的功能性TCR结构的成功计算设计。我们的研究结果为计算TCR设计奠定了基础,揭示了当前数据可用性和方法的局限性,为通过减少新靶点的搜索空间来加速个性化免疫治疗提供了框架。
{"title":"Conditional generation of real antigen-specific T cell receptor sequences","authors":"Dhuvarakesh Karthikeyan, Sarah N. Bennett, Amy G. Reynolds, Benjamin G. Vincent, Alex Rubinsteyn","doi":"10.1038/s42256-025-01096-6","DOIUrl":"10.1038/s42256-025-01096-6","url":null,"abstract":"Despite recent advances in T cell receptor (TCR) engineering, designing functional TCRs against arbitrary targets remains challenging due to complex rules governing cross-reactivity and limited paired data. Here we present TCR-TRANSLATE, a sequence-to-sequence framework that adapts low-resource machine translation techniques to generate antigen-specific TCR sequences against unseen epitopes. By evaluating 12 model variants of the BART and T5 model architectures, we identified key factors affecting performance and utility, revealing discordances between these objectives. Our flagship model, TCRT5, outperforms existing approaches on computational benchmarks, prioritizing functionally relevant sequences at higher ranks. Most significantly, we experimentally validated a computationally designed TCR against Wilms’ tumour antigen, a therapeutically relevant target in leukaemia, excluded from our training and validation sets. Although the identified TCR shows cross-reactivity with pathogen-derived peptides, highlighting limitations in specificity, our work represents the successful computational design of a functional TCR construct against a non-viral epitope from the target sequence alone. Our findings establish a foundation for computational TCR design and reveal current limitations in data availability and methodology, providing a framework for accelerating personalized immunotherapy by reducing the search space for novel targets. TCR-TRANSLATE, a deep learning framework adapting machine translation to immune design, demonstrates the successful generation of a functional T cell receptor sequence for a cancer epitope from the target sequence alone.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 9","pages":"1494-1509"},"PeriodicalIF":23.9,"publicationDate":"2025-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.comhttps://www.nature.com/articles/s42256-025-01096-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145009024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards compute-efficient Byzantine-robust federated learning with fully homomorphic encryption 面向计算效率高、具有完全同态加密的拜占庭鲁棒联邦学习
IF 23.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-08 DOI: 10.1038/s42256-025-01107-6
Siyang Jiang, Hao Yang, Qipeng Xie, Chuan Ma, Sen Wang, Zhe Liu, Tao Xiang, Guoliang Xing
In highly regulated domains such as finance and healthcare, where stringent data-sharing constraints pose substantial obstacles, federated learning (FL) has emerged as a transformative paradigm in distributed machine learning, facilitating collaborative model training, preserving data decentralization and upholding governance standards. Despite its advantages, FL is vulnerable to poisoning attacks during central model aggregation, prompting the development of Byzantine-robust FL systems that use robust aggregation rules to counter malicious attacks. However, neural network models in such systems are susceptible to unintentionally memorizing and revealing individual training instances, thereby introducing substantial information leakage risks, as adversaries may exploit this vulnerability to reconstruct sensitive data through model outputs transmitted over the air. Existing solutions fall short of providing a viable Byzantine-robust FL system that is completely secure against information leakage and is computationally efficient. To address these concerns, we propose Lancelot, an efficient and effective Byzantine-robust FL framework that uses fully homomorphic encryption to safeguard against malicious client activities. Lancelot introduces a mask-based encrypted sorting mechanism that overcomes the limitations of multiplication depth in ciphertext sorting with zero information leakage. It incorporates cryptographic enhancements like lazy relinearization, dynamic hoisting and GPU acceleration to ensure practical computational efficiency. Extensive experiments demonstrate that Lancelot surpasses existing approaches, achieving a 20-fold enhancement in processing speed. Lancelot, a compute-efficient federated learning framework using homomorphic encryption to prevent information leakage, is presented, achieving 20 times faster processing speeds through advanced cryptographic and encrypted sorting techniques.
在金融和医疗保健等高度监管的领域,严格的数据共享限制构成了重大障碍,联邦学习(FL)已成为分布式机器学习的一种变革性范例,促进了协作模型培训,保持了数据去中心化并维护了治理标准。尽管有这些优点,但在中心模型聚合过程中,FL很容易受到中毒攻击,这促使了拜占庭鲁棒FL系统的发展,该系统使用鲁棒聚合规则来对抗恶意攻击。然而,这种系统中的神经网络模型容易在无意中记忆和泄露单个训练实例,从而引入大量信息泄露风险,因为攻击者可能利用这一漏洞通过空中传输的模型输出重建敏感数据。现有的解决方案无法提供一个可行的拜占庭鲁棒FL系统,该系统可以完全防止信息泄漏,并且计算效率高。为了解决这些问题,我们提出了Lancelot,这是一个高效且有效的拜占庭健壮的FL框架,它使用完全同态加密来防止恶意客户端活动。Lancelot引入了一种基于掩码的加密排序机制,克服了密文排序中乘法深度的限制,实现了零信息泄漏。它结合了延迟线性化、动态提升和GPU加速等加密增强功能,以确保实际的计算效率。大量的实验表明,Lancelot超越了现有的方法,将处理速度提高了20倍。
{"title":"Towards compute-efficient Byzantine-robust federated learning with fully homomorphic encryption","authors":"Siyang Jiang, Hao Yang, Qipeng Xie, Chuan Ma, Sen Wang, Zhe Liu, Tao Xiang, Guoliang Xing","doi":"10.1038/s42256-025-01107-6","DOIUrl":"10.1038/s42256-025-01107-6","url":null,"abstract":"In highly regulated domains such as finance and healthcare, where stringent data-sharing constraints pose substantial obstacles, federated learning (FL) has emerged as a transformative paradigm in distributed machine learning, facilitating collaborative model training, preserving data decentralization and upholding governance standards. Despite its advantages, FL is vulnerable to poisoning attacks during central model aggregation, prompting the development of Byzantine-robust FL systems that use robust aggregation rules to counter malicious attacks. However, neural network models in such systems are susceptible to unintentionally memorizing and revealing individual training instances, thereby introducing substantial information leakage risks, as adversaries may exploit this vulnerability to reconstruct sensitive data through model outputs transmitted over the air. Existing solutions fall short of providing a viable Byzantine-robust FL system that is completely secure against information leakage and is computationally efficient. To address these concerns, we propose Lancelot, an efficient and effective Byzantine-robust FL framework that uses fully homomorphic encryption to safeguard against malicious client activities. Lancelot introduces a mask-based encrypted sorting mechanism that overcomes the limitations of multiplication depth in ciphertext sorting with zero information leakage. It incorporates cryptographic enhancements like lazy relinearization, dynamic hoisting and GPU acceleration to ensure practical computational efficiency. Extensive experiments demonstrate that Lancelot surpasses existing approaches, achieving a 20-fold enhancement in processing speed. Lancelot, a compute-efficient federated learning framework using homomorphic encryption to prevent information leakage, is presented, achieving 20 times faster processing speeds through advanced cryptographic and encrypted sorting techniques.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"7 10","pages":"1657-1668"},"PeriodicalIF":23.9,"publicationDate":"2025-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145009023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Nature Machine Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1