首页 > 最新文献

Nature Machine Intelligence最新文献

英文 中文
Learning efficient backprojections across cortical hierarchies in real time 实时学习大脑皮层中的高效反向推演
IF 18.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-06 DOI: 10.1038/s42256-024-00845-3
Kevin Max, Laura Kriener, Garibaldi Pineda García, Thomas Nowotny, Ismael Jaras, Walter Senn, Mihai A. Petrovici
Models of sensory processing and learning in the cortex need to efficiently assign credit to synapses in all areas. In deep learning, a known solution is error backpropagation, which requires biologically implausible weight transport from feed-forwards to feedback paths. We introduce phaseless alignment learning, a bio-plausible method to learn efficient feedback weights in layered cortical hierarchies. This is achieved by exploiting the noise naturally found in biophysical systems as an additional carrier of information. In our dynamical system, all weights are learned simultaneously with always-on plasticity and using only information locally available to the synapses. Our method is completely phase-free (no forwards and backwards passes or phased learning) and allows for efficient error propagation across multi-layer cortical hierarchies, while maintaining biologically plausible signal transport and learning. Our method is applicable to a wide class of models and improves on previously known biologically plausible ways of credit assignment: compared to random synaptic feedback, it can solve complex tasks with fewer neurons and learn more useful latent representations. We demonstrate this on various classification tasks using a cortical microcircuit model with prospective coding. The credit assignment problem involves assigning credit to synapses in a neural network so that weights are updated appropriately and the circuit learns. Max et al. developed an efficient solution to the weight transport problem in networks of biophysical neurons. The method exploits noise as an information carrier and enables networks to learn to solve a task efficiently.
大脑皮层的感官处理和学习模型需要有效地为所有区域的突触分配信用。在深度学习中,一种已知的解决方案是误差反向传播,这需要从前馈到反馈路径的生物学上难以置信的权重传输。我们引入了无相位对齐学习,这是一种在分层皮层中学习高效反馈权重的生物拟合方法。这是通过利用生物物理系统中自然存在的噪声作为额外的信息载体来实现的。在我们的动态系统中,所有权重的学习都是同时进行的,具有始终开启的可塑性,并且只使用突触局部可用的信息。我们的方法完全无阶段性(无前后传递或阶段性学习),允许在多层皮质层次结构中进行高效的误差传播,同时保持生物学上合理的信号传输和学习。我们的方法适用于各类模型,并改进了之前已知的生物学上合理的信用分配方式:与随机突触反馈相比,它能用更少的神经元解决复杂的任务,并学习到更有用的潜在表征。我们利用具有前瞻性编码的皮层微电路模型,在各种分类任务中证明了这一点。
{"title":"Learning efficient backprojections across cortical hierarchies in real time","authors":"Kevin Max, Laura Kriener, Garibaldi Pineda García, Thomas Nowotny, Ismael Jaras, Walter Senn, Mihai A. Petrovici","doi":"10.1038/s42256-024-00845-3","DOIUrl":"10.1038/s42256-024-00845-3","url":null,"abstract":"Models of sensory processing and learning in the cortex need to efficiently assign credit to synapses in all areas. In deep learning, a known solution is error backpropagation, which requires biologically implausible weight transport from feed-forwards to feedback paths. We introduce phaseless alignment learning, a bio-plausible method to learn efficient feedback weights in layered cortical hierarchies. This is achieved by exploiting the noise naturally found in biophysical systems as an additional carrier of information. In our dynamical system, all weights are learned simultaneously with always-on plasticity and using only information locally available to the synapses. Our method is completely phase-free (no forwards and backwards passes or phased learning) and allows for efficient error propagation across multi-layer cortical hierarchies, while maintaining biologically plausible signal transport and learning. Our method is applicable to a wide class of models and improves on previously known biologically plausible ways of credit assignment: compared to random synaptic feedback, it can solve complex tasks with fewer neurons and learn more useful latent representations. We demonstrate this on various classification tasks using a cortical microcircuit model with prospective coding. The credit assignment problem involves assigning credit to synapses in a neural network so that weights are updated appropriately and the circuit learns. Max et al. developed an efficient solution to the weight transport problem in networks of biophysical neurons. The method exploits noise as an information carrier and enables networks to learn to solve a task efficiently.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 6","pages":"619-630"},"PeriodicalIF":18.8,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141264732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generic protein–ligand interaction scoring by integrating physical prior knowledge and data augmentation modelling 通过整合物理先验知识和数据增强建模进行通用蛋白质配体相互作用评分
IF 18.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-06 DOI: 10.1038/s42256-024-00849-z
Duanhua Cao, Geng Chen, Jiaxin Jiang, Jie Yu, Runze Zhang, Mingan Chen, Wei Zhang, Lifan Chen, Feisheng Zhong, Yingying Zhang, Chenghao Lu, Xutong Li, Xiaomin Luo, Sulin Zhang, Mingyue Zheng
Developing robust methods for evaluating protein–ligand interactions has been a long-standing problem. Data-driven methods may memorize ligand and protein training data rather than learning protein–ligand interactions. Here we show a scoring approach called EquiScore, which utilizes a heterogeneous graph neural network to integrate physical prior knowledge and characterize protein–ligand interactions in equivariant geometric space. EquiScore is trained based on a new dataset constructed with multiple data augmentation strategies and a stringent redundancy-removal scheme. On two large external test sets, EquiScore consistently achieved top-ranking performance compared to 21 other methods. When EquiScore is used alongside different docking methods, it can effectively enhance the screening ability of these docking methods. EquiScore also showed good performance on the activity-ranking task of a series of structural analogues, indicating its potential to guide lead compound optimization. Finally, we investigated different levels of interpretability of EquiScore, which may provide more insights into structure-based drug design. Machine learning can improve scoring methods to evaluate protein–ligand interactions, but achieving good generalization is an outstanding challenge. Cao et al. introduce EquiScore, which is based on a graph neural network that integrates physical knowledge and is shown to have robust capabilities when applied to unseen protein targets.
开发稳健的蛋白质配体相互作用评估方法是一个长期存在的问题。数据驱动的方法可能会记忆配体和蛋白质的训练数据,而不是学习蛋白质配体之间的相互作用。在这里,我们展示了一种名为 EquiScore 的评分方法,它利用异构图神经网络整合物理先验知识,在等变几何空间中描述蛋白质-配体的相互作用。EquiScore 基于一个新的数据集进行训练,该数据集采用了多种数据增强策略和严格的冗余去除方案。在两个大型外部测试集上,与其他 21 种方法相比,EquiScore 的性能始终名列前茅。当 EquiScore 与不同的对接方法一起使用时,它能有效地提高这些对接方法的筛选能力。EquiScore 还在一系列结构类似物的活性排名任务中表现出良好的性能,这表明它具有指导先导化合物优化的潜力。最后,我们研究了 EquiScore 不同层次的可解释性,这可能会为基于结构的药物设计提供更多启示。
{"title":"Generic protein–ligand interaction scoring by integrating physical prior knowledge and data augmentation modelling","authors":"Duanhua Cao, Geng Chen, Jiaxin Jiang, Jie Yu, Runze Zhang, Mingan Chen, Wei Zhang, Lifan Chen, Feisheng Zhong, Yingying Zhang, Chenghao Lu, Xutong Li, Xiaomin Luo, Sulin Zhang, Mingyue Zheng","doi":"10.1038/s42256-024-00849-z","DOIUrl":"10.1038/s42256-024-00849-z","url":null,"abstract":"Developing robust methods for evaluating protein–ligand interactions has been a long-standing problem. Data-driven methods may memorize ligand and protein training data rather than learning protein–ligand interactions. Here we show a scoring approach called EquiScore, which utilizes a heterogeneous graph neural network to integrate physical prior knowledge and characterize protein–ligand interactions in equivariant geometric space. EquiScore is trained based on a new dataset constructed with multiple data augmentation strategies and a stringent redundancy-removal scheme. On two large external test sets, EquiScore consistently achieved top-ranking performance compared to 21 other methods. When EquiScore is used alongside different docking methods, it can effectively enhance the screening ability of these docking methods. EquiScore also showed good performance on the activity-ranking task of a series of structural analogues, indicating its potential to guide lead compound optimization. Finally, we investigated different levels of interpretability of EquiScore, which may provide more insights into structure-based drug design. Machine learning can improve scoring methods to evaluate protein–ligand interactions, but achieving good generalization is an outstanding challenge. Cao et al. introduce EquiScore, which is based on a graph neural network that integrates physical knowledge and is shown to have robust capabilities when applied to unseen protein targets.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 6","pages":"688-700"},"PeriodicalIF":18.8,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141264733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed constrained combinatorial optimization leveraging hypergraph neural networks 利用超图神经网络进行分布式约束组合优化
IF 18.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-30 DOI: 10.1038/s42256-024-00833-7
Nasimeh Heydaribeni, Xinrui Zhan, Ruisi Zhang, Tina Eliassi-Rad, Farinaz Koushanfar
Scalable addressing of high-dimensional constrained combinatorial optimization problems is a challenge that arises in several science and engineering disciplines. Recent work introduced novel applications of graph neural networks for solving quadratic-cost combinatorial optimization problems. However, effective utilization of models such as graph neural networks to address general problems with higher-order constraints is an unresolved challenge. This paper presents a framework, HypOp, that advances the state of the art for solving combinatorial optimization problems in several aspects: (1) it generalizes the prior results to higher-order constrained problems with arbitrary cost functions by leveraging hypergraph neural networks; (2) it enables scalability to larger problems by introducing a new distributed and parallel training architecture; (3) it demonstrates generalizability across different problem formulations by transferring knowledge within the same hypergraph; (4) it substantially boosts the solution accuracy compared with the prior art by suggesting a fine-tuning step using simulated annealing; and (5) it shows remarkable progress on numerous benchmark examples, including hypergraph MaxCut, satisfiability and resource allocation problems, with notable run-time improvements using a combination of fine-tuning and distributed training techniques. We showcase the application of HypOp in scientific discovery by solving a hypergraph MaxCut problem on a National Drug Code drug-substance hypergraph. Through extensive experimentation on various optimization problems, HypOp demonstrates superiority over existing unsupervised-learning-based solvers and generic optimization methods. Bolstering the broad and deep applicability of graph neural networks, Heydaribeni et al. introduce HypOp, a framework that uses hypergraph neural networks to solve general constrained combinatorial optimization problems. The presented method scales and generalizes well, improves accuracy and outperforms existing solvers on various benchmarking examples.
可扩展地解决高维约束组合优化问题是多个科学和工程学科面临的挑战。最近的工作介绍了图神经网络在解决二次成本组合优化问题中的新应用。然而,如何有效利用图神经网络等模型来解决具有高阶约束的一般问题是一个尚未解决的难题。本文提出了一个名为 HypOp 的框架,从几个方面推进了组合优化问题的解决技术:(1) 它利用超图神经网络,将先前的成果推广到具有任意成本函数的高阶约束问题;(2) 它通过引入新的分布式并行训练架构,实现了对更大问题的可扩展性;(3) 它通过在同一超图中传递知识,展示了在不同问题表述中的通用性;(4) 通过建议使用模拟退火进行微调步骤,与现有技术相比,它大大提高了求解的准确性;以及 (5) 它在众多基准示例(包括超图 MaxCut、可满足性和资源分配问题)上取得了显著进展,结合使用微调和分布式训练技术,在运行时间上有了明显改善。我们通过解决国家药品编码药物物质超图上的超图 MaxCut 问题,展示了 HypOp 在科学发现领域的应用。通过对各种优化问题的广泛实验,HypOp 证明了其优于现有的基于无监督学习的求解器和通用优化方法。
{"title":"Distributed constrained combinatorial optimization leveraging hypergraph neural networks","authors":"Nasimeh Heydaribeni, Xinrui Zhan, Ruisi Zhang, Tina Eliassi-Rad, Farinaz Koushanfar","doi":"10.1038/s42256-024-00833-7","DOIUrl":"10.1038/s42256-024-00833-7","url":null,"abstract":"Scalable addressing of high-dimensional constrained combinatorial optimization problems is a challenge that arises in several science and engineering disciplines. Recent work introduced novel applications of graph neural networks for solving quadratic-cost combinatorial optimization problems. However, effective utilization of models such as graph neural networks to address general problems with higher-order constraints is an unresolved challenge. This paper presents a framework, HypOp, that advances the state of the art for solving combinatorial optimization problems in several aspects: (1) it generalizes the prior results to higher-order constrained problems with arbitrary cost functions by leveraging hypergraph neural networks; (2) it enables scalability to larger problems by introducing a new distributed and parallel training architecture; (3) it demonstrates generalizability across different problem formulations by transferring knowledge within the same hypergraph; (4) it substantially boosts the solution accuracy compared with the prior art by suggesting a fine-tuning step using simulated annealing; and (5) it shows remarkable progress on numerous benchmark examples, including hypergraph MaxCut, satisfiability and resource allocation problems, with notable run-time improvements using a combination of fine-tuning and distributed training techniques. We showcase the application of HypOp in scientific discovery by solving a hypergraph MaxCut problem on a National Drug Code drug-substance hypergraph. Through extensive experimentation on various optimization problems, HypOp demonstrates superiority over existing unsupervised-learning-based solvers and generic optimization methods. Bolstering the broad and deep applicability of graph neural networks, Heydaribeni et al. introduce HypOp, a framework that uses hypergraph neural networks to solve general constrained combinatorial optimization problems. The presented method scales and generalizes well, improves accuracy and outperforms existing solvers on various benchmarking examples.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 6","pages":"664-672"},"PeriodicalIF":18.8,"publicationDate":"2024-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141177288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Empathic AI can’t get under the skin 感同身受的人工智能无法深入人心
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-24 DOI: 10.1038/s42256-024-00850-6
Personalized LLMs built with the capacity for emulating empathy are right around the corner. The effects on individual users needs careful consideration.
具有移情能力的个性化 LLM 即将问世。需要仔细考虑对个人用户的影响。
{"title":"Empathic AI can’t get under the skin","authors":"","doi":"10.1038/s42256-024-00850-6","DOIUrl":"10.1038/s42256-024-00850-6","url":null,"abstract":"Personalized LLMs built with the capacity for emulating empathy are right around the corner. The effects on individual users needs careful consideration.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"495-495"},"PeriodicalIF":23.8,"publicationDate":"2024-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.com/articles/s42256-024-00850-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141096488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Accurate and robust protein sequence design with CarbonDesign 利用 CarbonDesign 进行精确、稳健的蛋白质序列设计
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-23 DOI: 10.1038/s42256-024-00838-2
Milong Ren, Chungong Yu, Dongbo Bu, Haicang Zhang
Protein sequence design is critically important for protein engineering. Despite recent advancements in deep learning-based methods, achieving accurate and robust sequence design remains a challenge. Here we present CarbonDesign, an approach that draws inspiration from successful ingredients of AlphaFold and which has been developed specifically for protein sequence design. At its core, CarbonDesign introduces Inverseformer, which learns representations from backbone structures and an amortized Markov random fields model for sequence decoding. Moreover, we incorporate other essential AlphaFold concepts into CarbonDesign: an end-to-end network recycling technique to leverage evolutionary constraints from protein language models and a multitask learning technique for generating side-chain structures alongside designed sequences. CarbonDesign outperforms other methods on independent test sets including the 15th Critical Assessment of protein Structure Prediction (CASP15) dataset, the Continuous Automated Model Evaluation (CAMEO) dataset and de novo proteins from RFDiffusion. Furthermore, it supports zero-shot prediction of the functional effects of sequence variants, making it a promising tool for applications in bioengineering. Deep learning has led to great advances in predicting protein structure from sequences. Ren and colleagues present here a method for the inverse problem of finding a sequence that results in a desired protein structure, which is inspired by various components of AlphaFold combined with Markov random fields to decode sequences more efficiently.
蛋白质序列设计对蛋白质工程至关重要。尽管基于深度学习的方法取得了最新进展,但实现准确、稳健的序列设计仍然是一项挑战。在此,我们介绍 CarbonDesign,这是一种从 AlphaFold 的成功要素中汲取灵感,专门为蛋白质序列设计而开发的方法。CarbonDesign 的核心是引入反演器(Inverseformer),它能从骨架结构和摊销马尔可夫随机场模型中学习序列解码的表征。此外,我们还在 CarbonDesign 中融入了 AlphaFold 的其他基本概念:利用蛋白质语言模型的进化约束的端到端网络再循环技术,以及在设计序列的同时生成侧链结构的多任务学习技术。CarbonDesign在独立测试集上的表现优于其他方法,这些测试集包括第15次蛋白质结构预测关键评估(CASP15)数据集、连续自动模型评估(CAMEO)数据集和RFDiffusion的全新蛋白质。此外,它还支持对序列变异的功能效应进行零点预测,使其成为生物工程领域的一个前景广阔的应用工具。
{"title":"Accurate and robust protein sequence design with CarbonDesign","authors":"Milong Ren, Chungong Yu, Dongbo Bu, Haicang Zhang","doi":"10.1038/s42256-024-00838-2","DOIUrl":"10.1038/s42256-024-00838-2","url":null,"abstract":"Protein sequence design is critically important for protein engineering. Despite recent advancements in deep learning-based methods, achieving accurate and robust sequence design remains a challenge. Here we present CarbonDesign, an approach that draws inspiration from successful ingredients of AlphaFold and which has been developed specifically for protein sequence design. At its core, CarbonDesign introduces Inverseformer, which learns representations from backbone structures and an amortized Markov random fields model for sequence decoding. Moreover, we incorporate other essential AlphaFold concepts into CarbonDesign: an end-to-end network recycling technique to leverage evolutionary constraints from protein language models and a multitask learning technique for generating side-chain structures alongside designed sequences. CarbonDesign outperforms other methods on independent test sets including the 15th Critical Assessment of protein Structure Prediction (CASP15) dataset, the Continuous Automated Model Evaluation (CAMEO) dataset and de novo proteins from RFDiffusion. Furthermore, it supports zero-shot prediction of the functional effects of sequence variants, making it a promising tool for applications in bioengineering. Deep learning has led to great advances in predicting protein structure from sequences. Ren and colleagues present here a method for the inverse problem of finding a sequence that results in a desired protein structure, which is inspired by various components of AlphaFold combined with Markov random fields to decode sequences more efficiently.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"536-547"},"PeriodicalIF":23.8,"publicationDate":"2024-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141085633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quantum circuit synthesis with diffusion models 利用扩散模型进行量子电路合成
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-20 DOI: 10.1038/s42256-024-00831-9
Florian Fürrutter, Gorka Muñoz-Gil, Hans J. Briegel
Quantum computing has recently emerged as a transformative technology. Yet, its promised advantages rely on efficiently translating quantum operations into viable physical realizations. Here we use generative machine learning models, specifically denoising diffusion models (DMs), to facilitate this transformation. Leveraging text conditioning, we steer the model to produce desired quantum operations within gate-based quantum circuits. Notably, DMs allow to sidestep during training the exponential overhead inherent in the classical simulation of quantum dynamics—a consistent bottleneck in preceding machine learning techniques. We demonstrate the model’s capabilities across two tasks: entanglement generation and unitary compilation. The model excels at generating new circuits and supports typical DM extensions such as masking and editing to, for instance, align the circuit generation to the constraints of the targeted quantum device. Given their flexibility and generalization abilities, we envision DMs as pivotal in quantum circuit synthesis, both enhancing practical applications and providing insights into theoretical quantum computation. Achieving the promised advantages of quantum computing relies on translating quantum operations into physical realizations. Fürrutter and colleagues use diffusion models to create quantum circuits that are based on user specifications and tailored to experimental constraints.
量子计算最近已成为一项变革性技术。然而,它所承诺的优势有赖于有效地将量子操作转化为可行的物理实现。在这里,我们使用生成式机器学习模型,特别是去噪扩散模型(DMs)来促进这种转变。利用文本调节,我们引导模型在基于门的量子电路中产生所需的量子操作。值得注意的是,DM 可以在训练过程中避开量子动力学经典模拟中固有的指数级开销--这是之前机器学习技术的一贯瓶颈。我们在纠缠生成和单元编译这两项任务中展示了该模型的能力。该模型擅长生成新电路,并支持典型的 DM 扩展,如屏蔽和编辑,以便根据目标量子设备的约束条件调整电路生成。鉴于 DM 的灵活性和泛化能力,我们认为 DM 在量子电路合成中具有举足轻重的作用,既能增强实际应用,又能为理论量子计算提供见解。
{"title":"Quantum circuit synthesis with diffusion models","authors":"Florian Fürrutter, Gorka Muñoz-Gil, Hans J. Briegel","doi":"10.1038/s42256-024-00831-9","DOIUrl":"10.1038/s42256-024-00831-9","url":null,"abstract":"Quantum computing has recently emerged as a transformative technology. Yet, its promised advantages rely on efficiently translating quantum operations into viable physical realizations. Here we use generative machine learning models, specifically denoising diffusion models (DMs), to facilitate this transformation. Leveraging text conditioning, we steer the model to produce desired quantum operations within gate-based quantum circuits. Notably, DMs allow to sidestep during training the exponential overhead inherent in the classical simulation of quantum dynamics—a consistent bottleneck in preceding machine learning techniques. We demonstrate the model’s capabilities across two tasks: entanglement generation and unitary compilation. The model excels at generating new circuits and supports typical DM extensions such as masking and editing to, for instance, align the circuit generation to the constraints of the targeted quantum device. Given their flexibility and generalization abilities, we envision DMs as pivotal in quantum circuit synthesis, both enhancing practical applications and providing insights into theoretical quantum computation. Achieving the promised advantages of quantum computing relies on translating quantum operations into physical realizations. Fürrutter and colleagues use diffusion models to create quantum circuits that are based on user specifications and tailored to experimental constraints.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"515-524"},"PeriodicalIF":23.8,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141073903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient learning of accurate surrogates for simulations of complex systems 高效学习复杂系统模拟的精确代用指标
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-17 DOI: 10.1038/s42256-024-00839-1
A. Diaw, M. McKerns, I. Sagert, L. G. Stanton, M. S. Murillo
Machine learning methods are increasingly deployed to construct surrogate models for complex physical systems at a reduced computational cost. However, the predictive capability of these surrogates degrades in the presence of noisy, sparse or dynamic data. We introduce an online learning method empowered by optimizer-driven sampling that has two advantages over current approaches: it ensures that all local extrema (including endpoints) of the model response surface are included in the training data, and it employs a continuous validation and update process in which surrogates undergo retraining when their performance falls below a validity threshold. We find, using benchmark functions, that optimizer-directed sampling generally outperforms traditional sampling methods in terms of accuracy around local extrema even when the scoring metric is biased towards assessing overall accuracy. Finally, the application to dense nuclear matter demonstrates that highly accurate surrogates for a nuclear equation-of-state model can be reliably autogenerated from expensive calculations using few model evaluations. Machine learning-based surrogate models are important to model complex systems at a reduced computational cost; however, they must often be re-evaluated and adapted for validity on future data. Diaw and colleagues propose an online training method leveraging optimizer-directed sampling to produce surrogate models that can be applied to any future data and demonstrate the approach on a dense nuclear-matter equation of state containing a phase transition.
人们越来越多地采用机器学习方法来构建复杂物理系统的代用模型,从而降低计算成本。然而,在存在噪声、稀疏或动态数据的情况下,这些代用模型的预测能力会下降。我们引入了一种在线学习方法,该方法由优化器驱动采样,与目前的方法相比有两个优势:它能确保模型响应面的所有局部极值(包括端点)都包含在训练数据中;它还采用了持续验证和更新过程,当代理模型的性能低于有效阈值时,就会对其进行再训练。我们利用基准函数发现,即使评分标准偏向于评估整体准确性,优化器定向采样在局部极值附近的准确性方面也普遍优于传统采样方法。最后,对致密核物质的应用表明,核状态方程模型的高精度替代物可以通过昂贵的计算,使用少量模型评估,可靠地自动生成。
{"title":"Efficient learning of accurate surrogates for simulations of complex systems","authors":"A. Diaw, M. McKerns, I. Sagert, L. G. Stanton, M. S. Murillo","doi":"10.1038/s42256-024-00839-1","DOIUrl":"10.1038/s42256-024-00839-1","url":null,"abstract":"Machine learning methods are increasingly deployed to construct surrogate models for complex physical systems at a reduced computational cost. However, the predictive capability of these surrogates degrades in the presence of noisy, sparse or dynamic data. We introduce an online learning method empowered by optimizer-driven sampling that has two advantages over current approaches: it ensures that all local extrema (including endpoints) of the model response surface are included in the training data, and it employs a continuous validation and update process in which surrogates undergo retraining when their performance falls below a validity threshold. We find, using benchmark functions, that optimizer-directed sampling generally outperforms traditional sampling methods in terms of accuracy around local extrema even when the scoring metric is biased towards assessing overall accuracy. Finally, the application to dense nuclear matter demonstrates that highly accurate surrogates for a nuclear equation-of-state model can be reliably autogenerated from expensive calculations using few model evaluations. Machine learning-based surrogate models are important to model complex systems at a reduced computational cost; however, they must often be re-evaluated and adapted for validity on future data. Diaw and colleagues propose an online training method leveraging optimizer-directed sampling to produce surrogate models that can be applied to any future data and demonstrate the approach on a dense nuclear-matter equation of state containing a phase transition.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"568-577"},"PeriodicalIF":23.8,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140953368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Back to basics to open the black box 返璞归真,打开黑匣子
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-17 DOI: 10.1038/s42256-024-00842-6
Diego Marcondes, Adilson Simonis, Junior Barrera
Most research efforts in machine learning focus on performance and are detached from an explanation of the behaviour of the model. We call for going back to basics of machine learning methods, with more focus on the development of a basic understanding grounded in statistical theory.
机器学习方面的大多数研究工作都侧重于性能,而脱离了对模型行为的解释。我们呼吁回归机器学习方法的基本原理,更加注重发展以统计理论为基础的基本认识。
{"title":"Back to basics to open the black box","authors":"Diego Marcondes, Adilson Simonis, Junior Barrera","doi":"10.1038/s42256-024-00842-6","DOIUrl":"10.1038/s42256-024-00842-6","url":null,"abstract":"Most research efforts in machine learning focus on performance and are detached from an explanation of the behaviour of the model. We call for going back to basics of machine learning methods, with more focus on the development of a basic understanding grounded in statistical theory.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"498-501"},"PeriodicalIF":23.8,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140953389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Does it matter if empathic AI has no empathy? 移情人工智能没有移情能力重要吗?
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-15 DOI: 10.1038/s42256-024-00841-7
Garriy Shteynberg, Jodi Halpern, Amir Sadovnik, Jon Garthoff, Anat Perry, Jessica Hay, Carlos Montemayor, Michael A. Olson, Tim L. Hulsey, Abrol Fairweather
{"title":"Does it matter if empathic AI has no empathy?","authors":"Garriy Shteynberg, Jodi Halpern, Amir Sadovnik, Jon Garthoff, Anat Perry, Jessica Hay, Carlos Montemayor, Michael A. Olson, Tim L. Hulsey, Abrol Fairweather","doi":"10.1038/s42256-024-00841-7","DOIUrl":"10.1038/s42256-024-00841-7","url":null,"abstract":"","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"496-497"},"PeriodicalIF":23.8,"publicationDate":"2024-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140925029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-purpose RNA language modelling with motif-aware pretraining and type-guided fine-tuning 利用主题感知预训练和类型指导微调建立多用途 RNA 语言模型
IF 23.8 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-13 DOI: 10.1038/s42256-024-00836-4
Ning Wang, Jiang Bian, Yuchen Li, Xuhong Li, Shahid Mumtaz, Linghe Kong, Haoyi Xiong
Pretrained language models have shown promise in analysing nucleotide sequences, yet a versatile model excelling across diverse tasks with a single pretrained weight set remains elusive. Here we introduce RNAErnie, an RNA-focused pretrained model built upon the transformer architecture, employing two simple yet effective strategies. First, RNAErnie enhances pretraining by incorporating RNA motifs as biological priors and introducing motif-level random masking in addition to masked language modelling at base/subsequence levels. It also tokenizes RNA types (for example, miRNA, lnRNA) as stop words, appending them to sequences during pretraining. Second, subject to out-of-distribution tasks with RNA sequences not seen during the pretraining phase, RNAErnie proposes a type-guided fine-tuning strategy that first predicts possible RNA types using an RNA sequence and then appends the predicted type to the tail of sequence to refine feature embedding in a post hoc way. Our extensive evaluation across seven datasets and five tasks demonstrates the superiority of RNAErnie in both supervised and unsupervised learning. It surpasses baselines with up to 1.8% higher accuracy in classification, 2.2% greater accuracy in interaction prediction and 3.3% improved F1 score in structure prediction, showcasing its robustness and adaptability with a unified pretrained foundation. Despite the existence of various pretrained language models for nucleotide sequence analysis, achieving good performance on a broad range of downstream tasks using a single model is challenging. Wang and colleagues develop a pretrained language model specifically optimized for RNA sequence analysis and show that it can outperform state-of-the-art methods in a diverse set of downstream tasks.
预训练语言模型在分析核苷酸序列方面已显示出良好的前景,但在不同任务中使用单一预训练权重集的多功能模型仍未出现。在这里,我们介绍 RNAErnie,它是一种以 RNA 为重点的预训练模型,建立在转换器架构之上,采用了两种简单而有效的策略。首先,RNAErnie 将 RNA 主题作为生物学先验,并在基序/子序列级别的屏蔽语言建模之外引入了主题级别的随机屏蔽,从而增强了预训练效果。它还将 RNA 类型(如 miRNA、lnRNA)标记为停止词,并在预训练期间将其添加到序列中。其次,针对在预训练阶段未见过的 RNA 序列的分布外任务,RNAErnie 提出了一种类型引导的微调策略,首先使用 RNA 序列预测可能的 RNA 类型,然后将预测的类型附加到序列尾部,以事后方式完善特征嵌入。我们在七个数据集和五项任务中进行了广泛的评估,结果表明 RNAErnie 在监督和非监督学习方面都具有优势。它超越了基线,分类准确率提高了 1.8%,相互作用预测准确率提高了 2.2%,结构预测的 F1 分数提高了 3.3%,展示了它在统一预训练基础上的鲁棒性和适应性。
{"title":"Multi-purpose RNA language modelling with motif-aware pretraining and type-guided fine-tuning","authors":"Ning Wang, Jiang Bian, Yuchen Li, Xuhong Li, Shahid Mumtaz, Linghe Kong, Haoyi Xiong","doi":"10.1038/s42256-024-00836-4","DOIUrl":"10.1038/s42256-024-00836-4","url":null,"abstract":"Pretrained language models have shown promise in analysing nucleotide sequences, yet a versatile model excelling across diverse tasks with a single pretrained weight set remains elusive. Here we introduce RNAErnie, an RNA-focused pretrained model built upon the transformer architecture, employing two simple yet effective strategies. First, RNAErnie enhances pretraining by incorporating RNA motifs as biological priors and introducing motif-level random masking in addition to masked language modelling at base/subsequence levels. It also tokenizes RNA types (for example, miRNA, lnRNA) as stop words, appending them to sequences during pretraining. Second, subject to out-of-distribution tasks with RNA sequences not seen during the pretraining phase, RNAErnie proposes a type-guided fine-tuning strategy that first predicts possible RNA types using an RNA sequence and then appends the predicted type to the tail of sequence to refine feature embedding in a post hoc way. Our extensive evaluation across seven datasets and five tasks demonstrates the superiority of RNAErnie in both supervised and unsupervised learning. It surpasses baselines with up to 1.8% higher accuracy in classification, 2.2% greater accuracy in interaction prediction and 3.3% improved F1 score in structure prediction, showcasing its robustness and adaptability with a unified pretrained foundation. Despite the existence of various pretrained language models for nucleotide sequence analysis, achieving good performance on a broad range of downstream tasks using a single model is challenging. Wang and colleagues develop a pretrained language model specifically optimized for RNA sequence analysis and show that it can outperform state-of-the-art methods in a diverse set of downstream tasks.","PeriodicalId":48533,"journal":{"name":"Nature Machine Intelligence","volume":"6 5","pages":"548-557"},"PeriodicalIF":23.8,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.com/articles/s42256-024-00836-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140919490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Nature Machine Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1