首页 > 最新文献

Nature computational science最新文献

英文 中文
Predicting drug responses of unseen cell types through transfer learning with foundation models 基于基础模型的迁移学习预测未知细胞类型的药物反应。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-10-03 DOI: 10.1038/s43588-025-00887-6
Yixuan Wang, Xinyuan Liu, Yimin Fan, Binghui Xie, James Cheng, Kam Chung Wong, Peter Cheung, Irwin King, Yu Li
Drug repurposing through single-cell perturbation response prediction provides a cost-effective approach for drug development, but accurately predicting responses in unseen cell types that emerge during disease progression remains challenging. Existing methods struggle to achieve generalizable cell-type-specific predictions. To address these limitations, we introduce the cell-type-specific drug perturbatIon responses predictor (CRISP), a framework for predicting perturbation responses in previously unseen cell types at single-cell resolution. CRISP leverages foundation models and cell-type-specific learning strategies to enable effective transfer of information from control to perturbed states even with limited empirical data. Through systematic evaluation across increasingly challenging scenarios, from unseen cell types to cross-platform predictions, CRISP shows generalizability and performance improvements. We demonstrate CRISP’s drug repurposing potential through zero-shot prediction from solid tumor data to sorafenib’s therapeutic effects in chronic myeloid leukemia. The predicted anti-tumor mechanisms, including CXCR4 pathway inhibition, are supported by independent studies as an effective therapeutic strategy in chronic myeloid leukemia, aligning with past studies and clinical trials. This work develops CRISP, a framework using foundation models to predict drug responses in previously unseen cell types at single-cell resolution, advancing drug repurposing and drug screening capabilities.
通过单细胞扰动反应预测进行药物重新利用为药物开发提供了一种经济有效的方法,但准确预测疾病进展过程中出现的未见细胞类型的反应仍然具有挑战性。现有的方法难以实现可推广的细胞类型特异性预测。为了解决这些限制,我们引入了细胞类型特异性药物扰动反应预测器(CRISP),这是一个在单细胞分辨率下预测以前未见过的细胞类型的扰动反应的框架。CRISP利用基础模型和细胞类型特定的学习策略,即使在有限的经验数据下,也能有效地将信息从控制状态转移到受扰状态。通过对越来越具有挑战性的场景进行系统评估,从看不见的细胞类型到跨平台预测,CRISP显示了通用性和性能改进。我们通过从实体瘤数据到索拉非尼治疗慢性髓性白血病的零shot预测,证明了CRISP的药物再利用潜力。预测的抗肿瘤机制,包括CXCR4通路抑制,作为慢性髓性白血病的有效治疗策略得到了独立研究的支持,与过去的研究和临床试验一致。
{"title":"Predicting drug responses of unseen cell types through transfer learning with foundation models","authors":"Yixuan Wang, Xinyuan Liu, Yimin Fan, Binghui Xie, James Cheng, Kam Chung Wong, Peter Cheung, Irwin King, Yu Li","doi":"10.1038/s43588-025-00887-6","DOIUrl":"10.1038/s43588-025-00887-6","url":null,"abstract":"Drug repurposing through single-cell perturbation response prediction provides a cost-effective approach for drug development, but accurately predicting responses in unseen cell types that emerge during disease progression remains challenging. Existing methods struggle to achieve generalizable cell-type-specific predictions. To address these limitations, we introduce the cell-type-specific drug perturbatIon responses predictor (CRISP), a framework for predicting perturbation responses in previously unseen cell types at single-cell resolution. CRISP leverages foundation models and cell-type-specific learning strategies to enable effective transfer of information from control to perturbed states even with limited empirical data. Through systematic evaluation across increasingly challenging scenarios, from unseen cell types to cross-platform predictions, CRISP shows generalizability and performance improvements. We demonstrate CRISP’s drug repurposing potential through zero-shot prediction from solid tumor data to sorafenib’s therapeutic effects in chronic myeloid leukemia. The predicted anti-tumor mechanisms, including CXCR4 pathway inhibition, are supported by independent studies as an effective therapeutic strategy in chronic myeloid leukemia, aligning with past studies and clinical trials. This work develops CRISP, a framework using foundation models to predict drug responses in previously unseen cell types at single-cell resolution, advancing drug repurposing and drug screening capabilities.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"6 1","pages":"39-52"},"PeriodicalIF":18.3,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145226337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Proteoform search from protein database with top-down mass spectra 自顶向下质谱法在蛋白质数据库中搜索蛋白质形态。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-10-03 DOI: 10.1038/s43588-025-00880-z
Kunyi Li, Baozhen Shan, Lei Xin, Ming Li, Lusheng Wang
Here we propose a search algorithm for proteoform identification that computes the largest-size error-correction alignments between a protein mass graph and a spectrum mass graph. Our combined method uses a filtering algorithm to identify candidates and then applies a search algorithm to report the final results. Our exact searching method is 3.9 to 9.0 times faster than popular methods such as TopMG and TopPIC. Our combined method can further speed-up the running time of sTopMG without affecting the search accuracy. We develop a pipeline for generating simulated top-down spectra on the basis of input protein sequences with modifications. Experiments on simulated datasets show that our combined method has 95% accuracy, which exceeds existing methods. Experiments on real annotated datasets show that our method has ≥97.1% accuracy using deconvolution method FLASHDeconv. An algorithm for proteoform identification with top-down mass spectra is proposed, and a pipeline is developed for generating simulated top-down spectra on the basis of input protein sequences with modifications.
在这里,我们提出了一种用于蛋白质形态识别的搜索算法,该算法计算蛋白质质量图和谱质量图之间的最大尺寸误差校正比对。我们的组合方法使用过滤算法来识别候选对象,然后应用搜索算法来报告最终结果。我们的精确搜索方法比流行的TopMG和TopPIC等方法快3.9到9.0倍。我们的组合方法可以在不影响搜索精度的情况下进一步加快sTopMG的运行时间。我们开发了一个管道来生成模拟自顶向下的光谱的基础上的输入蛋白序列的修改。在模拟数据集上的实验表明,该方法的准确率达到95%,超过了现有的方法。在真实标注数据集上的实验表明,使用反卷积方法FLASHDeconv,我们的方法准确率≥97.1%。
{"title":"Proteoform search from protein database with top-down mass spectra","authors":"Kunyi Li, Baozhen Shan, Lei Xin, Ming Li, Lusheng Wang","doi":"10.1038/s43588-025-00880-z","DOIUrl":"10.1038/s43588-025-00880-z","url":null,"abstract":"Here we propose a search algorithm for proteoform identification that computes the largest-size error-correction alignments between a protein mass graph and a spectrum mass graph. Our combined method uses a filtering algorithm to identify candidates and then applies a search algorithm to report the final results. Our exact searching method is 3.9 to 9.0 times faster than popular methods such as TopMG and TopPIC. Our combined method can further speed-up the running time of sTopMG without affecting the search accuracy. We develop a pipeline for generating simulated top-down spectra on the basis of input protein sequences with modifications. Experiments on simulated datasets show that our combined method has 95% accuracy, which exceeds existing methods. Experiments on real annotated datasets show that our method has ≥97.1% accuracy using deconvolution method FLASHDeconv. An algorithm for proteoform identification with top-down mass spectra is proposed, and a pipeline is developed for generating simulated top-down spectra on the basis of input protein sequences with modifications.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 11","pages":"998-1009"},"PeriodicalIF":18.3,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145226292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Boosting power for time-to-event GWAS analysis affected by case ascertainment 增强受案例确定影响的时间到事件GWAS分析的能力。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-10-02 DOI: 10.1038/s43588-025-00892-9
We propose a computationally efficient genome-wide association study (GWAS) method, WtCoxG, for time-to-event (TTE) traits in the presence of case ascertainment— a form of oversampling bias. WtCoxG addresses case ascertainment bias by applying a weighted Cox proportional hazard model, and outperforms existing approaches when incorporating information on external allele frequencies.
我们提出了一种计算效率高的全基因组关联研究(GWAS)方法,WtCoxG,用于病例确定(一种过抽样偏差)存在的事件时间(TTE)特征。WtCoxG通过应用加权Cox比例风险模型来解决病例确定偏差,并且在结合外部等位基因频率信息时优于现有方法。
{"title":"Boosting power for time-to-event GWAS analysis affected by case ascertainment","authors":"","doi":"10.1038/s43588-025-00892-9","DOIUrl":"10.1038/s43588-025-00892-9","url":null,"abstract":"We propose a computationally efficient genome-wide association study (GWAS) method, WtCoxG, for time-to-event (TTE) traits in the presence of case ascertainment— a form of oversampling bias. WtCoxG addresses case ascertainment bias by applying a weighted Cox proportional hazard model, and outperforms existing approaches when incorporating information on external allele frequencies.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 11","pages":"996-997"},"PeriodicalIF":18.3,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145214719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-driving labs for biotechnology 生物技术的自动驾驶实验室。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-10-01 DOI: 10.1038/s43588-025-00885-8
Evan Collins, Robert Langer, Daniel G. Anderson
Self-driving laboratories that integrate robotic production with artificial intelligence have the potential to accelerate innovation in biotechnology. Because self-driving labs can be complex and not universally applicable, it is useful to consider their suitable use cases for successful integration into discovery workflows. Here, we review strategies for assessing the suitability of self-driving labs for biochemical design problems.
将机器人生产与人工智能相结合的自动驾驶实验室有可能加速生物技术的创新。因为自动驾驶实验室可能是复杂的,并且不是普遍适用的,所以考虑它们的合适用例以成功地集成到发现工作流中是有用的。在这里,我们回顾了评估自动驾驶实验室对生化设计问题的适用性的策略。
{"title":"Self-driving labs for biotechnology","authors":"Evan Collins, Robert Langer, Daniel G. Anderson","doi":"10.1038/s43588-025-00885-8","DOIUrl":"10.1038/s43588-025-00885-8","url":null,"abstract":"Self-driving laboratories that integrate robotic production with artificial intelligence have the potential to accelerate innovation in biotechnology. Because self-driving labs can be complex and not universally applicable, it is useful to consider their suitable use cases for successful integration into discovery workflows. Here, we review strategies for assessing the suitability of self-driving labs for biochemical design problems.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 11","pages":"976-979"},"PeriodicalIF":18.3,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145208389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting the regulatory impacts of noncoding variants on gene expression through epigenomic integration across tissues and single-cell landscapes 通过跨组织和单细胞景观的表观基因组整合预测非编码变异对基因表达的调控影响。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-26 DOI: 10.1038/s43588-025-00878-7
Zhe Liu  (, ), Yihang Bao  (, ), An Gu  (, ), Weichen Song  (, ), Guan Ning Lin  (, )
Noncoding mutations play a critical role in regulating gene expression, yet predicting their effects across diverse tissues and cell types remains a challenge. Here we present EMO, a transformer-based model that integrates DNA sequence with chromatin accessibility data (assay for transposase-accessible chromatin with sequencing) to predict the regulatory impact of noncoding single nucleotide polymorphisms on gene expression. A key component of EMO is its ability to incorporate personalized functional genomic profiles, enabling individual-level and disease-contextual predictions and addressing critical limitations of current approaches. EMO generalizes across tissues and cell types by modeling both short- and long-range regulatory interactions and capturing dynamic gene expression changes associated with disease progression. In benchmark evaluations, the pretraining-based EMO framework outperformed existing models, with fine-tuning small-sample tissues enhancing the model’s ability to fit target tissues. In single-cell contexts, EMO accurately identified cell-type-specific regulatory patterns and successfully captured the effects of disease-associated single nucleotide polymorphisms in conditions, linking genetic variation to disease-relevant pathways. EMO integrates DNA sequence and chromatin accessibility data to predict how noncoding variants regulate gene expression across tissues and single cells, enabling context-aware personalized insights into genetic effects for precision medicine.
非编码突变在调节基因表达中起着至关重要的作用,然而预测它们在不同组织和细胞类型中的作用仍然是一个挑战。在这里,我们提出了EMO,这是一种基于转换器的模型,将DNA序列与染色质可接近性数据(转座酶可接近染色质的测定与测序)整合在一起,以预测非编码单核苷酸多态性对基因表达的调节影响。EMO的一个关键组成部分是其整合个性化功能基因组图谱的能力,能够实现个体水平和疾病背景预测,并解决当前方法的关键局限性。EMO通过模拟短期和长期的调节相互作用以及捕捉与疾病进展相关的动态基因表达变化,从而推广到各种组织和细胞类型。在基准评估中,基于预训练的EMO框架优于现有模型,微调小样本组织增强了模型拟合目标组织的能力。在单细胞环境下,EMO准确识别细胞类型特异性调节模式,并成功捕获疾病相关单核苷酸多态性在条件下的影响,将遗传变异与疾病相关途径联系起来。
{"title":"Predicting the regulatory impacts of noncoding variants on gene expression through epigenomic integration across tissues and single-cell landscapes","authors":"Zhe Liu \u0000 (, ), Yihang Bao \u0000 (, ), An Gu \u0000 (, ), Weichen Song \u0000 (, ), Guan Ning Lin \u0000 (, )","doi":"10.1038/s43588-025-00878-7","DOIUrl":"10.1038/s43588-025-00878-7","url":null,"abstract":"Noncoding mutations play a critical role in regulating gene expression, yet predicting their effects across diverse tissues and cell types remains a challenge. Here we present EMO, a transformer-based model that integrates DNA sequence with chromatin accessibility data (assay for transposase-accessible chromatin with sequencing) to predict the regulatory impact of noncoding single nucleotide polymorphisms on gene expression. A key component of EMO is its ability to incorporate personalized functional genomic profiles, enabling individual-level and disease-contextual predictions and addressing critical limitations of current approaches. EMO generalizes across tissues and cell types by modeling both short- and long-range regulatory interactions and capturing dynamic gene expression changes associated with disease progression. In benchmark evaluations, the pretraining-based EMO framework outperformed existing models, with fine-tuning small-sample tissues enhancing the model’s ability to fit target tissues. In single-cell contexts, EMO accurately identified cell-type-specific regulatory patterns and successfully captured the effects of disease-associated single nucleotide polymorphisms in conditions, linking genetic variation to disease-relevant pathways. EMO integrates DNA sequence and chromatin accessibility data to predict how noncoding variants regulate gene expression across tissues and single cells, enabling context-aware personalized insights into genetic effects for precision medicine.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 10","pages":"927-939"},"PeriodicalIF":18.3,"publicationDate":"2025-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145180816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rhythm-based hierarchical predictive computations support acoustic−semantic transformation in speech processing 基于节奏的分层预测计算支持语音处理中的声-语义转换。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-26 DOI: 10.1038/s43588-025-00876-9
Olesia Dogonasheva, Keith B. Doelling, Denis Zakharov, Anne-Lise Giraud, Boris Gutkin
Unraveling how humans understand speech despite distortions has long intrigued researchers. A prominent hypothesis highlights the role of multiple endogenous brain rhythms in forming the computational context to predict speech structure and content. Yet how neural processes may implement rhythm-based context formation remains unclear. Here we propose the brain rhythm-based inference model (BRyBI) as a possible neural implementation of speech processing in the auditory cortex based on the interaction of endogenous brain rhythms in a predictive coding framework. BRyBI encodes key rhythmic processes for parsing spectro-temporal representations of the speech signal into phoneme sequences and to govern the formation of the phrasal context. BRyBI matches patterns of human performance in speech recognition tasks and explains contradictory experimental observations of rhythms during speech listening and their dependence on the informational aspect of speech (uncertainty and surprise). This work highlights the computational role of multiscale brain rhythms in predictive speech processing. This study presents a brain rhythm-based inference model (BRyBI) for speech processing in the auditory cortex. BRyBI shows how rhythmic neural activity enables robust speech processing by dynamically predicting context and elucidates mechanistic principles that allow robust speech parsing in the brain.
长期以来,研究人员一直好奇人类是如何在扭曲的情况下理解语言的。一个突出的假设强调了多种内源性大脑节律在形成预测语音结构和内容的计算环境中的作用。然而,神经过程如何实现基于节奏的上下文形成仍不清楚。在此,我们提出基于脑节奏的推理模型(BRyBI)作为一种基于内源性脑节奏在预测编码框架中的相互作用的听觉皮层语音处理的可能神经实现。BRyBI对关键的节奏过程进行编码,以将语音信号的光谱时间表征解析为音素序列,并控制短语上下文的形成。BRyBI与人类在语音识别任务中的表现模式相匹配,并解释了语音听力过程中节奏的矛盾实验观察及其对语音信息方面(不确定性和惊喜)的依赖。这项工作强调了多尺度脑节律在预测语音处理中的计算作用。
{"title":"Rhythm-based hierarchical predictive computations support acoustic−semantic transformation in speech processing","authors":"Olesia Dogonasheva, Keith B. Doelling, Denis Zakharov, Anne-Lise Giraud, Boris Gutkin","doi":"10.1038/s43588-025-00876-9","DOIUrl":"10.1038/s43588-025-00876-9","url":null,"abstract":"Unraveling how humans understand speech despite distortions has long intrigued researchers. A prominent hypothesis highlights the role of multiple endogenous brain rhythms in forming the computational context to predict speech structure and content. Yet how neural processes may implement rhythm-based context formation remains unclear. Here we propose the brain rhythm-based inference model (BRyBI) as a possible neural implementation of speech processing in the auditory cortex based on the interaction of endogenous brain rhythms in a predictive coding framework. BRyBI encodes key rhythmic processes for parsing spectro-temporal representations of the speech signal into phoneme sequences and to govern the formation of the phrasal context. BRyBI matches patterns of human performance in speech recognition tasks and explains contradictory experimental observations of rhythms during speech listening and their dependence on the informational aspect of speech (uncertainty and surprise). This work highlights the computational role of multiscale brain rhythms in predictive speech processing. This study presents a brain rhythm-based inference model (BRyBI) for speech processing in the auditory cortex. BRyBI shows how rhythmic neural activity enables robust speech processing by dynamically predicting context and elucidates mechanistic principles that allow robust speech parsing in the brain.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 10","pages":"915-926"},"PeriodicalIF":18.3,"publicationDate":"2025-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145180835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The rise of large language models 大型语言模型的兴起
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-24 DOI: 10.1038/s43588-025-00890-x
This issue of Nature Computational Science features a Focus that highlights both the promises and perils of large language models, their emerging applications across diverse scientific domains, and the opportunities to overcome the challenges that lie ahead.
这一期的《自然计算科学》重点介绍了大型语言模型的前景和危险,它们在不同科学领域的新兴应用,以及克服未来挑战的机会。
{"title":"The rise of large language models","authors":"","doi":"10.1038/s43588-025-00890-x","DOIUrl":"10.1038/s43588-025-00890-x","url":null,"abstract":"This issue of Nature Computational Science features a Focus that highlights both the promises and perils of large language models, their emerging applications across diverse scientific domains, and the opportunities to overcome the challenges that lie ahead.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 9","pages":"689-690"},"PeriodicalIF":18.3,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.comhttps://www.nature.com/articles/s43588-025-00890-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145129541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuromorphic principles in self-attention hardware for efficient transformers 高效变压器自注意硬件的神经形态原理。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-16 DOI: 10.1038/s43588-025-00868-9
Nathan Leroux, Jan Finkbeiner, Emre Neftci
Strong barriers remain between neuromorphic engineering and machine learning, especially with regard to recent large language models (LLMs) and transformers. This Comment makes the case that neuromorphic engineering may hold the keys to more efficient inference with transformer-like models.
神经形态工程和机器学习之间仍然存在很大的障碍,特别是在最近的大型语言模型(llm)和变形器方面。这篇评论表明,神经形态工程可能是对类似变形器的模型进行更有效推理的关键。
{"title":"Neuromorphic principles in self-attention hardware for efficient transformers","authors":"Nathan Leroux, Jan Finkbeiner, Emre Neftci","doi":"10.1038/s43588-025-00868-9","DOIUrl":"10.1038/s43588-025-00868-9","url":null,"abstract":"Strong barriers remain between neuromorphic engineering and machine learning, especially with regard to recent large language models (LLMs) and transformers. This Comment makes the case that neuromorphic engineering may hold the keys to more efficient inference with transformer-like models.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 9","pages":"708-710"},"PeriodicalIF":18.3,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145076704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the compatibility of generative AI and generative linguistics 论生成人工智能与生成语言学的兼容性。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-16 DOI: 10.1038/s43588-025-00861-2
Eva Portelance, Masoud Jasbi
Chomsky’s generative linguistics has made substantial contributions to cognitive science and symbolic artificial intelligence. With the rise of neural language models, however, the compatibility between generative artificial intelligence and generative linguistics has come under debate. Here we outline three ways in which generative artificial intelligence aligns with and supports the core ideas of generative linguistics. In turn, generative linguistics can provide criteria to evaluate and improve neural language models as models of human language and cognition. This Perspective discusses that generative AI aligns with generative linguistics by showing that neural language models (NLMs) are formal generative models. Furthermore, generative linguistics offers a framework for evaluating and improving NLMs.
乔姆斯基的生成语言学对认知科学和符号人工智能做出了重大贡献。然而,随着神经语言模型的兴起,生成式人工智能和生成式语言学之间的兼容性已经成为争论的焦点。在这里,我们概述了生成人工智能与生成语言学的核心思想相一致和支持的三种方式。反过来,生成语言学可以为评估和改进作为人类语言和认知模型的神经语言模型提供标准。
{"title":"On the compatibility of generative AI and generative linguistics","authors":"Eva Portelance, Masoud Jasbi","doi":"10.1038/s43588-025-00861-2","DOIUrl":"10.1038/s43588-025-00861-2","url":null,"abstract":"Chomsky’s generative linguistics has made substantial contributions to cognitive science and symbolic artificial intelligence. With the rise of neural language models, however, the compatibility between generative artificial intelligence and generative linguistics has come under debate. Here we outline three ways in which generative artificial intelligence aligns with and supports the core ideas of generative linguistics. In turn, generative linguistics can provide criteria to evaluate and improve neural language models as models of human language and cognition. This Perspective discusses that generative AI aligns with generative linguistics by showing that neural language models (NLMs) are formal generative models. Furthermore, generative linguistics offers a framework for evaluating and improving NLMs.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 9","pages":"745-753"},"PeriodicalIF":18.3,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145076692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Increasing alignment of large language models with language processing in the human brain 大型语言模型与人类大脑中语言处理的日益一致。
IF 18.3 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2025-09-16 DOI: 10.1038/s43588-025-00863-0
Changjiang Gao, Zhengwu Ma, Jiajun Chen, Ping Li, Shujian Huang, Jixing Li
Transformer-based large language models (LLMs) have considerably advanced our understanding of how meaning is represented in the human brain; however, the validity of increasingly large LLMs is being questioned due to their extensive training data and their ability to access context thousands of words long. In this study we investigated whether instruction tuning—another core technique in recent LLMs that goes beyond mere scaling—can enhance models’ ability to capture linguistic information in the human brain. We compared base and instruction-tuned LLMs of varying sizes against human behavioral and brain activity measured with eye-tracking and functional magnetic resonance imaging during naturalistic reading. We show that simply making LLMs larger leads to a closer match with the human brain than fine-tuning them with instructions. These finding have substantial implications for understanding the cognitive plausibility of LLMs and their role in studying naturalistic language comprehension. Larger LLMs’ self-attention more accurately predicts readers’ regressive saccades and fMRI responses in language regions, whereas instruction tuning adds no benefit.
基于变压器的大型语言模型(llm)大大提高了我们对人类大脑中意义如何表示的理解;然而,越来越多的大型法学硕士的有效性受到质疑,因为它们有大量的训练数据和数千字长的上下文访问能力。在这项研究中,我们调查了指令调整——最近法学硕士的另一项核心技术,超越了单纯的缩放——是否能提高模型在人脑中捕捉语言信息的能力。在自然阅读过程中,我们通过眼动追踪和功能性磁共振成像来测量人类的行为和大脑活动。我们表明,简单地将llm变大比用指令对其进行微调更接近人类大脑。这些发现对于理解法学硕士的认知合理性及其在研究自然语言理解中的作用具有重要意义。
{"title":"Increasing alignment of large language models with language processing in the human brain","authors":"Changjiang Gao, Zhengwu Ma, Jiajun Chen, Ping Li, Shujian Huang, Jixing Li","doi":"10.1038/s43588-025-00863-0","DOIUrl":"10.1038/s43588-025-00863-0","url":null,"abstract":"Transformer-based large language models (LLMs) have considerably advanced our understanding of how meaning is represented in the human brain; however, the validity of increasingly large LLMs is being questioned due to their extensive training data and their ability to access context thousands of words long. In this study we investigated whether instruction tuning—another core technique in recent LLMs that goes beyond mere scaling—can enhance models’ ability to capture linguistic information in the human brain. We compared base and instruction-tuned LLMs of varying sizes against human behavioral and brain activity measured with eye-tracking and functional magnetic resonance imaging during naturalistic reading. We show that simply making LLMs larger leads to a closer match with the human brain than fine-tuning them with instructions. These finding have substantial implications for understanding the cognitive plausibility of LLMs and their role in studying naturalistic language comprehension. Larger LLMs’ self-attention more accurately predicts readers’ regressive saccades and fMRI responses in language regions, whereas instruction tuning adds no benefit.","PeriodicalId":74246,"journal":{"name":"Nature computational science","volume":"5 11","pages":"1080-1090"},"PeriodicalIF":18.3,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.comhttps://www.nature.com/articles/s43588-025-00863-0.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145076701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Nature computational science
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1