首页 > 最新文献

Cognitive Neuroscience最新文献

英文 中文
Non-Markovian systems, phenomenology, and the challenges of capturing meaning and context - comment on Parr, Pezzulo, and Friston (2025). 非马尔可夫系统,现象学,以及捕捉意义和上下文的挑战——评论Parr, Pezzulo, and Friston(2025)。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-07-04 DOI: 10.1080/17588928.2025.2523889
Mahault Albarracin, Dalton A R Sakthivadivel

Parr, et al., explore the problem of non-Markovian pro cesses, in which the future state of a system depends not only on its present state but also on its past states. The authors suggest that the success of transformer networks in dealing with sequential data, such as language, stems from their ability to address this non-Markovian nature through the use of attention mechanisms. This commentary builds on their discussion, aiming to link it to some notions in Husserlian phenomenology and explore the implications for understanding meaning, context, and the nature of knowledge.

Parr等人探讨了非马尔可夫过程的问题,其中系统的未来状态不仅取决于它的当前状态,还取决于它的过去状态。作者认为,变压器网络在处理顺序数据(如语言)方面的成功源于它们通过使用注意力机制来处理这种非马尔可夫性质的能力。本评论以他们的讨论为基础,旨在将其与胡塞尔现象学中的一些概念联系起来,并探索理解意义、背景和知识本质的含义。
{"title":"Non-Markovian systems, phenomenology, and the challenges of capturing meaning and context - comment on Parr, Pezzulo, and Friston (2025).","authors":"Mahault Albarracin, Dalton A R Sakthivadivel","doi":"10.1080/17588928.2025.2523889","DOIUrl":"10.1080/17588928.2025.2523889","url":null,"abstract":"<p><p>Parr, et al., explore the problem of non-Markovian pro cesses, in which the future state of a system depends not only on its present state but also on its past states. The authors suggest that the success of transformer networks in dealing with sequential data, such as language, stems from their ability to address this non-Markovian nature through the use of attention mechanisms. This commentary builds on their discussion, aiming to link it to some notions in Husserlian phenomenology and explore the implications for understanding meaning, context, and the nature of knowledge.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"35-36"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144559411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ROSE: A Universal Neural Grammar. 柔丝:通用神经语法。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-07-14 DOI: 10.1080/17588928.2025.2523875
Elliot Murphy

Processing natural language syntax requires a negotiation between symbolic and subsymbolic representations. Building on the recent representation, operation, structure, encoding (ROSE) neurocomputational architecture for syntax that scales from single units to inter-areal dynamics, I discuss the prospects of reconciling the neural code for hierarchical syntax with predictive processes. Here, the higher levels of ROSE provide instructions for symbolic phrase structure representations (S/E), while the lower levels provide probabilistic aspects of linguistic processing (R/O), with different types of cross-frequency coupling being hypothesized to interface these domains. I argue that ROSE provides a possible infrastructure for flexibly implementing distinct types of minimalist grammar parsers for the real-time processing of language. This perspective helps furnish a more restrictive 'core language network' in the brain than contemporary approaches that isolate general sentence composition. I define the language network as being critically involved in executing specific parsing operations (i.e. establishing phrasal categories, tree-structure depth, resolving dependencies, and retrieving proprietary lexical representations), capturing these network-defining operations jointly with probabilistic aspects of parsing. ROSE offers a 'mesoscopic protectorate' for natural language; an intermediate level of emergent organizational complexity that demands multi-scale modeling. By drawing principled relations across computational, algorithmic and implementational Marrian levels, ROSE offers new constraints on what a unified neurocomputational settlement for natural language syntax might look like, providing a tentative scaffold for a 'Universal Neural Grammar' - a species-specific format for neurally organizing the construction of compositional syntactic structures, which matures in accordance with a genetically determined biological matrix.

处理自然语言语法需要在符号表示和子符号表示之间进行协商。在最近的表示、操作、结构、编码(ROSE)神经计算架构的基础上,我讨论了将层次语法的神经代码与预测过程协调起来的前景。在这里,更高级别的ROSE提供符号短语结构表示(S/E)的指令,而较低级别提供语言处理(R/O)的概率方面,并假设不同类型的交叉频率耦合来连接这些域。我认为ROSE提供了一种可能的基础设施,可以灵活地实现用于语言实时处理的不同类型的极简语法解析器。这种观点有助于在大脑中提供一个更严格的“核心语言网络”,而不是孤立一般句子组成的当代方法。我将语言网络定义为关键地参与执行特定的解析操作(例如,建立短语类别、树结构深度、解析依赖关系和检索专有词汇表示),并将这些网络定义操作与解析的概率方面结合起来。ROSE为自然语言提供了一个“中观保护国”;需要多尺度建模的紧急组织复杂性的中间层次。通过在计算、算法和实现的marian水平上绘制原则关系,ROSE为自然语言语法的统一神经计算解决方案提供了新的约束,为“通用神经语法”提供了一个初步的框架,“通用神经语法”是一种特定于物种的格式,用于神经组织组合句法结构的构建,它根据基因决定的生物矩阵成熟。
{"title":"ROSE: A Universal Neural Grammar.","authors":"Elliot Murphy","doi":"10.1080/17588928.2025.2523875","DOIUrl":"10.1080/17588928.2025.2523875","url":null,"abstract":"<p><p>Processing natural language syntax requires a negotiation between symbolic and subsymbolic representations. Building on the recent representation, operation, structure, encoding (ROSE) neurocomputational architecture for syntax that scales from single units to inter-areal dynamics, I discuss the prospects of reconciling the neural code for hierarchical syntax with predictive processes. Here, the higher levels of ROSE provide instructions for symbolic phrase structure representations (S/E), while the lower levels provide probabilistic aspects of linguistic processing (R/O), with different types of cross-frequency coupling being hypothesized to interface these domains. I argue that ROSE provides a possible infrastructure for flexibly implementing distinct types of minimalist grammar parsers for the real-time processing of language. This perspective helps furnish a more restrictive 'core language network' in the brain than contemporary approaches that isolate general sentence composition. I define the language network as being critically involved in executing specific parsing operations (i.e. establishing phrasal categories, tree-structure depth, resolving dependencies, and retrieving proprietary lexical representations), capturing these network-defining operations jointly with probabilistic aspects of parsing. ROSE offers a 'mesoscopic protectorate' for natural language; an intermediate level of emergent organizational complexity that demands multi-scale modeling. By drawing principled relations across computational, algorithmic and implementational Marrian levels, ROSE offers new constraints on what a unified neurocomputational settlement for natural language syntax might look like, providing a tentative scaffold for a 'Universal Neural Grammar' - a species-specific format for neurally organizing the construction of compositional syntactic structures, which matures in accordance with a genetically determined biological matrix.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"49-80"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144625415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Auditory facilitation in deterministic versus stochastic worlds. 确定性与随机世界中的听觉促进。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-04-29 DOI: 10.1080/17588928.2025.2497762
Berfin Bastug, Urte Roeber, Erich Schröger

The brain learns statistical regularities in sensory sequences, enhancing behavioral performance for predictable stimuli while impairing behavioral performance for unpredictable stimuli. While previous research has shown that violations of non-informative regularities hinder task performance, it remains unclear whether predictable but task-irrelevant structures can facilitate performance. In a tone duration discrimination task, we manipulated the task-irrelevant pitch dimension by varying transition probabilities (TP) between successive tone frequencies. Participants judged duration, while pitch sequences were either deterministic (a rule-governed pitch pattern, TP = 1) or stochastic (no discernible pitch pattern, TP = 1/number of pitch levels). The tone pitch was task-irrelevant and it did not predict duration. Results showed that reaction times (RTs) were significantly faster for deterministic sequences, suggesting that predictability in a task-irrelevant dimension still facilitates task performance. RTs were also faster in two-tone sequences compared to eight-tone sequences, likely due to reduced memory load. These findings suggest that statistical learning benefits extend beyond task-relevant dimensions, supporting a predictive coding framework in which the brain integrates predictable sensory input to optimize cognitive processing.

大脑在感觉序列中学习统计规律,在可预测的刺激下增强行为表现,而在不可预测的刺激下削弱行为表现。虽然先前的研究表明,违反非信息性规则会阻碍任务绩效,但目前尚不清楚可预测但与任务无关的结构是否能促进绩效。在音时识别任务中,我们通过改变连续音调频率之间的过渡概率来操纵任务无关音高维度。参与者判断持续时间,而音高序列要么是确定性的(规则控制的音高模式,TP = 1),要么是随机的(没有可辨别的音高模式,TP = 1/音高水平的数量)。音调与任务无关,也不能预测持续时间。结果表明,确定性序列的反应时间(RTs)明显更快,这表明在任务无关维度的可预测性仍然有助于任务表现。RTs在双音序列中也比在八音序列中更快,这可能是由于内存负载减少。这些发现表明,统计学习的好处超出了任务相关的维度,支持了一个预测编码框架,在这个框架中,大脑整合了可预测的感觉输入,以优化认知处理。
{"title":"Auditory facilitation in deterministic versus stochastic worlds.","authors":"Berfin Bastug, Urte Roeber, Erich Schröger","doi":"10.1080/17588928.2025.2497762","DOIUrl":"10.1080/17588928.2025.2497762","url":null,"abstract":"<p><p>The brain learns statistical regularities in sensory sequences, enhancing behavioral performance for predictable stimuli while impairing behavioral performance for unpredictable stimuli. While previous research has shown that violations of non-informative regularities hinder task performance, it remains unclear whether predictable but task-irrelevant structures can facilitate performance. In a tone duration discrimination task, we manipulated the task-irrelevant pitch dimension by varying transition probabilities (TP) between successive tone frequencies. Participants judged duration, while pitch sequences were either deterministic (a rule-governed pitch pattern, TP = 1) or stochastic (no discernible pitch pattern, TP = 1/number of pitch levels). The tone pitch was task-irrelevant and it did not predict duration. Results showed that reaction times (RTs) were significantly faster for deterministic sequences, suggesting that predictability in a task-irrelevant dimension still facilitates task performance. RTs were also faster in two-tone sequences compared to eight-tone sequences, likely due to reduced memory load. These findings suggest that statistical learning benefits extend beyond task-relevant dimensions, supporting a predictive coding framework in which the brain integrates predictable sensory input to optimize cognitive processing.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"93-99"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143969935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond individuals: Collective predictive coding for memory, attention, and the emergence of language. 超越个体:记忆、注意力和语言出现的集体预测编码。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-06-17 DOI: 10.1080/17588928.2025.2518942
Tadahiro Taniguchi
{"title":"Beyond individuals: Collective predictive coding for memory, attention, and the emergence of language.","authors":"Tadahiro Taniguchi","doi":"10.1080/17588928.2025.2518942","DOIUrl":"10.1080/17588928.2025.2518942","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"41-42"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144316040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How the brain recycled memory circuits for language: An evolutionary perspective on the ROSE model. 大脑如何为语言循环记忆回路:ROSE模型的进化视角。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-09-18 DOI: 10.1080/17588928.2025.2561587
Edward Ruoyang Shi
{"title":"How the brain recycled memory circuits for language: An evolutionary perspective on the ROSE model.","authors":"Edward Ruoyang Shi","doi":"10.1080/17588928.2025.2561587","DOIUrl":"10.1080/17588928.2025.2561587","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"88-90"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145079817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond prediction: comments on the format of natural intelligence. 超越预测:评论自然智能的形式。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-06-18 DOI: 10.1080/17588928.2025.2521403
Elliot Murphy
{"title":"Beyond prediction: comments on the format of natural intelligence.","authors":"Elliot Murphy","doi":"10.1080/17588928.2025.2521403","DOIUrl":"10.1080/17588928.2025.2521403","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"37-40"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144324646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamical motifs for computations in language. 语言计算的动态基元。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-10-30 DOI: 10.1080/17588928.2025.2581573
Katarína Labancová, Nina Kazanina
{"title":"Dynamical motifs for computations in language.","authors":"Katarína Labancová, Nina Kazanina","doi":"10.1080/17588928.2025.2581573","DOIUrl":"10.1080/17588928.2025.2581573","url":null,"abstract":"","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"83-84"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145400132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictive coding of cognitive processes in natural and artificial systems. 自然和人工系统中认知过程的预测编码。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-11-17 DOI: 10.1080/17588928.2025.2584209
Joseph B Hopfinger, Scott D Slotnick

With recent developments in artificial intelligence (AI), there is great interest in how mechanisms of human cognitive processing may be instantiated in those models and how those models may help us better understand human cognitive and neural processes. Recent research suggests predictive coding theories and associated generative models may help explain the processes of visual perception and language production, while newer AI models include mechanisms akin to human memory and attention. This special issue of Cognitive Neuroscience: Current Debates, Research & Reports presents 16 new papers that highlight important topics and present exciting new data, models, and controversies. The articles include a new discussion paper by Parr, Pezzulo, and Friston exploring how transformer architectures utilize non-Markovian generative models and how an attention-like process is critical for processing complex sequential data. This is followed by seven insightful commentaries and a reply from the authors. A discussion paper on a new neurocomputational model of syntax is provided by Murphy, in which predictive processes are integrated in a multi-level, hierarchical syntax architecture. This is followed by five commentaries suggesting important evolutionary and developmental perspectives and ways to explore and test the model. Finally, an empirical article by Bastug, Roeber, and Schröger on auditory perception presents new evidence suggesting that distracting information requires less cognitive processing when it is predictable. The topics of this special issue are evolving rapidly and promise to be at the heart of future developments in artificial learning systems and theories of the brain mechanisms that mediate cognitive processes.

随着人工智能(AI)的最新发展,人们对如何在这些模型中实例化人类认知处理的机制以及这些模型如何帮助我们更好地理解人类认知和神经过程产生了极大的兴趣。最近的研究表明,预测编码理论和相关的生成模型可能有助于解释视觉感知和语言产生的过程,而较新的人工智能模型包括类似于人类记忆和注意力的机制。本期《认知神经科学:当前的争论、研究和报告》特刊介绍了16篇新的论文,这些论文突出了重要的主题,并提出了令人兴奋的新数据、模型和争议。这些文章包括Parr、Pezzulo和Friston的一篇新的讨论论文,探讨了变压器架构如何利用非马尔可夫生成模型,以及类注意力过程如何对处理复杂的顺序数据至关重要。接下来是七篇深刻的评论和作者的回复。Murphy提供了一种新的神经计算语法模型,其中预测过程集成在一个多层次、分层的语法体系结构中。接下来是五篇评论,提出了重要的进化和发展观点以及探索和测试模型的方法。最后,Bastug, Roeber和Schröger关于听觉感知的一篇实证文章提出了新的证据,表明当可预测的信息分散注意力时,需要较少的认知处理。本期特刊的主题正在迅速发展,并有望成为人工学习系统和调解认知过程的大脑机制理论未来发展的核心。
{"title":"Predictive coding of cognitive processes in natural and artificial systems.","authors":"Joseph B Hopfinger, Scott D Slotnick","doi":"10.1080/17588928.2025.2584209","DOIUrl":"https://doi.org/10.1080/17588928.2025.2584209","url":null,"abstract":"<p><p>With recent developments in artificial intelligence (AI), there is great interest in how mechanisms of human cognitive processing may be instantiated in those models and how those models may help us better understand human cognitive and neural processes. Recent research suggests predictive coding theories and associated generative models may help explain the processes of visual perception and language production, while newer AI models include mechanisms akin to human memory and attention. This special issue of <i>Cognitive Neuroscience: Current Debates, Research & Reports</i> presents 16 new papers that highlight important topics and present exciting new data, models, and controversies. The articles include a new discussion paper by Parr, Pezzulo, and Friston exploring how transformer architectures utilize non-Markovian generative models and how an attention-like process is critical for processing complex sequential data. This is followed by seven insightful commentaries and a reply from the authors. A discussion paper on a new neurocomputational model of syntax is provided by Murphy, in which predictive processes are integrated in a multi-level, hierarchical syntax architecture. This is followed by five commentaries suggesting important evolutionary and developmental perspectives and ways to explore and test the model. Finally, an empirical article by Bastug, Roeber, and Schröger on auditory perception presents new evidence suggesting that distracting information requires less cognitive processing when it is predictable. The topics of this special issue are evolving rapidly and promise to be at the heart of future developments in artificial learning systems and theories of the brain mechanisms that mediate cognitive processes.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":"16 1-4","pages":"1-4"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145539234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Embeddings as Dirichlet counts: Attention is the tip of the iceberg. Dirichlet计算的嵌入:注意力是冰山一角。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-07-09 DOI: 10.1080/17588928.2025.2530430
Alexander Bernard Kiefer

Despite the overtly discrete nature of language, the use of semantic embedding spaces is pervasive in modern computational linguistics and machine learning for natural language. I argue that this is intelligible if language is viewed as an interface into a general-purpose system of concepts, in which metric spaces capture rich relationships. At the same time, language embeddings can be regarded, at least heuristically, as equivalent to parameters of distributions over word-word relationships.

尽管语言具有明显的离散性,但语义嵌入空间的使用在现代计算语言学和自然语言的机器学习中普遍存在。我认为,如果把语言看作是进入通用概念系统的接口,那么这是可以理解的,在这个系统中,度量空间捕获了丰富的关系。同时,语言嵌入可以被视为,至少在启发式上,相当于词-词关系上分布的参数。
{"title":"Embeddings as Dirichlet counts: Attention is the tip of the iceberg.","authors":"Alexander Bernard Kiefer","doi":"10.1080/17588928.2025.2530430","DOIUrl":"10.1080/17588928.2025.2530430","url":null,"abstract":"<p><p>Despite the overtly discrete nature of language, the use of semantic embedding spaces is pervasive in modern computational linguistics and machine learning for natural language. I argue that this is intelligible if language is viewed as an interface into a general-purpose system of concepts, in which metric spaces capture rich relationships. At the same time, language embeddings can be regarded, at least heuristically, as equivalent to parameters of distributions over word-word relationships.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"29-31"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144590540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dissociating model architectures from inference computations. 从推理计算中分离模型架构。
IF 2.2 4区 医学 Q3 NEUROSCIENCES Pub Date : 2025-01-01 Epub Date: 2025-07-17 DOI: 10.1080/17588928.2025.2532604
Noor Sajid, Johan Medrano

Parr et al., 2025 examines how auto-regressive and deep temporal models differ in their treatment of non-Markovian sequence modelling. Building on this, we highlight the need for dissociating model architectures-i.e., how the predictive distribution factorises-from the computations invoked at inference. We demonstrate that deep temporal computations are mimicked by autoregressive models by structuring context access during iterative inference. Using a transformer trained on next-token prediction, we show that inducing hierarchical temporal factorisation during iterative inference maintains predictive capacity while instantiating fewer computations. This emphasises that processes for constructing and refining predictions are not necessarily bound to their underlying model architectures.

Parr等人,2025研究了自回归模型和深度时间模型在处理非马尔可夫序列模型方面的差异。在此基础上,我们强调了分离模型体系结构的必要性。预测分布是如何从推理中调用的计算中分解出来的。我们证明了深度时间计算是由自回归模型通过在迭代推理过程中构建上下文访问来模拟的。使用在下一个令牌预测上训练的转换器,我们表明在迭代推理期间诱导分层时间分解保持预测能力,同时实例化更少的计算。这强调了构建和精炼预测的过程不一定绑定到它们的底层模型体系结构。
{"title":"Dissociating model architectures from inference computations.","authors":"Noor Sajid, Johan Medrano","doi":"10.1080/17588928.2025.2532604","DOIUrl":"10.1080/17588928.2025.2532604","url":null,"abstract":"<p><p>Parr et al., 2025 examines how auto-regressive and deep temporal models differ in their treatment of non-Markovian sequence modelling. Building on this, we highlight the need for dissociating model architectures-i.e., how the predictive distribution factorises-from the computations invoked at inference. We demonstrate that deep temporal computations are mimicked by autoregressive models by structuring context access during iterative inference. Using a transformer trained on next-token prediction, we show that inducing hierarchical temporal factorisation during iterative inference maintains predictive capacity while instantiating fewer computations. This emphasises that processes for constructing and refining predictions are not necessarily bound to their underlying model architectures.</p>","PeriodicalId":10413,"journal":{"name":"Cognitive Neuroscience","volume":" ","pages":"26-28"},"PeriodicalIF":2.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144648756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Cognitive Neuroscience
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1