首页 > 最新文献

Neural Computation最新文献

英文 中文
Decision Threshold Learning in the Basal Ganglia for Multiple Alternatives 多选项下基底神经节的决策阈值学习。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-17 DOI: 10.1162/neco_a_01760
Thom Griffith;Sophie-Anne Baker;Nathan F. Lepora
In recent years, researchers have integrated the historically separate, reinforcement learning (RL), and evidence-accumulation-to-bound approaches to decision modeling. A particular outcome of these efforts has been the RL-DDM, a model that combines value learning through reinforcement with a diffusion decision model (DDM). While the RL-DDM is a conceptually elegant extension of the original DDM, it faces a similar problem to the DDM in that it does not scale well to decisions with more than two options. Furthermore, in its current form, the RL-DDM lacks flexibility when it comes to adapting to rapid, context-cued changes in the reward environment. The question of how to best extend combined RL and DDM models so they can handle multiple choices remains open. Moreover, it is currently unclear how these algorithmic solutions should map to neurophysical processes in the brain, particularly in relation to so-called go/no-go-type models of decision making in the basal ganglia. Here, we propose a solution that addresses these issues by combining a previously proposed decision model based on the multichoice sequential probability ratio test (MSPRT), with a dual-pathway model of decision threshold learning in the basal ganglia region of the brain. Our model learns decision thresholds to optimize the trade-off between time cost and the cost of errors and so efficiently allocates the amount of time for decision deliberation. In addition, the model is context dependent and hence flexible to changes to the speed-accuracy trade-off (SAT) in the environment. Furthermore, the model reproduces the magnitude effect, a phenomenon seen experimentally in value-based decisions and is agnostic to the types of evidence and so can be used on perceptual decisions, value-based decisions, and other types of modeled evidence. The broader significance of the model is that it contributes to the active research area of how learning systems interact by linking the previously separate models of RL-DDM to dopaminergic models of motivation and risk taking in the basal ganglia, as well as scaling to multiple alternatives.
近年来,研究人员将历史上分离的强化学习(RL)和证据积累到边界的方法集成到决策建模中。这些努力的一个特别成果是RL-DDM,一个将价值学习通过强化与扩散决策模型(DDM)相结合的模型。虽然RL-DDM在概念上是原始DDM的优雅扩展,但它面临着与DDM类似的问题,即它不能很好地扩展到具有两个以上选项的决策。此外,在目前的形式下,RL-DDM在适应奖励环境的快速、情境变化方面缺乏灵活性。如何最好地扩展RL和DDM组合模型,使它们能够处理多种选择的问题仍然没有解决。此外,目前尚不清楚这些算法解决方案如何映射到大脑中的神经物理过程,特别是与基底神经节中所谓的go/no-go型决策模型有关。在这里,我们提出了一种解决方案,通过将先前提出的基于多选择顺序概率比检验(MSPRT)的决策模型与大脑基底神经节区域的决策阈值学习双通路模型相结合来解决这些问题。我们的模型通过学习决策阈值来优化时间成本和错误成本之间的权衡,从而有效地分配决策审议的时间。此外,该模型依赖于上下文,因此可以灵活地适应环境中速度-精度权衡(SAT)的变化。此外,该模型再现了量级效应,这是一种在基于价值的决策中实验观察到的现象,与证据类型无关,因此可以用于感知决策、基于价值的决策和其他类型的建模证据。该模型更广泛的意义在于,它通过将之前分离的RL-DDM模型与基底神经节中动机和风险承担的多巴胺能模型联系起来,以及扩展到多个替代模型,为学习系统如何相互作用的活跃研究领域做出了贡献。
{"title":"Decision Threshold Learning in the Basal Ganglia for Multiple Alternatives","authors":"Thom Griffith;Sophie-Anne Baker;Nathan F. Lepora","doi":"10.1162/neco_a_01760","DOIUrl":"10.1162/neco_a_01760","url":null,"abstract":"In recent years, researchers have integrated the historically separate, reinforcement learning (RL), and evidence-accumulation-to-bound approaches to decision modeling. A particular outcome of these efforts has been the RL-DDM, a model that combines value learning through reinforcement with a diffusion decision model (DDM). While the RL-DDM is a conceptually elegant extension of the original DDM, it faces a similar problem to the DDM in that it does not scale well to decisions with more than two options. Furthermore, in its current form, the RL-DDM lacks flexibility when it comes to adapting to rapid, context-cued changes in the reward environment. The question of how to best extend combined RL and DDM models so they can handle multiple choices remains open. Moreover, it is currently unclear how these algorithmic solutions should map to neurophysical processes in the brain, particularly in relation to so-called go/no-go-type models of decision making in the basal ganglia. Here, we propose a solution that addresses these issues by combining a previously proposed decision model based on the multichoice sequential probability ratio test (MSPRT), with a dual-pathway model of decision threshold learning in the basal ganglia region of the brain. Our model learns decision thresholds to optimize the trade-off between time cost and the cost of errors and so efficiently allocates the amount of time for decision deliberation. In addition, the model is context dependent and hence flexible to changes to the speed-accuracy trade-off (SAT) in the environment. Furthermore, the model reproduces the magnitude effect, a phenomenon seen experimentally in value-based decisions and is agnostic to the types of evidence and so can be used on perceptual decisions, value-based decisions, and other types of modeled evidence. The broader significance of the model is that it contributes to the active research area of how learning systems interact by linking the previously separate models of RL-DDM to dopaminergic models of motivation and risk taking in the basal ganglia, as well as scaling to multiple alternatives.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 7","pages":"1256-1287"},"PeriodicalIF":2.7,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144163728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Survey on Artificial Neural Networks in Human—Robot Interaction 人工神经网络在人机交互中的研究进展。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-17 DOI: 10.1162/neco_a_01764
Aleksandra Świetlicka
Artificial neural networks (ANNs) have shown great potential in enhancing human-robot interaction (HRI). ANNs are computational models inspired by the structure and function of biological neural networks in the brain, which can learn from examples and generalize to new situations. ANNs can be used to enable robots to interact with humans in a more natural and intuitive way by allowing them to recognize human gestures and expressions, understand natural language, and adapt to the environment. ANNs can also be used to improve robot autonomy, allowing robots to learn from their interactions with humans and to make more informed decisions. However, there are also challenges to using ANNs in HRI, including the need for large amounts of training data, issues with explainability, and the potential for bias. This review explores the current state of research on ANNs in HRI, highlighting both the opportunities and challenges of this approach and discussing potential directions for future research. The AI contribution involves applying ANNs to various aspects of HRI, while the application in engineering involves using ANNs to develop more interactive and intuitive robotic systems.
人工神经网络(ann)在增强人机交互(HRI)方面显示出巨大的潜力。人工神经网络是受大脑中生物神经网络的结构和功能启发的计算模型,它可以从例子中学习并推广到新的情况。人工神经网络可以通过让机器人识别人类的手势和表情,理解自然语言,并适应环境,从而使机器人以更自然、更直观的方式与人类互动。人工神经网络还可用于提高机器人的自主性,使机器人能够从与人类的互动中学习,并做出更明智的决定。然而,在人力资源研究中使用人工神经网络也存在挑战,包括需要大量的训练数据、可解释性问题以及潜在的偏见。本文探讨了人工神经网络在HRI中的研究现状,强调了这种方法的机遇和挑战,并讨论了未来研究的潜在方向。人工智能的贡献包括将人工神经网络应用于人力资源研究所的各个方面,而在工程上的应用则涉及使用人工神经网络开发更具交互性和直觉性的机器人系统。
{"title":"A Survey on Artificial Neural Networks in Human—Robot Interaction","authors":"Aleksandra Świetlicka","doi":"10.1162/neco_a_01764","DOIUrl":"10.1162/neco_a_01764","url":null,"abstract":"Artificial neural networks (ANNs) have shown great potential in enhancing human-robot interaction (HRI). ANNs are computational models inspired by the structure and function of biological neural networks in the brain, which can learn from examples and generalize to new situations. ANNs can be used to enable robots to interact with humans in a more natural and intuitive way by allowing them to recognize human gestures and expressions, understand natural language, and adapt to the environment. ANNs can also be used to improve robot autonomy, allowing robots to learn from their interactions with humans and to make more informed decisions. However, there are also challenges to using ANNs in HRI, including the need for large amounts of training data, issues with explainability, and the potential for bias. This review explores the current state of research on ANNs in HRI, highlighting both the opportunities and challenges of this approach and discussing potential directions for future research. The AI contribution involves applying ANNs to various aspects of HRI, while the application in engineering involves using ANNs to develop more interactive and intuitive robotic systems.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 7","pages":"1193-1255"},"PeriodicalIF":2.7,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144163722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Closed-Loop Multistep Planning 闭环多步骤规划。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-17 DOI: 10.1162/neco_a_01761
Giulia Lafratta;Bernd Porr;Christopher Chandler;Alice Miller
Living organisms interact with their surroundings in a closed-loop fashion, where sensory inputs dictate the initiation and termination of behaviors. Even simple animals are able to develop and execute complex plans, which has not yet been replicated in robotics using pure closed-loop input control. We propose a solution to this problem by defining a set of discrete and temporary closed-loop controllers, called “Tasks,” each representing a closed-loop behavior. We further introduce a supervisory module that has an innate understanding of physics and causality, through which it can simulate the execution of Task sequences over time and store the results in a model of the environment. On the basis of this model, plans can be made by chaining temporary closed-loop controllers. Our proposed framework was implemented for a robot and tested in two scenarios as proof of concept.
生物体以闭环方式与周围环境相互作用,其中感官输入决定行为的开始和结束。即使是简单的动物也能够制定和执行复杂的计划,这在使用纯闭环输入控制的机器人中还没有被复制。我们提出了一种解决方案,通过定义一组离散和临时的闭环控制器,称为“任务”,每个任务代表一个闭环行为。我们进一步介绍了一个对物理和因果关系具有天生理解的监督模块,通过它可以模拟任务序列随时间的执行,并将结果存储在环境模型中。在此模型的基础上,可以通过串联临时闭环控制器来制定计划。我们提出的框架在机器人上实现,并在两个场景中进行了测试,作为概念验证。
{"title":"Closed-Loop Multistep Planning","authors":"Giulia Lafratta;Bernd Porr;Christopher Chandler;Alice Miller","doi":"10.1162/neco_a_01761","DOIUrl":"10.1162/neco_a_01761","url":null,"abstract":"Living organisms interact with their surroundings in a closed-loop fashion, where sensory inputs dictate the initiation and termination of behaviors. Even simple animals are able to develop and execute complex plans, which has not yet been replicated in robotics using pure closed-loop input control. We propose a solution to this problem by defining a set of discrete and temporary closed-loop controllers, called “Tasks,” each representing a closed-loop behavior. We further introduce a supervisory module that has an innate understanding of physics and causality, through which it can simulate the execution of Task sequences over time and store the results in a model of the environment. On the basis of this model, plans can be made by chaining temporary closed-loop controllers. Our proposed framework was implemented for a robot and tested in two scenarios as proof of concept.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 7","pages":"1288-1319"},"PeriodicalIF":2.7,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144163725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Excitation–Inhibition Balance Controls Synchronization in a Simple Model of Coupled Phase Oscillators 激励-抑制平衡控制耦合相位振荡器的简单模型同步。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-17 DOI: 10.1162/neco_a_01763
Satoshi Kuroki;Kenji Mizuseki
Collective neuronal activity in the brain synchronizes during rest and desynchronizes during active behaviors, influencing cognitive processes such as memory consolidation, knowledge abstraction, and creative thinking. These states involve significant modulation of inhibition, which alters the excitation–inhibition (EI) balance of synaptic inputs. However, the influence of the EI balance on collective neuronal oscillation remains only partially understood. In this study, we introduce the EI-Kuramoto model, a modified version of the Kuramoto model, in which oscillators are categorized into excitatory and inhibitory groups with four distinct interaction types: excitatory–excitatory, excitatory–inhibitory, inhibitory–excitatory, and inhibitory–inhibitory. Numerical simulations identify three dynamic states—synchronized, bistable, and desynchronized—that can be controlled by adjusting the strength of the four interaction types. Theoretical analysis further demonstrates that the balance among these interactions plays a critical role in determining the dynamic states. This study provides valuable insights into the role of EI balance in synchronizing coupled oscillators and neurons.
大脑中的集体神经元活动在休息时同步,在活动时不同步,影响记忆巩固、知识抽象和创造性思维等认知过程。这些状态涉及抑制的显著调节,这改变了突触输入的兴奋-抑制(EI)平衡。然而,EI平衡对集体神经元振荡的影响尚不完全清楚。在本研究中,我们引入了EI-Kuramoto模型,这是Kuramoto模型的改进版本,其中振荡子被分为兴奋性和抑制性组,具有四种不同的相互作用类型:兴奋-兴奋、兴奋-抑制、抑制-兴奋和抑制-抑制。数值模拟确定了同步、双稳态和非同步三种动态状态,这三种动态状态可以通过调整四种相互作用类型的强度来控制。理论分析进一步表明,这些相互作用之间的平衡在决定动态状态方面起着关键作用。这项研究为EI平衡在同步耦合振荡器和神经元中的作用提供了有价值的见解。
{"title":"Excitation–Inhibition Balance Controls Synchronization in a Simple Model of Coupled Phase Oscillators","authors":"Satoshi Kuroki;Kenji Mizuseki","doi":"10.1162/neco_a_01763","DOIUrl":"10.1162/neco_a_01763","url":null,"abstract":"Collective neuronal activity in the brain synchronizes during rest and desynchronizes during active behaviors, influencing cognitive processes such as memory consolidation, knowledge abstraction, and creative thinking. These states involve significant modulation of inhibition, which alters the excitation–inhibition (EI) balance of synaptic inputs. However, the influence of the EI balance on collective neuronal oscillation remains only partially understood. In this study, we introduce the EI-Kuramoto model, a modified version of the Kuramoto model, in which oscillators are categorized into excitatory and inhibitory groups with four distinct interaction types: excitatory–excitatory, excitatory–inhibitory, inhibitory–excitatory, and inhibitory–inhibitory. Numerical simulations identify three dynamic states—synchronized, bistable, and desynchronized—that can be controlled by adjusting the strength of the four interaction types. Theoretical analysis further demonstrates that the balance among these interactions plays a critical role in determining the dynamic states. This study provides valuable insights into the role of EI balance in synchronizing coupled oscillators and neurons.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 7","pages":"1353-1372"},"PeriodicalIF":2.7,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11048764","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144163731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reformulation of RBM to Unify Linear and Nonlinear Dimensionality Reduction 统一线性降维和非线性降维的RBM的重新表述。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1162/neco_a_01751
Jiangsheng You;Chun-Yen Liu
A restricted Boltzmann machine (RBM) is a two-layer neural network with shared weights and has been extensively studied for dimensionality reduction, data representation, and recommendation systems in the literature. The traditional RBM requires a probabilistic interpretation of the values on both layers and a Markov chain Monte Carlo (MCMC) procedure to generate samples during the training. The contrastive divergence (CD) is efficient to train the RBM, but its convergence has not been proved mathematically. In this letter, we investigate the RBM by using a maximum a posteriori (MAP) estimate and the expectation–maximization (EM) algorithm. We show that the CD algorithm without MCMC is convergent for the conditional likelihood object function. Another key contribution in this letter is the reformulation of the RBM into a deterministic model. Within the reformulated RBM, the CD algorithm without MCMC approximates the gradient descent (GD) method. This reformulated RBM can take the continuous scalar and vector variables on the nodes with flexibility in choosing the activation functions. Numerical experiments show its capability in both linear and nonlinear dimensionality reduction, and for the nonlinear dimensionality reduction, the reformulated RBM can outperform principal component analysis (PCA) by choosing the proper activation functions. Finally, we demonstrate its application to vector-valued nodes for the CIFAR-10 data set (color images) and the multivariate sequence data, which cannot be configured naturally with the traditional RBM. This work not only provides theoretical insights regarding the traditional RBM but also unifies the linear and nonlinear dimensionality reduction for scalar and vector variables.
受限玻尔兹曼机(RBM)是一种具有共享权重的双层神经网络,在文献中已被广泛用于降维、数据表示和推荐系统。传统的 RBM 需要对两层的值进行概率解释,并在训练过程中使用马尔可夫链蒙特卡罗(MCMC)程序生成样本。对比发散(CD)是训练 RBM 的有效方法,但其收敛性尚未得到数学证明。在这封信中,我们使用最大后验(MAP)估计和期望最大化(EM)算法研究了 RBM。我们证明,对于条件似然对象函数,不使用 MCMC 的 CD 算法是收敛的。这封信的另一个重要贡献是将 RBM 重构为确定性模型。在重构的 RBM 中,无 MCMC 的 CD 算法近似于梯度下降(GD)方法。这种重构的 RBM 可以采用节点上的连续标量和矢量变量,并能灵活选择激活函数。数值实验显示了它在线性和非线性降维方面的能力,在非线性降维方面,通过选择适当的激活函数,重构的 RBM 可以优于主成分分析(PCA)。最后,我们演示了它在 CIFAR-10 数据集(彩色图像)和多元序列数据的向量值节点上的应用,传统的 RBM 无法自然配置这些节点。这项工作不仅提供了有关传统 RBM 的理论见解,还统一了标量变量和矢量变量的线性和非线性降维。
{"title":"Reformulation of RBM to Unify Linear and Nonlinear Dimensionality Reduction","authors":"Jiangsheng You;Chun-Yen Liu","doi":"10.1162/neco_a_01751","DOIUrl":"10.1162/neco_a_01751","url":null,"abstract":"A restricted Boltzmann machine (RBM) is a two-layer neural network with shared weights and has been extensively studied for dimensionality reduction, data representation, and recommendation systems in the literature. The traditional RBM requires a probabilistic interpretation of the values on both layers and a Markov chain Monte Carlo (MCMC) procedure to generate samples during the training. The contrastive divergence (CD) is efficient to train the RBM, but its convergence has not been proved mathematically. In this letter, we investigate the RBM by using a maximum a posteriori (MAP) estimate and the expectation–maximization (EM) algorithm. We show that the CD algorithm without MCMC is convergent for the conditional likelihood object function. Another key contribution in this letter is the reformulation of the RBM into a deterministic model. Within the reformulated RBM, the CD algorithm without MCMC approximates the gradient descent (GD) method. This reformulated RBM can take the continuous scalar and vector variables on the nodes with flexibility in choosing the activation functions. Numerical experiments show its capability in both linear and nonlinear dimensionality reduction, and for the nonlinear dimensionality reduction, the reformulated RBM can outperform principal component analysis (PCA) by choosing the proper activation functions. Finally, we demonstrate its application to vector-valued nodes for the CIFAR-10 data set (color images) and the multivariate sequence data, which cannot be configured naturally with the traditional RBM. This work not only provides theoretical insights regarding the traditional RBM but also unifies the linear and nonlinear dimensionality reduction for scalar and vector variables.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 5","pages":"1034-1055"},"PeriodicalIF":2.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143671836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Generalized Time Rescaling Theorem for Temporal Point Processes 时间点过程的广义时间重标定理。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1162/neco_a_01745
Xi Zhang;Akshay Aravamudan;Georgios C. Anagnostopoulos
Temporal point processes are essential for modeling event dynamics in fields such as neuroscience and social media. The time rescaling theorem is commonly used to assess model fit by transforming a point process into a homogeneous Poisson process. However, this approach requires that the process be nonterminating and that complete (hence, unbounded) realizations are observed—conditions that are often unmet in practice. This article introduces a generalized time-rescaling theorem to address these limitations and, as such, facilitates a more widely applicable evaluation framework for point process models in diverse real-world scenarios.
时间点过程对于神经科学和社交媒体等领域的事件动态建模至关重要。时间重定标定理通常通过将点过程转换为同质泊松过程来评估模型拟合度。然而,这种方法要求过程是非终结的,并且能观察到完整的(因此是无界的)实现--这些条件在实践中往往无法满足。本文引入了一个广义的时间缩放定理来解决这些局限性,从而为点过程模型在现实世界的各种场景中提供了一个更广泛适用的评估框架。
{"title":"A Generalized Time Rescaling Theorem for Temporal Point Processes","authors":"Xi Zhang;Akshay Aravamudan;Georgios C. Anagnostopoulos","doi":"10.1162/neco_a_01745","DOIUrl":"10.1162/neco_a_01745","url":null,"abstract":"Temporal point processes are essential for modeling event dynamics in fields such as neuroscience and social media. The time rescaling theorem is commonly used to assess model fit by transforming a point process into a homogeneous Poisson process. However, this approach requires that the process be nonterminating and that complete (hence, unbounded) realizations are observed—conditions that are often unmet in practice. This article introduces a generalized time-rescaling theorem to address these limitations and, as such, facilitates a more widely applicable evaluation framework for point process models in diverse real-world scenarios.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 5","pages":"871-885"},"PeriodicalIF":2.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10979820","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143607145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adding Space to Random Networks of Spiking Neurons: A Method Based on Scaling the Network Size 给脉冲神经元随机网络增加空间:一种基于网络大小缩放的方法。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1162/neco_a_01747
Cecilia Romaro;Jose Roberto Castilho Piqueira;A. C. Roque
Many spiking neural network models are based on random graphs that do not include topological and structural properties featured in real brain networks. To turn these models into spatial networks that describe the topographic arrangement of connections is a challenging task because one has to deal with neurons at the spatial network boundary. Addition of space may generate spurious network behavior like oscillations introduced by periodic boundary conditions or unbalanced neuronal spiking due to lack or excess of connections. Here, we introduce a boundary solution method for networks with added spatial extension that prevents the occurrence of spurious spiking behavior. The method is based on a recently proposed technique for scaling the network size that preserves first- and second-order statistics.
许多脉冲神经网络模型是基于随机图的,不包括真实大脑网络的拓扑和结构特性。将这些模型转化为描述连接的地形排列的空间网络是一项具有挑战性的任务,因为人们必须处理空间网络边界上的神经元。空间的增加可能会产生虚假的网络行为,如由周期性边界条件引入的振荡或由于缺乏或过多的连接而导致的不平衡神经元尖峰。在这里,我们引入了一种边界求解方法,用于增加空间扩展的网络,以防止虚假尖峰行为的发生。该方法基于最近提出的一种扩展网络大小的技术,该技术保留了一阶和二阶统计量。
{"title":"Adding Space to Random Networks of Spiking Neurons: A Method Based on Scaling the Network Size","authors":"Cecilia Romaro;Jose Roberto Castilho Piqueira;A. C. Roque","doi":"10.1162/neco_a_01747","DOIUrl":"10.1162/neco_a_01747","url":null,"abstract":"Many spiking neural network models are based on random graphs that do not include topological and structural properties featured in real brain networks. To turn these models into spatial networks that describe the topographic arrangement of connections is a challenging task because one has to deal with neurons at the spatial network boundary. Addition of space may generate spurious network behavior like oscillations introduced by periodic boundary conditions or unbalanced neuronal spiking due to lack or excess of connections. Here, we introduce a boundary solution method for networks with added spatial extension that prevents the occurrence of spurious spiking behavior. The method is based on a recently proposed technique for scaling the network size that preserves first- and second-order statistics.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 5","pages":"957-986"},"PeriodicalIF":2.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143671831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Elucidating the Theoretical Underpinnings of Surrogate Gradient Learning in Spiking Neural Networks 阐明尖峰神经网络中替代梯度学习的理论基础
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1162/neco_a_01752
Julia Gygax;Friedemann Zenke
Training spiking neural networks to approximate universal functions is essential for studying information processing in the brain and for neuromorphic computing. Yet the binary nature of spikes poses a challenge for direct gradient-based training. Surrogate gradients have been empirically successful in circumventing this problem, but their theoretical foundation remains elusive. Here, we investigate the relation of surrogate gradients to two theoretically well-founded approaches. On the one hand, we consider smoothed probabilistic models, which, due to the lack of support for automatic differentiation, are impractical for training multilayer spiking neural networks but provide derivatives equivalent to surrogate gradients for single neurons. On the other hand, we investigate stochastic automatic differentiation, which is compatible with discrete randomness but has not yet been used to train spiking neural networks. We find that the latter gives surrogate gradients a theoretical basis in stochastic spiking neural networks, where the surrogate derivative matches the derivative of the neuronal escape noise function. This finding supports the effectiveness of surrogate gradients in practice and suggests their suitability for stochastic spiking neural networks. However, surrogate gradients are generally not gradients of a surrogate loss despite their relation to stochastic automatic differentiation. Nevertheless, we empirically confirm the effectiveness of surrogate gradients in stochastic multilayer spiking neural networks and discuss their relation to deterministic networks as a special case. Our work gives theoretical support to surrogate gradients and the choice of a suitable surrogate derivative in stochastic spiking neural networks.
训练脉冲神经网络来近似通用函数对于研究大脑中的信息处理和神经形态计算是必不可少的。然而,尖峰的二元性对直接基于梯度的训练提出了挑战。替代梯度在经验上已经成功地规避了这个问题,但它们的理论基础仍然难以捉摸。在这里,我们研究了代理梯度与两种理论上有充分根据的方法的关系。一方面,我们考虑平滑概率模型,由于缺乏对自动分化的支持,它对于训练多层尖峰神经网络是不切实际的,但它为单个神经元提供了相当于代理梯度的导数。另一方面,我们研究了随机自动微分,它与离散随机性兼容,但尚未用于训练尖峰神经网络。我们发现后者为随机尖峰神经网络中的代理梯度提供了理论基础,其中代理导数与神经元逃逸噪声函数的导数相匹配。这一发现支持了替代梯度在实践中的有效性,并表明它们适合于随机尖峰神经网络。然而,代理梯度通常不是代理损失的梯度,尽管它们与随机自动微分有关。然而,我们在经验上证实了代理梯度在随机多层脉冲神经网络中的有效性,并作为一个特例讨论了它们与确定性网络的关系。我们的工作为随机尖峰神经网络的代理梯度和合适的代理导数的选择提供了理论支持。
{"title":"Elucidating the Theoretical Underpinnings of Surrogate Gradient Learning in Spiking Neural Networks","authors":"Julia Gygax;Friedemann Zenke","doi":"10.1162/neco_a_01752","DOIUrl":"10.1162/neco_a_01752","url":null,"abstract":"Training spiking neural networks to approximate universal functions is essential for studying information processing in the brain and for neuromorphic computing. Yet the binary nature of spikes poses a challenge for direct gradient-based training. Surrogate gradients have been empirically successful in circumventing this problem, but their theoretical foundation remains elusive. Here, we investigate the relation of surrogate gradients to two theoretically well-founded approaches. On the one hand, we consider smoothed probabilistic models, which, due to the lack of support for automatic differentiation, are impractical for training multilayer spiking neural networks but provide derivatives equivalent to surrogate gradients for single neurons. On the other hand, we investigate stochastic automatic differentiation, which is compatible with discrete randomness but has not yet been used to train spiking neural networks. We find that the latter gives surrogate gradients a theoretical basis in stochastic spiking neural networks, where the surrogate derivative matches the derivative of the neuronal escape noise function. This finding supports the effectiveness of surrogate gradients in practice and suggests their suitability for stochastic spiking neural networks. However, surrogate gradients are generally not gradients of a surrogate loss despite their relation to stochastic automatic differentiation. Nevertheless, we empirically confirm the effectiveness of surrogate gradients in stochastic multilayer spiking neural networks and discuss their relation to deterministic networks as a special case. Our work gives theoretical support to surrogate gradients and the choice of a suitable surrogate derivative in stochastic spiking neural networks.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 5","pages":"886-925"},"PeriodicalIF":2.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10979826","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143671833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed Synaptic Connection Strength Changes Dynamics in a Population Firing Rate Model in Response to Continuous External Stimuli 响应连续外部刺激的群体放电率模型中分布式突触连接强度的动态变化。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1162/neco_a_01749
Masato Sugino;Mai Tanaka;Kenta Shimba;Kiyoshi Kotani;Yasuhiko Jimbo
Neural network complexity allows for diverse neuronal population dynamics and realizes higherorder brain functions such as cognition and memory. Complexity is enhanced through chemical synapses with exponentially decaying conductance and greater variation in the neuronal connection strength due to synaptic plasticity. However, in the macroscopic neuronal population model, synaptic connections are often described by spike connections, and connection strengths within the population are assumed to be uniform. Thus, the effects of synaptic connections variation on network synchronization remain unclear. Based on recent advances in mean field theory for the quadratic integrate-and-fire neuronal network model, we introduce synaptic conductance and variation of connection strength into the excitatory and inhibitory neuronal population model and derive the macroscopic firing rate equations for faithful modeling. We then introduce a heuristic switching rule of the dynamic system with respect to the mean membrane potentials to avoid divergences in the computation caused by variations in the neuronal connection strength. We show that the switching rule agrees with the numerical computation of the microscopic level model. In the derived model, variations in synaptic conductance and connection strength strongly alter the stability of the solutions to the equations, which is related to the mechanism of synchronous firing. When we apply physiologically plausible values from layer 4 of the mammalian primary visual cortex to the derived model, we observe event-related desynchronization at the alpha and beta frequencies and event-related synchronization at the gamma frequency over a wide range of balanced external currents. Our results show that the introduction of complex synaptic connections and physiologically valid numerical values into the low-dimensional mean field equations reproduces dynamic changes such as eventrelated (de)synchronization, and provides a unique mathematical insight into the relationship between synaptic strength variation and oscillatory mechanism.
神经网络的复杂性允许不同的神经元种群动态,实现更高阶的大脑功能,如认知和记忆。由于突触的可塑性,化学突触的电导呈指数衰减,神经元连接强度的变化更大,从而增强了复杂性。然而,在宏观神经元群体模型中,突触连接通常被描述为尖峰连接,并且假设群体内的连接强度是均匀的。因此,突触连接变化对网络同步的影响尚不清楚。基于二次积分-放电神经网络模型平均场理论的最新进展,我们将突触电导和连接强度的变化引入兴奋性和抑制性神经元群模型中,并推导出宏观放电率方程,以忠实地建模。然后,我们引入了动态系统相对于平均膜电位的启发式切换规则,以避免由于神经元连接强度的变化而导致的计算分歧。结果表明,该开关规则与微观水平模型的数值计算一致。在推导的模型中,突触电导和连接强度的变化强烈地改变了方程解的稳定性,这与同步放电的机制有关。当我们将哺乳动物初级视觉皮层第4层的生理合理值应用于衍生模型时,我们观察到在广泛的平衡外部电流范围内,α和β频率上的事件相关非同步和γ频率上的事件相关同步。我们的研究结果表明,在低维平均场方程中引入复杂的突触连接和生理上有效的数值再现了事件相关(去)同步等动态变化,并为突触强度变化与振荡机制之间的关系提供了独特的数学见解。
{"title":"Distributed Synaptic Connection Strength Changes Dynamics in a Population Firing Rate Model in Response to Continuous External Stimuli","authors":"Masato Sugino;Mai Tanaka;Kenta Shimba;Kiyoshi Kotani;Yasuhiko Jimbo","doi":"10.1162/neco_a_01749","DOIUrl":"10.1162/neco_a_01749","url":null,"abstract":"Neural network complexity allows for diverse neuronal population dynamics and realizes higherorder brain functions such as cognition and memory. Complexity is enhanced through chemical synapses with exponentially decaying conductance and greater variation in the neuronal connection strength due to synaptic plasticity. However, in the macroscopic neuronal population model, synaptic connections are often described by spike connections, and connection strengths within the population are assumed to be uniform. Thus, the effects of synaptic connections variation on network synchronization remain unclear. Based on recent advances in mean field theory for the quadratic integrate-and-fire neuronal network model, we introduce synaptic conductance and variation of connection strength into the excitatory and inhibitory neuronal population model and derive the macroscopic firing rate equations for faithful modeling. We then introduce a heuristic switching rule of the dynamic system with respect to the mean membrane potentials to avoid divergences in the computation caused by variations in the neuronal connection strength. We show that the switching rule agrees with the numerical computation of the microscopic level model. In the derived model, variations in synaptic conductance and connection strength strongly alter the stability of the solutions to the equations, which is related to the mechanism of synchronous firing. When we apply physiologically plausible values from layer 4 of the mammalian primary visual cortex to the derived model, we observe event-related desynchronization at the alpha and beta frequencies and event-related synchronization at the gamma frequency over a wide range of balanced external currents. Our results show that the introduction of complex synaptic connections and physiologically valid numerical values into the low-dimensional mean field equations reproduces dynamic changes such as eventrelated (de)synchronization, and provides a unique mathematical insight into the relationship between synaptic strength variation and oscillatory mechanism.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 5","pages":"987-1009"},"PeriodicalIF":2.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143671832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multilevel Data Representation for Training Deep Helmholtz Machines 训练深度亥姆霍兹机的多层数据表示。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1162/neco_a_01748
Jose Miguel Ramos;Luis Sa-Couto;Andreas Wichert
A vast majority of the current research in the field of machine learning is done using algorithms with strong arguments pointing to their biological implausibility such as backpropagation, deviating the field’s focus from understanding its original organic inspiration to a compulsive search for optimal performance. Yet there have been a few proposed models that respect most of the biological constraints present in the human brain and are valid candidates for mimicking some of its properties and mechanisms. In this letter, we focus on guiding the learning of a biologically plausible generative model called the Helmholtz machine in complex search spaces using a heuristic based on the human image perception mechanism. We hypothesize that this model’s learning algorithm is not fit for deep networks due to its Hebbian-like local update rule, rendering it incapable of taking full advantage of the compositional properties that multilayer networks provide. We propose to overcome this problem by providing the network’s hidden layers with visual queues at different resolutions using multilevel data representation. The results on several image data sets showed that the model was able to not only obtain better overall quality but also a wider diversity in the generated images, corroborating our intuition that using our proposed heuristic allows the model to take more advantage of the network’s depth growth. More important, they show the unexplored possibilities underlying brain-inspired models and techniques.
目前机器学习领域的绝大多数研究都是使用算法来完成的,这些算法有很强的论据,指出它们在生物学上是不可信的,比如反向传播,偏离了该领域的重点,从理解其原始的有机灵感转向了对最佳性能的强制搜索。然而,已经提出了一些模型,这些模型尊重人类大脑中存在的大多数生物限制,并且是模仿其某些特性和机制的有效候选者。在这封信中,我们专注于使用基于人类图像感知机制的启发式方法指导复杂搜索空间中称为亥姆霍兹机的生物学上合理的生成模型的学习。我们假设该模型的学习算法不适合深度网络,因为它的hebbian式局部更新规则,使其无法充分利用多层网络提供的组合特性。我们建议通过使用多层数据表示为网络的隐藏层提供不同分辨率的可视化队列来克服这个问题。在几个图像数据集上的结果表明,该模型不仅能够获得更好的整体质量,而且生成的图像具有更大的多样性,这证实了我们的直觉,即使用我们提出的启发式方法可以让模型更多地利用网络的深度增长。更重要的是,它们展示了大脑启发模型和技术背后未被探索的可能性。
{"title":"Multilevel Data Representation for Training Deep Helmholtz Machines","authors":"Jose Miguel Ramos;Luis Sa-Couto;Andreas Wichert","doi":"10.1162/neco_a_01748","DOIUrl":"10.1162/neco_a_01748","url":null,"abstract":"A vast majority of the current research in the field of machine learning is done using algorithms with strong arguments pointing to their biological implausibility such as backpropagation, deviating the field’s focus from understanding its original organic inspiration to a compulsive search for optimal performance. Yet there have been a few proposed models that respect most of the biological constraints present in the human brain and are valid candidates for mimicking some of its properties and mechanisms. In this letter, we focus on guiding the learning of a biologically plausible generative model called the Helmholtz machine in complex search spaces using a heuristic based on the human image perception mechanism. We hypothesize that this model’s learning algorithm is not fit for deep networks due to its Hebbian-like local update rule, rendering it incapable of taking full advantage of the compositional properties that multilayer networks provide. We propose to overcome this problem by providing the network’s hidden layers with visual queues at different resolutions using multilevel data representation. The results on several image data sets showed that the model was able to not only obtain better overall quality but also a wider diversity in the generated images, corroborating our intuition that using our proposed heuristic allows the model to take more advantage of the network’s depth growth. More important, they show the unexplored possibilities underlying brain-inspired models and techniques.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 5","pages":"1010-1033"},"PeriodicalIF":2.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143671835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1