首页 > 最新文献

Neural Computation最新文献

英文 中文
Dynamics and Bifurcation Structure of a Mean-Field Model of Adaptive Exponential Integrate-and-Fire Networks 自适应指数型积分网络平均场模型的动力学和分岔结构。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-14 DOI: 10.1162/neco_a_01758
Lionel Kusch;Damien Depannemaecker;Alain Destexhe;Viktor Jirsa
The study of brain activity spans diverse scales and levels of description and requires the development of computational models alongside experimental investigations to explore integrations across scales. The high dimensionality of spiking networks presents challenges for understanding their dynamics. To tackle this, a mean-field formulation offers a potential approach for dimensionality reduction while retaining essential elements. Here, we focus on a previously developed mean-field model of adaptive exponential integrate and fire (AdEx) networks used in various research work. We observe qualitative similarities in the bifurcation structure but quantitative differences in mean firing rates between the mean-field model and AdEx spiking network simulations. Even if the mean-field model does not accurately predict phase shift during transients and oscillatory input, it generally captures the qualitative dynamics of the spiking network’s response to both constant and varying inputs. Finally, we offer an overview of the dynamical properties of the AdExMF to assist future users in interpreting their results of simulations.
大脑活动的研究跨越了不同的尺度和描述水平,需要在实验调查的基础上发展计算模型来探索跨尺度的整合。尖峰网络的高维性给理解其动力学带来了挑战。为了解决这个问题,平均场公式提供了一种潜在的降维方法,同时保留了基本元素。在这里,我们将重点关注先前开发的自适应指数积分和火焰(AdEx)网络的平均场模型,该模型用于各种研究工作。我们观察到平均场模型和AdEx尖峰网络模拟在分岔结构上的定性相似性,但在平均发射率上的定量差异。即使平均场模型不能准确地预测瞬态和振荡输入期间的相移,它通常也能捕捉到脉冲网络对恒定和变化输入响应的定性动态。最后,我们概述了AdExMF的动态特性,以帮助未来的用户解释他们的模拟结果。
{"title":"Dynamics and Bifurcation Structure of a Mean-Field Model of Adaptive Exponential Integrate-and-Fire Networks","authors":"Lionel Kusch;Damien Depannemaecker;Alain Destexhe;Viktor Jirsa","doi":"10.1162/neco_a_01758","DOIUrl":"10.1162/neco_a_01758","url":null,"abstract":"The study of brain activity spans diverse scales and levels of description and requires the development of computational models alongside experimental investigations to explore integrations across scales. The high dimensionality of spiking networks presents challenges for understanding their dynamics. To tackle this, a mean-field formulation offers a potential approach for dimensionality reduction while retaining essential elements. Here, we focus on a previously developed mean-field model of adaptive exponential integrate and fire (AdEx) networks used in various research work. We observe qualitative similarities in the bifurcation structure but quantitative differences in mean firing rates between the mean-field model and AdEx spiking network simulations. Even if the mean-field model does not accurately predict phase shift during transients and oscillatory input, it generally captures the qualitative dynamics of the spiking network’s response to both constant and varying inputs. Finally, we offer an overview of the dynamical properties of the AdExMF to assist future users in interpreting their results of simulations.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 6","pages":"1102-1123"},"PeriodicalIF":2.7,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144045166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Memory States From Almost Nothing: Representing and Computing in a Nonassociative Algebra 记忆状态:非结合代数的表示与计算。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-14 DOI: 10.1162/neco_a_01755
Stefan Reimann
This letter presents a nonassociative algebraic framework for the representation and computation of information items in high-dimensional space. This framework is consistent with the principles of spatial computing and with the empirical findings in cognitive science about memory. Computations are performed through a process of multiplication-like binding and nonassociative interference-like bundling. Models that rely on associative bundling typically lose order information, which necessitates the use of auxiliary order structures, such as position markers, to represent sequential information that is important for cognitive tasks. In contrast, the nonassociative bundling proposed allows the construction of sparse representations of arbitrarily long sequences that maintain their temporal structure across arbitrary lengths. In this operation, noise is a constituent element of the representation of order information rather than a means of obscuring it. The nonassociative nature of the proposed framework results in the representation of a single sequence by two distinct states. The L-state, generated through left-associative bundling, continuously updates and emphasizes a recency effect, while the R-state, formed through right-associative bundling, encodes finite sequences or chunks, capturing a primacy effect. The construction of these states may be associated with activity in the prefrontal cortex in relation to short-term memory and hippocampal encoding in long-term memory, respectively. The accuracy of retrieval is contingent on a decision-making process that is based on the mutual information between the memory states and the cue. The model is able to replicate the serial position curve, which reflects the empirical recency and primacy effects observed in cognitive experiments.
这封信提出了高维空间中信息项的表示和计算的非关联代数框架。这一框架与空间计算原理和认知科学关于记忆的实证研究结果是一致的。计算是通过类似乘法的绑定和类似非关联干涉的绑定过程来完成的。依赖于关联捆绑的模型通常会丢失顺序信息,这就需要使用辅助顺序结构(如位置标记)来表示对认知任务很重要的顺序信息。相反,提出的非关联捆绑允许构建任意长序列的稀疏表示,这些序列在任意长度上保持其时间结构。在这个操作中,噪声是表示有序信息的一个组成元素,而不是使其模糊的一种手段。所提出的框架的非关联性质导致用两个不同的状态表示单个序列。左结合捆绑形成的l态不断更新,强调近因效应;右结合捆绑形成的r态编码有限序列或块,捕捉质数效应。这些状态的构建可能分别与与短期记忆有关的前额叶皮层活动和与长期记忆有关的海马编码有关。检索的准确性取决于基于记忆状态和线索之间相互信息的决策过程。该模型能够复制序列位置曲线,反映了认知实验中观察到的经验近因效应和因因效应。
{"title":"Memory States From Almost Nothing: Representing and Computing in a Nonassociative Algebra","authors":"Stefan Reimann","doi":"10.1162/neco_a_01755","DOIUrl":"10.1162/neco_a_01755","url":null,"abstract":"This letter presents a nonassociative algebraic framework for the representation and computation of information items in high-dimensional space. This framework is consistent with the principles of spatial computing and with the empirical findings in cognitive science about memory. Computations are performed through a process of multiplication-like binding and nonassociative interference-like bundling. Models that rely on associative bundling typically lose order information, which necessitates the use of auxiliary order structures, such as position markers, to represent sequential information that is important for cognitive tasks. In contrast, the nonassociative bundling proposed allows the construction of sparse representations of arbitrarily long sequences that maintain their temporal structure across arbitrary lengths. In this operation, noise is a constituent element of the representation of order information rather than a means of obscuring it. The nonassociative nature of the proposed framework results in the representation of a single sequence by two distinct states. The L-state, generated through left-associative bundling, continuously updates and emphasizes a recency effect, while the R-state, formed through right-associative bundling, encodes finite sequences or chunks, capturing a primacy effect. The construction of these states may be associated with activity in the prefrontal cortex in relation to short-term memory and hippocampal encoding in long-term memory, respectively. The accuracy of retrieval is contingent on a decision-making process that is based on the mutual information between the memory states and the cue. The model is able to replicate the serial position curve, which reflects the empirical recency and primacy effects observed in cognitive experiments.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 6","pages":"1154-1170"},"PeriodicalIF":2.7,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144057372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Low-Rank, High-Order Tensor Completion via t- Product-Induced Tucker (tTucker) Decomposition 通过t积诱导塔克(tTucker)分解的低秩、高阶张量补全。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-14 DOI: 10.1162/neco_a_01756
Yaodong Li;Jun Tan;Peilin Yang;Guoxu Zhou;Qibin Zhao
Recently, tensor singular value decomposition (t-SVD)–based methods were proposed to solve the low-rank tensor completion (LRTC) problem, which has achieved unprecedented success on image and video inpainting tasks. The t-SVD is limited to process third-order tensors. When faced with higher-order tensors, it reshapes them into third-order tensors, leading to the destruction of interdimensional correlations. To address this limitation, this letter introduces a tproductinduced Tucker decomposition (tTucker) model that replaces the mode product in Tucker decomposition with t-product, which jointly extends the ideas of t-SVD and high-order SVD. This letter defines the rank of the tTucker decomposition and presents an LRTC model that minimizes the induced Schatten-p norm. An efficient alternating direction multiplier method (ADMM) algorithm is developed to optimize the proposed LRTC model, and its effectiveness is demonstrated through experiments conducted on both synthetic and real data sets, showcasing excellent performance.
近年来,基于张量奇异值分解(t-SVD)的方法被提出用于解决低秩张量补全(LRTC)问题,并在图像和视频补全任务中取得了前所未有的成功。t-SVD仅限于处理三阶张量。当面对高阶张量时,它将其重塑为三阶张量,导致维间相关性的破坏。为了解决这一局限性,本文引入了一种t-product Tucker分解(tTucker)模型,该模型用t-product代替Tucker分解中的模态积,它共同扩展了t-SVD和高阶SVD的思想。这封信定义了塔克分解的秩,并提出了一个最小化诱导schattenp范数的LRTC模型。为了优化LRTC模型,提出了一种高效的交变方向乘子算法(ADMM),并在合成数据集和真实数据集上进行了实验,证明了该算法的有效性。
{"title":"Low-Rank, High-Order Tensor Completion via t- Product-Induced Tucker (tTucker) Decomposition","authors":"Yaodong Li;Jun Tan;Peilin Yang;Guoxu Zhou;Qibin Zhao","doi":"10.1162/neco_a_01756","DOIUrl":"10.1162/neco_a_01756","url":null,"abstract":"Recently, tensor singular value decomposition (t-SVD)–based methods were proposed to solve the low-rank tensor completion (LRTC) problem, which has achieved unprecedented success on image and video inpainting tasks. The t-SVD is limited to process third-order tensors. When faced with higher-order tensors, it reshapes them into third-order tensors, leading to the destruction of interdimensional correlations. To address this limitation, this letter introduces a tproductinduced Tucker decomposition (tTucker) model that replaces the mode product in Tucker decomposition with t-product, which jointly extends the ideas of t-SVD and high-order SVD. This letter defines the rank of the tTucker decomposition and presents an LRTC model that minimizes the induced Schatten-p norm. An efficient alternating direction multiplier method (ADMM) algorithm is developed to optimize the proposed LRTC model, and its effectiveness is demonstrated through experiments conducted on both synthetic and real data sets, showcasing excellent performance.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 6","pages":"1171-1192"},"PeriodicalIF":2.7,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144029302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Replay as a Basis for Backpropagation Through Time in the Brain 回放作为大脑中穿越时间反向传播的基础。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-14 DOI: 10.1162/neco_a_01735
Huzi Cheng;Joshua W. Brown
How episodic memories are formed in the brain is a continuing puzzle for the neuroscience community. The brain areas that are critical for episodic learning (e.g., the hippocampus) are characterized by recurrent connectivity and generate frequent offline replay events. The function of the replay events is a subject of active debate. Recurrent connectivity, computational simulations show, enables sequence learning when combined with a suitable learning algorithm such as backpropagation through time (BPTT). BPTT, however, is not biologically plausible. We describe here, for the first time, a biologically plausible variant of BPTT in a reversible recurrent neural network, R2N2, that critically leverages offline replay to support episodic learning. The model uses forward and backward offline replay to transfer information between two recurrent neural networks, a cache and a consolidator, that perform rapid one-shot learning and statistical learning, respectively. Unlike replay in standard BPTT, this architecture requires no artificial external memory store. This approach outperforms existing solutions like random feedback local online learning and reservoir network. It also accounts for the functional significance of hippocampal replay events. We demonstrate the R2N2 network properties using benchmark tests from computer science and simulate the rodent delayed alternation T-maze task.
情景记忆是如何在大脑中形成的,这一直是神经科学界的一个谜。对情景学习至关重要的大脑区域(如海马体)的特点是反复连接,并产生频繁的离线重播事件。重播事件的功能是一个积极争论的主题。计算模拟表明,当与适当的学习算法(如时间反向传播(BPTT))相结合时,循环连接可以实现序列学习。然而,BPTT在生物学上并不可信。我们首次在可逆循环神经网络R2N2中描述了一种生物学上合理的BPTT变体,该变体主要利用离线重播来支持情景学习。该模型使用向前和向后离线重放在两个循环神经网络(缓存和整合)之间传递信息,分别进行快速的一次性学习和统计学习。与标准BPTT中的重放不同,这种体系结构不需要人工的外部存储器存储。这种方法优于随机反馈本地在线学习和水库网络等现有解决方案。这也解释了海马体回放事件的功能意义。我们使用计算机科学的基准测试来证明R2N2网络的特性,并模拟啮齿动物的延迟交替t -迷宫任务。
{"title":"Replay as a Basis for Backpropagation Through Time in the Brain","authors":"Huzi Cheng;Joshua W. Brown","doi":"10.1162/neco_a_01735","DOIUrl":"10.1162/neco_a_01735","url":null,"abstract":"How episodic memories are formed in the brain is a continuing puzzle for the neuroscience community. The brain areas that are critical for episodic learning (e.g., the hippocampus) are characterized by recurrent connectivity and generate frequent offline replay events. The function of the replay events is a subject of active debate. Recurrent connectivity, computational simulations show, enables sequence learning when combined with a suitable learning algorithm such as backpropagation through time (BPTT). BPTT, however, is not biologically plausible. We describe here, for the first time, a biologically plausible variant of BPTT in a reversible recurrent neural network, R2N2, that critically leverages offline replay to support episodic learning. The model uses forward and backward offline replay to transfer information between two recurrent neural networks, a cache and a consolidator, that perform rapid one-shot learning and statistical learning, respectively. Unlike replay in standard BPTT, this architecture requires no artificial external memory store. This approach outperforms existing solutions like random feedback local online learning and reservoir network. It also accounts for the functional significance of hippocampal replay events. We demonstrate the R2N2 network properties using benchmark tests from computer science and simulate the rodent delayed alternation T-maze task.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 3","pages":"403-436"},"PeriodicalIF":2.7,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gradual Domain Adaptation via Normalizing Flows 通过规范化流程逐步适应领域。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-14 DOI: 10.1162/neco_a_01734
Shogo Sagawa;Hideitsu Hino
Standard domain adaptation methods do not work well when a large gap exists between the source and target domains. Gradual domain adaptation is one of the approaches used to address the problem. It involves leveraging the intermediate domain, which gradually shifts from the source domain to the target domain. In previous work, it is assumed that the number of intermediate domains is large and the distance between adjacent domains is small; hence, the gradual domain adaptation algorithm, involving self-training with unlabeled data sets, is applicable. In practice, however, gradual self-training will fail because the number of intermediate domains is limited and the distance between adjacent domains is large. We propose the use of normalizing flows to deal with this problem while maintaining the framework of unsupervised domain adaptation. The proposed method learns a transformation from the distribution of the target domains to the gaussian mixture distribution via the source domain. We evaluate our proposed method by experiments using real-world data sets and confirm that it mitigates the problem we have explained and improves the classification performance.
当源域和目标域之间存在较大差距时,标准的域自适应方法不能很好地工作。逐步的领域适应是解决这个问题的方法之一。它涉及到利用中间域,中间域逐渐从源域转移到目标域。在以前的工作中,假设中间域的数量很大,相邻域之间的距离很小;因此,适用于使用无标记数据集进行自训练的渐进式域适应算法。然而,在实践中,由于中间域的数量有限,相邻域之间的距离较大,逐渐的自我训练将会失败。我们建议使用规范化流来处理这个问题,同时保持无监督域自适应的框架。该方法通过源域学习从目标域的分布到高斯混合分布的转换。我们通过使用真实数据集的实验来评估我们提出的方法,并确认它减轻了我们解释的问题并提高了分类性能。
{"title":"Gradual Domain Adaptation via Normalizing Flows","authors":"Shogo Sagawa;Hideitsu Hino","doi":"10.1162/neco_a_01734","DOIUrl":"10.1162/neco_a_01734","url":null,"abstract":"Standard domain adaptation methods do not work well when a large gap exists between the source and target domains. Gradual domain adaptation is one of the approaches used to address the problem. It involves leveraging the intermediate domain, which gradually shifts from the source domain to the target domain. In previous work, it is assumed that the number of intermediate domains is large and the distance between adjacent domains is small; hence, the gradual domain adaptation algorithm, involving self-training with unlabeled data sets, is applicable. In practice, however, gradual self-training will fail because the number of intermediate domains is limited and the distance between adjacent domains is large. We propose the use of normalizing flows to deal with this problem while maintaining the framework of unsupervised domain adaptation. The proposed method learns a transformation from the distribution of the target domains to the gaussian mixture distribution via the source domain. We evaluate our proposed method by experiments using real-world data sets and confirm that it mitigates the problem we have explained and improves the classification performance.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 3","pages":"522-568"},"PeriodicalIF":2.7,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncovering Dynamical Equations of Stochastic Decision Models Using Data-Driven SINDy Algorithm 用数据驱动SINDy算法揭示随机决策模型的动力学方程。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-14 DOI: 10.1162/neco_a_01736
Brendan Lenfesty;Saugat Bhattacharyya;KongFatt Wong-Lin
Decision formation in perceptual decision making involves sensory evidence accumulation instantiated by the temporal integration of an internal decision variable toward some decision criterion or threshold, as described by sequential sampling theoretical models. The decision variable can be represented in the form of experimentally observable neural activities. Hence, elucidating the appropriate theoretical model becomes crucial to understanding the mechanisms underlying perceptual decision formation. Existing computational methods are limited to either fitting of choice behavioral data or linear model estimation from neural activity data. In this work, we made use of sparse identification of nonlinear dynamics (SINDy), a data-driven approach, to elucidate the deterministic linear and nonlinear components of often-used stochastic decision models within reaction time task paradigms. Based on the simulated decision variable activities of the models and assuming the noise coefficient term is known beforehand, SINDy, enhanced with approaches using multiple trials, could readily estimate the deterministic terms in the dynamical equations, choice accuracy, and decision time of the models across a range of signal-to-noise ratio values. In particular, SINDy performed the best using the more memory-intensive multi-trial approach while trial-averaging of parameters performed more moderately. The single-trial approach, although expectedly not performing as well, may be useful for real-time modeling. Taken together, our work offers alternative approaches for SINDy to uncover the dynamics in perceptual decision making and, more generally, for first-passage time problems.
在知觉决策中,决策的形成涉及到感官证据的积累,这种积累由内部决策变量对某些决策标准或阈值的时间整合实例化,如顺序抽样理论模型所描述的那样。决策变量可以用实验观察到的神经活动的形式来表示。因此,阐明适当的理论模型对于理解感知决策形成的机制至关重要。现有的计算方法要么局限于选择行为数据的拟合,要么局限于神经活动数据的线性模型估计。在这项工作中,我们利用非线性动力学的稀疏识别(SINDy),一种数据驱动的方法,来阐明在反应时间任务范式中经常使用的随机决策模型的确定性线性和非线性成分。基于模型的模拟决策变量活动,假设噪声系数项事先已知,SINDy通过使用多次试验的方法增强,可以很容易地估计动态方程中的确定性项、模型的选择精度和决策时间在一系列信噪比值上。特别是,SINDy在使用更多内存密集型的多试验方法时表现最好,而参数的试验平均表现更为温和。单次试验的方法,虽然预期表现不佳,但可能对实时建模有用。综上所述,我们的工作为SINDy提供了另一种方法来揭示感知决策的动态,更一般地说,是首次通过时间问题。
{"title":"Uncovering Dynamical Equations of Stochastic Decision Models Using Data-Driven SINDy Algorithm","authors":"Brendan Lenfesty;Saugat Bhattacharyya;KongFatt Wong-Lin","doi":"10.1162/neco_a_01736","DOIUrl":"10.1162/neco_a_01736","url":null,"abstract":"Decision formation in perceptual decision making involves sensory evidence accumulation instantiated by the temporal integration of an internal decision variable toward some decision criterion or threshold, as described by sequential sampling theoretical models. The decision variable can be represented in the form of experimentally observable neural activities. Hence, elucidating the appropriate theoretical model becomes crucial to understanding the mechanisms underlying perceptual decision formation. Existing computational methods are limited to either fitting of choice behavioral data or linear model estimation from neural activity data. In this work, we made use of sparse identification of nonlinear dynamics (SINDy), a data-driven approach, to elucidate the deterministic linear and nonlinear components of often-used stochastic decision models within reaction time task paradigms. Based on the simulated decision variable activities of the models and assuming the noise coefficient term is known beforehand, SINDy, enhanced with approaches using multiple trials, could readily estimate the deterministic terms in the dynamical equations, choice accuracy, and decision time of the models across a range of signal-to-noise ratio values. In particular, SINDy performed the best using the more memory-intensive multi-trial approach while trial-averaging of parameters performed more moderately. The single-trial approach, although expectedly not performing as well, may be useful for real-time modeling. Taken together, our work offers alternative approaches for SINDy to uncover the dynamics in perceptual decision making and, more generally, for first-passage time problems.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 3","pages":"569-587"},"PeriodicalIF":2.7,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10908352","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward a Free-Response Paradigm of Decision Making in Spiking Neural Networks 基于脉冲神经网络的自由反应决策模式研究。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-14 DOI: 10.1162/neco_a_01733
Zhichao Zhu;Yang Qi;Wenlian Lu;Zhigang Wang;Lu Cao;Jianfeng Feng
Spiking neural networks (SNNs) have attracted significant interest in the development of brain-inspired computing systems due to their energy efficiency and similarities to biological information processing. In contrast to continuous-valued artificial neural networks, which produce results in a single step, SNNs require multiple steps during inference to achieve a desired accuracy level, resulting in a burden in real-time response and energy efficiency. Inspired by the tradeoff between speed and accuracy in human and animal decision-making processes, which exhibit correlations among reaction times, task complexity, and decision confidence, an inquiry emerges regarding how an SNN model can benefit by implementing these attributes. Here, we introduce a theory of decision making in SNNs by untangling the interplay between signal and noise. Under this theory, we introduce a new learning objective that trains an SNN not only to make the correct decisions but also to shape its confidence. Numerical experiments demonstrate that SNNs trained in this way exhibit improved confidence expression, reduced trial-to-trial variability, and shorter latency to reach the desired accuracy. We then introduce a stopping policy that can stop inference in a way that further enhances the time efficiency of SNNs. The stopping time can serve as an indicator to whether a decision is correct, akin to the reaction time in animal behavior experiments. By integrating stochasticity into decision making, this study opens up new possibilities to explore the capabilities of SNNs and advance SNNs and their applications in complex decision-making scenarios where model performance is limited.
脉冲神经网络(snn)由于其能量效率和与生物信息处理的相似性,在脑启发计算系统的发展中引起了极大的兴趣。连续值人工神经网络只需要一步就能得到结果,而snn在推理过程中需要多个步骤才能达到理想的精度水平,这给实时响应和能源效率带来了负担。受人类和动物决策过程中速度和准确性之间的权衡(反应时间、任务复杂性和决策置信度之间存在相关性)的启发,人们开始研究SNN模型如何通过实现这些属性而受益。在这里,我们通过解开信号和噪声之间的相互作用,介绍了一种snn决策理论。在这个理论下,我们引入了一个新的学习目标,训练SNN不仅做出正确的决策,而且塑造它的信心。数值实验表明,以这种方式训练的snn表现出更好的置信度表达,减少了试验间的可变性,并缩短了达到所需精度的延迟。然后,我们引入了一个停止策略,该策略可以以进一步提高snn时间效率的方式停止推理。停止时间可以作为一个决定是否正确的指标,类似于动物行为实验中的反应时间。通过将随机性整合到决策中,本研究为探索snn的能力开辟了新的可能性,并推进了snn及其在模型性能有限的复杂决策场景中的应用。
{"title":"Toward a Free-Response Paradigm of Decision Making in Spiking Neural Networks","authors":"Zhichao Zhu;Yang Qi;Wenlian Lu;Zhigang Wang;Lu Cao;Jianfeng Feng","doi":"10.1162/neco_a_01733","DOIUrl":"10.1162/neco_a_01733","url":null,"abstract":"Spiking neural networks (SNNs) have attracted significant interest in the development of brain-inspired computing systems due to their energy efficiency and similarities to biological information processing. In contrast to continuous-valued artificial neural networks, which produce results in a single step, SNNs require multiple steps during inference to achieve a desired accuracy level, resulting in a burden in real-time response and energy efficiency. Inspired by the tradeoff between speed and accuracy in human and animal decision-making processes, which exhibit correlations among reaction times, task complexity, and decision confidence, an inquiry emerges regarding how an SNN model can benefit by implementing these attributes. Here, we introduce a theory of decision making in SNNs by untangling the interplay between signal and noise. Under this theory, we introduce a new learning objective that trains an SNN not only to make the correct decisions but also to shape its confidence. Numerical experiments demonstrate that SNNs trained in this way exhibit improved confidence expression, reduced trial-to-trial variability, and shorter latency to reach the desired accuracy. We then introduce a stopping policy that can stop inference in a way that further enhances the time efficiency of SNNs. The stopping time can serve as an indicator to whether a decision is correct, akin to the reaction time in animal behavior experiments. By integrating stochasticity into decision making, this study opens up new possibilities to explore the capabilities of SNNs and advance SNNs and their applications in complex decision-making scenarios where model performance is limited.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 3","pages":"481-521"},"PeriodicalIF":2.7,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10908351","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Recall in Sparse Associative Memories That Use Neurogenesis 利用神经发生提高稀疏联想记忆的回忆准确性。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-02-14 DOI: 10.1162/neco_a_01732
Katy Warr;Jonathon Hare;David Thomas
The creation of future low-power neuromorphic solutions requires specialist spiking neural network (SNN) algorithms that are optimized for neuromorphic settings. One such algorithmic challenge is the ability to recall learned patterns from their noisy variants. Solutions to this problem may be required to memorize vast numbers of patterns based on limited training data and subsequently recall the patterns in the presence of noise. To solve this problem, previous work has explored sparse associative memory (SAM)—associative memory neural models that exploit the principle of sparse neural coding observed in the brain. Research into a subcategory of SAM has been inspired by the biological process of adult neurogenesis, whereby new neurons are generated to facilitate adaptive and effective lifelong learning. Although these neurogenesis models have been demonstrated in previous research, they have limitations in terms of recall memory capacity and robustness to noise. In this article, we provide a unifying framework for characterizing a type of SAM network that has been pretrained using a learning strategy that incorporated a simple neurogenesis model. Using this characterization, we formally define network topology and threshold optimization methods to empirically demonstrate greater than 104 times improvement in memory capacity compared to previous work. We show that these optimizations can facilitate the development of networks that have reduced interneuron connectivity while maintaining high recall efficacy. This paves the way for ongoing research into fast, effective, low-power realizations of associative memory on neuromorphic platforms.
未来的低功耗神经形态解决方案需要专门针对神经形态设置进行优化的峰值神经网络(SNN)算法。其中一个算法挑战就是从嘈杂的变体中回忆已学习模式的能力。这个问题的解决方案可能需要基于有限的训练数据记忆大量的模式,然后在存在噪声的情况下回忆模式。为了解决这个问题,以前的工作已经探索了稀疏联想记忆(SAM)-利用在大脑中观察到的稀疏神经编码原理的联想记忆神经模型。对SAM的一个子类的研究受到成人神经发生的生物学过程的启发,在这个过程中,新的神经元产生以促进适应性和有效的终身学习。虽然这些神经发生模型已经在以前的研究中得到证实,但它们在回忆记忆能力和对噪声的鲁棒性方面存在局限性。在这封信中,我们提供了一个统一的框架来描述一种SAM网络,该网络使用一种包含简单神经发生模型的学习策略进行预训练。使用这种特性,我们正式定义了网络拓扑和阈值优化方法,以经验证明与以前的工作相比,内存容量提高了10$^{{4}}$。我们表明,这些优化可以促进网络的发展,减少神经元之间的连接,同时保持高回忆效率。这为在神经形态平台上快速、有效、低功耗地实现联想记忆铺平了道路。
{"title":"Improving Recall in Sparse Associative Memories That Use Neurogenesis","authors":"Katy Warr;Jonathon Hare;David Thomas","doi":"10.1162/neco_a_01732","DOIUrl":"10.1162/neco_a_01732","url":null,"abstract":"The creation of future low-power neuromorphic solutions requires specialist spiking neural network (SNN) algorithms that are optimized for neuromorphic settings. One such algorithmic challenge is the ability to recall learned patterns from their noisy variants. Solutions to this problem may be required to memorize vast numbers of patterns based on limited training data and subsequently recall the patterns in the presence of noise. To solve this problem, previous work has explored sparse associative memory (SAM)—associative memory neural models that exploit the principle of sparse neural coding observed in the brain. Research into a subcategory of SAM has been inspired by the biological process of adult neurogenesis, whereby new neurons are generated to facilitate adaptive and effective lifelong learning. Although these neurogenesis models have been demonstrated in previous research, they have limitations in terms of recall memory capacity and robustness to noise. In this article, we provide a unifying framework for characterizing a type of SAM network that has been pretrained using a learning strategy that incorporated a simple neurogenesis model. Using this characterization, we formally define network topology and threshold optimization methods to empirically demonstrate greater than 104 times improvement in memory capacity compared to previous work. We show that these optimizations can facilitate the development of networks that have reduced interneuron connectivity while maintaining high recall efficacy. This paves the way for ongoing research into fast, effective, low-power realizations of associative memory on neuromorphic platforms.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 3","pages":"437-480"},"PeriodicalIF":2.7,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Fast Algorithm for the Real-Valued Combinatorial Pure Exploration of the Multi-Armed Bandit 多臂强盗实值组合纯探索的快速算法。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-21 DOI: 10.1162/neco_a_01728
Shintaro Nakamura;Masashi Sugiyama
We study the real-valued combinatorial pure exploration problem in the stochastic multi-armed bandit (R-CPE-MAB). We study the case where the size of the action set is polynomial with respect to the number of arms. In such a case, the R-CPE-MAB can be seen as a special case of the so-called transductive linear bandits. We introduce the combinatorial gap-based exploration (CombGapE) algorithm, whose sample complexity upper-bound-matches the lower bound up to a problem-dependent constant factor. We numerically show that the CombGapE algorithm outperforms existing methods significantly in both synthetic and real-world data sets.
研究随机多臂土匪(R-CPE-MAB)中的实值组合纯勘探问题。我们研究了动作集的大小是关于臂数的多项式的情况。在这种情况下,R-CPE-MAB可以被视为所谓的转导线性强盗的特殊情况。提出了一种基于组合间隙的探索算法(CombGapE),该算法的样本复杂度上界与下界匹配到一个与问题相关的常数因子。数值结果表明,在合成数据集和真实数据集中,CombGapE算法都明显优于现有方法。
{"title":"A Fast Algorithm for the Real-Valued Combinatorial Pure Exploration of the Multi-Armed Bandit","authors":"Shintaro Nakamura;Masashi Sugiyama","doi":"10.1162/neco_a_01728","DOIUrl":"10.1162/neco_a_01728","url":null,"abstract":"We study the real-valued combinatorial pure exploration problem in the stochastic multi-armed bandit (R-CPE-MAB). We study the case where the size of the action set is polynomial with respect to the number of arms. In such a case, the R-CPE-MAB can be seen as a special case of the so-called transductive linear bandits. We introduce the combinatorial gap-based exploration (CombGapE) algorithm, whose sample complexity upper-bound-matches the lower bound up to a problem-dependent constant factor. We numerically show that the CombGapE algorithm outperforms existing methods significantly in both synthetic and real-world data sets.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 2","pages":"294-310"},"PeriodicalIF":2.7,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142774828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Compressive Power of Autoencoders With Linear and ReLU Activation Functions 具有线性和ReLU激活函数的自编码器的压缩能力。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-21 DOI: 10.1162/neco_a_01729
Liangjie Sun;Chenyao Wu;Wai-Ki Ching;Tatsuya Akutsu
In this article, we mainly study the depth and width of autoencoders consisting of rectified linear unit (ReLU) activation functions. An autoencoder is a layered neural network consisting of an encoder, which compresses an input vector to a lower-dimensional vector, and a decoder, which transforms the low-dimensional vector back to the original input vector exactly (or approximately). In a previous study, Melkman et al. (2023) studied the depth and width of autoencoders using linear threshold activation functions with binary input and output vectors. We show that similar theoretical results hold if autoencoders using ReLU activation functions with real input and output vectors are used. Furthermore, we show that it is possible to compress input vectors to one-dimensional vectors using ReLU activation functions, although the size of compressed vectors is trivially Ω(log n) for autoencoders with linear threshold activation functions, where n is the number of input vectors. We also study the cases of linear activation functions. The results suggest that the compressive power of autoencoders using linear activation functions is considerably limited compared with those using ReLU activation functions.
本文主要研究了由整流线性单元(ReLU)激活函数组成的自编码器的深度和宽度。自编码器是一个分层神经网络,由编码器和解码器组成,编码器将输入向量压缩为低维向量,解码器将低维向量精确(或近似)转换回原始输入向量。在先前的研究中,Melkman等人(2023)使用具有二进制输入和输出向量的线性阈值激活函数研究了自编码器的深度和宽度。如果使用具有真实输入和输出向量的ReLU激活函数的自编码器,我们证明了类似的理论结果。此外,我们表明可以使用ReLU激活函数将输入向量压缩为一维向量,尽管对于具有线性阈值激活函数的自编码器,压缩向量的大小是微不足道的Ω(log n),其中n是输入向量的数量。我们还研究了线性激活函数的情况。结果表明,与使用ReLU激活函数的自编码器相比,使用线性激活函数的自编码器的压缩能力明显有限。
{"title":"On the Compressive Power of Autoencoders With Linear and ReLU Activation Functions","authors":"Liangjie Sun;Chenyao Wu;Wai-Ki Ching;Tatsuya Akutsu","doi":"10.1162/neco_a_01729","DOIUrl":"10.1162/neco_a_01729","url":null,"abstract":"In this article, we mainly study the depth and width of autoencoders consisting of rectified linear unit (ReLU) activation functions. An autoencoder is a layered neural network consisting of an encoder, which compresses an input vector to a lower-dimensional vector, and a decoder, which transforms the low-dimensional vector back to the original input vector exactly (or approximately). In a previous study, Melkman et al. (2023) studied the depth and width of autoencoders using linear threshold activation functions with binary input and output vectors. We show that similar theoretical results hold if autoencoders using ReLU activation functions with real input and output vectors are used. Furthermore, we show that it is possible to compress input vectors to one-dimensional vectors using ReLU activation functions, although the size of compressed vectors is trivially Ω(log n) for autoencoders with linear threshold activation functions, where n is the number of input vectors. We also study the cases of linear activation functions. The results suggest that the compressive power of autoencoders using linear activation functions is considerably limited compared with those using ReLU activation functions.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":"37 2","pages":"235-259"},"PeriodicalIF":2.7,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142774849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1