首页 > 最新文献

Neural Computation最新文献

英文 中文
Active Predictive Coding: A Unifying Neural Model for Active Perception, Compositional Learning, and Hierarchical Planning 主动预测编码:主动感知、组合学习和分层规划的统一神经模型
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-12-12 DOI: 10.1162/neco_a_01627
Rajesh P. N. Rao;Dimitrios C. Gklezakos;Vishwas Sathish
There is growing interest in predictive coding as a model of how the brain learns through predictions and prediction errors. Predictive coding models have traditionally focused on sensory coding and perception. Here we introduce active predictive coding (APC) as a unifying model for perception, action, and cognition. The APC model addresses important open problems in cognitive science and AI, including (1) how we learn compositional representations (e.g., part-whole hierarchies for equivariant vision) and (2) how we solve large-scale planning problems, which are hard for traditional reinforcement learning, by composing complex state dynamics and abstract actions from simpler dynamics and primitive actions. By using hypernetworks, self-supervised learning, and reinforcement learning, APC learns hierarchical world models by combining task-invariant state transition networks and task-dependent policy networks at multiple abstraction levels. We illustrate the applicability of the APC model to active visual perception and hierarchical planning. Our results represent, to our knowledge, the first proof-of-concept demonstration of a unified approach to addressing the part-whole learning problem in vision, the nested reference frames learning problem in cognition, and the integrated state-action hierarchy learning problem in reinforcement learning.
预测编码作为大脑如何通过预测和预测错误进行学习的模型,越来越受到人们的关注。预测编码模型历来侧重于感官编码和感知。在这里,我们将主动预测编码(APC)作为感知、行动和认知的统一模型加以介绍。主动预测编码模型解决了认知科学和人工智能领域的重要开放性问题,包括:(1)我们如何学习组合表征(如等变视觉的部分-整体层次结构);(2)我们如何通过将复杂的状态动态和抽象动作与较简单的动态和原始动作组合起来,解决传统强化学习难以解决的大规模规划问题。通过使用超网络、自监督学习和强化学习,APC 在多个抽象层级上结合任务不变状态转换网络和任务相关策略网络,学习分层世界模型。我们说明了 APC 模型在主动视觉感知和分层规划方面的适用性。据我们所知,我们的研究结果代表了首次概念验证,展示了一种统一的方法来解决视觉中的部分-整体学习问题、认知中的嵌套参照系学习问题以及强化学习中的综合状态-行动层次学习问题。
{"title":"Active Predictive Coding: A Unifying Neural Model for Active Perception, Compositional Learning, and Hierarchical Planning","authors":"Rajesh P. N. Rao;Dimitrios C. Gklezakos;Vishwas Sathish","doi":"10.1162/neco_a_01627","DOIUrl":"10.1162/neco_a_01627","url":null,"abstract":"There is growing interest in predictive coding as a model of how the brain learns through predictions and prediction errors. Predictive coding models have traditionally focused on sensory coding and perception. Here we introduce active predictive coding (APC) as a unifying model for perception, action, and cognition. The APC model addresses important open problems in cognitive science and AI, including (1) how we learn compositional representations (e.g., part-whole hierarchies for equivariant vision) and (2) how we solve large-scale planning problems, which are hard for traditional reinforcement learning, by composing complex state dynamics and abstract actions from simpler dynamics and primitive actions. By using hypernetworks, self-supervised learning, and reinforcement learning, APC learns hierarchical world models by combining task-invariant state transition networks and task-dependent policy networks at multiple abstraction levels. We illustrate the applicability of the APC model to active visual perception and hierarchical planning. Our results represent, to our knowledge, the first proof-of-concept demonstration of a unified approach to addressing the part-whole learning problem in vision, the nested reference frames learning problem in cognition, and the integrated state-action hierarchy learning problem in reinforcement learning.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138489040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Filter Model of Cerebellum for Biological Muscle Control With Spike Train Inputs 小脑自适应滤波模型用于具有尖峰训练输入的生物肌肉控制。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-11-07 DOI: 10.1162/neco_a_01617
Emma Wilson
Prior applications of the cerebellar adaptive filter model have included a range of tasks within simulated and robotic systems. However, this has been limited to systems driven by continuous signals. Here, the adaptive filter model of the cerebellum is applied to the control of a system driven by spiking inputs by considering the problem of controlling muscle force. The performance of the standard adaptive filter algorithm is compared with the algorithm with a modified learning rule that minimizes inputs and a simple proportional-integral-derivative (PID) controller. Control performance is evaluated in terms of the number of spikes, the accuracy of spike input locations, and the accuracy of muscle force output. Results show that the cerebellar adaptive filter model can be applied without change to the control of systems driven by spiking inputs. The cerebellar algorithm results in good agreement between input spikes and force outputs and significantly improves on a PID controller. Input minimization can be used to reduce the number of spike inputs, but at the expense of a decrease in accuracy of spike input location and force output. This work extends the applications of the cerebellar algorithm and demonstrates the potential of the adaptive filter model to be used to improve functional electrical stimulation muscle control.
小脑自适应滤波器模型的先前应用包括模拟和机器人系统中的一系列任务。然而,这仅限于由连续信号驱动的系统。这里,通过考虑控制肌肉力量的问题,将小脑的自适应滤波器模型应用于由尖峰输入驱动的系统的控制。将标准自适应滤波器算法的性能与具有最小化输入的改进学习规则和简单比例积分微分(PID)控制器的算法进行了比较。控制性能是根据尖峰的数量、尖峰输入位置的准确性和肌肉力量输出的准确性来评估的。结果表明,小脑自适应滤波器模型可以在不改变尖峰输入驱动系统控制的情况下应用。小脑算法在输入尖峰和力输出之间产生良好的一致性,并显著改进了PID控制器。输入最小化可用于减少尖峰输入的数量,但以降低尖峰输入位置和力输出的准确性为代价。这项工作扩展了小脑算法的应用,并证明了自适应滤波器模型用于改善功能性电刺激肌肉控制的潜力。
{"title":"Adaptive Filter Model of Cerebellum for Biological Muscle Control With Spike Train Inputs","authors":"Emma Wilson","doi":"10.1162/neco_a_01617","DOIUrl":"10.1162/neco_a_01617","url":null,"abstract":"Prior applications of the cerebellar adaptive filter model have included a range of tasks within simulated and robotic systems. However, this has been limited to systems driven by continuous signals. Here, the adaptive filter model of the cerebellum is applied to the control of a system driven by spiking inputs by considering the problem of controlling muscle force. The performance of the standard adaptive filter algorithm is compared with the algorithm with a modified learning rule that minimizes inputs and a simple proportional-integral-derivative (PID) controller. Control performance is evaluated in terms of the number of spikes, the accuracy of spike input locations, and the accuracy of muscle force output. Results show that the cerebellar adaptive filter model can be applied without change to the control of systems driven by spiking inputs. The cerebellar algorithm results in good agreement between input spikes and force outputs and significantly improves on a PID controller. Input minimization can be used to reduce the number of spike inputs, but at the expense of a decrease in accuracy of spike input location and force output. This work extends the applications of the cerebellar algorithm and demonstrates the potential of the adaptive filter model to be used to improve functional electrical stimulation muscle control.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41240925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Training a Hyperdimensional Computing Classifier Using a Threshold on Its Confidence 使用置信度阈值训练超维计算分类器。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-11-07 DOI: 10.1162/neco_a_01618
Laura Smets;Werner Van Leekwijck;Ing Jyh Tsang;Steven Latré
Hyperdimensional computing (HDC) has become popular for light-weight and energy-efficient machine learning, suitable for wearable Internet-of-Things devices and near-sensor or on-device processing. HDC is computationally less complex than traditional deep learning algorithms and achieves moderate to good classification performance. This letter proposes to extend the training procedure in HDC by taking into account not only wrongly classified samples but also samples that are correctly classified by the HDC model but with low confidence. We introduce a confidence threshold that can be tuned for each data set to achieve the best classification accuracy. The proposed training procedure is tested on UCIHAR, CTG, ISOLET, and HAND data sets for which the performance consistently improves compared to the baseline across a range of confidence threshold values. The extended training procedure also results in a shift toward higher confidence values of the correctly classified samples, making the classifier not only more accurate but also more confident about its predictions.
超维计算(HDC)在轻量级和节能的机器学习中变得很流行,适用于可穿戴物联网设备和近传感器或设备上处理。HDC在计算上比传统的深度学习算法复杂,并且实现了中等到良好的分类性能。这封信建议扩展HDC中的训练程序,不仅要考虑错误分类的样本,还要考虑HDC模型正确分类但置信度低的样本。我们引入了一个置信阈值,可以对每个数据集进行调整,以实现最佳的分类精度。所提出的训练程序在UCIHAR、CTG、ISOLET和HAND数据集上进行了测试,在一系列置信阈值范围内,与基线相比,这些数据集的性能不断提高。扩展的训练过程还导致正确分类的样本向更高置信度值的转变,使分类器不仅更准确,而且对其预测更有信心。
{"title":"Training a Hyperdimensional Computing Classifier Using a Threshold on Its Confidence","authors":"Laura Smets;Werner Van Leekwijck;Ing Jyh Tsang;Steven Latré","doi":"10.1162/neco_a_01618","DOIUrl":"10.1162/neco_a_01618","url":null,"abstract":"Hyperdimensional computing (HDC) has become popular for light-weight and energy-efficient machine learning, suitable for wearable Internet-of-Things devices and near-sensor or on-device processing. HDC is computationally less complex than traditional deep learning algorithms and achieves moderate to good classification performance. This letter proposes to extend the training procedure in HDC by taking into account not only wrongly classified samples but also samples that are correctly classified by the HDC model but with low confidence. We introduce a confidence threshold that can be tuned for each data set to achieve the best classification accuracy. The proposed training procedure is tested on UCIHAR, CTG, ISOLET, and HAND data sets for which the performance consistently improves compared to the baseline across a range of confidence threshold values. The extended training procedure also results in a shift toward higher confidence values of the correctly classified samples, making the classifier not only more accurate but also more confident about its predictions.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41240929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Robustness to Transformations Across Categories: Is Robustness Driven by Invariant Neural Representations? 跨类别变换的鲁棒性:鲁棒性是由不变的神经表示驱动的吗?
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-11-07 DOI: 10.1162/neco_a_01621
Hojin Jang;Syed Suleman Abbas Zaidi;Xavier Boix;Neeraj Prasad;Sharon Gilad-Gutnick;Shlomit Ben-Ami;Pawan Sinha
Deep convolutional neural networks (DCNNs) have demonstrated impressive robustness to recognize objects under transformations (e.g., blur or noise) when these transformations are included in the training set. A hypothesis to explain such robustness is that DCNNs develop invariant neural representations that remain unaltered when the image is transformed. However, to what extent this hypothesis holds true is an outstanding question, as robustness to transformations could be achieved with properties different from invariance; for example, parts of the network could be specialized to recognize either transformed or nontransformed images. This article investigates the conditions under which invariant neural representations emerge by leveraging that they facilitate robustness to transformations beyond the training distribution. Concretely, we analyze a training paradigm in which only some object categories are seen transformed during training and evaluate whether the DCNN is robust to transformations across categories not seen transformed. Our results with state-of-the-art DCNNs indicate that invariant neural representations do not always drive robustness to transformations, as networks show robustness for categories seen transformed during training even in the absence of invariant neural representations. Invariance emerges only as the number of transformed categories in the training set is increased. This phenomenon is much more prominent with local transformations such as blurring and high-pass filtering than geometric transformations such as rotation and thinning, which entail changes in the spatial arrangement of the object. Our results contribute to a better understanding of invariant neural representations in deep learning and the conditions under which it spontaneously emerges.
深度卷积神经网络(DCNN)在将变换(例如模糊或噪声)包括在训练集中时,在识别变换下的对象方面表现出了令人印象深刻的鲁棒性。解释这种稳健性的一个假设是,DCNN开发出不变的神经表示,当图像被变换时,这些神经表示保持不变。然而,这一假设在多大程度上成立是一个悬而未决的问题,因为对变换的鲁棒性可以通过不同于不变性的性质来实现;例如,网络的一部分可以专门用于识别变换的图像或未变换的图像。本文研究了不变神经表示出现的条件,利用它们促进了对训练分布之外的变换的鲁棒性。具体来说,我们分析了一种训练范式,其中在训练过程中只看到一些对象类别被转换,并评估DCNN是否对未被转换的类别之间的转换具有鲁棒性。我们对最先进的DCNN的结果表明,不变神经表示并不总是驱动对变换的鲁棒性,因为即使在没有不变神经表示的情况下,网络对训练过程中变换的类别也表现出鲁棒性。只有当训练集中变换类别的数量增加时,不变性才会出现。这种现象在局部变换(如模糊和高通滤波)中比几何变换(如旋转和细化)更为突出,后者需要改变对象的空间排列。我们的研究结果有助于更好地理解深度学习中的不变神经表征及其自发出现的条件。
{"title":"Robustness to Transformations Across Categories: Is Robustness Driven by Invariant Neural Representations?","authors":"Hojin Jang;Syed Suleman Abbas Zaidi;Xavier Boix;Neeraj Prasad;Sharon Gilad-Gutnick;Shlomit Ben-Ami;Pawan Sinha","doi":"10.1162/neco_a_01621","DOIUrl":"10.1162/neco_a_01621","url":null,"abstract":"Deep convolutional neural networks (DCNNs) have demonstrated impressive robustness to recognize objects under transformations (e.g., blur or noise) when these transformations are included in the training set. A hypothesis to explain such robustness is that DCNNs develop invariant neural representations that remain unaltered when the image is transformed. However, to what extent this hypothesis holds true is an outstanding question, as robustness to transformations could be achieved with properties different from invariance; for example, parts of the network could be specialized to recognize either transformed or nontransformed images. This article investigates the conditions under which invariant neural representations emerge by leveraging that they facilitate robustness to transformations beyond the training distribution. Concretely, we analyze a training paradigm in which only some object categories are seen transformed during training and evaluate whether the DCNN is robust to transformations across categories not seen transformed. Our results with state-of-the-art DCNNs indicate that invariant neural representations do not always drive robustness to transformations, as networks show robustness for categories seen transformed during training even in the absence of invariant neural representations. Invariance emerges only as the number of transformed categories in the training set is increased. This phenomenon is much more prominent with local transformations such as blurring and high-pass filtering than geometric transformations such as rotation and thinning, which entail changes in the spatial arrangement of the object. Our results contribute to a better understanding of invariant neural representations in deep learning and the conditions under which it spontaneously emerges.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41240928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Predictive Coding as a Neuromorphic Alternative to Backpropagation: A Critical Evaluation 预测编码作为反向传播的一种神经形态替代:一项关键评估。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-11-07 DOI: 10.1162/neco_a_01620
Umais Zahid;Qinghai Guo;Zafeirios Fountas
Backpropagation has rapidly become the workhorse credit assignment algorithm for modern deep learning methods. Recently, modified forms of predictive coding (PC), an algorithm with origins in computational neuroscience, have been shown to result in approximately or exactly equal parameter updates to those under backpropagation. Due to this connection, it has been suggested that PC can act as an alternative to backpropagation with desirable properties that may facilitate implementation in neuromorphic systems. Here, we explore these claims using the different contemporary PC variants proposed in the literature. We obtain time complexity bounds for these PC variants, which we show are lower bounded by backpropagation. We also present key properties of these variants that have implications for neurobiological plausibility and their interpretations, particularly from the perspective of standard PC as a variational Bayes algorithm for latent probabilistic models. Our findings shed new light on the connection between the two learning frameworks and suggest that in its current forms, PC may have more limited potential as a direct replacement of backpropagation than previously envisioned.
反向传播已迅速成为现代深度学习方法的主要学分分配算法。最近,预测编码(PC)的改进形式,一种起源于计算神经科学的算法,已被证明会导致与反向传播下的参数更新大致或完全相等的参数更新。由于这种联系,有人建议PC可以作为反向传播的替代品,具有理想的特性,有助于在神经形态系统中实现。在这里,我们使用文献中提出的不同的当代PC变体来探讨这些说法。我们获得了这些PC变体的时间复杂度边界,我们通过反向传播证明了其是下界。我们还介绍了这些变体的关键特性,这些特性对神经生物学的合理性及其解释具有启示意义,特别是从标准PC作为潜在概率模型的变分贝叶斯算法的角度来看。我们的发现为这两种学习框架之间的联系提供了新的线索,并表明在目前的形式下,PC作为反向传播的直接替代品的潜力可能比以前设想的更为有限。
{"title":"Predictive Coding as a Neuromorphic Alternative to Backpropagation: A Critical Evaluation","authors":"Umais Zahid;Qinghai Guo;Zafeirios Fountas","doi":"10.1162/neco_a_01620","DOIUrl":"10.1162/neco_a_01620","url":null,"abstract":"Backpropagation has rapidly become the workhorse credit assignment algorithm for modern deep learning methods. Recently, modified forms of predictive coding (PC), an algorithm with origins in computational neuroscience, have been shown to result in approximately or exactly equal parameter updates to those under backpropagation. Due to this connection, it has been suggested that PC can act as an alternative to backpropagation with desirable properties that may facilitate implementation in neuromorphic systems. Here, we explore these claims using the different contemporary PC variants proposed in the literature. We obtain time complexity bounds for these PC variants, which we show are lower bounded by backpropagation. We also present key properties of these variants that have implications for neurobiological plausibility and their interpretations, particularly from the perspective of standard PC as a variational Bayes algorithm for latent probabilistic models. Our findings shed new light on the connection between the two learning frameworks and suggest that in its current forms, PC may have more limited potential as a direct replacement of backpropagation than previously envisioned.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41240927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Generalized Low-Rank Update: Model Parameter Bounds for Low-Rank Training Data Modifications 广义低秩更新:低秩训练数据修改的模型参数界。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-11-07 DOI: 10.1162/neco_a_01619
Hiroyuki Hanada;Noriaki Hashimoto;Kouichi Taji;Ichiro Takeuchi
In this study, we have developed an incremental machine learning (ML) method that efficiently obtains the optimal model when a small number of instances or features are added or removed. This problem holds practical importance in model selection, such as cross-validation (CV) and feature selection. Among the class of ML methods known as linear estimators, there exists an efficient model update framework, the low-rank update, that can effectively handle changes in a small number of rows and columns within the data matrix. However, for ML methods beyond linear estimators, there is currently no comprehensive framework available to obtain knowledge about the updated solution within a specific computational complexity. In light of this, our study introduces a the generalized low-rank update (GLRU) method, which extends the low-rank update framework of linear estimators to ML methods formulated as a certain class of regularized empirical risk minimization, including commonly used methods such as support vector machines and logistic regression. The proposed GLRU method not only expands the range of its applicability but also provides information about the updated solutions with a computational complexity proportional to the number of data set changes. To demonstrate the effectiveness of the GLRU method, we conduct experiments showcasing its efficiency in performing cross-validation and feature selection compared to other baseline methods.
在这项研究中,我们开发了一种增量机器学习(ML)方法,当添加或删除少量实例或特征时,该方法可以有效地获得最优模型。这个问题在模型选择中具有实际意义,例如交叉验证(CV)和特征选择。在被称为线性估计量的ML方法中,存在一种高效的模型更新框架,即低秩更新,它可以有效地处理数据矩阵中少量行和列的变化。然而,对于线性估计量之外的ML方法,目前还没有一个全面的框架可以在特定的计算复杂度内获得关于更新解决方案的知识。有鉴于此,我们的研究引入了一种广义低秩更新(GLRU)方法,该方法将线性估计量的低秩更新框架扩展到ML方法,该ML方法被公式化为一类正则化经验风险最小化,包括常用的方法,如支持向量机和逻辑回归。所提出的GLRU方法不仅扩大了其适用范围,而且还提供了关于更新的解决方案的信息,其计算复杂度与数据集变化的数量成比例。为了证明GLRU方法的有效性,我们进行了实验,展示了与其他基线方法相比,GLRU方法在执行交叉验证和特征选择方面的效率。
{"title":"Generalized Low-Rank Update: Model Parameter Bounds for Low-Rank Training Data Modifications","authors":"Hiroyuki Hanada;Noriaki Hashimoto;Kouichi Taji;Ichiro Takeuchi","doi":"10.1162/neco_a_01619","DOIUrl":"10.1162/neco_a_01619","url":null,"abstract":"In this study, we have developed an incremental machine learning (ML) method that efficiently obtains the optimal model when a small number of instances or features are added or removed. This problem holds practical importance in model selection, such as cross-validation (CV) and feature selection. Among the class of ML methods known as linear estimators, there exists an efficient model update framework, the low-rank update, that can effectively handle changes in a small number of rows and columns within the data matrix. However, for ML methods beyond linear estimators, there is currently no comprehensive framework available to obtain knowledge about the updated solution within a specific computational complexity. In light of this, our study introduces a the generalized low-rank update (GLRU) method, which extends the low-rank update framework of linear estimators to ML methods formulated as a certain class of regularized empirical risk minimization, including commonly used methods such as support vector machines and logistic regression. The proposed GLRU method not only expands the range of its applicability but also provides information about the updated solutions with a computational complexity proportional to the number of data set changes. To demonstrate the effectiveness of the GLRU method, we conduct experiments showcasing its efficiency in performing cross-validation and feature selection compared to other baseline methods.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41240926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Organization of Nonlinearly Coupled Neural Fluctuations Into Synergistic Population Codes 非线性耦合神经波动到协同总体码中的自组织。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-10-10 DOI: 10.1162/neco_a_01612
Hengyuan Ma;Yang Qi;Pulin Gong;Jie Zhang;Wen-lian Lu;Jianfeng Feng
Neural activity in the brain exhibits correlated fluctuations that may strongly influence the properties of neural population coding. However, how such correlated neural fluctuations may arise from the intrinsic neural circuit dynamics and subsequently affect the computational properties of neural population activity remains poorly understood. The main difficulty lies in resolving the nonlinear coupling between correlated fluctuations with the overall dynamics of the system. In this study, we investigate the emergence of synergistic neural population codes from the intrinsic dynamics of correlated neural fluctuations in a neural circuit model capturing realistic nonlinear noise coupling of spiking neurons. We show that a rich repertoire of spatial correlation patterns naturally emerges in a bump attractor network and further reveals the dynamical regime under which the interplay between differential and noise correlations leads to synergistic codes. Moreover, we find that negative correlations may induce stable bound states between two bumps, a phenomenon previously unobserved in firing rate models. These noise-induced effects of bump attractors lead to a number of computational advantages including enhanced working memory capacity and efficient spatiotemporal multiplexing and can account for a range of cognitive and behavioral phenomena related to working memory. This study offers a dynamical approach to investigating realistic correlated neural fluctuations and insights to their roles in cortical computations.
大脑中的神经活动表现出相关的波动,这可能会强烈影响神经群体编码的特性。然而,这种相关的神经波动是如何从内在的神经回路动力学中产生并随后影响神经群体活动的计算特性的,目前还知之甚少。主要困难在于解决相关波动与系统整体动力学之间的非线性耦合。在这项研究中,我们研究了在捕捉尖峰神经元的真实非线性噪声耦合的神经电路模型中,从相关神经波动的内在动力学中出现的协同神经群体代码。我们表明,在凸点吸引器网络中自然会出现丰富的空间相关模式,并进一步揭示了微分和噪声相关性之间的相互作用导致协同代码的动力学机制。此外,我们发现负相关性可能会在两个凸点之间诱导稳定的束缚态,这是以前在发射率模型中未观察到的现象。凸点吸引器的这些噪声诱导效应带来了许多计算优势,包括增强的工作记忆容量和高效的时空复用,并可以解释与工作记忆相关的一系列认知和行为现象。这项研究为研究现实的相关神经波动提供了一种动力学方法,并深入了解了它们在皮层计算中的作用。
{"title":"Self-Organization of Nonlinearly Coupled Neural Fluctuations Into Synergistic Population Codes","authors":"Hengyuan Ma;Yang Qi;Pulin Gong;Jie Zhang;Wen-lian Lu;Jianfeng Feng","doi":"10.1162/neco_a_01612","DOIUrl":"10.1162/neco_a_01612","url":null,"abstract":"Neural activity in the brain exhibits correlated fluctuations that may strongly influence the properties of neural population coding. However, how such correlated neural fluctuations may arise from the intrinsic neural circuit dynamics and subsequently affect the computational properties of neural population activity remains poorly understood. The main difficulty lies in resolving the nonlinear coupling between correlated fluctuations with the overall dynamics of the system. In this study, we investigate the emergence of synergistic neural population codes from the intrinsic dynamics of correlated neural fluctuations in a neural circuit model capturing realistic nonlinear noise coupling of spiking neurons. We show that a rich repertoire of spatial correlation patterns naturally emerges in a bump attractor network and further reveals the dynamical regime under which the interplay between differential and noise correlations leads to synergistic codes. Moreover, we find that negative correlations may induce stable bound states between two bumps, a phenomenon previously unobserved in firing rate models. These noise-induced effects of bump attractors lead to a number of computational advantages including enhanced working memory capacity and efficient spatiotemporal multiplexing and can account for a range of cognitive and behavioral phenomena related to working memory. This study offers a dynamical approach to investigating realistic correlated neural fluctuations and insights to their roles in cortical computations.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41166103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Winning the Lottery With Neural Connectivity Constraints: Faster Learning Across Cognitive Tasks With Spatially Constrained Sparse RNNs 利用神经连接约束赢得彩票:利用空间约束稀疏RNN在认知任务中更快地学习。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-10-10 DOI: 10.1162/neco_a_01613
Mikail Khona;Sarthak Chandra;Joy J. Ma;Ila R. Fiete
Recurrent neural networks (RNNs) are often used to model circuits in the brain and can solve a variety of difficult computational problems requiring memory, error correction, or selection (Hopfield, 1982; Maass et al., 2002; Maass, 2011). However, fully connected RNNs contrast structurally with their biological counterparts, which are extremely sparse (about 0.1%). Motivated by the neocortex, where neural connectivity is constrained by physical distance along cortical sheets and other synaptic wiring costs, we introduce locality masked RNNs (LM-RNNs) that use task-agnostic predetermined graphs with sparsity as low as 4%. We study LM-RNNs in a multitask learning setting relevant to cognitive systems neuroscience with a commonly used set of tasks, 20-Cog-tasks (Yang et al., 2019). We show through reductio ad absurdum that 20-Cog-tasks can be solved by a small pool of separated autapses that we can mechanistically analyze and understand. Thus, these tasks fall short of the goal of inducing complex recurrent dynamics and modular structure in RNNs. We next contribute a new cognitive multitask battery, Mod-Cog, consisting of up to 132 tasks that expands by about seven-fold the number of tasks and task complexity of 20-Cog-tasks. Importantly, while autapses can solve the simple 20-Cog-tasks, the expanded task set requires richer neural architectures and continuous attractor dynamics. On these tasks, we show that LM-RNNs with an optimal sparsity result in faster training and better data efficiency than fully connected networks.
递归神经网络(RNN)通常用于对大脑中的电路进行建模,可以解决各种需要记忆、纠错或选择的计算难题(Hopfield,1982;Maas等人,2002年;Maas,2011年)。然而,完全连接的RNN在结构上与极为稀疏(约0.1%)的生物对应物形成对比。受新皮质的启发,神经连接受到沿皮质片的物理距离和其他突触布线成本的限制,我们引入了局部掩蔽RNN(LM RNN),它使用稀疏度低至4%的任务不可知的预定图。我们在与认知系统神经科学相关的多任务学习环境中研究LM RNN,使用一组常用的任务,即20个Cog任务(Yang et al.,2019)。我们通过荒谬的还原表明,20个Cog任务可以通过我们可以机械地分析和理解的一小部分分离的自闭症来解决。因此,这些任务没有达到在RNN中引入复杂的递归动力学和模块化结构的目标。接下来,我们贡献了一个新的认知多任务组Mod Cog,它由多达132个任务组成,任务数量和任务复杂性是20个Cog任务的7倍。重要的是,虽然自闭症可以解决简单的20个Cog任务,但扩展的任务集需要更丰富的神经结构和连续的吸引子动力学。在这些任务中,我们表明,与完全连接的网络相比,具有最佳稀疏性的LM RNN可以获得更快的训练和更好的数据效率。
{"title":"Winning the Lottery With Neural Connectivity Constraints: Faster Learning Across Cognitive Tasks With Spatially Constrained Sparse RNNs","authors":"Mikail Khona;Sarthak Chandra;Joy J. Ma;Ila R. Fiete","doi":"10.1162/neco_a_01613","DOIUrl":"10.1162/neco_a_01613","url":null,"abstract":"Recurrent neural networks (RNNs) are often used to model circuits in the brain and can solve a variety of difficult computational problems requiring memory, error correction, or selection (Hopfield, 1982; Maass et al., 2002; Maass, 2011). However, fully connected RNNs contrast structurally with their biological counterparts, which are extremely sparse (about 0.1%). Motivated by the neocortex, where neural connectivity is constrained by physical distance along cortical sheets and other synaptic wiring costs, we introduce locality masked RNNs (LM-RNNs) that use task-agnostic predetermined graphs with sparsity as low as 4%. We study LM-RNNs in a multitask learning setting relevant to cognitive systems neuroscience with a commonly used set of tasks, 20-Cog-tasks (Yang et al., 2019). We show through reductio ad absurdum that 20-Cog-tasks can be solved by a small pool of separated autapses that we can mechanistically analyze and understand. Thus, these tasks fall short of the goal of inducing complex recurrent dynamics and modular structure in RNNs. We next contribute a new cognitive multitask battery, Mod-Cog, consisting of up to 132 tasks that expands by about seven-fold the number of tasks and task complexity of 20-Cog-tasks. Importantly, while autapses can solve the simple 20-Cog-tasks, the expanded task set requires richer neural architectures and continuous attractor dynamics. On these tasks, we show that LM-RNNs with an optimal sparsity result in faster training and better data efficiency than fully connected networks.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41175279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Reducing Catastrophic Forgetting With Associative Learning: A Lesson From Fruit Flies 用联想学习减少灾难性遗忘:果蝇的经验教训。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-10-10 DOI: 10.1162/neco_a_01615
Yang Shen;Sanjoy Dasgupta;Saket Navlakha
Catastrophic forgetting remains an outstanding challenge in continual learning. Recently, methods inspired by the brain, such as continual representation learning and memory replay, have been used to combat catastrophic forgetting. Associative learning (retaining associations between inputs and outputs, even after good representations are learned) plays an important function in the brain; however, its role in continual learning has not been carefully studied. Here, we identified a two-layer neural circuit in the fruit fly olfactory system that performs continual associative learning between odors and their associated valences. In the first layer, inputs (odors) are encoded using sparse, high-dimensional representations, which reduces memory interference by activating nonoverlapping populations of neurons for different odors. In the second layer, only the synapses between odor-activated neurons and the odor’s associated output neuron are modified during learning; the rest of the weights are frozen to prevent unrelated memories from being overwritten. We prove theoretically that these two perceptron-like layers help reduce catastrophic forgetting compared to the original perceptron algorithm, under continual learning. We then show empirically on benchmark data sets that this simple and lightweight architecture outperforms other popular neural-inspired algorithms when also using a two-layer feedforward architecture. Overall, fruit flies evolved an efficient continual associative learning algorithm, and circuit mechanisms from neuroscience can be translated to improve machine computation.
灾难性遗忘仍然是持续学习中的一个突出挑战。最近,受大脑启发的方法,如持续表征学习和记忆重放,已被用于对抗灾难性遗忘。联想学习(保持输入和输出之间的关联,即使在学习了良好的表征之后)在大脑中发挥着重要作用;然而,它在持续学习中的作用并没有得到认真的研究。在这里,我们在果蝇嗅觉系统中发现了一个两层神经回路,它在气味及其相关价态之间进行连续的联想学习。在第一层中,使用稀疏的高维表示对输入(气味)进行编码,这通过激活不同气味的不重叠神经元群体来减少记忆干扰。在第二层中,只有气味激活神经元和气味相关输出神经元之间的突触在学习过程中被修改;其余的权重被冻结以防止不相关的存储器被重写。我们从理论上证明,在连续学习下,与原始感知器算法相比,这两个感知器样层有助于减少灾难性遗忘。然后,我们在基准数据集上实证表明,当同样使用三层前馈架构时,这种简单轻便的架构优于其他流行的中性启发算法。总的来说,果蝇进化出了一种高效的连续联想学习算法,神经科学中的电路机制可以转化为改进机器计算。
{"title":"Reducing Catastrophic Forgetting With Associative Learning: A Lesson From Fruit Flies","authors":"Yang Shen;Sanjoy Dasgupta;Saket Navlakha","doi":"10.1162/neco_a_01615","DOIUrl":"10.1162/neco_a_01615","url":null,"abstract":"Catastrophic forgetting remains an outstanding challenge in continual learning. Recently, methods inspired by the brain, such as continual representation learning and memory replay, have been used to combat catastrophic forgetting. Associative learning (retaining associations between inputs and outputs, even after good representations are learned) plays an important function in the brain; however, its role in continual learning has not been carefully studied. Here, we identified a two-layer neural circuit in the fruit fly olfactory system that performs continual associative learning between odors and their associated valences. In the first layer, inputs (odors) are encoded using sparse, high-dimensional representations, which reduces memory interference by activating nonoverlapping populations of neurons for different odors. In the second layer, only the synapses between odor-activated neurons and the odor’s associated output neuron are modified during learning; the rest of the weights are frozen to prevent unrelated memories from being overwritten. We prove theoretically that these two perceptron-like layers help reduce catastrophic forgetting compared to the original perceptron algorithm, under continual learning. We then show empirically on benchmark data sets that this simple and lightweight architecture outperforms other popular neural-inspired algorithms when also using a two-layer feedforward architecture. Overall, fruit flies evolved an efficient continual associative learning algorithm, and circuit mechanisms from neuroscience can be translated to improve machine computation.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41151453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Tutorial on the Spectral Theory of Markov Chains 马尔可夫链谱理论教程。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2023-10-10 DOI: 10.1162/neco_a_01611
Eddie Seabrook;Laurenz Wiskott
Markov chains are a class of probabilistic models that have achieved widespread application in the quantitative sciences. This is in part due to their versatility, but is compounded by the ease with which they can be probed analytically. This tutorial provides an in-depth introduction to Markov chains and explores their connection to graphs and random walks. We use tools from linear algebra and graph theory to describe the transition matrices of different types of Markov chains, with a particular focus on exploring properties of the eigenvalues and eigenvectors corresponding to these matrices. The results presented are relevant to a number of methods in machine learning and data mining, which we describe at various stages. Rather than being a novel academic study in its own right, this text presents a collection of known results, together with some new concepts. Moreover, the tutorial focuses on offering intuition to readers rather than formal understanding and only assumes basic exposure to concepts from linear algebra and probability theory. It is therefore accessible to students and researchers from a wide variety of disciplines.
马尔可夫链是一类在定量科学中得到广泛应用的概率模型。这在一定程度上是由于它们的多功能性,但由于它们可以很容易地进行分析研究,这一点更加复杂。本教程深入介绍了马尔可夫链,并探讨了它们与图和随机游动的联系。我们使用线性代数和图论中的工具来描述不同类型马尔可夫链的转移矩阵,特别关注于探索与这些矩阵相对应的特征值和特征向量的性质。所给出的结果与我们在不同阶段描述的机器学习和数据挖掘中的许多方法有关。本文不是一项新颖的学术研究,而是一组已知的结果,以及一些新的概念。此外,本教程侧重于向读者提供直觉,而不是形式上的理解,并且只假设基本了解线性代数和概率论的概念。因此,来自不同学科的学生和研究人员都可以使用它。
{"title":"A Tutorial on the Spectral Theory of Markov Chains","authors":"Eddie Seabrook;Laurenz Wiskott","doi":"10.1162/neco_a_01611","DOIUrl":"10.1162/neco_a_01611","url":null,"abstract":"Markov chains are a class of probabilistic models that have achieved widespread application in the quantitative sciences. This is in part due to their versatility, but is compounded by the ease with which they can be probed analytically. This tutorial provides an in-depth introduction to Markov chains and explores their connection to graphs and random walks. We use tools from linear algebra and graph theory to describe the transition matrices of different types of Markov chains, with a particular focus on exploring properties of the eigenvalues and eigenvectors corresponding to these matrices. The results presented are relevant to a number of methods in machine learning and data mining, which we describe at various stages. Rather than being a novel academic study in its own right, this text presents a collection of known results, together with some new concepts. Moreover, the tutorial focuses on offering intuition to readers rather than formal understanding and only assumes basic exposure to concepts from linear algebra and probability theory. It is therefore accessible to students and researchers from a wide variety of disciplines.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2023-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41174469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Neural Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1