首页 > 最新文献

Neural Computation最新文献

英文 中文
CA3 Circuit Model Compressing Sequential Information in Theta Oscillation and Replay 在 Theta 振荡和重放中压缩序列信息的 CA3 电路模型。
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01641
Satoshi Kuroki;Kenji Mizuseki
The hippocampus plays a critical role in the compression and retrieval of sequential information. During wakefulness, it achieves this through theta phase precession and theta sequences. Subsequently, during periods of sleep or rest, the compressed information reactivates through sharp-wave ripple events, manifesting as memory replay. However, how these sequential neuronal activities are generated and how they store information about the external environment remain unknown. We developed a hippocampal cornu ammonis 3 (CA3) computational model based on anatomical and electrophysiological evidence from the biological CA3 circuit to address these questions. The model comprises theta rhythm inhibition, place input, and CA3-CA3 plastic recurrent connection. The model can compress the sequence of the external inputs, reproduce theta phase precession and replay, learn additional sequences, and reorganize previously learned sequences. A gradual increase in synaptic inputs, controlled by interactions between theta-paced inhibition and place inputs, explained the mechanism of sequence acquisition. This model highlights the crucial role of plasticity in the CA3 recurrent connection and theta oscillational dynamics and hypothesizes how the CA3 circuit acquires, compresses, and replays sequential information.
海马体在序列信息的压缩和检索中发挥着至关重要的作用。在清醒状态下,海马体通过θ相位前冲和θ序列实现这一功能。随后,在睡眠或休息期间,被压缩的信息通过锐波波纹事件重新激活,表现为记忆重放。然而,这些连续的神经元活动是如何产生的,它们又是如何存储外部环境信息的,这些仍然是未知数。为了解决这些问题,我们基于生物 CA3 电路的解剖学和电生理学证据,建立了一个海马角弓 3(CA3)计算模型。该模型包括θ节律抑制、位置输入和CA3-CA3可塑性递归连接。该模型可以压缩外部输入的序列,重现θ相位前冲和重放,学习额外的序列,并重组以前学习过的序列。在θ步抑制和位置输入的相互作用控制下,突触输入的逐渐增加解释了序列习得的机制。该模型强调了可塑性在CA3递归连接和θ振荡动态中的关键作用,并假设了CA3回路是如何获取、压缩和重放序列信息的。
{"title":"CA3 Circuit Model Compressing Sequential Information in Theta Oscillation and Replay","authors":"Satoshi Kuroki;Kenji Mizuseki","doi":"10.1162/neco_a_01641","DOIUrl":"10.1162/neco_a_01641","url":null,"abstract":"The hippocampus plays a critical role in the compression and retrieval of sequential information. During wakefulness, it achieves this through theta phase precession and theta sequences. Subsequently, during periods of sleep or rest, the compressed information reactivates through sharp-wave ripple events, manifesting as memory replay. However, how these sequential neuronal activities are generated and how they store information about the external environment remain unknown. We developed a hippocampal cornu ammonis 3 (CA3) computational model based on anatomical and electrophysiological evidence from the biological CA3 circuit to address these questions. The model comprises theta rhythm inhibition, place input, and CA3-CA3 plastic recurrent connection. The model can compress the sequence of the external inputs, reproduce theta phase precession and replay, learn additional sequences, and reorganize previously learned sequences. A gradual increase in synaptic inputs, controlled by interactions between theta-paced inhibition and place inputs, explained the mechanism of sequence acquisition. This model highlights the crucial role of plasticity in the CA3 recurrent connection and theta oscillational dynamics and hypothesizes how the CA3 circuit acquires, compresses, and replays sequential information.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10535082","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Column Row Convolutional Neural Network: Reducing Parameters for Efficient Image Processing 列行卷积神经网络:减少参数,实现高效图像处理
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01653
Seongil Im;Jae-Seung Jeong;Junseo Lee;Changhwan Shin;Jeong Ho Cho;Hyunsu Ju
Recent advancements in deep learning have achieved significant progress by increasing the number of parameters in a given model. However, this comes at the cost of computing resources, prompting researchers to explore model compression techniques that reduce the number of parameters while maintaining or even improving performance. Convolutional neural networks (CNN) have been recognized as more efficient and effective than fully connected (FC) networks. We propose a column row convolutional neural network (CRCNN) in this letter that applies 1D convolution to image data, significantly reducing the number of learning parameters and operational steps. The CRCNN uses column and row local receptive fields to perform data abstraction, concatenating each direction's feature before connecting it to an FC layer. Experimental results demonstrate that the CRCNN maintains comparable accuracy while reducing the number of parameters and compared to prior work. Moreover, the CRCNN is employed for one-class anomaly detection, demonstrating its feasibility for various applications.
通过增加给定模型中的参数数量,深度学习最近取得了重大进展。然而,这是以计算资源为代价的,这促使研究人员探索模型压缩技术,以减少参数数量,同时保持甚至提高性能。卷积神经网络(CNN)已被公认为比全连接(FC)网络更高效、更有效。我们在这封信中提出了一种列行卷积神经网络(CRCNN),它将一维卷积应用于图像数据,大大减少了学习参数和操作步骤的数量。CRCNN 利用列和行局部感受野进行数据抽象,在将每个方向的特征连接到 FC 层之前将其串联起来。实验结果表明,与之前的研究相比,CRCNN 在减少参数数量的同时保持了相当的准确性。此外,CRCNN 被用于单类异常检测,证明了它在各种应用中的可行性。
{"title":"Column Row Convolutional Neural Network: Reducing Parameters for Efficient Image Processing","authors":"Seongil Im;Jae-Seung Jeong;Junseo Lee;Changhwan Shin;Jeong Ho Cho;Hyunsu Ju","doi":"10.1162/neco_a_01653","DOIUrl":"10.1162/neco_a_01653","url":null,"abstract":"Recent advancements in deep learning have achieved significant progress by increasing the number of parameters in a given model. However, this comes at the cost of computing resources, prompting researchers to explore model compression techniques that reduce the number of parameters while maintaining or even improving performance. Convolutional neural networks (CNN) have been recognized as more efficient and effective than fully connected (FC) networks. We propose a column row convolutional neural network (CRCNN) in this letter that applies 1D convolution to image data, significantly reducing the number of learning parameters and operational steps. The CRCNN uses column and row local receptive fields to perform data abstraction, concatenating each direction's feature before connecting it to an FC layer. Experimental results demonstrate that the CRCNN maintains comparable accuracy while reducing the number of parameters and compared to prior work. Moreover, the CRCNN is employed for one-class anomaly detection, demonstrating its feasibility for various applications.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Frequency Propagation: Multimechanism Learning in Nonlinear Physical Networks 频率传播:非线性物理网络中的多机制学习
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01648
Vidyesh Rao Anisetti;Ananth Kandala;Benjamin Scellier;J. M. Schwarz
We introduce frequency propagation, a learning algorithm for nonlinear physical networks. In a resistive electrical circuit with variable resistors, an activation current is applied at a set of input nodes at one frequency and an error current is applied at a set of output nodes at another frequency. The voltage response of the circuit to these boundary currents is the superposition of an activation signal and an error signal whose coefficients can be read in different frequencies of the frequency domain. Each conductance is updated proportionally to the product of the two coefficients. The learning rule is local and proved to perform gradient descent on a loss function. We argue that frequency propagation is an instance of a multimechanism learning strategy for physical networks, be it resistive, elastic, or flow networks. Multimechanism learning strategies incorporate at least two physical quantities, potentially governed by independent physical mechanisms, to act as activation and error signals in the training process. Locally available information about these two signals is then used to update the trainable parameters to perform gradient descent. We demonstrate how earlier work implementing learning via chemical signaling in flow networks (Anisetti, Scellier, et al., 2023) also falls under the rubric of multimechanism learning.
我们介绍一种非线性物理网络的学习算法--频率传播。在一个带有可变电阻的电阻电路中,一组输入节点上施加一个频率的激活电流,一组输出节点上施加一个频率的误差电流。电路对这些边界电流的电压响应是激活信号和误差信号的叠加,这两个信号的系数可在频域的不同频率下读取。每个电导的更新都与这两个系数的乘积成比例。学习规则是局部的,并被证明可在损失函数上执行梯度下降。我们认为,频率传播是物理网络(无论是电阻网络、弹性网络还是流动网络)多机制学习策略的一个实例。多机制学习策略包含至少两个物理量,可能由独立的物理机制控制,作为训练过程中的激活信号和误差信号。关于这两个信号的局部可用信息随后被用于更新可训练参数,以执行梯度下降。我们展示了早先在流网络中通过化学信号进行学习的工作(Anisetti, Scellier, et al.
{"title":"Frequency Propagation: Multimechanism Learning in Nonlinear Physical Networks","authors":"Vidyesh Rao Anisetti;Ananth Kandala;Benjamin Scellier;J. M. Schwarz","doi":"10.1162/neco_a_01648","DOIUrl":"10.1162/neco_a_01648","url":null,"abstract":"We introduce frequency propagation, a learning algorithm for nonlinear physical networks. In a resistive electrical circuit with variable resistors, an activation current is applied at a set of input nodes at one frequency and an error current is applied at a set of output nodes at another frequency. The voltage response of the circuit to these boundary currents is the superposition of an activation signal and an error signal whose coefficients can be read in different frequencies of the frequency domain. Each conductance is updated proportionally to the product of the two coefficients. The learning rule is local and proved to perform gradient descent on a loss function. We argue that frequency propagation is an instance of a multimechanism learning strategy for physical networks, be it resistive, elastic, or flow networks. Multimechanism learning strategies incorporate at least two physical quantities, potentially governed by independent physical mechanisms, to act as activation and error signals in the training process. Locally available information about these two signals is then used to update the trainable parameters to perform gradient descent. We demonstrate how earlier work implementing learning via chemical signaling in flow networks (Anisetti, Scellier, et al., 2023) also falls under the rubric of multimechanism learning.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Korobov Functions by Correntropy and Convolutional Neural Networks 通过熵和卷积神经网络学习 Korobov 函数
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01650
Zhiying Fang;Tong Mao;Jun Fan
Combining information-theoretic learning with deep learning has gained significant attention in recent years, as it offers a promising approach to tackle the challenges posed by big data. However, the theoretical understanding of convolutional structures, which are vital to many structured deep learning models, remains incomplete. To partially bridge this gap, this letter aims to develop generalization analysis for deep convolutional neural network (CNN) algorithms using learning theory. Specifically, we focus on investigating robust regression using correntropy-induced loss functions derived from information-theoretic learning. Our analysis demonstrates an explicit convergence rate for deep CNN-based robust regression algorithms when the target function resides in the Korobov space. This study sheds light on the theoretical underpinnings of CNNs and provides a framework for understanding their performance and limitations.
近年来,信息理论学习与深度学习的结合受到了广泛关注,因为它为应对大数据带来的挑战提供了一种前景广阔的方法。然而,对许多结构化深度学习模型至关重要的卷积结构的理论理解仍然不完整。为了部分弥补这一差距,这封信旨在利用学习理论对深度卷积神经网络(CNN)算法进行泛化分析。具体来说,我们重点研究了使用从信息论学习中得出的熵诱导损失函数进行鲁棒回归的问题。我们的分析表明,当目标函数位于 Korobov 空间时,基于深度 CNN 的鲁棒回归算法具有明确的收敛率。这项研究揭示了 CNN 的理论基础,并为了解其性能和局限性提供了一个框架。
{"title":"Learning Korobov Functions by Correntropy and Convolutional Neural Networks","authors":"Zhiying Fang;Tong Mao;Jun Fan","doi":"10.1162/neco_a_01650","DOIUrl":"10.1162/neco_a_01650","url":null,"abstract":"Combining information-theoretic learning with deep learning has gained significant attention in recent years, as it offers a promising approach to tackle the challenges posed by big data. However, the theoretical understanding of convolutional structures, which are vital to many structured deep learning models, remains incomplete. To partially bridge this gap, this letter aims to develop generalization analysis for deep convolutional neural network (CNN) algorithms using learning theory. Specifically, we focus on investigating robust regression using correntropy-induced loss functions derived from information-theoretic learning. Our analysis demonstrates an explicit convergence rate for deep CNN-based robust regression algorithms when the target function resides in the Korobov space. This study sheds light on the theoretical underpinnings of CNNs and provides a framework for understanding their performance and limitations.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lateral Connections Improve Generalizability of Learning in a Simple Neural Network 侧向连接提高了简单神经网络学习的通用性
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01640
Garrett Crutcher
To navigate the world around us, neural circuits rapidly adapt to their environment learning generalizable strategies to decode information. When modeling these learning strategies, network models find the optimal solution to satisfy one task condition but fail when introduced to a novel task or even a different stimulus in the same space. In the experiments described in this letter, I investigate the role of lateral gap junctions in learning generalizable strategies to process information. Lateral gap junctions are formed by connexin proteins creating an open pore that allows for direct electrical signaling between two neurons. During neural development, the rate of gap junctions is high, and daughter cells that share similar tuning properties are more likely to be connected by these junctions. Gap junctions are highly plastic and get heavily pruned throughout development. I hypothesize that they mediate generalized learning by imprinting the weighting structure within a layer to avoid overfitting to one task condition. To test this hypothesis, I implemented a feedforward probabilistic neural network mimicking a cortical fast spiking neuron circuit that is heavily involved in movement. Many of these cells are tuned to speeds that I used as the input stimulus for the network to estimate. When training this network using a delta learning rule, both a laterally connected network and an unconnected network can estimate a single speed. However, when asking the network to estimate two or more speeds, alternated in training, an unconnected network either cannot learn speed or optimizes to a singular speed, while the laterally connected network learns the generalizable strategy and can estimate both speeds. These results suggest that lateral gap junctions between neurons enable generalized learning, which may help explain learning differences across life span.
为了引导我们周围的世界,神经回路会迅速适应环境,学习可通用的信息解码策略。在对这些学习策略进行建模时,网络模型会找到满足某一任务条件的最优解,但当引入新任务甚至同一空间的不同刺激时,网络模型就会失效。在这封信所描述的实验中,我研究了横向间隙连接在学习处理信息的通用策略中的作用。侧向间隙连接是由连接蛋白形成的一个开放孔,它允许两个神经元之间直接进行电信号传递。在神经发育过程中,间隙连接的发生率很高,具有相似调谐特性的子细胞更有可能通过这些连接点连接起来。间隙连接具有高度可塑性,在整个发育过程中会被大量修剪。我的假设是,间隙连接通过印记层内的加权结构来介导泛化学习,以避免过度适应某一任务条件。为了验证这一假设,我模仿皮质快速尖峰神经元回路,建立了一个前馈概率神经网络,该神经网络在很大程度上参与了运动。这些细胞中有许多都被调谐到速度上,我将其作为输入刺激供网络估计。当使用三角学习规则训练该网络时,横向连接的网络和非连接的网络都能估计出单一速度。然而,当要求网络在训练中交替估计两个或更多速度时,未连接的网络要么无法学习速度,要么优化为单一速度,而横向连接的网络则能学习通用策略,并能估计两个速度。这些结果表明,神经元之间的横向间隙连接实现了泛化学习,这可能有助于解释不同生命期的学习差异。
{"title":"Lateral Connections Improve Generalizability of Learning in a Simple Neural Network","authors":"Garrett Crutcher","doi":"10.1162/neco_a_01640","DOIUrl":"10.1162/neco_a_01640","url":null,"abstract":"To navigate the world around us, neural circuits rapidly adapt to their environment learning generalizable strategies to decode information. When modeling these learning strategies, network models find the optimal solution to satisfy one task condition but fail when introduced to a novel task or even a different stimulus in the same space. In the experiments described in this letter, I investigate the role of lateral gap junctions in learning generalizable strategies to process information. Lateral gap junctions are formed by connexin proteins creating an open pore that allows for direct electrical signaling between two neurons. During neural development, the rate of gap junctions is high, and daughter cells that share similar tuning properties are more likely to be connected by these junctions. Gap junctions are highly plastic and get heavily pruned throughout development. I hypothesize that they mediate generalized learning by imprinting the weighting structure within a layer to avoid overfitting to one task condition. To test this hypothesis, I implemented a feedforward probabilistic neural network mimicking a cortical fast spiking neuron circuit that is heavily involved in movement. Many of these cells are tuned to speeds that I used as the input stimulus for the network to estimate. When training this network using a delta learning rule, both a laterally connected network and an unconnected network can estimate a single speed. However, when asking the network to estimate two or more speeds, alternated in training, an unconnected network either cannot learn speed or optimizes to a singular speed, while the laterally connected network learns the generalizable strategy and can estimate both speeds. These results suggest that lateral gap junctions between neurons enable generalized learning, which may help explain learning differences across life span.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Probing the Structure and Functional Properties of the Dropout-Induced Correlated Variability in Convolutional Neural Networks 探究卷积神经网络中由辍学引发的相关变异性的结构和功能特性
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01652
Xu Pan;Ruben Coen-Cagli;Odelia Schwartz
Computational neuroscience studies have shown that the structure of neural variability to an unchanged stimulus affects the amount of information encoded. Some artificial deep neural networks, such as those with Monte Carlo dropout layers, also have variable responses when the input is fixed. However, the structure of the trial-by-trial neural covariance in neural networks with dropout has not been studied, and its role in decoding accuracy is unknown. We studied the above questions in a convolutional neural network model with dropout in both the training and testing phases. We found that trial-by-trial correlation between neurons (i.e., noise correlation) is positive and low dimensional. Neurons that are close in a feature map have larger noise correlation. These properties are surprisingly similar to the findings in the visual cortex. We further analyzed the alignment of the main axes of the covariance matrix. We found that different images share a common trial-by-trial noise covariance subspace, and they are aligned with the global signal covariance. This evidence that the noise covariance is aligned with signal covariance suggests that noise covariance in dropout neural networks reduces network accuracy, which we further verified directly with a trial-shuffling procedure commonly used in neuroscience. These findings highlight a previously overlooked aspect of dropout layers that can affect network performance. Such dropout networks could also potentially be a computational model of neural variability.
计算神经科学研究表明,神经对不变刺激的变化结构会影响编码信息的数量。一些人工深度神经网络,如具有蒙特卡洛剔除层的网络,在输入固定时也具有可变的响应。然而,人们还没有研究过带剔除层的神经网络中逐次试验的神经协方差结构,也不知道它在解码准确性中的作用。我们在一个训练和测试阶段都有丢弃的卷积神经网络模型中研究了上述问题。我们发现,神经元之间的逐次试验相关性(即噪声相关性)是正的,且维度较低。在特征图中距离较近的神经元具有较大的噪声相关性。这些特性与视觉皮层的发现惊人地相似。我们进一步分析了协方差矩阵主轴的排列。我们发现,不同图像共享一个共同的逐次试验噪声协方差子空间,并且它们与全局信号协方差对齐。噪声协方差与信号协方差一致的这一证据表明,辍学神经网络中的噪声协方差会降低网络的准确性,我们通过神经科学中常用的试验洗牌程序进一步直接验证了这一点。这些发现凸显了以前被忽视的、可能影响网络性能的 "剔除层 "的一个方面。这种剔除网络也有可能成为神经变异性的计算模型。
{"title":"Probing the Structure and Functional Properties of the Dropout-Induced Correlated Variability in Convolutional Neural Networks","authors":"Xu Pan;Ruben Coen-Cagli;Odelia Schwartz","doi":"10.1162/neco_a_01652","DOIUrl":"10.1162/neco_a_01652","url":null,"abstract":"Computational neuroscience studies have shown that the structure of neural variability to an unchanged stimulus affects the amount of information encoded. Some artificial deep neural networks, such as those with Monte Carlo dropout layers, also have variable responses when the input is fixed. However, the structure of the trial-by-trial neural covariance in neural networks with dropout has not been studied, and its role in decoding accuracy is unknown. We studied the above questions in a convolutional neural network model with dropout in both the training and testing phases. We found that trial-by-trial correlation between neurons (i.e., noise correlation) is positive and low dimensional. Neurons that are close in a feature map have larger noise correlation. These properties are surprisingly similar to the findings in the visual cortex. We further analyzed the alignment of the main axes of the covariance matrix. We found that different images share a common trial-by-trial noise covariance subspace, and they are aligned with the global signal covariance. This evidence that the noise covariance is aligned with signal covariance suggests that noise covariance in dropout neural networks reduces network accuracy, which we further verified directly with a trial-shuffling procedure commonly used in neuroscience. These findings highlight a previously overlooked aspect of dropout layers that can affect network performance. Such dropout networks could also potentially be a computational model of neural variability.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vector Symbolic Finite State Machines in Attractor Neural Networks 吸引器神经网络中的矢量符号有限状态机
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01638
Madison Cotteret;Hugh Greatorex;Martin Ziegler;Elisabetta Chicca
Hopfield attractor networks are robust distributed models of human memory, but they lack a general mechanism for effecting state-dependent attractor transitions in response to input. We propose construction rules such that an attractor network may implement an arbitrary finite state machine (FSM), where states and stimuli are represented by high-dimensional random vectors and all state transitions are enacted by the attractor network's dynamics. Numerical simulations show the capacity of the model, in terms of the maximum size of implementable FSM, to be linear in the size of the attractor network for dense bipolar state vectors and approximately quadratic for sparse binary state vectors. We show that the model is robust to imprecise and noisy weights, and so a prime candidate for implementation with high-density but unreliable devices. By endowing attractor networks with the ability to emulate arbitrary FSMs, we propose a plausible path by which FSMs could exist as a distributed computational primitive in biological neural networks.
Hopfield 吸引子网络是人类记忆的稳健分布式模型,但它们缺乏一种通用机制来实现与输入相关的吸引子转换。我们提出了构建规则,使吸引子网络可以实现任意的有限状态机(FSM),其中状态和刺激由高维随机向量表示,所有状态转换都由吸引子网络的动力学来实现。数值模拟显示,就可实现的 FSM 的最大规模而言,该模型的容量与密集双极状态向量的吸引子网络规模呈线性关系,而对于稀疏二进制状态向量则近似二次方关系。我们的研究表明,该模型对不精确和有噪声的权重具有很强的鲁棒性,因此是使用高密度但不可靠的设备实现的最佳候选模型。通过赋予吸引子网络模拟任意 FSM 的能力,我们提出了一条可行的途径,使 FSM 可以作为分布式计算基元存在于生物神经网络中。
{"title":"Vector Symbolic Finite State Machines in Attractor Neural Networks","authors":"Madison Cotteret;Hugh Greatorex;Martin Ziegler;Elisabetta Chicca","doi":"10.1162/neco_a_01638","DOIUrl":"10.1162/neco_a_01638","url":null,"abstract":"Hopfield attractor networks are robust distributed models of human memory, but they lack a general mechanism for effecting state-dependent attractor transitions in response to input. We propose construction rules such that an attractor network may implement an arbitrary finite state machine (FSM), where states and stimuli are represented by high-dimensional random vectors and all state transitions are enacted by the attractor network's dynamics. Numerical simulations show the capacity of the model, in terms of the maximum size of implementable FSM, to be linear in the size of the attractor network for dense bipolar state vectors and approximately quadratic for sparse binary state vectors. We show that the model is robust to imprecise and noisy weights, and so a prime candidate for implementation with high-density but unreliable devices. By endowing attractor networks with the ability to emulate arbitrary FSMs, we propose a plausible path by which FSMs could exist as a distributed computational primitive in biological neural networks.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10535093","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mathematical Modeling of PI3K/Akt Pathway in Microglia 小胶质细胞中 PI3K/Akt 通路的数学建模
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01643
Alireza Poshtkohi;John Wade;Liam McDaid;Junxiu Liu;Mark L. Dallas;Angela Bithell
The motility of microglia involves intracellular signaling pathways that are predominantly controlled by changes in cytosolic Ca2+ and activation of PI3K/Akt (phosphoinositide-3-kinase/protein kinase B). In this letter, we develop a novel biophysical model for cytosolic Ca2+ activation of the PI3K/Akt pathway in microglia where Ca2+ influx is mediated by both P2Y purinergic receptors (P2YR) and P2X purinergic receptors (P2XR). The model parameters are estimated by employing optimization techniques to fit the model to phosphorylated Akt (pAkt) experimental modeling/in vitro data. The integrated model supports the hypothesis that Ca2+ influx via P2YR and P2XR can explain the experimentally reported biphasic transient responses in measuring pAkt levels. Our predictions reveal new quantitative insights into P2Rs on how they regulate Ca2+ and Akt in terms of physiological interactions and transient responses. It is shown that the upregulation of P2X receptors through a repetitive application of agonist results in a continual increase in the baseline [Ca2+], which causes the biphasic response to become a monophasic response which prolongs elevated levels of pAkt.
小胶质细胞的运动涉及细胞内信号通路,这些通路主要受控于细胞膜 Ca2+ 的变化和 PI3K/Akt(磷脂酰肌醇-3-激酶/蛋白激酶 B)的激活。在这封信中,我们为小胶质细胞中 PI3K/Akt 通路的细胞膜 Ca2+ 激活建立了一个新的生物物理模型,其中 Ca2+ 的流入由 P2Y 嘌呤能受体(P2YR)和 P2X 嘌呤能受体(P2XR)介导。模型参数的估算采用了优化技术,使模型与磷酸化 Akt(pAkt)实验模型/体外数据相匹配。综合模型支持这样的假设,即通过 P2YR 和 P2XR 流入的 Ca2+ 可以解释实验报告中测量 pAkt 水平的双相瞬时反应。我们的预测揭示了 P2R 在生理相互作用和瞬时反应方面如何调控 Ca2+ 和 Akt 的新定量见解。研究表明,通过重复应用激动剂上调 P2X 受体会导致基线[Ca2+]持续增加,从而使双相反应变成单相反应,延长 pAkt 水平升高的时间。
{"title":"Mathematical Modeling of PI3K/Akt Pathway in Microglia","authors":"Alireza Poshtkohi;John Wade;Liam McDaid;Junxiu Liu;Mark L. Dallas;Angela Bithell","doi":"10.1162/neco_a_01643","DOIUrl":"10.1162/neco_a_01643","url":null,"abstract":"The motility of microglia involves intracellular signaling pathways that are predominantly controlled by changes in cytosolic Ca2+ and activation of PI3K/Akt (phosphoinositide-3-kinase/protein kinase B). In this letter, we develop a novel biophysical model for cytosolic Ca2+ activation of the PI3K/Akt pathway in microglia where Ca2+ influx is mediated by both P2Y purinergic receptors (P2YR) and P2X purinergic receptors (P2XR). The model parameters are estimated by employing optimization techniques to fit the model to phosphorylated Akt (pAkt) experimental modeling/in vitro data. The integrated model supports the hypothesis that Ca2+ influx via P2YR and P2XR can explain the experimentally reported biphasic transient responses in measuring pAkt levels. Our predictions reveal new quantitative insights into P2Rs on how they regulate Ca2+ and Akt in terms of physiological interactions and transient responses. It is shown that the upregulation of P2X receptors through a repetitive application of agonist results in a continual increase in the baseline [Ca2+], which causes the biphasic response to become a monophasic response which prolongs elevated levels of pAkt.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Object-Centric Scene Representations Using Active Inference 使用主动推理进行以对象为中心的场景表示
IF 2.9 4区 计算机科学 Q1 Arts and Humanities Pub Date : 2024-03-21 DOI: 10.1162/neco_a_01637
Toon Van de Maele;Tim Verbelen;Pietro Mazzaglia;Stefano Ferraro;Bart Dhoedt
Representing a scene and its constituent objects from raw sensory data is a core ability for enabling robots to interact with their environment. In this letter, we propose a novel approach for scene understanding, leveraging an object-centric generative model that enables an agent to infer object category and pose in an allocentric reference frame using active inference, a neuro-inspired framework for action and perception. For evaluating the behavior of an active vision agent, we also propose a new benchmark where, given a target viewpoint of a particular object, the agent needs to find the best matching viewpoint given a workspace with randomly positioned objects in 3D. We demonstrate that our active inference agent is able to balance epistemic foraging and goal-driven behavior, and quantitatively outperforms both supervised and reinforcement learning baselines by more than a factor of two in terms of success rate.
从原始感官数据中呈现场景及其组成物体是机器人与环境互动的核心能力。在这封信中,我们提出了一种新颖的场景理解方法,利用以物体为中心的生成模型,使代理能够使用主动推理(一种用于行动和感知的神经启发框架)在以分配为中心的参考框架中推断出物体的类别和姿势。为了评估主动视觉代理的行为,我们还提出了一个新的基准,即在给定特定物体的目标视角后,代理需要在三维随机定位物体的工作空间中找到最佳匹配视角。我们证明,我们的主动推理代理能够在认识论觅食和目标驱动行为之间取得平衡,并且在成功率方面定量优于监督学习和强化学习基线超过两倍。
{"title":"Object-Centric Scene Representations Using Active Inference","authors":"Toon Van de Maele;Tim Verbelen;Pietro Mazzaglia;Stefano Ferraro;Bart Dhoedt","doi":"10.1162/neco_a_01637","DOIUrl":"10.1162/neco_a_01637","url":null,"abstract":"Representing a scene and its constituent objects from raw sensory data is a core ability for enabling robots to interact with their environment. In this letter, we propose a novel approach for scene understanding, leveraging an object-centric generative model that enables an agent to infer object category and pose in an allocentric reference frame using active inference, a neuro-inspired framework for action and perception. For evaluating the behavior of an active vision agent, we also propose a new benchmark where, given a target viewpoint of a particular object, the agent needs to find the best matching viewpoint given a workspace with randomly positioned objects in 3D. We demonstrate that our active inference agent is able to balance epistemic foraging and goal-driven behavior, and quantitatively outperforms both supervised and reinforcement learning baselines by more than a factor of two in terms of success rate.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.9,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140066269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Determining Role of Covariances in Large Networks of Stochastic Neurons 随机神经元大型网络中协方差的决定性作用
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-03-10 DOI: 10.1162/neco_a_01656
Vincent Painchaud;Patrick Desrosiers;Nicolas Doyon
Biological neural networks are notoriously hard to model due to their stochastic behavior and high dimensionality. We tackle this problem by constructing a dynamical model of both the expectations and covariances of the fractions of active and refractory neurons in the network’s populations. We do so by describing the evolution of the states of individual neurons with a continuous-time Markov chain, from which we formally derive a low-dimensional dynamical system. This is done by solving a moment closure problem in a way that is compatible with the nonlinearity and boundedness of the activation function. Our dynamical system captures the behavior of the high-dimensional stochastic model even in cases where the mean-field approximation fails to do so. Taking into account the second-order moments modifies the solutions that would be obtained with the mean-field approximation and can lead to the appearance or disappearance of fixed points and limit cycles. We moreover perform numerical experiments where the mean-field approximation leads to periodically oscillating solutions, while the solutions of the second-order model can be interpreted as an average taken over many realizations of the stochastic model. Altogether, our results highlight the importance of including higher moments when studying stochastic networks and deepen our understanding of correlated neuronal activity.
生物神经网络因其随机行为和高维度而难以建模。我们通过构建一个动态模型来解决这一问题,该模型既包含对网络种群中活跃神经元和折射神经元的期望值,也包含它们的协方差。为此,我们用连续时间马尔可夫链描述单个神经元状态的演变,并从中正式推导出一个低维动态系统。这是通过解决矩闭合问题来实现的,该问题与激活函数的非线性和有界性相兼容。即使在均值场近似无法捕捉到高维随机模型行为的情况下,我们的动力系统也能捕捉到。考虑到二阶矩会改变用均值场近似得到的解,并可能导致定点和极限循环的出现或消失。此外,我们还进行了数值实验,发现均场近似会导致周期性振荡解,而二阶模型的解可以解释为随机模型多次实现后的平均值。总之,我们的研究结果强调了在研究随机网络时加入高阶矩的重要性,并加深了我们对相关神经元活动的理解。
{"title":"The Determining Role of Covariances in Large Networks of Stochastic Neurons","authors":"Vincent Painchaud;Patrick Desrosiers;Nicolas Doyon","doi":"10.1162/neco_a_01656","DOIUrl":"10.1162/neco_a_01656","url":null,"abstract":"Biological neural networks are notoriously hard to model due to their stochastic behavior and high dimensionality. We tackle this problem by constructing a dynamical model of both the expectations and covariances of the fractions of active and refractory neurons in the network’s populations. We do so by describing the evolution of the states of individual neurons with a continuous-time Markov chain, from which we formally derive a low-dimensional dynamical system. This is done by solving a moment closure problem in a way that is compatible with the nonlinearity and boundedness of the activation function. Our dynamical system captures the behavior of the high-dimensional stochastic model even in cases where the mean-field approximation fails to do so. Taking into account the second-order moments modifies the solutions that would be obtained with the mean-field approximation and can lead to the appearance or disappearance of fixed points and limit cycles. We moreover perform numerical experiments where the mean-field approximation leads to periodically oscillating solutions, while the solutions of the second-order model can be interpreted as an average taken over many realizations of the stochastic model. Altogether, our results highlight the importance of including higher moments when studying stochastic networks and deepen our understanding of correlated neuronal activity.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140805835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1