首页 > 最新文献

Neural Computation最新文献

英文 中文
Top-Down Priors Disambiguate Target and Distractor Features in Simulated Covert Visual Search. 在模拟隐蔽视觉搜索中,自上而下的先验信息能区分目标和干扰特征。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1162/neco_a_01700
Justin D Theiss, Michael A Silver

Several models of visual search consider visual attention as part of a perceptual inference process, in which top-down priors disambiguate bottom-up sensory information. Many of these models have focused on gaze behavior, but there are relatively fewer models of covert spatial attention, in which attention is directed to a peripheral location in visual space without a shift in gaze direction. Here, we propose a biologically plausible model of covert attention during visual search that helps to bridge the gap between Bayesian modeling and neurophysiological modeling by using (1) top-down priors over target features that are acquired through Hebbian learning, and (2) spatial resampling of modeled cortical receptive fields to enhance local spatial resolution of image representations for downstream target classification. By training a simple generative model using a Hebbian update rule, top-down priors for target features naturally emerge without the need for hand-tuned or predetermined priors. Furthermore, the implementation of covert spatial attention in our model is based on a known neurobiological mechanism, providing a plausible process through which Bayesian priors could locally enhance the spatial resolution of image representations. We validate this model during simulated visual search for handwritten digits among nondigit distractors, demonstrating that top-down priors improve accuracy for estimation of target location and classification, relative to bottom-up signals alone. Our results support previous reports in the literature that demonstrated beneficial effects of top-down priors on visual search performance, while extending this literature to incorporate known neural mechanisms of covert spatial attention.

有几种视觉搜索模型将视觉注意力视为知觉推理过程的一部分,在这一过程中,自上而下的先验信息会对自下而上的感官信息产生歧义。这些模型中有很多都侧重于注视行为,但关于隐蔽空间注意的模型却相对较少,在这种模型中,注意力被引导到视觉空间的外围位置,而不改变注视方向。在这里,我们提出了一种在视觉搜索过程中隐蔽注意力的生物学合理模型,该模型有助于弥合贝叶斯建模和神经生理学建模之间的差距,它使用通过希比学习和建模皮层感受野的空间重采样获得的目标特征自上而下的先验,以提高图像表征的局部空间分辨率,从而进行下游目标分类。通过使用希比更新规则训练一个简单的生成模型,目标特征的自上而下先验条件就会自然出现,而无需人工调整或预先确定先验条件。此外,我们模型中隐蔽空间注意力的实现是基于已知的神经生物学机制,提供了一个合理的过程,通过这个过程,贝叶斯先验可以局部增强图像表征的空间分辨率。我们在模拟视觉搜索非数字干扰物中手写数字的过程中验证了这一模型,结果表明自上而下的前置条件提高了估计目标位置和分类的准确性,而自下而上的信号则相对较弱。我们的研究结果支持了之前的文献报道,这些报道证明了自上而下先验对视觉搜索性能的有利影响,同时也扩展了这些文献,将已知的隐蔽空间注意力神经机制纳入其中。
{"title":"Top-Down Priors Disambiguate Target and Distractor Features in Simulated Covert Visual Search.","authors":"Justin D Theiss, Michael A Silver","doi":"10.1162/neco_a_01700","DOIUrl":"10.1162/neco_a_01700","url":null,"abstract":"<p><p>Several models of visual search consider visual attention as part of a perceptual inference process, in which top-down priors disambiguate bottom-up sensory information. Many of these models have focused on gaze behavior, but there are relatively fewer models of covert spatial attention, in which attention is directed to a peripheral location in visual space without a shift in gaze direction. Here, we propose a biologically plausible model of covert attention during visual search that helps to bridge the gap between Bayesian modeling and neurophysiological modeling by using (1) top-down priors over target features that are acquired through Hebbian learning, and (2) spatial resampling of modeled cortical receptive fields to enhance local spatial resolution of image representations for downstream target classification. By training a simple generative model using a Hebbian update rule, top-down priors for target features naturally emerge without the need for hand-tuned or predetermined priors. Furthermore, the implementation of covert spatial attention in our model is based on a known neurobiological mechanism, providing a plausible process through which Bayesian priors could locally enhance the spatial resolution of image representations. We validate this model during simulated visual search for handwritten digits among nondigit distractors, demonstrating that top-down priors improve accuracy for estimation of target location and classification, relative to bottom-up signals alone. Our results support previous reports in the literature that demonstrated beneficial effects of top-down priors on visual search performance, while extending this literature to incorporate known neural mechanisms of covert spatial attention.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11430503/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141984036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mechanism of Duration Perception in Artificial Brains Suggests New Model of Attentional Entrainment. 人工大脑的时长感知机制提出了注意力牵制的新模式
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1162/neco_a_01699
Ali Tehrani-Saleh, J Devin McAuley, Christoph Adami

While cognitive theory has advanced several candidate frameworks to explain attentional entrainment, the neural basis for the temporal allocation of attention is unknown. Here we present a new model of attentional entrainment guided by empirical evidence obtained using a cohort of 50 artificial brains. These brains were evolved in silico to perform a duration judgment task similar to one where human subjects perform duration judgments in auditory oddball paradigms. We found that the artificial brains display psychometric characteristics remarkably similar to those of human listeners and exhibit similar patterns of distortions of perception when presented with out-of-rhythm oddballs. A detailed analysis of mechanisms behind the duration distortion suggests that attention peaks at the end of the tone, which is inconsistent with previous attentional entrainment models. Instead, the new model of entrainment emphasizes increased attention to those aspects of the stimulus that the brain expects to be highly informative.

尽管认知理论已经提出了几种候选框架来解释注意力诱导,但注意力的时间分配的神经基础尚不清楚。在这里,我们通过使用一组 50 个人工大脑所获得的经验证据,提出了一个新的注意力诱导模型。这些人工大脑是在硅学基础上进化而来的,可以完成类似于人类受试者在听觉怪人范式中进行持续时间判断的任务。我们发现,人工大脑显示出与人类听者极为相似的心理测量特征,并在遇到节奏失常的怪音时表现出相似的感知失真模式。对时长失真背后机制的详细分析表明,注意力在音调结束时达到顶峰,这与之前的注意力夹带模型不一致。相反,新的注意力诱导模型强调的是,大脑会增加对刺激信息量大的那些方面的注意。
{"title":"Mechanism of Duration Perception in Artificial Brains Suggests New Model of Attentional Entrainment.","authors":"Ali Tehrani-Saleh, J Devin McAuley, Christoph Adami","doi":"10.1162/neco_a_01699","DOIUrl":"10.1162/neco_a_01699","url":null,"abstract":"<p><p>While cognitive theory has advanced several candidate frameworks to explain attentional entrainment, the neural basis for the temporal allocation of attention is unknown. Here we present a new model of attentional entrainment guided by empirical evidence obtained using a cohort of 50 artificial brains. These brains were evolved in silico to perform a duration judgment task similar to one where human subjects perform duration judgments in auditory oddball paradigms. We found that the artificial brains display psychometric characteristics remarkably similar to those of human listeners and exhibit similar patterns of distortions of perception when presented with out-of-rhythm oddballs. A detailed analysis of mechanisms behind the duration distortion suggests that attention peaks at the end of the tone, which is inconsistent with previous attentional entrainment models. Instead, the new model of entrainment emphasizes increased attention to those aspects of the stimulus that the brain expects to be highly informative.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Active Inference and Reinforcement Learning: A Unified Inference on Continuous State and Action Spaces Under Partial Observability. 主动推理与强化学习:部分可观测性下连续状态和行动空间的统一推理》(A Unified Inference on Continuous State and Action Spaces under Partial Observability.
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1162/neco_a_01698
Parvin Malekzadeh, Konstantinos N Plataniotis

Reinforcement learning (RL) has garnered significant attention for developing decision-making agents that aim to maximize rewards, specified by an external supervisor, within fully observable environments. However, many real-world problems involve partial or noisy observations, where agents cannot access complete and accurate information about the environment. These problems are commonly formulated as partially observable Markov decision processes (POMDPs). Previous studies have tackled RL in POMDPs by either incorporating the memory of past actions and observations or by inferring the true state of the environment from observed data. Nevertheless, aggregating observations and actions over time becomes impractical in problems with large decision-making time horizons and high-dimensional spaces. Furthermore, inference-based RL approaches often require many environmental samples to perform well, as they focus solely on reward maximization and neglect uncertainty in the inferred state. Active inference (AIF) is a framework naturally formulated in POMDPs and directs agents to select actions by minimizing a function called expected free energy (EFE). This supplies reward-maximizing (or exploitative) behavior, as in RL, with information-seeking (or exploratory) behavior. Despite this exploratory behavior of AIF, its use is limited to problems with small time horizons and discrete spaces due to the computational challenges associated with EFE. In this article, we propose a unified principle that establishes a theoretical connection between AIF and RL, enabling seamless integration of these two approaches and overcoming their limitations in continuous space POMDP settings. We substantiate our findings with rigorous theoretical analysis, providing novel perspectives for using AIF in designing and implementing artificial agents. Experimental results demonstrate the superior learning capabilities of our method compared to other alternative RL approaches in solving partially observable tasks with continuous spaces. Notably, our approach harnesses information-seeking exploration, enabling it to effectively solve reward-free problems and rendering explicit task reward design by an external supervisor optional.

强化学习(RL)在开发决策代理方面备受关注,这些代理的目标是在完全可观测的环境中最大限度地提高外部监督者指定的奖励。然而,现实世界中的许多问题都涉及部分或嘈杂的观测,在这些问题中,代理无法获得有关环境的完整而准确的信息。这些问题通常被表述为部分可观测马尔可夫决策过程(POMDP)。以往的研究通过结合对过去行动和观察结果的记忆,或通过从观察数据推断环境的真实状态,来解决 POMDPs 中的 RL 问题。然而,在决策时间跨度大、空间维度高的问题中,随着时间的推移汇总观察结果和行动是不切实际的。此外,基于推理的 RL 方法往往需要许多环境样本才能取得良好效果,因为它们只关注报酬最大化,而忽略了推理状态的不确定性。主动推理(AIF)是在 POMDPs 中自然形成的一个框架,它指导代理通过最小化称为期望自由能(EFE)的函数来选择行动。这将 RL 中的报酬最大化(或利用)行为与信息搜索(或探索)行为结合起来。尽管 AIF 具有探索行为,但由于 EFE 带来的计算挑战,它的应用仅限于小时间跨度和离散空间的问题。在本文中,我们提出了一个统一的原则,在 AIF 和 RL 之间建立了理论联系,实现了这两种方法的无缝集成,克服了它们在连续空间 POMDP 设置中的局限性。我们通过严谨的理论分析证实了我们的发现,为使用 AIF 设计和实现人工代理提供了新的视角。实验结果表明,在解决连续空间的部分可观测任务时,与其他可供选择的 RL 方法相比,我们的方法具有更强的学习能力。值得注意的是,我们的方法利用了信息搜索探索,使其能够有效地解决无奖励问题,并使外部监督者的明确任务奖励设计变得可有可无。
{"title":"Active Inference and Reinforcement Learning: A Unified Inference on Continuous State and Action Spaces Under Partial Observability.","authors":"Parvin Malekzadeh, Konstantinos N Plataniotis","doi":"10.1162/neco_a_01698","DOIUrl":"10.1162/neco_a_01698","url":null,"abstract":"<p><p>Reinforcement learning (RL) has garnered significant attention for developing decision-making agents that aim to maximize rewards, specified by an external supervisor, within fully observable environments. However, many real-world problems involve partial or noisy observations, where agents cannot access complete and accurate information about the environment. These problems are commonly formulated as partially observable Markov decision processes (POMDPs). Previous studies have tackled RL in POMDPs by either incorporating the memory of past actions and observations or by inferring the true state of the environment from observed data. Nevertheless, aggregating observations and actions over time becomes impractical in problems with large decision-making time horizons and high-dimensional spaces. Furthermore, inference-based RL approaches often require many environmental samples to perform well, as they focus solely on reward maximization and neglect uncertainty in the inferred state. Active inference (AIF) is a framework naturally formulated in POMDPs and directs agents to select actions by minimizing a function called expected free energy (EFE). This supplies reward-maximizing (or exploitative) behavior, as in RL, with information-seeking (or exploratory) behavior. Despite this exploratory behavior of AIF, its use is limited to problems with small time horizons and discrete spaces due to the computational challenges associated with EFE. In this article, we propose a unified principle that establishes a theoretical connection between AIF and RL, enabling seamless integration of these two approaches and overcoming their limitations in continuous space POMDP settings. We substantiate our findings with rigorous theoretical analysis, providing novel perspectives for using AIF in designing and implementing artificial agents. Experimental results demonstrate the superior learning capabilities of our method compared to other alternative RL approaches in solving partially observable tasks with continuous spaces. Notably, our approach harnesses information-seeking exploration, enabling it to effectively solve reward-free problems and rendering explicit task reward design by an external supervisor optional.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal and Multifactor Branching Time Active Inference. 多模态和多因素分支时间主动推理
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1162/neco_a_01703
Théophile Champion, Marek Grześ, Howard Bowman

Active inference is a state-of-the-art framework for modeling the brain that explains a wide range of mechanisms. Recently, two versions of branching time active inference (BTAI) have been developed to handle the exponential (space and time) complexity class that occurs when computing the prior over all possible policies up to the time horizon. However, those two versions of BTAI still suffer from an exponential complexity class with regard to the number of observed and latent variables being modeled. We resolve this limitation by allowing each observation to have its own likelihood mapping and each latent variable to have its own transition mapping. The implicit mean field approximation was tested in terms of its efficiency and computational cost using a dSprites environment in which the metadata of the dSprites data set was used as input to the model. In this setting, earlier implementations of branching time active inference (namely, BTAIVMP and BTAIBF) underperformed in relation to the mean field approximation (BTAI3MF) in terms of performance and computational efficiency. Specifically, BTAIVMP was able to solve 96.9% of the task in 5.1 seconds, and BTAIBF was able to solve 98.6% of the task in 17.5 seconds. Our new approach outperformed both of its predecessors by solving the task completely (100%) in only 2.559 seconds.

主动推理是最先进的大脑建模框架,可以解释各种机制。最近,人们开发了两个版本的分支时间主动推理(BTAI),以处理在计算时间范围内所有可能策略的先验时出现的指数级(空间和时间)复杂性。然而,这两个版本的 BTAI 仍然受到与被建模的观察变量和潜在变量数量有关的指数复杂性的影响。我们允许每个观测值都有自己的似然映射,每个潜变量都有自己的转换映射,从而解决了这一限制。我们使用 dSprites 环境测试了隐式均值场近似的效率和计算成本,在该环境中,dSprites 数据集的元数据被用作模型的输入。在这种情况下,分支时间主动推理的早期实现(即 BTAIVMP 和 BTAIBF)在性能和计算效率方面都不如均值场近似(BTAI3MF)。具体来说,BTAIVMP 能在 5.1 秒内解决 96.9% 的任务,而 BTAIBF 能在 17.5 秒内解决 98.6% 的任务。我们的新方法仅用了 2.559 秒就完全(100%)解决了任务,表现优于前两者。
{"title":"Multimodal and Multifactor Branching Time Active Inference.","authors":"Théophile Champion, Marek Grześ, Howard Bowman","doi":"10.1162/neco_a_01703","DOIUrl":"https://doi.org/10.1162/neco_a_01703","url":null,"abstract":"<p><p>Active inference is a state-of-the-art framework for modeling the brain that explains a wide range of mechanisms. Recently, two versions of branching time active inference (BTAI) have been developed to handle the exponential (space and time) complexity class that occurs when computing the prior over all possible policies up to the time horizon. However, those two versions of BTAI still suffer from an exponential complexity class with regard to the number of observed and latent variables being modeled. We resolve this limitation by allowing each observation to have its own likelihood mapping and each latent variable to have its own transition mapping. The implicit mean field approximation was tested in terms of its efficiency and computational cost using a dSprites environment in which the metadata of the dSprites data set was used as input to the model. In this setting, earlier implementations of branching time active inference (namely, BTAIVMP and BTAIBF) underperformed in relation to the mean field approximation (BTAI3MF) in terms of performance and computational efficiency. Specifically, BTAIVMP was able to solve 96.9% of the task in 5.1 seconds, and BTAIBF was able to solve 98.6% of the task in 17.5 seconds. Our new approach outperformed both of its predecessors by solving the task completely (100%) in only 2.559 seconds.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142114870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prototype Analysis in Hopfield Networks with Hebbian Learning. 采用 Hebbian 学习的 Hopfield 网络中的原型分析
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1162/neco_a_01704
Hayden McAlister, Anthony Robins, Lech Szymanski

We discuss prototype formation in the Hopfield network. Typically, Hebbian learning with highly correlated states leads to degraded memory performance. We show that this type of learning can lead to prototype formation, where unlearned states emerge as representatives of large correlated subsets of states, alleviating capacity woes. This process has similarities to prototype learning in human cognition. We provide a substantial literature review of prototype learning in associative memories, covering contributions from psychology, statistical physics, and computer science. We analyze prototype formation from a theoretical perspective and derive a stability condition for these states based on the number of examples of the prototype presented for learning, the noise in those examples, and the number of nonexample states presented. The stability condition is used to construct a probability of stability for a prototype state as the factors of stability change. We also note similarities to traditional network analysis, allowing us to find a prototype capacity. We corroborate these expectations of prototype formation with experiments using a simple Hopfield network with standard Hebbian learning. We extend our experiments to a Hopfield network trained on data with multiple prototypes and find the network is capable of stabilizing multiple prototypes concurrently. We measure the basins of attraction of the multiple prototype states, finding attractor strength grows with the number of examples and the agreement of examples. We link the stability and dominance of prototype states to the energy profile of these states, particularly when comparing the profile shape to target states or other spurious states.

我们讨论了 Hopfield 网络中的原型形成。通常,具有高度相关状态的海比学习会导致记忆性能下降。我们的研究表明,这种类型的学习会导致原型形成,即未学习的状态会作为大相关状态子集的代表出现,从而缓解容量问题。这一过程与人类认知中的原型学习有相似之处。我们对联想记忆中的原型学习进行了大量的文献综述,涵盖了心理学、统计物理学和计算机科学的研究成果。我们从理论角度分析了原型的形成,并根据为学习而呈现的原型示例数量、这些示例中的噪声以及呈现的非示例状态数量,推导出了这些状态的稳定条件。随着稳定因素的变化,稳定条件被用来构建原型状态的稳定概率。我们还注意到与传统网络分析的相似之处,这使我们能够找到原型容量。我们通过使用标准海比学习的简单 Hopfield 网络进行实验,证实了对原型形成的这些预期。我们将实验扩展到在具有多个原型的数据上训练的 Hopfield 网络,发现该网络能够同时稳定多个原型。我们测量了多个原型状态的吸引盆地,发现吸引强度会随着示例数量和示例一致性的增加而增加。我们将原型状态的稳定性和主导性与这些状态的能量曲线联系起来,特别是在将曲线形状与目标状态或其他虚假状态进行比较时。
{"title":"Prototype Analysis in Hopfield Networks with Hebbian Learning.","authors":"Hayden McAlister, Anthony Robins, Lech Szymanski","doi":"10.1162/neco_a_01704","DOIUrl":"https://doi.org/10.1162/neco_a_01704","url":null,"abstract":"<p><p>We discuss prototype formation in the Hopfield network. Typically, Hebbian learning with highly correlated states leads to degraded memory performance. We show that this type of learning can lead to prototype formation, where unlearned states emerge as representatives of large correlated subsets of states, alleviating capacity woes. This process has similarities to prototype learning in human cognition. We provide a substantial literature review of prototype learning in associative memories, covering contributions from psychology, statistical physics, and computer science. We analyze prototype formation from a theoretical perspective and derive a stability condition for these states based on the number of examples of the prototype presented for learning, the noise in those examples, and the number of nonexample states presented. The stability condition is used to construct a probability of stability for a prototype state as the factors of stability change. We also note similarities to traditional network analysis, allowing us to find a prototype capacity. We corroborate these expectations of prototype formation with experiments using a simple Hopfield network with standard Hebbian learning. We extend our experiments to a Hopfield network trained on data with multiple prototypes and find the network is capable of stabilizing multiple prototypes concurrently. We measure the basins of attraction of the multiple prototype states, finding attractor strength grows with the number of examples and the agreement of examples. We link the stability and dominance of prototype states to the energy profile of these states, particularly when comparing the profile shape to target states or other spurious states.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142114872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictive Representations: Building Blocks of Intelligence. 预测表征:智能的基石
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-30 DOI: 10.1162/neco_a_01705
Wilka Carvalho, Momchil S Tomov, William de Cothi, Caswell Barry, Samuel J Gershman

Adaptive behavior often requires predicting future events. The theory of reinforcement learning prescribes what kinds of predictive representations are useful and how to compute them. This review integrates these theoretical ideas with work on cognition and neuroscience. We pay special attention to the successor representation and its generalizations, which have been widely applied as both engineering tools and models of brain function. This convergence suggests that particular kinds of predictive representations may function as versatile building blocks of intelligence.

适应行为往往需要预测未来事件。强化学习理论规定了哪些类型的预测表征是有用的,以及如何计算它们。这篇综述将这些理论观点与认知和神经科学方面的工作相结合。我们特别关注继任表征及其泛化,它们已被广泛应用为工程工具和大脑功能模型。这种趋同性表明,特定类型的预测表征可以作为智能的多功能构件发挥作用。
{"title":"Predictive Representations: Building Blocks of Intelligence.","authors":"Wilka Carvalho, Momchil S Tomov, William de Cothi, Caswell Barry, Samuel J Gershman","doi":"10.1162/neco_a_01705","DOIUrl":"https://doi.org/10.1162/neco_a_01705","url":null,"abstract":"<p><p>Adaptive behavior often requires predicting future events. The theory of reinforcement learning prescribes what kinds of predictive representations are useful and how to compute them. This review integrates these theoretical ideas with work on cognition and neuroscience. We pay special attention to the successor representation and its generalizations, which have been widely applied as both engineering tools and models of brain function. This convergence suggests that particular kinds of predictive representations may function as versatile building blocks of intelligence.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142114871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spiking Neural Network Pressure Sensor. 尖峰神经网络压力传感器
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-23 DOI: 10.1162/neco_a_01706
Michał Markiewicz, Ireneusz Brzozowski, Szymon Janusz

Von Neumann architecture requires information to be encoded as numerical values. For that reason, artificial neural networks running on computers require the data coming from sensors to be discretized. Other network architectures that more closely mimic biological neural networks (e.g., spiking neural networks) can be simulated on von Neumann architecture, but more important, they can also be executed on dedicated electrical circuits having orders of magnitude less power consumption. Unfortunately, input signal conditioning and encoding are usually not supported by such circuits, so a separate module consisting of an analog-to-digital converter, encoder, and transmitter is required. The aim of this letter is to propose a sensor architecture, the output signal of which can be directly connected to the input of a spiking neural network. We demonstrate that the output signal is a valid spike source for the Izhikevich model neurons, ensuring the proper operation of a number of neurocomputational features. The advantages are clear: much lower power consumption, smaller area, and a less complex electronic circuit. The main disadvantage is that sensor characteristics somehow limit the parameters of applicable spiking neurons. The proposed architecture is illustrated by a case study involving a capacitive pressure sensor circuit, which is compatible with most of the neurocomputational properties of the Izhikevich neuron model. The sensor itself is characterized by very low power consumption: it draws only 3.49 μA at 3.3 V.

冯-诺依曼架构要求将信息编码为数值。因此,在计算机上运行的人工神经网络需要将来自传感器的数据离散化。其他更接近生物神经网络的网络架构(如尖峰神经网络)可以在冯-诺依曼架构上进行模拟,但更重要的是,它们也可以在专用电路上执行,功耗要低得多。遗憾的是,这类电路通常不支持输入信号调节和编码,因此需要一个由模数转换器、编码器和发射器组成的独立模块。本文旨在提出一种传感器结构,其输出信号可直接连接到尖峰神经网络的输入端。我们证明,输出信号是 Izhikevich 模型神经元的有效尖峰源,可确保一些神经计算功能的正常运行。其优点显而易见:功耗更低、占地面积更小、电子电路更简单。主要缺点是传感器特性在某种程度上限制了适用尖峰神经元的参数。我们通过一个涉及电容式压力传感器电路的案例研究来说明所提出的架构,该电路与 Izhikevich 神经元模型的大部分神经计算特性相兼容。传感器本身的特点是功耗极低:在 3.3 V 电压下仅消耗 3.49 μA 电流。
{"title":"Spiking Neural Network Pressure Sensor.","authors":"Michał Markiewicz, Ireneusz Brzozowski, Szymon Janusz","doi":"10.1162/neco_a_01706","DOIUrl":"https://doi.org/10.1162/neco_a_01706","url":null,"abstract":"<p><p>Von Neumann architecture requires information to be encoded as numerical values. For that reason, artificial neural networks running on computers require the data coming from sensors to be discretized. Other network architectures that more closely mimic biological neural networks (e.g., spiking neural networks) can be simulated on von Neumann architecture, but more important, they can also be executed on dedicated electrical circuits having orders of magnitude less power consumption. Unfortunately, input signal conditioning and encoding are usually not supported by such circuits, so a separate module consisting of an analog-to-digital converter, encoder, and transmitter is required. The aim of this letter is to propose a sensor architecture, the output signal of which can be directly connected to the input of a spiking neural network. We demonstrate that the output signal is a valid spike source for the Izhikevich model neurons, ensuring the proper operation of a number of neurocomputational features. The advantages are clear: much lower power consumption, smaller area, and a less complex electronic circuit. The main disadvantage is that sensor characteristics somehow limit the parameters of applicable spiking neurons. The proposed architecture is illustrated by a case study involving a capacitive pressure sensor circuit, which is compatible with most of the neurocomputational properties of the Izhikevich neuron model. The sensor itself is characterized by very low power consumption: it draws only 3.49 μA at 3.3 V.</p>","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Search for Data-Driven and Reproducible Schizophrenia Subtypes Using Resting State fMRI Data From Multiple Sites 利用多部位静息态 fMRI 数据寻找数据驱动和可重复的精神分裂症亚型。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-19 DOI: 10.1162/neco_a_01689
Lærke Gebser Krohne;Ingeborg Helbech Hansen;Kristoffer H. Madsen
For decades, fMRI data have been used to search for biomarkers for patients with schizophrenia. Still, firm conclusions are yet to be made, which is often attributed to the high internal heterogeneity of the disorder. A promising way to disentangle the heterogeneity is to search for subgroups of patients with more homogeneous biological profiles. We applied an unsupervised multiple co-clustering (MCC) method to identify subtypes using functional connectivity data from a multisite resting-state data set. We merged data from two publicly available databases and split the data into a discovery data set (143 patients and 143 healthy controls (HC)) and an external test data set (63 patients and 63 HC) from independent sites. On the discovery data, we investigated the stability of the clustering toward data splits and initializations. Subsequently we searched for cluster solutions, also called “views,” with a significant diagnosis association and evaluated these based on their subject and feature cluster separability, and correlation to clinical manifestations as measured with the positive and negative syndrome scale (PANSS). Finally, we validated our findings by testing the diagnosis association on the external test data. A major finding of our study was that the stability of the clustering was highly dependent on variations in the data set, and even across initializations, we found only a moderate subject clustering stability. Nevertheless, we still discovered one view with a significant diagnosis association. This view reproducibly showed an overrepresentation of schizophrenia patients in three subject clusters, and one feature cluster showed a continuous trend, ranging from positive to negative connectivity values, when sorted according to the proportions of patients with schizophrenia. When investigating all patients, none of the feature clusters in the view were associated with severity of positive, negative, and generalized symptoms, indicating that the cluster solutions reflect other disease related mechanisms.
几十年来,fMRI 数据一直被用于寻找精神分裂症患者的生物标志物。然而,由于精神分裂症具有高度的内部异质性,至今仍未得出确切的结论。消除异质性的一个可行方法是寻找生物特征更为同质的患者亚群。我们采用了一种无监督多重协同聚类(MCC)方法,利用来自多点静息态数据集的功能连接数据来识别亚型。我们合并了两个公开数据库中的数据,并将数据分为发现数据集(143 名患者和 143 名健康对照(HC))和外部测试数据集(63 名患者和 63 名健康对照)。在发现数据上,我们研究了聚类对数据分割和初始化的稳定性。随后,我们搜索了具有显著诊断关联性的聚类解决方案(也称为 "观点"),并根据其主题和特征聚类的可分离性,以及与临床表现的相关性进行了评估,这些临床表现是用正负综合征量表(PANSS)测量的。最后,我们在外部测试数据上测试了诊断关联,从而验证了我们的研究结果。我们研究的一个主要发现是,聚类的稳定性高度依赖于数据集的变化,即使在不同的初始化过程中,我们也只发现了中等程度的主体聚类稳定性。尽管如此,我们仍然发现了一个与诊断有显著关联的视图。在三个研究对象聚类中,该视图重复性地显示出精神分裂症患者的比例过高,而且根据精神分裂症患者的比例进行排序时,一个特征聚类显示出从正连接值到负连接值的连续趋势。在调查所有患者时,视图中没有一个特征群与阳性症状、阴性症状和全身症状的严重程度相关,这表明群解反映了其他疾病相关机制。
{"title":"On the Search for Data-Driven and Reproducible Schizophrenia Subtypes Using Resting State fMRI Data From Multiple Sites","authors":"Lærke Gebser Krohne;Ingeborg Helbech Hansen;Kristoffer H. Madsen","doi":"10.1162/neco_a_01689","DOIUrl":"10.1162/neco_a_01689","url":null,"abstract":"For decades, fMRI data have been used to search for biomarkers for patients with schizophrenia. Still, firm conclusions are yet to be made, which is often attributed to the high internal heterogeneity of the disorder. A promising way to disentangle the heterogeneity is to search for subgroups of patients with more homogeneous biological profiles. We applied an unsupervised multiple co-clustering (MCC) method to identify subtypes using functional connectivity data from a multisite resting-state data set. We merged data from two publicly available databases and split the data into a discovery data set (143 patients and 143 healthy controls (HC)) and an external test data set (63 patients and 63 HC) from independent sites. On the discovery data, we investigated the stability of the clustering toward data splits and initializations. Subsequently we searched for cluster solutions, also called “views,” with a significant diagnosis association and evaluated these based on their subject and feature cluster separability, and correlation to clinical manifestations as measured with the positive and negative syndrome scale (PANSS). Finally, we validated our findings by testing the diagnosis association on the external test data. A major finding of our study was that the stability of the clustering was highly dependent on variations in the data set, and even across initializations, we found only a moderate subject clustering stability. Nevertheless, we still discovered one view with a significant diagnosis association. This view reproducibly showed an overrepresentation of schizophrenia patients in three subject clusters, and one feature cluster showed a continuous trend, ranging from positive to negative connectivity values, when sorted according to the proportions of patients with schizophrenia. When investigating all patients, none of the feature clusters in the view were associated with severity of positive, negative, and generalized symptoms, indicating that the cluster solutions reflect other disease related mechanisms.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141898955","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spontaneous Emergence of Robustness to Light Variation in CNNs With a Precortically Inspired Module 带有前皮质启发模块的 CNN 对光线变化的自发鲁棒性。
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-19 DOI: 10.1162/neco_a_01691
J. Petkovic;R. Fioresi
The analogies between the mammalian primary visual cortex and the structure of CNNs used for image classification tasks suggest that the introduction of an additional preliminary convolutional module inspired by the mathematical modeling of the precortical neuronal circuits can improve robustness with respect to global light intensity and contrast variations in the input images. We validate this hypothesis using the popular databases MNIST, FashionMNIST, and SVHN for these variations once an extra module is added.
哺乳动物初级视觉皮层与用于图像分类任务的 CNN 结构之间的类比表明,在皮层前神经元电路数学建模的启发下,引入额外的初步卷积模块可以提高输入图像中全局光强度和对比度变化的鲁棒性。我们使用流行的数据库 MNIST、FashionMNIST 和 SVHN 验证了这一假设,即一旦添加了额外的模块,这些变化就会出现。
{"title":"Spontaneous Emergence of Robustness to Light Variation in CNNs With a Precortically Inspired Module","authors":"J. Petkovic;R. Fioresi","doi":"10.1162/neco_a_01691","DOIUrl":"10.1162/neco_a_01691","url":null,"abstract":"The analogies between the mammalian primary visual cortex and the structure of CNNs used for image classification tasks suggest that the introduction of an additional preliminary convolutional module inspired by the mathematical modeling of the precortical neuronal circuits can improve robustness with respect to global light intensity and contrast variations in the input images. We validate this hypothesis using the popular databases MNIST, FashionMNIST, and SVHN for these variations once an extra module is added.","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141898956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Hyperdimensional Computing With Spiking Phasors 利用尖峰相位进行高效超维计算
IF 2.7 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-19 DOI: 10.1162/neco_a_01693
Jeff Orchard;P. Michael Furlong;Kathryn Simone
Hyperdimensional (HD) computing (also referred to as vector symbolic architectures, VSAs) offers a method for encoding symbols into vectors, allowing for those symbols to be combined in different ways to form other vectors in the same vector space. The vectors and operators form a compositional algebra, such that composite vectors can be decomposed back to their constituent vectors. Many useful algorithms have implementations in HD computing, such as classification, spatial navigation, language modeling, and logic. In this letter, we propose a spiking implementation of Fourier holographic reduced representation (FHRR), one of the most versatile VSAs. The phase of each complex number of an FHRR vector is encoded as a spike time within a cycle. Neuron models derived from these spiking phasors can perform the requisite vector operations to implement an FHRR. We demonstrate the power and versatility of our spiking networks in a number of foundational problem domains, including symbol binding and unbinding, spatial representation, function representation, function integration, and memory (i.e., signal delay).
超维(HD)计算(也称为向量符号架构,VSA)提供了一种将符号编码成向量的方法,允许这些符号以不同的方式组合成同一向量空间中的其他向量。向量和运算符构成了一个组合代数,因此复合向量可以分解回其组成向量。许多有用的算法都可以在高清计算中实现,如分类、空间导航、语言建模和逻辑。在这封信中,我们提出了傅立叶全息还原表示法(FHRR)的尖峰实施方案,这是最通用的 VSA 之一。FHRR 向量每个复数的相位被编码为一个周期内的尖峰时间。从这些尖峰相位衍生出来的神经元模型可以执行必要的向量运算,从而实现 FHRR。我们在多个基础问题领域展示了我们的尖峰网络的强大功能和多功能性,包括符号绑定和解绑、空间表示、函数表示、函数整合和记忆(即信号延迟)。
{"title":"Efficient Hyperdimensional Computing With Spiking Phasors","authors":"Jeff Orchard;P. Michael Furlong;Kathryn Simone","doi":"10.1162/neco_a_01693","DOIUrl":"10.1162/neco_a_01693","url":null,"abstract":"Hyperdimensional (HD) computing (also referred to as vector symbolic architectures, VSAs) offers a method for encoding symbols into vectors, allowing for those symbols to be combined in different ways to form other vectors in the same vector space. The vectors and operators form a compositional algebra, such that composite vectors can be decomposed back to their constituent vectors. Many useful algorithms have implementations in HD computing, such as classification, spatial navigation, language modeling, and logic. In this letter, we propose a spiking implementation of Fourier holographic reduced representation (FHRR), one of the most versatile VSAs. The phase of each complex number of an FHRR vector is encoded as a spike time within a cycle. Neuron models derived from these spiking phasors can perform the requisite vector operations to implement an FHRR. We demonstrate the power and versatility of our spiking networks in a number of foundational problem domains, including symbol binding and unbinding, spatial representation, function representation, function integration, and memory (i.e., signal delay).","PeriodicalId":54731,"journal":{"name":"Neural Computation","volume":null,"pages":null},"PeriodicalIF":2.7,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141898952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Computation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1