Robust Coding of Eye Position in Posterior Parietal Cortex despite Context-Dependent Tuning.

IF 0.1 0 RELIGION International Review of Mission Pub Date : 2022-05-18 Epub Date: 2022-04-11 DOI:10.1523/JNEUROSCI.0674-21.2022
Jamie R McFadyen, Barbara Heider, Anushree N Karkhanis, Shaun L Cloherty, Fabian Muñoz, Ralph M Siegel, Adam P Morris
{"title":"Robust Coding of Eye Position in Posterior Parietal Cortex despite Context-Dependent Tuning.","authors":"Jamie R McFadyen, Barbara Heider, Anushree N Karkhanis, Shaun L Cloherty, Fabian Muñoz, Ralph M Siegel, Adam P Morris","doi":"10.1523/JNEUROSCI.0674-21.2022","DOIUrl":null,"url":null,"abstract":"<p><p>Neurons in posterior parietal cortex (PPC) encode many aspects of the sensory world (e.g., scene structure), the posture of the body, and plans for action. For a downstream computation, however, only some of these dimensions are relevant; the rest are \"nuisance variables\" because their influence on neural activity changes with sensory and behavioral context, potentially corrupting the read-out of relevant information. Here we show that a key postural variable for vision (eye position) is represented robustly in male macaque PPC across a range of contexts, although the tuning of single neurons depended strongly on context. Contexts were defined by different stages of a visually guided reaching task, including (1) a visually sparse epoch, (2) a visually rich epoch, (3) a \"go\" epoch in which the reach was cued, and (4) during the reach itself. Eye position was constant within trials but varied across trials in a 3 × 3 grid spanning 24° × 24°. Using demixed principal component analysis of neural spike-counts, we found that the subspace of the population response encoding eye position is orthogonal to that encoding task context. Accordingly, a context-naive (fixed-parameter) decoder was nevertheless able to estimate eye position reliably across contexts. Errors were small given the sample size (∼1.78°) and would likely be even smaller with larger populations. Moreover, they were comparable to that of decoders that were optimized for each context. Our results suggest that population codes in PPC shield encoded signals from crosstalk to support robust sensorimotor transformations across contexts.<b>SIGNIFICANCE STATEMENT</b> Neurons in posterior parietal cortex (PPC) which are sensitive to gaze direction are thought to play a key role in spatial perception and behavior (e.g., reaching, navigation), and provide a potential substrate for brain-controlled prosthetics. Many, however, change their tuning under different sensory and behavioral contexts, raising the prospect that they provide unreliable representations of egocentric space. Here, we analyze the structure of encoding dimensions for gaze direction and context in PPC during different stages of a visually guided reaching task. We use demixed dimensionality reduction and decoding techniques to show that the coding of gaze direction in PPC is mostly invariant to context. This suggests that PPC can provide reliable spatial information across sensory and behavioral contexts.</p>","PeriodicalId":54038,"journal":{"name":"International Review of Mission","volume":"90 1","pages":"4116-4130"},"PeriodicalIF":0.1000,"publicationDate":"2022-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9121829/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Review of Mission","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1523/JNEUROSCI.0674-21.2022","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2022/4/11 0:00:00","PubModel":"Epub","JCR":"0","JCRName":"RELIGION","Score":null,"Total":0}
引用次数: 0

Abstract

Neurons in posterior parietal cortex (PPC) encode many aspects of the sensory world (e.g., scene structure), the posture of the body, and plans for action. For a downstream computation, however, only some of these dimensions are relevant; the rest are "nuisance variables" because their influence on neural activity changes with sensory and behavioral context, potentially corrupting the read-out of relevant information. Here we show that a key postural variable for vision (eye position) is represented robustly in male macaque PPC across a range of contexts, although the tuning of single neurons depended strongly on context. Contexts were defined by different stages of a visually guided reaching task, including (1) a visually sparse epoch, (2) a visually rich epoch, (3) a "go" epoch in which the reach was cued, and (4) during the reach itself. Eye position was constant within trials but varied across trials in a 3 × 3 grid spanning 24° × 24°. Using demixed principal component analysis of neural spike-counts, we found that the subspace of the population response encoding eye position is orthogonal to that encoding task context. Accordingly, a context-naive (fixed-parameter) decoder was nevertheless able to estimate eye position reliably across contexts. Errors were small given the sample size (∼1.78°) and would likely be even smaller with larger populations. Moreover, they were comparable to that of decoders that were optimized for each context. Our results suggest that population codes in PPC shield encoded signals from crosstalk to support robust sensorimotor transformations across contexts.SIGNIFICANCE STATEMENT Neurons in posterior parietal cortex (PPC) which are sensitive to gaze direction are thought to play a key role in spatial perception and behavior (e.g., reaching, navigation), and provide a potential substrate for brain-controlled prosthetics. Many, however, change their tuning under different sensory and behavioral contexts, raising the prospect that they provide unreliable representations of egocentric space. Here, we analyze the structure of encoding dimensions for gaze direction and context in PPC during different stages of a visually guided reaching task. We use demixed dimensionality reduction and decoding techniques to show that the coding of gaze direction in PPC is mostly invariant to context. This suggests that PPC can provide reliable spatial information across sensory and behavioral contexts.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
在顶叶后皮层中对眼球位置进行稳健编码,尽管这取决于上下文的调谐。
后顶叶皮层(PPC)的神经元对感官世界的许多方面(如场景结构)、身体姿势和行动计划进行编码。然而,对于下游计算而言,这些维度中只有一些是相关的;其余的都是 "讨厌的变量",因为它们对神经活动的影响会随着感觉和行为背景的变化而变化,从而有可能破坏相关信息的读出。在这里,我们证明了视觉的一个关键姿势变量(眼睛位置)在雄性猕猴的大脑皮层中能在各种情境下被稳健地表示出来,尽管单个神经元的调谐在很大程度上取决于情境。情境是由视觉引导的伸手任务的不同阶段定义的,包括(1)视觉稀疏的时程,(2)视觉丰富的时程,(3)"开始 "时程,其中伸手任务受到提示,以及(4)伸手任务本身。眼球位置在各次试验中保持不变,但在跨度为 24° × 24° 的 3 × 3 网格中,各次试验的眼球位置各不相同。通过对神经尖峰计数进行混合主成分分析,我们发现编码眼球位置的群体反应子空间与编码任务情境的子空间是正交的。因此,与情境无关(固定参数)的解码器能够在不同情境下可靠地估计眼球位置。从样本量来看,误差很小(1.78°),如果样本量更大,误差可能会更小。此外,这些误差与针对每种情境进行优化的解码器的误差相当。我们的研究结果表明,顶叶后皮层(PPC)中对凝视方向敏感的神经元群编码能够屏蔽来自串扰的编码信号,从而支持跨情境的稳健感觉运动转换。然而,在不同的感官和行为环境下,许多神经元的调谐会发生变化,从而导致它们对以自我为中心的空间提供不可靠的表征。在此,我们分析了在视觉引导的伸手任务的不同阶段,PPC 中注视方向和上下文的编码维度结构。我们使用去混合降维和解码技术表明,PPC 中的注视方向编码大多与上下文无关。这表明,在不同的感官和行为情境下,PPC 可以提供可靠的空间信息。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
0.20
自引率
0.00%
发文量
19
期刊最新文献
Issue Information Editorial Negotiating Popular Culture and Public Theology in the Indonesian Context Reparations and the Ministry of Planetary Peace Troubling (Public) Theologies: Spaces, Bodies, Technologies. Edited by Jione Havea. Minneapolis: Fortress Academic, 2023. 238 pp.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1