{"title":"The roles of vision and proprioception in spatial tuning of sensory attenuation.","authors":"Clara Fritz, Manuel Bayer, Eckart Zimmermann","doi":"10.1007/s00221-024-06982-w","DOIUrl":null,"url":null,"abstract":"<p><p>When we touch ourselves, the pressure appears weaker compared to when someone else touches us, an effect known as sensory attenuation. Sensory attenuation is spatially tuned and does only occur if the positions of the touching and the touched body-party spatially coincide. Here, we ask about the contribution of visual or proprioceptive signals to determine self-touch. By using a 3D arm model in a virtual reality environment, we dissociated the visual from the proprioceptive arm signal. When a virtual arm was visible indicating self-touch, we found that sensory attenuation generalized across different locations. When no virtual arm was visible, we found sensory attenuation to be strongest when subjects pointed to the position where they felt their arm to be located. We conclude that the spatial tuning of tactile attenuation depends on which signal determines the occurrence of self-touch. When observers can see their hand, the visual signal dominates the proprioceptive determining self-touch in a single visual snapshot. When only the proprioceptive signal is available, the positions of the touching and the touched body-part must be separately estimated and subsequently compared if they overlap in anatomical space.</p>","PeriodicalId":12268,"journal":{"name":"Experimental Brain Research","volume":"243 1","pages":"42"},"PeriodicalIF":1.7000,"publicationDate":"2025-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11724775/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Experimental Brain Research","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s00221-024-06982-w","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
When we touch ourselves, the pressure appears weaker compared to when someone else touches us, an effect known as sensory attenuation. Sensory attenuation is spatially tuned and does only occur if the positions of the touching and the touched body-party spatially coincide. Here, we ask about the contribution of visual or proprioceptive signals to determine self-touch. By using a 3D arm model in a virtual reality environment, we dissociated the visual from the proprioceptive arm signal. When a virtual arm was visible indicating self-touch, we found that sensory attenuation generalized across different locations. When no virtual arm was visible, we found sensory attenuation to be strongest when subjects pointed to the position where they felt their arm to be located. We conclude that the spatial tuning of tactile attenuation depends on which signal determines the occurrence of self-touch. When observers can see their hand, the visual signal dominates the proprioceptive determining self-touch in a single visual snapshot. When only the proprioceptive signal is available, the positions of the touching and the touched body-part must be separately estimated and subsequently compared if they overlap in anatomical space.
期刊介绍:
Founded in 1966, Experimental Brain Research publishes original contributions on many aspects of experimental research of the central and peripheral nervous system. The focus is on molecular, physiology, behavior, neurochemistry, developmental, cellular and molecular neurobiology, and experimental pathology relevant to general problems of cerebral function. The journal publishes original papers, reviews, and mini-reviews.