After removing a virtual reality headset, people can be surprised to find that they are facing a different direction than expected. Here, we investigated if people can maintain spatial representations of one environment while immersed in another. In the first three experiments, stationary participants were asked to point to previously seen targets in one environment, either the real world or a virtual environment, while in the other environment. We varied the amount of misalignment between the two environments (detectable or undetectable), the virtual environment itself (lab or kitchen), and the instructions (general or egocentric priming). Pointing endpoints were based primarily on the locations of objects in the currently seen environment, suggesting a strong reliance on allocentric cues. In the fourth experiment, participants moved in virtual reality while keeping track of an unseen real-world target. We confirmed that the pointing errors were due to a reliance on the currently seen environment. It appears that people hardly ever keep track of object positions in a previously seen environment and instead primarily rely on currently available spatial information to plan their actions.
{"title":"Allocentric spatial representations dominate when switching between real and virtual worlds.","authors":"Meaghan McManus, Franziska Seifert, Immo Schütz, Katja Fiehler","doi":"10.1167/jov.25.13.7","DOIUrl":"10.1167/jov.25.13.7","url":null,"abstract":"<p><p>After removing a virtual reality headset, people can be surprised to find that they are facing a different direction than expected. Here, we investigated if people can maintain spatial representations of one environment while immersed in another. In the first three experiments, stationary participants were asked to point to previously seen targets in one environment, either the real world or a virtual environment, while in the other environment. We varied the amount of misalignment between the two environments (detectable or undetectable), the virtual environment itself (lab or kitchen), and the instructions (general or egocentric priming). Pointing endpoints were based primarily on the locations of objects in the currently seen environment, suggesting a strong reliance on allocentric cues. In the fourth experiment, participants moved in virtual reality while keeping track of an unseen real-world target. We confirmed that the pointing errors were due to a reliance on the currently seen environment. It appears that people hardly ever keep track of object positions in a previously seen environment and instead primarily rely on currently available spatial information to plan their actions.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"7"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12629136/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145507142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Benjamin Balas, Myra Morton, Molly Setchfield, Lily Roshau, Emily Westrick
Face pareidolia, the phenomenon of seeing face-like patterns in non-face images, has a dual nature: Pareidolic patterns are experienced as face-like, even while observers can recognize the true nature of the stimulus (Stuart et al., 2025). Although pareidolic faces seem to result largely from the canonical arrangement of eye spots and a mouth, we hypothesized that competition between veridical and face-like interpretations of pareidolic patterns may constrain face pareidolia in natural scenes and textures. Specifically, we predicted that contrast negation, which disrupts multiple aspects of mid- to high-level recognition, may increase rates of face pareidolia in complex natural textures by weakening the veridical, non-face stimulus interpretation. We presented adult participants (n = 27) and 5- to 12-year-old children (n = 67) with a series of natural images depicting textures such as grass, leaves, shells, and rocks. We asked participants to circle any patterns in each image that looked face-like, with no constraints on response time or pattern size, position, and orientation. We found that, across our adult and child samples, contrast-negated images yielded more pareidolic face detections than positive images. We conclude that disrupting veridical object and texture recognition enhances pareidolia in children and adults by compromising half of the dual nature of a pareidolic pattern.
面部幻想性视错觉,即在非人脸图像中看到类似人脸的图案的现象,具有双重性质:即使观察者可以识别刺激的真实性质,但幻想性模式也被体验为类似人脸的模式(Stuart et al., 2025)。尽管空想面孔似乎主要是由眼斑和嘴巴的规范排列造成的,但我们假设,对空想模式的真实解释和面部解释之间的竞争可能会限制自然场景和纹理中的面部空想。具体来说,我们预测对比否定会破坏中高级识别的多个方面,通过削弱真实的、非面部刺激的解释,可能会增加复杂自然纹理中面部空想性视错觉的发生率。我们向成人参与者(n = 27)和5- 12岁的儿童(n = 67)展示了一系列描绘草、树叶、贝壳和岩石等纹理的自然图像。我们要求参与者圈出每张图片中任何看起来像脸的图案,没有反应时间、图案大小、位置和方向的限制。我们发现,在我们的成人和儿童样本中,对比阴性图像比阳性图像产生更多的空想性面部检测。我们得出的结论是,干扰物体和纹理识别会增强儿童和成人的幻想性视错觉,因为它损害了幻想性视错觉模式的一半双重性质。
{"title":"Contrast negation increases face pareidolia rates in natural scenes.","authors":"Benjamin Balas, Myra Morton, Molly Setchfield, Lily Roshau, Emily Westrick","doi":"10.1167/jov.25.13.5","DOIUrl":"10.1167/jov.25.13.5","url":null,"abstract":"<p><p>Face pareidolia, the phenomenon of seeing face-like patterns in non-face images, has a dual nature: Pareidolic patterns are experienced as face-like, even while observers can recognize the true nature of the stimulus (Stuart et al., 2025). Although pareidolic faces seem to result largely from the canonical arrangement of eye spots and a mouth, we hypothesized that competition between veridical and face-like interpretations of pareidolic patterns may constrain face pareidolia in natural scenes and textures. Specifically, we predicted that contrast negation, which disrupts multiple aspects of mid- to high-level recognition, may increase rates of face pareidolia in complex natural textures by weakening the veridical, non-face stimulus interpretation. We presented adult participants (n = 27) and 5- to 12-year-old children (n = 67) with a series of natural images depicting textures such as grass, leaves, shells, and rocks. We asked participants to circle any patterns in each image that looked face-like, with no constraints on response time or pattern size, position, and orientation. We found that, across our adult and child samples, contrast-negated images yielded more pareidolic face detections than positive images. We conclude that disrupting veridical object and texture recognition enhances pareidolia in children and adults by compromising half of the dual nature of a pareidolic pattern.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"5"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12617666/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145507220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Karla Matic, Issam Tafech, Peter König, John-Dylan Haynes
After the offset of complex visual stimuli, rich stimulus information remains briefly available to the observer, reflecting a rapidly decaying iconic memory trace. Here we found that even if the cues are presented in the final stage of the stimulus presentation, the reportable information already starts decaying. Using closely spaced readout cues and a theoretical model of information availability, we observed that a cue has to be presented around 10 to 30 milliseconds before stimulus offset to access the full sensory information. We suggest that this does not reflect an early loss in sensory encoding, but instead it is a consequence of a latency in the processing of the cue that postpones the readout of the sensory representation by 10 to 30 milliseconds. Our analysis also shows that spatial proximity of items in complex arrays impacts sensory representation during both perceptual encoding and initial memory decay. Overall, these results provide a theoretical and empirical characterization of the readout from visual representations and offer a detailed insight into the transition from perception into iconic memory.
{"title":"Temporal dynamics and readout latency in perception and iconic memory.","authors":"Karla Matic, Issam Tafech, Peter König, John-Dylan Haynes","doi":"10.1167/jov.25.13.3","DOIUrl":"10.1167/jov.25.13.3","url":null,"abstract":"<p><p>After the offset of complex visual stimuli, rich stimulus information remains briefly available to the observer, reflecting a rapidly decaying iconic memory trace. Here we found that even if the cues are presented in the final stage of the stimulus presentation, the reportable information already starts decaying. Using closely spaced readout cues and a theoretical model of information availability, we observed that a cue has to be presented around 10 to 30 milliseconds before stimulus offset to access the full sensory information. We suggest that this does not reflect an early loss in sensory encoding, but instead it is a consequence of a latency in the processing of the cue that postpones the readout of the sensory representation by 10 to 30 milliseconds. Our analysis also shows that spatial proximity of items in complex arrays impacts sensory representation during both perceptual encoding and initial memory decay. Overall, these results provide a theoretical and empirical characterization of the readout from visual representations and offer a detailed insight into the transition from perception into iconic memory.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"3"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12603959/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145460314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jacob M Morris, Esteban Fernández-Juricic, Caryn E Plummer, Bret A Moore
To describe the visual field of three common model species in vision science to understand the organization of their visual perceptual experience and contribute to continued studies of visual processing. Visual fields were measured using an ophthalmoscopic reflex technique in four common ferrets, four albino rats, and six northern tree shrews. Animals were anesthetized to avoid stress and the midpoint between their eyes was centered inside a spherical space. A rotating perimetric arm was manipulated in 10° increments around the head. At each increment, direct ophthalmoscopy was used to visualize limits of the retinal reflex for each eye, the overlap being the binocular visual field. Mean binocularity in the horizontal plane was 63.7 ± 5.1°, 79.1 ± 7.4°, and 53.6 ± 12.0° in the ferret, rat, and shrew, respectively. Maximum mean binocularity was 69.0 ± 1.6° in the ferret, 90.0 ± 3.1° in the rat, and 53.6 ± 12.2° in the shrew, located at 10° above, 40° above, and at the horizontal plane, respectively. Binocularity extended to 160°, 200°, and 180° in the sagittal plane in the ferret, rat, and shrew, respectively, from at least below the nose to above the head in all animals. Establishing the extent of the visual field accessible to the retina provides insight into the egocentric perceptual experience of animals. In describing the visual field, we provide a reference for the representation of the visual space in different cortical and retinal regions, many of which represent specific subregions of the visual field.
{"title":"Visual field of the ferret (Mustela putorius furo), rat (Rattus norvegicus), and tree shrew (Tupaia belangeri).","authors":"Jacob M Morris, Esteban Fernández-Juricic, Caryn E Plummer, Bret A Moore","doi":"10.1167/jov.25.13.8","DOIUrl":"10.1167/jov.25.13.8","url":null,"abstract":"<p><p>To describe the visual field of three common model species in vision science to understand the organization of their visual perceptual experience and contribute to continued studies of visual processing. Visual fields were measured using an ophthalmoscopic reflex technique in four common ferrets, four albino rats, and six northern tree shrews. Animals were anesthetized to avoid stress and the midpoint between their eyes was centered inside a spherical space. A rotating perimetric arm was manipulated in 10° increments around the head. At each increment, direct ophthalmoscopy was used to visualize limits of the retinal reflex for each eye, the overlap being the binocular visual field. Mean binocularity in the horizontal plane was 63.7 ± 5.1°, 79.1 ± 7.4°, and 53.6 ± 12.0° in the ferret, rat, and shrew, respectively. Maximum mean binocularity was 69.0 ± 1.6° in the ferret, 90.0 ± 3.1° in the rat, and 53.6 ± 12.2° in the shrew, located at 10° above, 40° above, and at the horizontal plane, respectively. Binocularity extended to 160°, 200°, and 180° in the sagittal plane in the ferret, rat, and shrew, respectively, from at least below the nose to above the head in all animals. Establishing the extent of the visual field accessible to the retina provides insight into the egocentric perceptual experience of animals. In describing the visual field, we provide a reference for the representation of the visual space in different cortical and retinal regions, many of which represent specific subregions of the visual field.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"8"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12629130/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145514104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sophie Meissner, Jochen Miksch, Lena Würbach, Sascha Feder, Sabine Grimm, Wolfgang Einhäuser, Jutta Billino
Gaze behavior during locomotion must balance the sampling of relevant information and the need for a stable gait. To maintain a safe gait in the light of declining resources, older adults might shift this balance toward the uptake of gait-related information. We investigated how violations of expectations affect gaze behavior and information uptake across age groups by asking younger and older adults to locomote through a virtual hallway, where they encountered expected and unexpected objects. We found that older adults look more on the floor, despite the translational locomotion, though not the rotational, being virtual. Dwell times on unexpected objects were increased in both age groups compared to expected objects. Although older adults showed shorter dwell times on expected objects, dwell times on unexpected objects were similar across age groups. Thus the difference between expected and unexpected objects was greater in older adults. Gaze distributions were more influenced by cognitive control capacities than by motor control capacities. Our findings indicate that unexpected information attracts attention during locomotion-particularly in older adults. However, during actual locomotion in the real world, increased information processing might come at the cost of reduced gait safety if processing resources are shifted away from stabilizing gait.
{"title":"Violated expectations during locomotion through virtual environments: Age effects on gaze guidance.","authors":"Sophie Meissner, Jochen Miksch, Lena Würbach, Sascha Feder, Sabine Grimm, Wolfgang Einhäuser, Jutta Billino","doi":"10.1167/jov.25.13.11","DOIUrl":"10.1167/jov.25.13.11","url":null,"abstract":"<p><p>Gaze behavior during locomotion must balance the sampling of relevant information and the need for a stable gait. To maintain a safe gait in the light of declining resources, older adults might shift this balance toward the uptake of gait-related information. We investigated how violations of expectations affect gaze behavior and information uptake across age groups by asking younger and older adults to locomote through a virtual hallway, where they encountered expected and unexpected objects. We found that older adults look more on the floor, despite the translational locomotion, though not the rotational, being virtual. Dwell times on unexpected objects were increased in both age groups compared to expected objects. Although older adults showed shorter dwell times on expected objects, dwell times on unexpected objects were similar across age groups. Thus the difference between expected and unexpected objects was greater in older adults. Gaze distributions were more influenced by cognitive control capacities than by motor control capacities. Our findings indicate that unexpected information attracts attention during locomotion-particularly in older adults. However, during actual locomotion in the real world, increased information processing might come at the cost of reduced gait safety if processing resources are shifted away from stabilizing gait.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"11"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12663891/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145566014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wiebke Nörenberg, Richard Schweitzer, Martin Rolfs
The accurate inference of causality between actions and their sensory outcomes requires determining their temporal relationship correctly despite variable delays within and across sensory modalities. Temporal recalibration-the perceptual realignment of actions with delayed sensory feedback-has been demonstrated across various sensorimotor domains. Here, we investigate whether this mechanism extends to saccadic eye movements and sensory events contingent on them. In three experiments, participants made horizontal saccades that triggered high-contrast flashes at varying delays. They then reported whether the flashes occurred during or after the saccade, allowing us to track perceived event timing. Exposure to consistent delays between saccade onset and the flash led to a shift in perceptual reports: flashes presented after saccade offset were more often judged as occurring during the movement. This recalibration effect was robust even when we manipulated relevant visual cues such as the presence of a structured background or the continuity of the saccade target. In a replay condition, we found a significant but much smaller recalibration effect between replayed saccades and flash, demonstrating the importance of action execution for visuomotor temporal recalibration. These findings highlight the visual system's remarkable adaptability to temporal delays between eye movements and their sensory consequences. A similar recalibration mechanism may support perceptual stability in natural vision by dynamically realigning saccades with their resulting visual input, even amid changing visual conditions.
{"title":"Temporal recalibration to delayed visual consequences of saccades.","authors":"Wiebke Nörenberg, Richard Schweitzer, Martin Rolfs","doi":"10.1167/jov.25.13.4","DOIUrl":"10.1167/jov.25.13.4","url":null,"abstract":"<p><p>The accurate inference of causality between actions and their sensory outcomes requires determining their temporal relationship correctly despite variable delays within and across sensory modalities. Temporal recalibration-the perceptual realignment of actions with delayed sensory feedback-has been demonstrated across various sensorimotor domains. Here, we investigate whether this mechanism extends to saccadic eye movements and sensory events contingent on them. In three experiments, participants made horizontal saccades that triggered high-contrast flashes at varying delays. They then reported whether the flashes occurred during or after the saccade, allowing us to track perceived event timing. Exposure to consistent delays between saccade onset and the flash led to a shift in perceptual reports: flashes presented after saccade offset were more often judged as occurring during the movement. This recalibration effect was robust even when we manipulated relevant visual cues such as the presence of a structured background or the continuity of the saccade target. In a replay condition, we found a significant but much smaller recalibration effect between replayed saccades and flash, demonstrating the importance of action execution for visuomotor temporal recalibration. These findings highlight the visual system's remarkable adaptability to temporal delays between eye movements and their sensory consequences. A similar recalibration mechanism may support perceptual stability in natural vision by dynamically realigning saccades with their resulting visual input, even amid changing visual conditions.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"4"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12603963/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145460373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
When a manual reaching target is selected from a number of alternatives, decision uncertainty can often result in curvature of movement trajectories toward a nonchosen alternative. This curvature in the two-dimensional object plane is typically attributed to competitive interactions between different movement goals. Several models of action selection assume an explicit link between the momentary position of the hand and the state of the underlying decision process. Under this assumption, tracking the position of the hand can be used to infer the temporal evolution of the decision. However, even without a selection requirement, movements show variable amounts of curvature due to motor noise. We assessed the relative contributions of decision uncertainty and motor noise to the variability in curvature in naturalistic reach-to-grasp actions. Participants had to pick up one of two blocks (the brighter/dimmer block) and we manipulated decision uncertainty by varying the luminance difference between the two blocks. Single target baseline reaches were included to model the variability in curvature without a choice requirement. We assessed to what extent this baseline model can account for the curvature distributions observed under choice conditions, and tested several modifications of the model to capture any effects of decision uncertainty. The best model of the curvature distributions under choice conditions involved a mixture of the baseline component along with a separate choice component. The weight of this choice component and analysis of the likelihood of observed reaches under the choice/baseline components, suggest that the majority of reaches were unaffected by decision uncertainty and were compatible with the natural variability in movement trajectories due to motor noise. Unless the variability induced by factors unrelated to the decision process is adequately accounted for, the role of decision uncertainty may be overstated when it is inferred from reach trajectories.
{"title":"Disentangling decision uncertainty and motor noise in curved movement trajectories.","authors":"William G Chapman, Casimir J H Ludwig","doi":"10.1167/jov.25.13.6","DOIUrl":"10.1167/jov.25.13.6","url":null,"abstract":"<p><p>When a manual reaching target is selected from a number of alternatives, decision uncertainty can often result in curvature of movement trajectories toward a nonchosen alternative. This curvature in the two-dimensional object plane is typically attributed to competitive interactions between different movement goals. Several models of action selection assume an explicit link between the momentary position of the hand and the state of the underlying decision process. Under this assumption, tracking the position of the hand can be used to infer the temporal evolution of the decision. However, even without a selection requirement, movements show variable amounts of curvature due to motor noise. We assessed the relative contributions of decision uncertainty and motor noise to the variability in curvature in naturalistic reach-to-grasp actions. Participants had to pick up one of two blocks (the brighter/dimmer block) and we manipulated decision uncertainty by varying the luminance difference between the two blocks. Single target baseline reaches were included to model the variability in curvature without a choice requirement. We assessed to what extent this baseline model can account for the curvature distributions observed under choice conditions, and tested several modifications of the model to capture any effects of decision uncertainty. The best model of the curvature distributions under choice conditions involved a mixture of the baseline component along with a separate choice component. The weight of this choice component and analysis of the likelihood of observed reaches under the choice/baseline components, suggest that the majority of reaches were unaffected by decision uncertainty and were compatible with the natural variability in movement trajectories due to motor noise. Unless the variability induced by factors unrelated to the decision process is adequately accounted for, the role of decision uncertainty may be overstated when it is inferred from reach trajectories.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"6"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12629137/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145507653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A fundamental challenge in motion perception lies in the fact that the global motion of any translating object will generate a heterogeneous set of local motion signals that vary depending on the oriented contour information within each local region. This so-called aperture problem illustrates how the visual system must integrate diverse types of motion signals to attain direction selectivity that is invariant to visual form. Here, we investigated how form-invariant motion selectivity emerges across the human visual pathway by using functional magnetic resonance imaging (fMRI) to measure direction-selective responses to drifting gratings and random-dot motion, and then testing for reliable generalization across stimulus types. All visual areas of interest showed highly reliable direction-selective classification performance for a given stimulus type, but early areas V1 and V2 showed chance-level generalization between motion types. Indeed, V1 responses tended to confuse random-dot motion as resembling orthogonal grating motion, implying that drifting random dots generated oriented motion-streak responses in V1. By contrast, motion-sensitive areas MT+ and V3A showed reliable cross-generalization performance in our fMRI experiments that tested both linear and spiral motion trajectories. Our findings provide compelling evidence that motion direction selectivity becomes more invariant to stimulus form in higher visual areas, particularly along the dorsal visual pathway.
{"title":"Emergence of form-independent direction selectivity in human V3A and MT.","authors":"Sang Wook Hong, Frank Tong","doi":"10.1167/jov.25.13.1","DOIUrl":"10.1167/jov.25.13.1","url":null,"abstract":"<p><p>A fundamental challenge in motion perception lies in the fact that the global motion of any translating object will generate a heterogeneous set of local motion signals that vary depending on the oriented contour information within each local region. This so-called aperture problem illustrates how the visual system must integrate diverse types of motion signals to attain direction selectivity that is invariant to visual form. Here, we investigated how form-invariant motion selectivity emerges across the human visual pathway by using functional magnetic resonance imaging (fMRI) to measure direction-selective responses to drifting gratings and random-dot motion, and then testing for reliable generalization across stimulus types. All visual areas of interest showed highly reliable direction-selective classification performance for a given stimulus type, but early areas V1 and V2 showed chance-level generalization between motion types. Indeed, V1 responses tended to confuse random-dot motion as resembling orthogonal grating motion, implying that drifting random dots generated oriented motion-streak responses in V1. By contrast, motion-sensitive areas MT+ and V3A showed reliable cross-generalization performance in our fMRI experiments that tested both linear and spiral motion trajectories. Our findings provide compelling evidence that motion direction selectivity becomes more invariant to stimulus form in higher visual areas, particularly along the dorsal visual pathway.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"1"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12594259/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145432755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
During self-movement, the visual system can identify scene-relative object motion via flow parsing and estimate the direction of self-movement (heading) from optic flow. However, the temporal dynamics of these two processes have not been examined and compared using matched displays. In this study, we examined how the accuracy of flow parsing and heading estimation changed with stimulus duration. Participants viewed a stereo optic flow display simulating forward translational self-movement through a cloud composed of wireframe objects with stimulus durations at 100, 200, 400, 700, and 1000 ms. In Experiment 1, a yellow dot probe moved vertically for 100 ms in the scene near the end of the trial. A nulling motion component was added through an adaptive staircase to the probe's image motion to determine when the probe was perceived to move vertically in the scene, which was then used to compute the accuracy of flow parsing. In Experiment 2, participants viewed the same optic flow display without the moving probe object. The simulated heading was randomly varied in each trial, and participants were asked to estimate heading at the end of the trial. As stimulus duration increased, the accuracy of flow parsing decreased, whereas the accuracy of heading estimation increased. These contrasting temporal dynamics suggest that despite both processes relying on optic flow, flow parsing and heading estimation involve distinct processing mechanisms with different temporal characteristics. This divergence, together with previous neurophysiological findings, led us to propose two potential neural mechanisms subserving these two processes to inspire future research.
{"title":"Distinct temporal dynamics of judging scene-relative object motion and estimating heading from optic flow.","authors":"Mingyang Xie, Li Li","doi":"10.1167/jov.25.13.10","DOIUrl":"10.1167/jov.25.13.10","url":null,"abstract":"<p><p>During self-movement, the visual system can identify scene-relative object motion via flow parsing and estimate the direction of self-movement (heading) from optic flow. However, the temporal dynamics of these two processes have not been examined and compared using matched displays. In this study, we examined how the accuracy of flow parsing and heading estimation changed with stimulus duration. Participants viewed a stereo optic flow display simulating forward translational self-movement through a cloud composed of wireframe objects with stimulus durations at 100, 200, 400, 700, and 1000 ms. In Experiment 1, a yellow dot probe moved vertically for 100 ms in the scene near the end of the trial. A nulling motion component was added through an adaptive staircase to the probe's image motion to determine when the probe was perceived to move vertically in the scene, which was then used to compute the accuracy of flow parsing. In Experiment 2, participants viewed the same optic flow display without the moving probe object. The simulated heading was randomly varied in each trial, and participants were asked to estimate heading at the end of the trial. As stimulus duration increased, the accuracy of flow parsing decreased, whereas the accuracy of heading estimation increased. These contrasting temporal dynamics suggest that despite both processes relying on optic flow, flow parsing and heading estimation involve distinct processing mechanisms with different temporal characteristics. This divergence, together with previous neurophysiological findings, led us to propose two potential neural mechanisms subserving these two processes to inspire future research.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"10"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12636965/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Locomotion poses a challenge to clear and stable vision. Reflexive head and eye movements act to stabilize the retinal image, but these do not act perfectly, so retinal image motion is increased during walking compared with standing. We nevertheless perceive the world as clear and stable during locomotion, suggesting that the visual system is well-adapted to meet the challenges posed by locomotion. To better understand these processes, we assessed dynamic contrast sensitivity during locomotion by presenting brief (24 ms) foveal Gabor targets (6°, 11 cpd) at threshold contrast to observers walking on a treadmill in an otherwise darkened room. Head and ankle motion were tracked, and presentation time was randomized, which allowed post hoc binning of responses according to stride-cycle timing to investigate how sensitivity is impacted by head motion and stride-cycle timing. Contrast sensitivity was improved during walking compared with standing over large portions of the stride cycle, except for epochs aligned with heel strikes, which drive large and unpredictable perturbations. This resulted in periodicity in contrast sensitivity at two cycles per stride, with additional oscillations observed at four and six cycles per stride. Pupil size was found to be moderately larger during walking compared with standing and also exhibited periodic fluctuations that were phase-locked to the stride cycle. Perceptual oscillations reflect the entrainment of visual processing by active behaviors. Robust contrast sensitivity during walking may be supported by action-contingent effects of locomotion on visual cortical activity that have been observed in several animal models.
{"title":"Dynamic contrast sensitivity during human locomotion.","authors":"Brian Szekely, Paul R MacNeilage","doi":"10.1167/jov.25.13.9","DOIUrl":"10.1167/jov.25.13.9","url":null,"abstract":"<p><p>Locomotion poses a challenge to clear and stable vision. Reflexive head and eye movements act to stabilize the retinal image, but these do not act perfectly, so retinal image motion is increased during walking compared with standing. We nevertheless perceive the world as clear and stable during locomotion, suggesting that the visual system is well-adapted to meet the challenges posed by locomotion. To better understand these processes, we assessed dynamic contrast sensitivity during locomotion by presenting brief (24 ms) foveal Gabor targets (6°, 11 cpd) at threshold contrast to observers walking on a treadmill in an otherwise darkened room. Head and ankle motion were tracked, and presentation time was randomized, which allowed post hoc binning of responses according to stride-cycle timing to investigate how sensitivity is impacted by head motion and stride-cycle timing. Contrast sensitivity was improved during walking compared with standing over large portions of the stride cycle, except for epochs aligned with heel strikes, which drive large and unpredictable perturbations. This resulted in periodicity in contrast sensitivity at two cycles per stride, with additional oscillations observed at four and six cycles per stride. Pupil size was found to be moderately larger during walking compared with standing and also exhibited periodic fluctuations that were phase-locked to the stride cycle. Perceptual oscillations reflect the entrainment of visual processing by active behaviors. Robust contrast sensitivity during walking may be supported by action-contingent effects of locomotion on visual cortical activity that have been observed in several animal models.</p>","PeriodicalId":49955,"journal":{"name":"Journal of Vision","volume":"25 13","pages":"9"},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12637006/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145515018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}