Andrea Gregor de Varda , Marco Petilli, Marco Marelli
{"title":"A distributional model of concepts grounded in the spatial organization of objects","authors":"Andrea Gregor de Varda , Marco Petilli, Marco Marelli","doi":"10.1016/j.jml.2025.104624","DOIUrl":null,"url":null,"abstract":"<div><div>Data-driven models of concepts are gaining popularity in Psychology and Cognitive Science. Distributional semantic models represent word meanings as abstract word co-occurrence patterns, and excel at capturing human meaning intuitions about conceptual relationships; however, they lack the explicit links to the physical world that humans acquire through perception. Computer vision neural networks, on the other hand, can produce representations of visually-grounded concepts, but they do not support the extraction of information about the relationships between objects. To bridge the gap between distributional semantic models and computer vision networks, we introduce SemanticScape, a model of semantic concepts grounded in the visual relationships between objects in natural images. The model captures the latent statistics of the spatial organization of objects in the visual environment. Its implementation is based on the calculation of the summed Euclidean distances between all object pairs in visual scenes, which are then abstracted by means of dimensionality reduction. We validate our model against human explicit intuitions on semantic and visual similarity, relatedness, analogical reasoning, and several semantic and visual implicit processing measurements. Our results show that SemanticScape explains variance in human responses in the semantic tasks above and beyond what can be accounted for by standard distributional semantic models and convolutional neural networks; however, it is not predictive of human performance in implicit perceptual tasks. Our findings highlight that implicit information about the objects’ spatial distribution in the environment has a specific impact on semantic processing, demonstrating the importance of this often neglected experiential source.</div></div>","PeriodicalId":16493,"journal":{"name":"Journal of memory and language","volume":"142 ","pages":"Article 104624"},"PeriodicalIF":2.9000,"publicationDate":"2025-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of memory and language","FirstCategoryId":"102","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0749596X25000178","RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"LINGUISTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Data-driven models of concepts are gaining popularity in Psychology and Cognitive Science. Distributional semantic models represent word meanings as abstract word co-occurrence patterns, and excel at capturing human meaning intuitions about conceptual relationships; however, they lack the explicit links to the physical world that humans acquire through perception. Computer vision neural networks, on the other hand, can produce representations of visually-grounded concepts, but they do not support the extraction of information about the relationships between objects. To bridge the gap between distributional semantic models and computer vision networks, we introduce SemanticScape, a model of semantic concepts grounded in the visual relationships between objects in natural images. The model captures the latent statistics of the spatial organization of objects in the visual environment. Its implementation is based on the calculation of the summed Euclidean distances between all object pairs in visual scenes, which are then abstracted by means of dimensionality reduction. We validate our model against human explicit intuitions on semantic and visual similarity, relatedness, analogical reasoning, and several semantic and visual implicit processing measurements. Our results show that SemanticScape explains variance in human responses in the semantic tasks above and beyond what can be accounted for by standard distributional semantic models and convolutional neural networks; however, it is not predictive of human performance in implicit perceptual tasks. Our findings highlight that implicit information about the objects’ spatial distribution in the environment has a specific impact on semantic processing, demonstrating the importance of this often neglected experiential source.
期刊介绍:
Articles in the Journal of Memory and Language contribute to the formulation of scientific issues and theories in the areas of memory, language comprehension and production, and cognitive processes. Special emphasis is given to research articles that provide new theoretical insights based on a carefully laid empirical foundation. The journal generally favors articles that provide multiple experiments. In addition, significant theoretical papers without new experimental findings may be published.
The Journal of Memory and Language is a valuable tool for cognitive scientists, including psychologists, linguists, and others interested in memory and learning, language, reading, and speech.
Research Areas include:
• Topics that illuminate aspects of memory or language processing
• Linguistics
• Neuropsychology.