{"title":"图像标注的视觉和视觉语言领域的几何敏感语义建模","authors":"Wencai Zhu, Zetao Jiang, Yuting He","doi":"10.1016/j.engappai.2025.110330","DOIUrl":null,"url":null,"abstract":"<div><div>Transformer-based models with grid features as visual representations perform well in image captioning. However, the division and flattening operations increase the difficulty of capturing objects and their relationships via pure semantic modeling. Furthermore, the natural language generated by the current Transformer model still suffers from semantic overconcentration. In this paper, we aim to improve the attention modules in two ways to solve the above issues. We first propose a Geometry-Sensitive Self-Attention (GSSA) module, subdivide geometric signals in the visual domain into relative position and distance, and assist the semantic modeling process according to their unique characteristics. It compensates for the lack of objects and their relationships in the grid features. Then, we propose a Geometry-Sensitive Cross-Attention (GSCA) module, which perceives the source neighboring relationships between images and text in the visual-language domain from a geometric perspective and uses these relationships to adjust the semantic correspondences between the two dynamically. It spreads overly focused attention to surrounding grids to improve understanding of full image content during captioning. To prove our designs, we apply GSSA and GSCA to a standard Transformer to construct a novel Geometry-Sensitive Transformer Network (GSTNet), which conducts geometry-sensitive semantic modeling in visual and visual-language domains. Extensive experiments are conducted to verify the effectiveness of our proposal. The results show that our GSTNet achieves superior performance compared to many state-of-the-art image captioning models on the Microsoft Common Objects in Context (MSCOCO) dataset. Besides, the generalization of GSTNet is also verified on the Flickr30k dataset.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"147 ","pages":"Article 110330"},"PeriodicalIF":8.0000,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Geometry-sensitive semantic modeling in visual and visual-language domains for image captioning\",\"authors\":\"Wencai Zhu, Zetao Jiang, Yuting He\",\"doi\":\"10.1016/j.engappai.2025.110330\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Transformer-based models with grid features as visual representations perform well in image captioning. However, the division and flattening operations increase the difficulty of capturing objects and their relationships via pure semantic modeling. Furthermore, the natural language generated by the current Transformer model still suffers from semantic overconcentration. In this paper, we aim to improve the attention modules in two ways to solve the above issues. We first propose a Geometry-Sensitive Self-Attention (GSSA) module, subdivide geometric signals in the visual domain into relative position and distance, and assist the semantic modeling process according to their unique characteristics. It compensates for the lack of objects and their relationships in the grid features. Then, we propose a Geometry-Sensitive Cross-Attention (GSCA) module, which perceives the source neighboring relationships between images and text in the visual-language domain from a geometric perspective and uses these relationships to adjust the semantic correspondences between the two dynamically. It spreads overly focused attention to surrounding grids to improve understanding of full image content during captioning. To prove our designs, we apply GSSA and GSCA to a standard Transformer to construct a novel Geometry-Sensitive Transformer Network (GSTNet), which conducts geometry-sensitive semantic modeling in visual and visual-language domains. Extensive experiments are conducted to verify the effectiveness of our proposal. The results show that our GSTNet achieves superior performance compared to many state-of-the-art image captioning models on the Microsoft Common Objects in Context (MSCOCO) dataset. Besides, the generalization of GSTNet is also verified on the Flickr30k dataset.</div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"147 \",\"pages\":\"Article 110330\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-05-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625003306\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/2/25 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625003306","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/25 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Geometry-sensitive semantic modeling in visual and visual-language domains for image captioning
Transformer-based models with grid features as visual representations perform well in image captioning. However, the division and flattening operations increase the difficulty of capturing objects and their relationships via pure semantic modeling. Furthermore, the natural language generated by the current Transformer model still suffers from semantic overconcentration. In this paper, we aim to improve the attention modules in two ways to solve the above issues. We first propose a Geometry-Sensitive Self-Attention (GSSA) module, subdivide geometric signals in the visual domain into relative position and distance, and assist the semantic modeling process according to their unique characteristics. It compensates for the lack of objects and their relationships in the grid features. Then, we propose a Geometry-Sensitive Cross-Attention (GSCA) module, which perceives the source neighboring relationships between images and text in the visual-language domain from a geometric perspective and uses these relationships to adjust the semantic correspondences between the two dynamically. It spreads overly focused attention to surrounding grids to improve understanding of full image content during captioning. To prove our designs, we apply GSSA and GSCA to a standard Transformer to construct a novel Geometry-Sensitive Transformer Network (GSTNet), which conducts geometry-sensitive semantic modeling in visual and visual-language domains. Extensive experiments are conducted to verify the effectiveness of our proposal. The results show that our GSTNet achieves superior performance compared to many state-of-the-art image captioning models on the Microsoft Common Objects in Context (MSCOCO) dataset. Besides, the generalization of GSTNet is also verified on the Flickr30k dataset.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.