{"title":"基于Deeplabv3+和注意机制的语义分割","authors":"Rongrong Liu, Dongzhi He","doi":"10.1109/IMCEC51613.2021.9482207","DOIUrl":null,"url":null,"abstract":"In this paper, we propose vertical attention and spatial attention network (VSANet), which is a semantic segmentation method based on Deeplabv3+ and attention module, for improving semantic segmentation effect for autonomous driving road scene images. The improvement of this paper is primarily on the following two aspects. One is that this paper introduces the spatial attention module (SAM) after the atrous convolution, which effectively obtains more spatial context information. Second, by studying the road scene image, it is found that there are considerable differences in the pixel-level distribution of the horizontal segmentation area in the image. For this reason, this paper introduces the vertical attention module (VAM), which can better segment the road scene image. A large number of experimental results indicate that the segmentation accuracy of the proposed model is improved by 1.94% compared with the Deeplabv3+ network model on the test dataset of Cityscapes dataset.","PeriodicalId":240400,"journal":{"name":"2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC)","volume":"142 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Semantic Segmentation Based on Deeplabv3+ and Attention Mechanism\",\"authors\":\"Rongrong Liu, Dongzhi He\",\"doi\":\"10.1109/IMCEC51613.2021.9482207\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we propose vertical attention and spatial attention network (VSANet), which is a semantic segmentation method based on Deeplabv3+ and attention module, for improving semantic segmentation effect for autonomous driving road scene images. The improvement of this paper is primarily on the following two aspects. One is that this paper introduces the spatial attention module (SAM) after the atrous convolution, which effectively obtains more spatial context information. Second, by studying the road scene image, it is found that there are considerable differences in the pixel-level distribution of the horizontal segmentation area in the image. For this reason, this paper introduces the vertical attention module (VAM), which can better segment the road scene image. A large number of experimental results indicate that the segmentation accuracy of the proposed model is improved by 1.94% compared with the Deeplabv3+ network model on the test dataset of Cityscapes dataset.\",\"PeriodicalId\":240400,\"journal\":{\"name\":\"2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC)\",\"volume\":\"142 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-06-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IMCEC51613.2021.9482207\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IMCEC51613.2021.9482207","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
摘要
为了提高自动驾驶道路场景图像的语义分割效果,本文提出了一种基于Deeplabv3+和注意力模块的语义分割方法——垂直注意力和空间注意力网络(vertical attention and spatial attention network, VSANet)。本文的改进主要体现在以下两个方面。一是在亚历斯卷积之后引入空间注意模块(SAM),有效地获取了更多的空间上下文信息。其次,通过对道路场景图像的研究,发现图像中水平分割区域的像素级分布存在较大差异。为此,本文引入了垂直关注模块(vertical attention module, VAM),该模块可以更好地分割道路场景图像。大量实验结果表明,在cityscape数据集的测试数据集上,与Deeplabv3+网络模型相比,该模型的分割精度提高了1.94%。
Semantic Segmentation Based on Deeplabv3+ and Attention Mechanism
In this paper, we propose vertical attention and spatial attention network (VSANet), which is a semantic segmentation method based on Deeplabv3+ and attention module, for improving semantic segmentation effect for autonomous driving road scene images. The improvement of this paper is primarily on the following two aspects. One is that this paper introduces the spatial attention module (SAM) after the atrous convolution, which effectively obtains more spatial context information. Second, by studying the road scene image, it is found that there are considerable differences in the pixel-level distribution of the horizontal segmentation area in the image. For this reason, this paper introduces the vertical attention module (VAM), which can better segment the road scene image. A large number of experimental results indicate that the segmentation accuracy of the proposed model is improved by 1.94% compared with the Deeplabv3+ network model on the test dataset of Cityscapes dataset.