{"title":"MSANet: A Multi-Scale Attention Module","authors":"Yucheng Huang, Wei Liu, Chao Li, Yongsheng Liang, Huoxiang Yang, Fanyang Meng","doi":"10.1109/iske47853.2019.9170354","DOIUrl":null,"url":null,"abstract":"Multi-scale representation ability is one of key criteria for measuring convolutional neural networks (CNNs) effectiveness. Recent studies have shown that multi-scale features can represent different semantic information of original images, and a combination of them would have positive influence on vision tasks. Many researchers are investigated in extract the multi-scale features in a layerwise manner and equipped with relatively inflexible receptive field. In this paper, we propose a multi-scale attention (MSA) module for CNNs, namely MSANet, where the residual block comprises hierarchical attention connections and skip connections. The MSANet improves the multi-scale representation power of the network by adaptively enriching the receptive fields of each convolutional branch. We insert the proposed MSANet block into several backbone CNN models and achieve consistent improvement over backbone models on CIFAR-100 dataset. To better verify the effectiveness of MSANet, the experimental results on major network details, i.e., scale, depth, further demonstrate the superiority of the MSANet over the Res2Net methods.","PeriodicalId":399084,"journal":{"name":"2019 IEEE 14th International Conference on Intelligent Systems and Knowledge Engineering (ISKE)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE 14th International Conference on Intelligent Systems and Knowledge Engineering (ISKE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/iske47853.2019.9170354","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Multi-scale representation ability is one of key criteria for measuring convolutional neural networks (CNNs) effectiveness. Recent studies have shown that multi-scale features can represent different semantic information of original images, and a combination of them would have positive influence on vision tasks. Many researchers are investigated in extract the multi-scale features in a layerwise manner and equipped with relatively inflexible receptive field. In this paper, we propose a multi-scale attention (MSA) module for CNNs, namely MSANet, where the residual block comprises hierarchical attention connections and skip connections. The MSANet improves the multi-scale representation power of the network by adaptively enriching the receptive fields of each convolutional branch. We insert the proposed MSANet block into several backbone CNN models and achieve consistent improvement over backbone models on CIFAR-100 dataset. To better verify the effectiveness of MSANet, the experimental results on major network details, i.e., scale, depth, further demonstrate the superiority of the MSANet over the Res2Net methods.