{"title":"基于轻量级网络的牙齿注意力机制图像分类","authors":"Zejun Zhang, Yulong He, Huabin He, Zhiming Cai","doi":"10.1145/3548608.3559211","DOIUrl":null,"url":null,"abstract":"In order to solve the problem that embedding attention mechanism module into lightweight convolutional neural network will increase the number of parameters and FLOPs with less accuracy improvement, a Teeth Attention Module (TAM) based on lightweight network was proposed. The feature map is enhanced from two aspects of channel attention and spatial attention. In the channel attention module, local channel interaction is carried out from height and width without dimensionality reduction through one-dimensional convolution, and it is divided into MP-ECAM and AP-ECAM according to pooling mode. In the spatial attention module, the average value and maximum value of each pixel point in the feature map are calculated according to the dimensions to construct ESAM. Finally, TAM is constructed in a channel-space-channel manner. TAM was embedded into the lightweight networks MobileNetV2, ShuffleNetV2 and EfficientNetV1 respectively, and the experiment was carried out on CIFAR-10 image classification datasets. Compared with Squeeze and Excitation Module (SE), Convolutional Block Attention Module (CBAM) and Efficient Channel Attention Module (ECA), this module can achieve higher accuracy with less increase of the number of parameters and FLOPs.","PeriodicalId":201434,"journal":{"name":"Proceedings of the 2022 2nd International Conference on Control and Intelligent Robotics","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Teeth attention mechanism image classification based on lightweight network\",\"authors\":\"Zejun Zhang, Yulong He, Huabin He, Zhiming Cai\",\"doi\":\"10.1145/3548608.3559211\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In order to solve the problem that embedding attention mechanism module into lightweight convolutional neural network will increase the number of parameters and FLOPs with less accuracy improvement, a Teeth Attention Module (TAM) based on lightweight network was proposed. The feature map is enhanced from two aspects of channel attention and spatial attention. In the channel attention module, local channel interaction is carried out from height and width without dimensionality reduction through one-dimensional convolution, and it is divided into MP-ECAM and AP-ECAM according to pooling mode. In the spatial attention module, the average value and maximum value of each pixel point in the feature map are calculated according to the dimensions to construct ESAM. Finally, TAM is constructed in a channel-space-channel manner. TAM was embedded into the lightweight networks MobileNetV2, ShuffleNetV2 and EfficientNetV1 respectively, and the experiment was carried out on CIFAR-10 image classification datasets. Compared with Squeeze and Excitation Module (SE), Convolutional Block Attention Module (CBAM) and Efficient Channel Attention Module (ECA), this module can achieve higher accuracy with less increase of the number of parameters and FLOPs.\",\"PeriodicalId\":201434,\"journal\":{\"name\":\"Proceedings of the 2022 2nd International Conference on Control and Intelligent Robotics\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-06-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2022 2nd International Conference on Control and Intelligent Robotics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3548608.3559211\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 2nd International Conference on Control and Intelligent Robotics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3548608.3559211","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Teeth attention mechanism image classification based on lightweight network
In order to solve the problem that embedding attention mechanism module into lightweight convolutional neural network will increase the number of parameters and FLOPs with less accuracy improvement, a Teeth Attention Module (TAM) based on lightweight network was proposed. The feature map is enhanced from two aspects of channel attention and spatial attention. In the channel attention module, local channel interaction is carried out from height and width without dimensionality reduction through one-dimensional convolution, and it is divided into MP-ECAM and AP-ECAM according to pooling mode. In the spatial attention module, the average value and maximum value of each pixel point in the feature map are calculated according to the dimensions to construct ESAM. Finally, TAM is constructed in a channel-space-channel manner. TAM was embedded into the lightweight networks MobileNetV2, ShuffleNetV2 and EfficientNetV1 respectively, and the experiment was carried out on CIFAR-10 image classification datasets. Compared with Squeeze and Excitation Module (SE), Convolutional Block Attention Module (CBAM) and Efficient Channel Attention Module (ECA), this module can achieve higher accuracy with less increase of the number of parameters and FLOPs.