{"title":"通过 RGB 角度旋转方法提高 CNN 模型分类性能","authors":"Yahya Dogan, Cuneyt Ozdemir, Yılmaz Kaya","doi":"10.1007/s00521-024-10232-z","DOIUrl":null,"url":null,"abstract":"<p>In recent years, convolutional neural networks have significantly advanced the field of computer vision by automatically extracting features from image data. CNNs enable the modeling of complex and abstract image features using learnable filters, eliminating the need for manual feature extraction. However, combining feature maps obtained from CNNs with different approaches can lead to more complex and interpretable inferences, thereby enhancing model performance and generalizability. In this study, we propose a new method called RGB angle rotation to effectively obtain feature maps from RGB images. Our method rotates color channels at different angles and uses the angle information between channels to generate new feature maps. We then investigate the effects of integrating models trained with these feature maps into an ensemble architecture. Experimental results on the CIFAR-10 dataset show that using the proposed method in the ensemble model results in performance increases of 9.10 and 8.42% for the B and R channels, respectively, compared to the original model, while the effect of the G channel is very limited. For the CIFAR-100 dataset, the proposed method resulted in a 17.09% improvement in ensemble model performance for the R channel, a 5.06% increase for the B channel, and no significant improvement for the G channel compared to the original model. Additionally, we compared our method with traditional feature extraction methods like scale-invariant feature transform and local binary pattern and observed higher performance. In conclusion, it has been observed that the proposed RGB angle rotation method significantly impacts model performance.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing CNN model classification performance through RGB angle rotation method\",\"authors\":\"Yahya Dogan, Cuneyt Ozdemir, Yılmaz Kaya\",\"doi\":\"10.1007/s00521-024-10232-z\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>In recent years, convolutional neural networks have significantly advanced the field of computer vision by automatically extracting features from image data. CNNs enable the modeling of complex and abstract image features using learnable filters, eliminating the need for manual feature extraction. However, combining feature maps obtained from CNNs with different approaches can lead to more complex and interpretable inferences, thereby enhancing model performance and generalizability. In this study, we propose a new method called RGB angle rotation to effectively obtain feature maps from RGB images. Our method rotates color channels at different angles and uses the angle information between channels to generate new feature maps. We then investigate the effects of integrating models trained with these feature maps into an ensemble architecture. Experimental results on the CIFAR-10 dataset show that using the proposed method in the ensemble model results in performance increases of 9.10 and 8.42% for the B and R channels, respectively, compared to the original model, while the effect of the G channel is very limited. For the CIFAR-100 dataset, the proposed method resulted in a 17.09% improvement in ensemble model performance for the R channel, a 5.06% increase for the B channel, and no significant improvement for the G channel compared to the original model. Additionally, we compared our method with traditional feature extraction methods like scale-invariant feature transform and local binary pattern and observed higher performance. In conclusion, it has been observed that the proposed RGB angle rotation method significantly impacts model performance.</p>\",\"PeriodicalId\":18925,\"journal\":{\"name\":\"Neural Computing and Applications\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Computing and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/s00521-024-10232-z\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Computing and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s00521-024-10232-z","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
近年来,卷积神经网络通过从图像数据中自动提取特征,极大地推动了计算机视觉领域的发展。卷积神经网络能够利用可学习的滤波器对复杂抽象的图像特征进行建模,从而消除了人工特征提取的需要。然而,将从 CNN 中获得的特征图与不同的方法相结合,可以得到更复杂、更可解释的推论,从而提高模型的性能和普适性。在本研究中,我们提出了一种名为 RGB 角度旋转的新方法,以有效地从 RGB 图像中获取特征图。我们的方法以不同的角度旋转颜色通道,并利用通道间的角度信息生成新的特征图。然后,我们研究了将使用这些特征图训练的模型集成到集合架构中的效果。在 CIFAR-10 数据集上的实验结果表明,与原始模型相比,在集合模型中使用所提出的方法,B 和 R 信道的性能分别提高了 9.10% 和 8.42%,而 G 信道的影响则非常有限。对于 CIFAR-100 数据集,与原始模型相比,建议方法使 R 信道的集合模型性能提高了 17.09%,B 信道提高了 5.06%,而 G 信道没有显著提高。此外,我们还将我们的方法与传统的特征提取方法(如尺度不变特征变换和局部二进制模式)进行了比较,发现我们的方法性能更高。总之,我们发现所提出的 RGB 角度旋转方法对模型性能有显著影响。
Enhancing CNN model classification performance through RGB angle rotation method
In recent years, convolutional neural networks have significantly advanced the field of computer vision by automatically extracting features from image data. CNNs enable the modeling of complex and abstract image features using learnable filters, eliminating the need for manual feature extraction. However, combining feature maps obtained from CNNs with different approaches can lead to more complex and interpretable inferences, thereby enhancing model performance and generalizability. In this study, we propose a new method called RGB angle rotation to effectively obtain feature maps from RGB images. Our method rotates color channels at different angles and uses the angle information between channels to generate new feature maps. We then investigate the effects of integrating models trained with these feature maps into an ensemble architecture. Experimental results on the CIFAR-10 dataset show that using the proposed method in the ensemble model results in performance increases of 9.10 and 8.42% for the B and R channels, respectively, compared to the original model, while the effect of the G channel is very limited. For the CIFAR-100 dataset, the proposed method resulted in a 17.09% improvement in ensemble model performance for the R channel, a 5.06% increase for the B channel, and no significant improvement for the G channel compared to the original model. Additionally, we compared our method with traditional feature extraction methods like scale-invariant feature transform and local binary pattern and observed higher performance. In conclusion, it has been observed that the proposed RGB angle rotation method significantly impacts model performance.