David Kornish, Soundararajan Ezekiel, Maria Scalzo-Cornacchia
{"title":"Fusion based Heterogeneous Convolutional Neural Networks Architecture","authors":"David Kornish, Soundararajan Ezekiel, Maria Scalzo-Cornacchia","doi":"10.1109/AIPR.2018.8707371","DOIUrl":null,"url":null,"abstract":"In recent years, Deep Convolutional Neural Networks (DCNNs) have gained lots of attention and won many competitions in machine learning, object detection, image classification, and pattern recognition. The breakthroughs in the development of graphical processing units have made it possible to train DCNNs quickly for state-of-the-art tasks such as image classification, speech recognition, and many others. However, to solve complex problems, these multilayered convolutional neural networks become increasingly large, complex, and abstract. We propose methods to improve the performance of neural networks while reducing their dimensionality, enabling a better understanding of the learning process. To leverage the extensive training, as well as strengths of several pretrained models, we explored new approaches for combining features from fully connected layers of models with heterogeneous architectures. The proposed approach combines features extracted from the penultimate fully connected layer from three different DCNNs. We merge the features of all three DCNNs together and apply principal component analysis or linear discriminant analysis. Our approach aims to reduce the dimensionality of the feature vector and find the smallest feature vector dimension that can maintain the classifier performance. For this task we use a linear Support Vector Machine as a classifier. We also investigate whether it is advantageous to fuse only penultimate fully connected layers, or to perform fusion based on other fully connected layers using multiple homogenous or heterogeneous networks. The results show that the fusion method outperformed both individual networks in terms of accuracy and computational time in all of our various trial sizes. Overall our fusion methods are faster and more accurate than individual networks in both training and testing. Finally, we compared heterogeneous with homogenous fusion methods and the results show heterogeneous methods outperform homogeneous methods.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AIPR.2018.8707371","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
In recent years, Deep Convolutional Neural Networks (DCNNs) have gained lots of attention and won many competitions in machine learning, object detection, image classification, and pattern recognition. The breakthroughs in the development of graphical processing units have made it possible to train DCNNs quickly for state-of-the-art tasks such as image classification, speech recognition, and many others. However, to solve complex problems, these multilayered convolutional neural networks become increasingly large, complex, and abstract. We propose methods to improve the performance of neural networks while reducing their dimensionality, enabling a better understanding of the learning process. To leverage the extensive training, as well as strengths of several pretrained models, we explored new approaches for combining features from fully connected layers of models with heterogeneous architectures. The proposed approach combines features extracted from the penultimate fully connected layer from three different DCNNs. We merge the features of all three DCNNs together and apply principal component analysis or linear discriminant analysis. Our approach aims to reduce the dimensionality of the feature vector and find the smallest feature vector dimension that can maintain the classifier performance. For this task we use a linear Support Vector Machine as a classifier. We also investigate whether it is advantageous to fuse only penultimate fully connected layers, or to perform fusion based on other fully connected layers using multiple homogenous or heterogeneous networks. The results show that the fusion method outperformed both individual networks in terms of accuracy and computational time in all of our various trial sizes. Overall our fusion methods are faster and more accurate than individual networks in both training and testing. Finally, we compared heterogeneous with homogenous fusion methods and the results show heterogeneous methods outperform homogeneous methods.