{"title":"基于计算机视觉的深度神经网络性能分析","authors":"Nidhi Sindhwani, Rohit Anand, M. S, Rati Shukla, Mahendra Pratap Yadav, Vikash Yadav","doi":"10.4108/eai.13-10-2021.171318","DOIUrl":null,"url":null,"abstract":"INTRODUCTION: In recent years, deep learning techniques have been made to outperform the earlier state-of-the-art machine learning techniques in many areas, with one of the most notable cases being computer vision. Deep learning is also employed to train the neural networks with the images and to perform the various tasks such as classification and segmentation using several different models. The size and depth of current deep learning models have increased to solve certain tasks as these models provide better accuracy. As pre-trained weights may be used for further training and prevent costly computing, transfer learning is therefore of vital importance. A brief account is given of their history, structure, benefits, and drawbacks, followed by a description of their applications in the different tasks of computer vision, such as object detection, face recognition etc. OBJECTIVE:. The purpose of this paper is to train a deep neural network to properly classify the images that it has never seen before, define techniques to enhance the efficiency of deep learning and deploy deep neural networks in various applications. METHOD: The proposed approach represents that after the reading of images, 256x256 pixel image’s random parts are extracted and noise, distortion, flip, or rotation transforms are applied. Multiple convolution and pooling steps are applied by controlling the stride lengths. RESULT: Data analysis and research findings showed that DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. The proposed work outperforms the previous techniques in predicting the dependent variables, learning rate, image count, image mean, performance analysis of loss rate and learning rate during training, performance Analysis of Loss with respect to Epoch for Training, Validation and Accuracy. CONCLUSION: This research encompasses a large variety of computer applications, from image recognition and machine translation to enhanced learning. DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. Extensive research has been conducted using the various deep architectures such as AlexNet, InceptionNet, etc. To the best of authors’ knowledge, this is the first work that presents a quantitative analysis of the deep architectures mentioned above.","PeriodicalId":33474,"journal":{"name":"EAI Endorsed Transactions on Industrial Networks and Intelligent Systems","volume":"17 1","pages":"e3"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"Performance Analysis of Deep Neural Networks Using Computer Vision\",\"authors\":\"Nidhi Sindhwani, Rohit Anand, M. S, Rati Shukla, Mahendra Pratap Yadav, Vikash Yadav\",\"doi\":\"10.4108/eai.13-10-2021.171318\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"INTRODUCTION: In recent years, deep learning techniques have been made to outperform the earlier state-of-the-art machine learning techniques in many areas, with one of the most notable cases being computer vision. Deep learning is also employed to train the neural networks with the images and to perform the various tasks such as classification and segmentation using several different models. The size and depth of current deep learning models have increased to solve certain tasks as these models provide better accuracy. As pre-trained weights may be used for further training and prevent costly computing, transfer learning is therefore of vital importance. A brief account is given of their history, structure, benefits, and drawbacks, followed by a description of their applications in the different tasks of computer vision, such as object detection, face recognition etc. OBJECTIVE:. The purpose of this paper is to train a deep neural network to properly classify the images that it has never seen before, define techniques to enhance the efficiency of deep learning and deploy deep neural networks in various applications. METHOD: The proposed approach represents that after the reading of images, 256x256 pixel image’s random parts are extracted and noise, distortion, flip, or rotation transforms are applied. Multiple convolution and pooling steps are applied by controlling the stride lengths. RESULT: Data analysis and research findings showed that DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. The proposed work outperforms the previous techniques in predicting the dependent variables, learning rate, image count, image mean, performance analysis of loss rate and learning rate during training, performance Analysis of Loss with respect to Epoch for Training, Validation and Accuracy. CONCLUSION: This research encompasses a large variety of computer applications, from image recognition and machine translation to enhanced learning. DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. Extensive research has been conducted using the various deep architectures such as AlexNet, InceptionNet, etc. To the best of authors’ knowledge, this is the first work that presents a quantitative analysis of the deep architectures mentioned above.\",\"PeriodicalId\":33474,\"journal\":{\"name\":\"EAI Endorsed Transactions on Industrial Networks and Intelligent Systems\",\"volume\":\"17 1\",\"pages\":\"e3\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"EAI Endorsed Transactions on Industrial Networks and Intelligent Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.4108/eai.13-10-2021.171318\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"Engineering\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"EAI Endorsed Transactions on Industrial Networks and Intelligent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4108/eai.13-10-2021.171318","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Engineering","Score":null,"Total":0}
引用次数: 8
摘要
近年来,深度学习技术已经在许多领域超越了早期最先进的机器学习技术,其中最著名的案例之一是计算机视觉。深度学习也被用于训练神经网络与图像,并执行各种任务,如分类和分割使用几个不同的模型。当前深度学习模型的规模和深度已经增加,以解决某些任务,因为这些模型提供了更好的准确性。由于预训练的权重可以用于进一步的训练,并且可以避免昂贵的计算,因此迁移学习至关重要。简要介绍了它们的历史、结构、优点和缺点,然后描述了它们在计算机视觉的不同任务中的应用,如物体检测、人脸识别等。目的:。本文的目的是训练深度神经网络对从未见过的图像进行正确分类,定义提高深度学习效率的技术,并将深度神经网络部署在各种应用中。方法:该方法是在读取图像后,提取256x256像素图像的随机部分,并对其进行噪声、失真、翻转或旋转变换。通过控制步长,应用了多个卷积和池化步骤。结果:数据分析和研究结果表明,DNN模型已经在三种主要的深度学习配置中实现:CNTK、MXNet和TensorFlow。所提出的工作在预测因变量、学习率、图像计数、图像均值、训练期间损失率和学习率的性能分析、loss相对于Epoch for training、Validation和Accuracy的性能分析方面优于先前的技术。结论:这项研究涵盖了大量的计算机应用,从图像识别和机器翻译到增强学习。DNN模型已经在深度学习的三种主要配置中实现:CNTK, MXNet和TensorFlow。使用各种深度架构(如AlexNet, InceptionNet等)进行了广泛的研究。据作者所知,这是第一本对上述深度架构进行定量分析的著作。
Performance Analysis of Deep Neural Networks Using Computer Vision
INTRODUCTION: In recent years, deep learning techniques have been made to outperform the earlier state-of-the-art machine learning techniques in many areas, with one of the most notable cases being computer vision. Deep learning is also employed to train the neural networks with the images and to perform the various tasks such as classification and segmentation using several different models. The size and depth of current deep learning models have increased to solve certain tasks as these models provide better accuracy. As pre-trained weights may be used for further training and prevent costly computing, transfer learning is therefore of vital importance. A brief account is given of their history, structure, benefits, and drawbacks, followed by a description of their applications in the different tasks of computer vision, such as object detection, face recognition etc. OBJECTIVE:. The purpose of this paper is to train a deep neural network to properly classify the images that it has never seen before, define techniques to enhance the efficiency of deep learning and deploy deep neural networks in various applications. METHOD: The proposed approach represents that after the reading of images, 256x256 pixel image’s random parts are extracted and noise, distortion, flip, or rotation transforms are applied. Multiple convolution and pooling steps are applied by controlling the stride lengths. RESULT: Data analysis and research findings showed that DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. The proposed work outperforms the previous techniques in predicting the dependent variables, learning rate, image count, image mean, performance analysis of loss rate and learning rate during training, performance Analysis of Loss with respect to Epoch for Training, Validation and Accuracy. CONCLUSION: This research encompasses a large variety of computer applications, from image recognition and machine translation to enhanced learning. DNN models have been implemented in three main configurations of deep learning: CNTK, MXNet and TensorFlow. Extensive research has been conducted using the various deep architectures such as AlexNet, InceptionNet, etc. To the best of authors’ knowledge, this is the first work that presents a quantitative analysis of the deep architectures mentioned above.