裁剪或填充输入量:什么对卷积神经网络有益?

U. M. Al-Saggaf, Abdelaziz Botalb, M. Moinuddin, S. Alfakeh, Syed Saad Azhar Ali, Tang Tong Boon
{"title":"裁剪或填充输入量:什么对卷积神经网络有益?","authors":"U. M. Al-Saggaf, Abdelaziz Botalb, M. Moinuddin, S. Alfakeh, Syed Saad Azhar Ali, Tang Tong Boon","doi":"10.1109/ICIAS49414.2021.9642661","DOIUrl":null,"url":null,"abstract":"Convolutional Neural Network (CNN) is the most popular method of deep learning in the machine learning field. Training a CNN has always been a demanding task compared to other machine learning paradigms, and this is due to its big space of hyper-parameters such as convolutional kernel size, number of strides, number of layers, pooling window size, etc. What makes the CNN’s huge hyper-parameters space optimization harder is that there is no universal robust theory supporting it, and any work flow proposed so far in literature is based on heuristics that are just rules of thumb and only depend on the dataset and problem at hand. In this work, it is empirically illustrated that the performance of a CNN is not linked only with the choice of the right hyper-parameters, but also linked to how some of the CNN operations are implemented. More specifically, the CNN performance is contrasted for two different implementations: cropping and padding the input volume. The results state that padding the input volume achieves higher accuracy and takes less time in training compared with cropping method.","PeriodicalId":212635,"journal":{"name":"2020 8th International Conference on Intelligent and Advanced Systems (ICIAS)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Either crop or pad the input volume: What is beneficial for Convolutional Neural Network?\",\"authors\":\"U. M. Al-Saggaf, Abdelaziz Botalb, M. Moinuddin, S. Alfakeh, Syed Saad Azhar Ali, Tang Tong Boon\",\"doi\":\"10.1109/ICIAS49414.2021.9642661\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Convolutional Neural Network (CNN) is the most popular method of deep learning in the machine learning field. Training a CNN has always been a demanding task compared to other machine learning paradigms, and this is due to its big space of hyper-parameters such as convolutional kernel size, number of strides, number of layers, pooling window size, etc. What makes the CNN’s huge hyper-parameters space optimization harder is that there is no universal robust theory supporting it, and any work flow proposed so far in literature is based on heuristics that are just rules of thumb and only depend on the dataset and problem at hand. In this work, it is empirically illustrated that the performance of a CNN is not linked only with the choice of the right hyper-parameters, but also linked to how some of the CNN operations are implemented. More specifically, the CNN performance is contrasted for two different implementations: cropping and padding the input volume. The results state that padding the input volume achieves higher accuracy and takes less time in training compared with cropping method.\",\"PeriodicalId\":212635,\"journal\":{\"name\":\"2020 8th International Conference on Intelligent and Advanced Systems (ICIAS)\",\"volume\":\"2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 8th International Conference on Intelligent and Advanced Systems (ICIAS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICIAS49414.2021.9642661\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 8th International Conference on Intelligent and Advanced Systems (ICIAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIAS49414.2021.9642661","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

卷积神经网络(CNN)是机器学习领域最流行的深度学习方法。与其他机器学习范例相比,训练CNN一直是一项艰巨的任务,这是由于它的超参数空间很大,如卷积核大小、跨步数量、层数、池化窗口大小等。CNN巨大的超参数空间优化之所以更加困难,是因为没有通用的强大理论支持它,而且迄今为止,文献中提出的任何工作流程都是基于经验法则的启发式,仅依赖于数据集和手头的问题。在这项工作中,经验表明,CNN的性能不仅与正确超参数的选择有关,而且还与一些CNN操作的实现方式有关。更具体地说,CNN的性能对比了两种不同的实现:裁剪和填充输入体积。结果表明,与裁剪方法相比,填充输入体积具有更高的准确率和更少的训练时间。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Either crop or pad the input volume: What is beneficial for Convolutional Neural Network?
Convolutional Neural Network (CNN) is the most popular method of deep learning in the machine learning field. Training a CNN has always been a demanding task compared to other machine learning paradigms, and this is due to its big space of hyper-parameters such as convolutional kernel size, number of strides, number of layers, pooling window size, etc. What makes the CNN’s huge hyper-parameters space optimization harder is that there is no universal robust theory supporting it, and any work flow proposed so far in literature is based on heuristics that are just rules of thumb and only depend on the dataset and problem at hand. In this work, it is empirically illustrated that the performance of a CNN is not linked only with the choice of the right hyper-parameters, but also linked to how some of the CNN operations are implemented. More specifically, the CNN performance is contrasted for two different implementations: cropping and padding the input volume. The results state that padding the input volume achieves higher accuracy and takes less time in training compared with cropping method.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Prospects and Techniques of Regenerative Current Breaking in DC Circuit Breaker Topology Fractional Stochastic Gradient Descent Based Learning Algorithm For Multi-layer Perceptron Neural Networks Domestic Electrical Energy Monitoring and Alerting Using SCADA and IoT Stochastic Approach for the Identification of Retinopathy of Prematurity Fault Classification and Location in Three-Phase Transmission Lines Using Wavelet-based Machine Learning
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1