{"title":"用代数表示法加快卷积神经网络的预测速度","authors":"Johnny Joyce, Jan Verschelde","doi":"arxiv-2408.07815","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks (CNNs) are a popular choice of model for tasks\nin computer vision. When CNNs are made with many layers, resulting in a deep\nneural network, skip connections may be added to create an easier gradient\noptimization problem while retaining model expressiveness. In this paper, we\nshow that arbitrarily complex, trained, linear CNNs with skip connections can\nbe simplified into a single-layer model, resulting in greatly reduced\ncomputational requirements during prediction time. We also present a method for\ntraining nonlinear models with skip connections that are gradually removed\nthroughout training, giving the benefits of skip connections without requiring\ncomputational overhead during during prediction time. These results are\ndemonstrated with practical examples on Residual Networks (ResNet)\narchitecture.","PeriodicalId":501033,"journal":{"name":"arXiv - CS - Symbolic Computation","volume":"7 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Algebraic Representations for Faster Predictions in Convolutional Neural Networks\",\"authors\":\"Johnny Joyce, Jan Verschelde\",\"doi\":\"arxiv-2408.07815\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Convolutional neural networks (CNNs) are a popular choice of model for tasks\\nin computer vision. When CNNs are made with many layers, resulting in a deep\\nneural network, skip connections may be added to create an easier gradient\\noptimization problem while retaining model expressiveness. In this paper, we\\nshow that arbitrarily complex, trained, linear CNNs with skip connections can\\nbe simplified into a single-layer model, resulting in greatly reduced\\ncomputational requirements during prediction time. We also present a method for\\ntraining nonlinear models with skip connections that are gradually removed\\nthroughout training, giving the benefits of skip connections without requiring\\ncomputational overhead during during prediction time. These results are\\ndemonstrated with practical examples on Residual Networks (ResNet)\\narchitecture.\",\"PeriodicalId\":501033,\"journal\":{\"name\":\"arXiv - CS - Symbolic Computation\",\"volume\":\"7 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Symbolic Computation\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.07815\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Symbolic Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.07815","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Algebraic Representations for Faster Predictions in Convolutional Neural Networks
Convolutional neural networks (CNNs) are a popular choice of model for tasks
in computer vision. When CNNs are made with many layers, resulting in a deep
neural network, skip connections may be added to create an easier gradient
optimization problem while retaining model expressiveness. In this paper, we
show that arbitrarily complex, trained, linear CNNs with skip connections can
be simplified into a single-layer model, resulting in greatly reduced
computational requirements during prediction time. We also present a method for
training nonlinear models with skip connections that are gradually removed
throughout training, giving the benefits of skip connections without requiring
computational overhead during during prediction time. These results are
demonstrated with practical examples on Residual Networks (ResNet)
architecture.