Minghui Yao , Dongzhou Cheng , Lei Zhang , LiangDong Liu , Shuangteng Song , Hao Wu , Aiguo Song
{"title":"A sparse diverse-branch large kernel convolutional neural network for human activity recognition using wearables","authors":"Minghui Yao , Dongzhou Cheng , Lei Zhang , LiangDong Liu , Shuangteng Song , Hao Wu , Aiguo Song","doi":"10.1016/j.asoc.2024.112444","DOIUrl":null,"url":null,"abstract":"<div><div>During the past decade, large convolutional kernels have long been under the shadow of small convolutional kernels since the introduction of VGG backbone network. It always remains mysterious whether one can design pure convolutional neural network (CNN) while plugging larger kernels to model long-range dependency for human activity recognition (HAR), which has been rarely explored in previous literatures. In this paper, we revive the usage of larger kernels in the context of HAR and attempt to eliminate the performance gap between large kernels and small kernels by strategically applying a large receptive field, without incurring high memory and computational footprints. Built on two recipes, i.e., Diverse-Branch and Dynamic Sparsity, we design a pure CNN architecture named SLK-Net for activity recognition, which is equipped with sparse diverse-branch larger kernels. To validate the effectiveness of our approach, we perform a series of extensive experiments on four public benchmarks including UCI-HAR, WISDM, UniMiB-SHAR and USC-HAD, which show that large kernels can benefit its ability to capture long-range dependency and consistently beat state-of-the-art small-kernel counterparts across a wide range of activity classification tasks. Real activity inference latency is measured on a mobile device, which reveals that such sparse diverse-branch kernels can lead to inference speedup than vanilla large kernels. We hope this work may further inspire relevant CNN-based studies in the HAR community.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"167 ","pages":"Article 112444"},"PeriodicalIF":7.2000,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494624012183","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
During the past decade, large convolutional kernels have long been under the shadow of small convolutional kernels since the introduction of VGG backbone network. It always remains mysterious whether one can design pure convolutional neural network (CNN) while plugging larger kernels to model long-range dependency for human activity recognition (HAR), which has been rarely explored in previous literatures. In this paper, we revive the usage of larger kernels in the context of HAR and attempt to eliminate the performance gap between large kernels and small kernels by strategically applying a large receptive field, without incurring high memory and computational footprints. Built on two recipes, i.e., Diverse-Branch and Dynamic Sparsity, we design a pure CNN architecture named SLK-Net for activity recognition, which is equipped with sparse diverse-branch larger kernels. To validate the effectiveness of our approach, we perform a series of extensive experiments on four public benchmarks including UCI-HAR, WISDM, UniMiB-SHAR and USC-HAD, which show that large kernels can benefit its ability to capture long-range dependency and consistently beat state-of-the-art small-kernel counterparts across a wide range of activity classification tasks. Real activity inference latency is measured on a mobile device, which reveals that such sparse diverse-branch kernels can lead to inference speedup than vanilla large kernels. We hope this work may further inspire relevant CNN-based studies in the HAR community.
期刊介绍:
Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities.
Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.