Crowd Counting via Enhanced Feature Channel Convolutional Neural Network

Yinlong Bian, Jiehong Shen, Xin Xiong, Ying Li, Wei-Ji He, Peng Li
{"title":"Crowd Counting via Enhanced Feature Channel Convolutional Neural Network","authors":"Yinlong Bian, Jiehong Shen, Xin Xiong, Ying Li, Wei-Ji He, Peng Li","doi":"10.1109/ICTAI.2019.00118","DOIUrl":null,"url":null,"abstract":"Accurate crowd counting is important for interpreting and understanding crowd, which has great practical significance in video monitoring, public safety, urban planning, the construction of intelligent shopping malls and so on. For accurate counting, many excellent algorithms have been proposed, but there are still some challenges in terms of scale variation, occlusion, inaccurate counting in various backgrounds and so forth. In this paper, we propose a new model EFCCNN (Enhanced Feature Channel Convolutional Neural Network) to deal with these challenges. The proposed EFCCNN model has three main contributions. We propose a new convolutional neural network, which can be trained by end-to-end, and it performs better than other crowd counting networks. Additionally, we use SENet (Squeeze-and-Excitation Network) structure to change the channel weight, which can enhance the significant channel, and we use residual structure to transmit the channel weight to improve the counting precision. The SENet structure is helpful to solve the problem of scale variation and occlusion. The EFCCNN model is the first crowd counting model using channel weight information. Furthermore, a new loss function focusing on the structural information of images is proposed, which reduces the mean absolute error of crowd counting, effectively solves the problem of inaccurate counting in various backgrounds, such as crowd miscounting in the tree and brush background, and improves the quality of the crowd density map on SSIM (Structural Similarity Index Measure). Experiments on ShanghaiTech, Mall, UCF_CC_50 dataset show that EFCCNN have a lower mean absolute error of crowd counting and a higher quality density map.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI.2019.00118","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Accurate crowd counting is important for interpreting and understanding crowd, which has great practical significance in video monitoring, public safety, urban planning, the construction of intelligent shopping malls and so on. For accurate counting, many excellent algorithms have been proposed, but there are still some challenges in terms of scale variation, occlusion, inaccurate counting in various backgrounds and so forth. In this paper, we propose a new model EFCCNN (Enhanced Feature Channel Convolutional Neural Network) to deal with these challenges. The proposed EFCCNN model has three main contributions. We propose a new convolutional neural network, which can be trained by end-to-end, and it performs better than other crowd counting networks. Additionally, we use SENet (Squeeze-and-Excitation Network) structure to change the channel weight, which can enhance the significant channel, and we use residual structure to transmit the channel weight to improve the counting precision. The SENet structure is helpful to solve the problem of scale variation and occlusion. The EFCCNN model is the first crowd counting model using channel weight information. Furthermore, a new loss function focusing on the structural information of images is proposed, which reduces the mean absolute error of crowd counting, effectively solves the problem of inaccurate counting in various backgrounds, such as crowd miscounting in the tree and brush background, and improves the quality of the crowd density map on SSIM (Structural Similarity Index Measure). Experiments on ShanghaiTech, Mall, UCF_CC_50 dataset show that EFCCNN have a lower mean absolute error of crowd counting and a higher quality density map.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于增强特征通道卷积神经网络的人群计数
准确的人群统计对于解读和了解人群具有重要意义,在视频监控、公共安全、城市规划、智能商场建设等方面都具有重要的现实意义。为了精确计数,已经提出了许多优秀的算法,但在尺度变化、遮挡、不同背景下计数不准确等方面仍然存在一些挑战。在本文中,我们提出了一个新的模型EFCCNN (Enhanced Feature Channel Convolutional Neural Network,增强特征通道卷积神经网络)来应对这些挑战。提出的EFCCNN模型有三个主要贡献。我们提出了一种新的卷积神经网络,它可以端到端训练,并且性能优于其他人群计数网络。采用SENet (Squeeze-and-Excitation Network)结构改变信道权值,增强有效信道;采用残差结构传递信道权值,提高计数精度。SENet结构有助于解决尺度变化和遮挡问题。EFCCNN模型是第一个使用信道权重信息的人群计数模型。在此基础上,提出了一种新的以图像结构信息为中心的损失函数,降低了人群计数的平均绝对误差,有效解决了树形背景和刷形背景下人群计数不准确的问题,提高了基于SSIM (structural Similarity Index Measure)的人群密度图的质量。在上海科技、Mall、UCF_CC_50数据集上的实验表明,EFCCNN具有较低的人群统计平均绝对误差和较高质量的密度图。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Monaural Music Source Separation using a ResNet Latent Separator Network Graph-Based Attention Networks for Aspect Level Sentiment Analysis A Multi-channel Neural Network for Imbalanced Emotion Recognition Scaling up Prediction of Psychosis by Natural Language Processing Improving Bandit-Based Recommendations with Spatial Context Reasoning: An Online Evaluation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1