首页 > 最新文献

International Conference on Computer Vision and Graphics最新文献

英文 中文
Energy Efficient Hardware Acceleration of Neural Networks with Power-of-Two Quantisation 二次幂量化神经网络的高能效硬件加速
Pub Date : 2022-09-30 DOI: 10.48550/arXiv.2209.15257
Dominika Przewlocka-Rus, T. Kryjak
Deep neural networks virtually dominate the domain of most modern vision systems, providing high performance at a cost of increased computational complexity.Since for those systems it is often required to operate both in real-time and with minimal energy consumption (e.g., for wearable devices or autonomous vehicles, edge Internet of Things (IoT), sensor networks), various network optimisation techniques are used, e.g., quantisation, pruning, or dedicated lightweight architectures. Due to the logarithmic distribution of weights in neural network layers, a method providing high performance with significant reduction in computational precision (for 4-bit weights and less) is the Power-of-Two (PoT) quantisation (and therefore also with a logarithmic distribution). This method introduces additional possibilities of replacing the typical for neural networks Multiply and ACcumulate (MAC -- performing, e.g., convolution operations) units, with more energy-efficient Bitshift and ACcumulate (BAC). In this paper, we show that a hardware neural network accelerator with PoT weights implemented on the Zynq UltraScale + MPSoC ZCU104 SoC FPGA can be at least $1.4x$ more energy efficient than the uniform quantisation version. To further reduce the actual power requirement by omitting part of the computation for zero weights, we also propose a new pruning method adapted to logarithmic quantisation.
深度神经网络几乎主导了大多数现代视觉系统的领域,以增加计算复杂性为代价提供高性能。由于这些系统通常需要实时运行并以最小的能耗运行(例如,可穿戴设备或自动驾驶汽车、边缘物联网(IoT)、传感器网络),因此使用了各种网络优化技术,例如量化、修剪或专用轻量级架构。由于神经网络层中权重的对数分布,一种提供高性能但计算精度显著降低(对于4位及以下权重)的方法是2的幂(PoT)量化(因此也具有对数分布)。这种方法引入了额外的可能性,用更节能的Bitshift和ACcumulate (BAC)取代神经网络中典型的Multiply和ACcumulate (MAC -执行,例如卷积操作)单元。在本文中,我们证明了在Zynq UltraScale + MPSoC ZCU104 SoC FPGA上实现的具有PoT权重的硬件神经网络加速器可以比统一量化版本至少节省1.4美元的能源效率。为了进一步降低实际功率需求,我们还提出了一种新的适合对数量化的剪枝方法。
{"title":"Energy Efficient Hardware Acceleration of Neural Networks with Power-of-Two Quantisation","authors":"Dominika Przewlocka-Rus, T. Kryjak","doi":"10.48550/arXiv.2209.15257","DOIUrl":"https://doi.org/10.48550/arXiv.2209.15257","url":null,"abstract":"Deep neural networks virtually dominate the domain of most modern vision systems, providing high performance at a cost of increased computational complexity.Since for those systems it is often required to operate both in real-time and with minimal energy consumption (e.g., for wearable devices or autonomous vehicles, edge Internet of Things (IoT), sensor networks), various network optimisation techniques are used, e.g., quantisation, pruning, or dedicated lightweight architectures. Due to the logarithmic distribution of weights in neural network layers, a method providing high performance with significant reduction in computational precision (for 4-bit weights and less) is the Power-of-Two (PoT) quantisation (and therefore also with a logarithmic distribution). This method introduces additional possibilities of replacing the typical for neural networks Multiply and ACcumulate (MAC -- performing, e.g., convolution operations) units, with more energy-efficient Bitshift and ACcumulate (BAC). In this paper, we show that a hardware neural network accelerator with PoT weights implemented on the Zynq UltraScale + MPSoC ZCU104 SoC FPGA can be at least $1.4x$ more energy efficient than the uniform quantisation version. To further reduce the actual power requirement by omitting part of the computation for zero weights, we also propose a new pruning method adapted to logarithmic quantisation.","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130712387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
PointPillars Backbone Type Selection For Fast and Accurate LiDAR Object Detection 快速准确的激光雷达目标检测点柱主干网类型选择
Pub Date : 2022-09-30 DOI: 10.48550/arXiv.2209.15252
K. Lis, T. Kryjak
3D object detection from LiDAR sensor data is an important topic in the context of autonomous cars and drones. In this paper, we present the results of experiments on the impact of backbone selection of a deep convolutional neural network on detection accuracy and computation speed. We chose the PointPillars network, which is characterised by a simple architecture, high speed, and modularity that allows for easy expansion. During the experiments, we paid particular attention to the change in detection efficiency (measured by the mAP metric) and the total number of multiply-addition operations needed to process one point cloud. We tested 10 different convolutional neural network architectures that are widely used in image-based detection problems. For a backbone like MobilenetV1, we obtained an almost 4x speedup at the cost of a 1.13% decrease in mAP. On the other hand, for CSPDarknet we got an acceleration of more than 1.5x at an increase in mAP of 0.33%. We have thus demonstrated that it is possible to significantly speed up a 3D object detector in LiDAR point clouds with a small decrease in detection efficiency. This result can be used when PointPillars or similar algorithms are implemented in embedded systems, including SoC FPGAs. The code is available at https://github.com/vision-agh/pointpillars_backbone.
在自动驾驶汽车和无人机的背景下,从激光雷达传感器数据中检测三维物体是一个重要的课题。本文给出了深度卷积神经网络主干选择对检测精度和计算速度影响的实验结果。我们选择了PointPillars网络,其特点是架构简单,速度快,模块化,易于扩展。在实验过程中,我们特别关注了检测效率的变化(通过mAP度量来衡量)和处理一个点云所需的乘法加法操作的总数。我们测试了10种不同的卷积神经网络架构,这些架构广泛用于基于图像的检测问题。对于像MobilenetV1这样的主干网,我们以降低1.13%的mAP为代价获得了近4倍的加速。另一方面,对于CSPDarknet,我们在mAP增加0.33%的情况下获得了超过1.5倍的加速。因此,我们已经证明了在激光雷达点云中显著加速3D物体探测器而检测效率略有下降是可能的。该结果可用于PointPillars或类似算法在嵌入式系统中实现,包括SoC fpga。该代码可从https://github.com/vision-agh/pointpillars_backbone获得。
{"title":"PointPillars Backbone Type Selection For Fast and Accurate LiDAR Object Detection","authors":"K. Lis, T. Kryjak","doi":"10.48550/arXiv.2209.15252","DOIUrl":"https://doi.org/10.48550/arXiv.2209.15252","url":null,"abstract":"3D object detection from LiDAR sensor data is an important topic in the context of autonomous cars and drones. In this paper, we present the results of experiments on the impact of backbone selection of a deep convolutional neural network on detection accuracy and computation speed. We chose the PointPillars network, which is characterised by a simple architecture, high speed, and modularity that allows for easy expansion. During the experiments, we paid particular attention to the change in detection efficiency (measured by the mAP metric) and the total number of multiply-addition operations needed to process one point cloud. We tested 10 different convolutional neural network architectures that are widely used in image-based detection problems. For a backbone like MobilenetV1, we obtained an almost 4x speedup at the cost of a 1.13% decrease in mAP. On the other hand, for CSPDarknet we got an acceleration of more than 1.5x at an increase in mAP of 0.33%. We have thus demonstrated that it is possible to significantly speed up a 3D object detector in LiDAR point clouds with a small decrease in detection efficiency. This result can be used when PointPillars or similar algorithms are implemented in embedded systems, including SoC FPGAs. The code is available at https://github.com/vision-agh/pointpillars_backbone.","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"129 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124020470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Traffic Sign Classification Using Deep and Quantum Neural Networks 基于深度和量子神经网络的交通标志分类
Pub Date : 2022-09-30 DOI: 10.48550/arXiv.2209.15251
Sylwia Kuros, T. Kryjak
Quantum Neural Networks (QNNs) are an emerging technology that can be used in many applications including computer vision. In this paper, we presented a traffic sign classification system implemented using a hybrid quantum-classical convolutional neural network. Experiments on the German Traffic Sign Recognition Benchmark dataset indicate that currently QNN do not outperform classical DCNN (Deep Convolutuional Neural Networks), yet still provide an accuracy of over 90% and are a definitely promising solution for advanced computer vision.
量子神经网络(QNNs)是一项新兴技术,可用于包括计算机视觉在内的许多应用。本文提出了一种基于量子-经典混合卷积神经网络的交通标志分类系统。在德国交通标志识别基准数据集上的实验表明,目前QNN的性能并不优于经典的DCNN(深度卷积神经网络),但仍然提供超过90%的准确率,是一种非常有前途的高级计算机视觉解决方案。
{"title":"Traffic Sign Classification Using Deep and Quantum Neural Networks","authors":"Sylwia Kuros, T. Kryjak","doi":"10.48550/arXiv.2209.15251","DOIUrl":"https://doi.org/10.48550/arXiv.2209.15251","url":null,"abstract":"Quantum Neural Networks (QNNs) are an emerging technology that can be used in many applications including computer vision. In this paper, we presented a traffic sign classification system implemented using a hybrid quantum-classical convolutional neural network. Experiments on the German Traffic Sign Recognition Benchmark dataset indicate that currently QNN do not outperform classical DCNN (Deep Convolutuional Neural Networks), yet still provide an accuracy of over 90% and are a definitely promising solution for advanced computer vision.","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"354 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124474801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Scene Recognition Using AlexNet to Recognize Significant Events Within Cricket Game Footage 场景识别使用AlexNet识别板球比赛镜头中的重要事件
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_9
Tevin Moodley, D. Haar
{"title":"Scene Recognition Using AlexNet to Recognize Significant Events Within Cricket Game Footage","authors":"Tevin Moodley, D. Haar","doi":"10.1007/978-3-030-59006-2_9","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_9","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125324432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A New Watermarking Method for Video Authentication with Tamper Localization 一种基于篡改定位的视频认证水印方法
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_18
Y. Vybornova
{"title":"A New Watermarking Method for Video Authentication with Tamper Localization","authors":"Y. Vybornova","doi":"10.1007/978-3-030-59006-2_18","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_18","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"301 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122800806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
RGB-D and Lidar Calibration Supported by GPU GPU支持RGB-D和激光雷达校准
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_19
A. Wilkowski, D. Mańkowski
{"title":"RGB-D and Lidar Calibration Supported by GPU","authors":"A. Wilkowski, D. Mańkowski","doi":"10.1007/978-3-030-59006-2_19","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_19","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132474153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Video Footage Highlight Detection in Formula 1 Through Vehicle Recognition with Faster R-CNN Trained on Game Footage 利用基于比赛录像的更快R-CNN训练的车辆识别来检测f1中的视频片段高光
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_16
Ruan Spijkerman, D. Haar
{"title":"Video Footage Highlight Detection in Formula 1 Through Vehicle Recognition with Faster R-CNN Trained on Game Footage","authors":"Ruan Spijkerman, D. Haar","doi":"10.1007/978-3-030-59006-2_16","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_16","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130172807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Performance Evaluation of Selected 3D Keypoint Detector-Descriptor Combinations 选定三维关键点检测器-描述子组合的性能评价
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_17
Paula Štancelová, E. Sikudová, Z. Černeková
{"title":"Performance Evaluation of Selected 3D Keypoint Detector-Descriptor Combinations","authors":"Paula Štancelová, E. Sikudová, Z. Černeková","doi":"10.1007/978-3-030-59006-2_17","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_17","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114261832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Tuberculosis Abnormality Detection in Chest X-Rays: A Deep Learning Approach 胸部x光中的结核异常检测:一种深度学习方法
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_11
M. Oloko-Oba, Serestina Viriri
{"title":"Tuberculosis Abnormality Detection in Chest X-Rays: A Deep Learning Approach","authors":"M. Oloko-Oba, Serestina Viriri","doi":"10.1007/978-3-030-59006-2_11","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_11","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121786944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Facial Age Estimation Using Compact Facial Features 基于紧凑面部特征的面部年龄估计
Pub Date : 2020-09-14 DOI: 10.1007/978-3-030-59006-2_1
J. D. Akinyemi, O. Onifade
{"title":"Facial Age Estimation Using Compact Facial Features","authors":"J. D. Akinyemi, O. Onifade","doi":"10.1007/978-3-030-59006-2_1","DOIUrl":"https://doi.org/10.1007/978-3-030-59006-2_1","url":null,"abstract":"","PeriodicalId":124003,"journal":{"name":"International Conference on Computer Vision and Graphics","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121322575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Conference on Computer Vision and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1