首页 > 最新文献

2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)最新文献

英文 中文
Blur vs. Block: Investigating the Effectiveness of Privacy-Enhancing Obfuscation for Images 模糊与块:调查图像隐私增强混淆的有效性
Yifang Li, Nishant Vishwamitra, Bart P. Knijnenburg, Hongxin Hu, Kelly E. Caine
Computer vision can lead to privacy issues such as unauthorized disclosure of private information and identity theft, but it may also be used to preserve user privacy. For example, using computer vision, we may be able to identify sensitive elements of an image and obfuscate those elements thereby protecting private information or identity. However, there is a lack of research investigating the effectiveness of applying obfuscation techniques to parts of images as a privacy enhancing technology. In particular, we know very little about how well obfuscation works for human viewers or users' attitudes towards using these mechanisms. In this paper, we report results from an online experiment with 53 participants that investigates the effectiveness two exemplar obfuscation techniques: "blurring" and "blocking", and explores users' perceptions of these obfuscations in terms of image satisfaction, information sufficiency, enjoyment, and social presence. Results show that although "blocking" is more effective at de-identification compared to "blurring" or leaving the image "as is", users' attitudes towards "blocking" are the most negative, which creates a conflict between privacy protection and users' experience. Future work should explore alternative obfuscation techniques that could protect users' privacy and also provide a good viewing experience.
计算机视觉可能导致隐私问题,如未经授权的私人信息泄露和身份盗窃,但它也可以用来保护用户隐私。例如,使用计算机视觉,我们可能能够识别图像中的敏感元素并对这些元素进行模糊处理,从而保护私人信息或身份。然而,对于将混淆技术应用于图像部分作为隐私增强技术的有效性,缺乏研究。特别是,我们对混淆对人类观众或用户使用这些机制的态度的效果知之甚少。在本文中,我们报告了53名参与者的在线实验结果,该实验调查了两种典型混淆技术的有效性:“模糊”和“阻塞”,并探讨了用户在图像满意度、信息充分性、享受和社会存在方面对这些混淆的看法。结果表明,虽然“屏蔽”在去识别方面比“模糊”或“保持原样”更有效,但用户对“屏蔽”的态度是最消极的,这造成了隐私保护与用户体验之间的冲突。未来的工作应该探索其他混淆技术,既能保护用户隐私,又能提供良好的观看体验。
{"title":"Blur vs. Block: Investigating the Effectiveness of Privacy-Enhancing Obfuscation for Images","authors":"Yifang Li, Nishant Vishwamitra, Bart P. Knijnenburg, Hongxin Hu, Kelly E. Caine","doi":"10.1109/CVPRW.2017.176","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.176","url":null,"abstract":"Computer vision can lead to privacy issues such as unauthorized disclosure of private information and identity theft, but it may also be used to preserve user privacy. For example, using computer vision, we may be able to identify sensitive elements of an image and obfuscate those elements thereby protecting private information or identity. However, there is a lack of research investigating the effectiveness of applying obfuscation techniques to parts of images as a privacy enhancing technology. In particular, we know very little about how well obfuscation works for human viewers or users' attitudes towards using these mechanisms. In this paper, we report results from an online experiment with 53 participants that investigates the effectiveness two exemplar obfuscation techniques: \"blurring\" and \"blocking\", and explores users' perceptions of these obfuscations in terms of image satisfaction, information sufficiency, enjoyment, and social presence. Results show that although \"blocking\" is more effective at de-identification compared to \"blurring\" or leaving the image \"as is\", users' attitudes towards \"blocking\" are the most negative, which creates a conflict between privacy protection and users' experience. Future work should explore alternative obfuscation techniques that could protect users' privacy and also provide a good viewing experience.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"65 1","pages":"1343-1351"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85714754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
The First Automatic Method for Mapping the Pothole in Seagrass 第一种海草坑洞自动测绘方法
M. Rahnemoonfar, M. Yari, Abdullah F. Rahman, Richard J. Kline
There is a vital need to map seagrass ecosystems in order to determine worldwide abundance and distribution. Currently there is no established method for mapping the pothole or scars in seagrass. Detection of seagrass with optical remote sensing is challenged by the fact that light is attenuated as it passes through the water column and reflects back from the benthos. Optical remote sensing of seagrass is only possible if the water is shallow and relatively clear. In reality, coastal waters are commonly turbid, and seagrasses can grow under 10 meters of water or even deeper. One of the most precise sensors to map the seagrass disturbance is side scan sonar. Underwater acoustics mapping produces a high definition, two-dimensional sonar image of seagrass ecosystems. This paper proposes a methodology which detects seagrass potholes in sonar images. Side scan sonar images usually contain speckle noise and uneven illumination across the image. Moreover, disturbance presents complex patterns where most segmentation techniques will fail. In this paper, the quality of image is improved in the first stage using adaptive thresholding and wavelet denoising techniques. In the next step, a novel level set technique is applied to identify the pothole patterns. Our method is robust to noise and uneven illumination. Moreover it can detect the complex pothole patterns. We tested our proposed approach on a collection of underwater sonar images taken from Laguna Madre in Texas. Experimental results in comparison with the ground-truth show the efficiency of the proposed method.
迫切需要绘制海草生态系统图,以便确定世界范围内的丰度和分布。目前还没有确定的方法来绘制海草中的坑或疤痕。由于光在穿过水柱并从底栖生物反射回来时被衰减,因此光学遥感对海草的探测面临挑战。只有在水较浅且相对清澈的情况下,才能对海草进行光学遥感。实际上,沿海水域通常是浑浊的,海草可以生长在10米以下甚至更深的水里。绘制海草扰动图最精确的传感器之一是侧扫声纳。水下声学测绘产生海草生态系统的高清晰度,二维声纳图像。本文提出了一种检测声纳图像中海草坑的方法。侧扫声纳图像通常包含散斑噪声和光照不均匀。此外,干扰呈现出复杂的模式,大多数分割技术将失败。本文首先采用自适应阈值和小波去噪技术提高图像质量。在接下来的步骤中,采用一种新的水平集技术来识别凹坑模式。该方法对噪声和光照不均匀具有较强的鲁棒性。此外,该方法还能探测复杂的坑穴模式。我们在德克萨斯州拉古纳马德雷的一组水下声纳图像上测试了我们提出的方法。实验结果与地面真值的比较表明了该方法的有效性。
{"title":"The First Automatic Method for Mapping the Pothole in Seagrass","authors":"M. Rahnemoonfar, M. Yari, Abdullah F. Rahman, Richard J. Kline","doi":"10.1109/CVPRW.2017.39","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.39","url":null,"abstract":"There is a vital need to map seagrass ecosystems in order to determine worldwide abundance and distribution. Currently there is no established method for mapping the pothole or scars in seagrass. Detection of seagrass with optical remote sensing is challenged by the fact that light is attenuated as it passes through the water column and reflects back from the benthos. Optical remote sensing of seagrass is only possible if the water is shallow and relatively clear. In reality, coastal waters are commonly turbid, and seagrasses can grow under 10 meters of water or even deeper. One of the most precise sensors to map the seagrass disturbance is side scan sonar. Underwater acoustics mapping produces a high definition, two-dimensional sonar image of seagrass ecosystems. This paper proposes a methodology which detects seagrass potholes in sonar images. Side scan sonar images usually contain speckle noise and uneven illumination across the image. Moreover, disturbance presents complex patterns where most segmentation techniques will fail. In this paper, the quality of image is improved in the first stage using adaptive thresholding and wavelet denoising techniques. In the next step, a novel level set technique is applied to identify the pothole patterns. Our method is robust to noise and uneven illumination. Moreover it can detect the complex pothole patterns. We tested our proposed approach on a collection of underwater sonar images taken from Laguna Madre in Texas. Experimental results in comparison with the ground-truth show the efficiency of the proposed method.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"16 1","pages":"267-274"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85996780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Vehicle Type Classification Using Bagging and Convolutional Neural Network on Multi View Surveillance Image 基于Bagging和卷积神经网络的多视点监控图像车型分类
Pyong-Kun Kim, Kil-Taek Lim
This paper aims to introduce a new vehicle type classification scheme on the images from multi-view surveillance camera. We propose four concepts to increase the performance on the images which have various resolutions from multi-view point. The Deep Learning method is essential to multi-view point image, bagging method makes system robust, data augmentation help to grow the classification capability, and post-processing compensate for imbalanced data. We combine these schemes and build a novel vehicle type classification system. Our system shows 97.84% classification accuracy on the 103,833 images in classification challenge dataset.
本文旨在介绍一种基于多视点监控摄像机图像的新型车辆分类方案。为了提高多视点不同分辨率图像的性能,我们提出了四个概念。深度学习方法是多视点图像的关键,套袋方法使系统具有鲁棒性,数据增强有助于提高分类能力,后处理有助于补偿数据不平衡。我们将这些方案结合起来,建立了一个新的车型分类系统。系统对分类挑战数据集中103833张图像的分类准确率为97.84%。
{"title":"Vehicle Type Classification Using Bagging and Convolutional Neural Network on Multi View Surveillance Image","authors":"Pyong-Kun Kim, Kil-Taek Lim","doi":"10.1109/CVPRW.2017.126","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.126","url":null,"abstract":"This paper aims to introduce a new vehicle type classification scheme on the images from multi-view surveillance camera. We propose four concepts to increase the performance on the images which have various resolutions from multi-view point. The Deep Learning method is essential to multi-view point image, bagging method makes system robust, data augmentation help to grow the classification capability, and post-processing compensate for imbalanced data. We combine these schemes and build a novel vehicle type classification system. Our system shows 97.84% classification accuracy on the 103,833 images in classification challenge dataset.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"38 1","pages":"914-919"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73096183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Optimizing the Lens Selection Process for Multi-focus Plenoptic Cameras and Numerical Evaluation 多焦全光相机镜头选择过程优化及数值评价
L. Palmieri, R. Koch
The last years have seen a quick rise of digital photography. Plenoptic cameras provide extended capabilities with respect to previous models. Multi-focus cameras enlarge the depth-of-field of the pictures using different focal lengths in the lens composing the array, but questions still arise on how to select and use these lenses. In this work a further insight on the lens selection was made, and a novel method was developed in order to choose the best available lens combination for the disparity estimation. We test different lens combinations, ranking them based on the error and the number of different lenses used, creating a mapping function that relates the virtual depth with the combination that achieves the best result. The results are then organized in a look up table that can be tuned to trade off between performances and accuracy. This allows for fast and accurate lens selection. Moreover, new synthetic images with respective ground truth are provided, in order to confirm that this work performs better than the current state of the art in efficiency and accuracy of the results.
近年来,数码摄影迅速兴起。全光相机相对于以前的型号提供了扩展的功能。多焦相机通过在组成阵列的镜头中使用不同焦距来扩大图像的景深,但是如何选择和使用这些镜头仍然存在一些问题。本文对视差估计的透镜选择问题进行了深入的研究,提出了一种选择最佳透镜组合进行视差估计的方法。我们测试了不同的镜头组合,根据误差和使用的不同镜头数量对它们进行排名,创建了一个映射函数,将虚拟深度与达到最佳效果的组合联系起来。然后将结果组织在一个查找表中,可以在性能和准确性之间进行调整。这允许快速和准确的镜头选择。此外,还提供了具有各自地面真值的新合成图像,以确认该工作在结果的效率和准确性方面优于当前的技术状态。
{"title":"Optimizing the Lens Selection Process for Multi-focus Plenoptic Cameras and Numerical Evaluation","authors":"L. Palmieri, R. Koch","doi":"10.1109/CVPRW.2017.223","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.223","url":null,"abstract":"The last years have seen a quick rise of digital photography. Plenoptic cameras provide extended capabilities with respect to previous models. Multi-focus cameras enlarge the depth-of-field of the pictures using different focal lengths in the lens composing the array, but questions still arise on how to select and use these lenses. In this work a further insight on the lens selection was made, and a novel method was developed in order to choose the best available lens combination for the disparity estimation. We test different lens combinations, ranking them based on the error and the number of different lenses used, creating a mapping function that relates the virtual depth with the combination that achieves the best result. The results are then organized in a look up table that can be tuned to trade off between performances and accuracy. This allows for fast and accurate lens selection. Moreover, new synthetic images with respective ground truth are provided, in order to confirm that this work performs better than the current state of the art in efficiency and accuracy of the results.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"33 1","pages":"1763-1774"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73695914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Investigating Nuisance Factors in Face Recognition with DCNN Representation 用DCNN表示研究人脸识别中的妨害因素
C. Ferrari, G. Lisanti, S. Berretti, A. Bimbo
Deep learning based approaches proved to be dramatically effective to address many computer vision applications, including "face recognition in the wild". It has been extensively demonstrated that methods exploiting Deep Convolutional Neural Networks (DCNN) are powerful enough to overcome to a great extent many problems that negatively affected computer vision algorithms based on hand-crafted features. These problems include variations in illumination, pose, expression and occlusion, to mention some. The DCNNs excellent discriminative power comes from the fact that they learn low-and high-level representations directly from the raw image data. Considering this, it can be assumed that the performance of a DCNN are influenced by the characteristics of the raw image data that are fed to the network. In this work, we evaluate the effect of different bounding box dimensions, alignment, positioning and data source on face recognition using DCNNs, and present a thorough evaluation on two well known, public DCNN architectures.
事实证明,基于深度学习的方法在解决许多计算机视觉应用(包括“野外人脸识别”)方面非常有效。已经广泛证明,利用深度卷积神经网络(DCNN)的方法足够强大,可以在很大程度上克服许多对基于手工特征的计算机视觉算法产生负面影响的问题。这些问题包括光照、姿势、表情和遮挡等方面的变化。DCNNs出色的判别能力来自于它们直接从原始图像数据中学习低级和高级表示。考虑到这一点,可以假设DCNN的性能受到输入到网络的原始图像数据的特征的影响。在这项工作中,我们评估了不同的边界盒尺寸、对齐、定位和数据源对使用DCNN进行人脸识别的影响,并对两种知名的公共DCNN架构进行了全面的评估。
{"title":"Investigating Nuisance Factors in Face Recognition with DCNN Representation","authors":"C. Ferrari, G. Lisanti, S. Berretti, A. Bimbo","doi":"10.1109/CVPRW.2017.86","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.86","url":null,"abstract":"Deep learning based approaches proved to be dramatically effective to address many computer vision applications, including \"face recognition in the wild\". It has been extensively demonstrated that methods exploiting Deep Convolutional Neural Networks (DCNN) are powerful enough to overcome to a great extent many problems that negatively affected computer vision algorithms based on hand-crafted features. These problems include variations in illumination, pose, expression and occlusion, to mention some. The DCNNs excellent discriminative power comes from the fact that they learn low-and high-level representations directly from the raw image data. Considering this, it can be assumed that the performance of a DCNN are influenced by the characteristics of the raw image data that are fed to the network. In this work, we evaluate the effect of different bounding box dimensions, alignment, positioning and data source on face recognition using DCNNs, and present a thorough evaluation on two well known, public DCNN architectures.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"19 1","pages":"583-591"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81111976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Leveraging Intra and Inter-Dataset Variations for Robust Face Alignment 利用数据集内部和数据集之间的变化进行稳健的人脸对齐
Wenyan Wu, Shuo Yang
Face alignment is a critical topic in the computer vision community. Numerous efforts have been made and various benchmark datasets have been released in recent decades. However, two significant issues remain in recent datasets, e.g., Intra-Dataset Variation and Inter-Dataset Variation. Inter-Dataset Variation refers to bias on expression, head pose, etc. inside one certain dataset, while Intra-Dataset Variation refers to different bias across different datasets. To address the mentioned problems, we proposed a novel Deep Variation Leveraging Network (DVLN), which consists of two strong coupling sub-networks, e.g., Dataset-Across Network (DA-Net) and Candidate-Decision Network (CD-Net). Extensive evaluations show that our approach demonstrates real-time performance and dramatically outperforms state-of-the-art methods on the challenging 300-W dataset.,,,,,, To address the mentioned problems, we proposed a novel Deep Variation Leveraging Network (DVLN), which consists of two strong coupling sub-networks, e.g., Dataset-Across Network (DA-Net) and Candidate-Decision Network (CD-Net). In particular, DA-Net takes advantage of different characteristics and distributions across different datasets, while CD-Net makes a final decision on candidate hypotheses given by DA-Net to leverage variations within one certain dataset. Extensive evaluations show that our approach demonstrates real-time performance and dramatically outperforms state-of-the-art methods on the challenging 300-W dataset.
人脸对齐是计算机视觉领域的一个重要课题。近几十年来,人们做了大量的努力,发布了各种基准数据集。然而,在最近的数据集中仍然存在两个重要的问题,即数据集内的变化和数据集间的变化。数据集间差异指的是同一数据集内部的表情、头部姿势等偏差,而数据集内差异指的是不同数据集之间的不同偏差。为了解决上述问题,我们提出了一种新的深度变化利用网络(DVLN),它由两个强耦合子网络组成,即数据集跨网络(DA-Net)和候选决策网络(CD-Net)。广泛的评估表明,我们的方法具有实时性,并且在具有挑战性的300-W数据集上显著优于最先进的方法。,,,,,,为了解决上述问题,我们提出了一种新的深度变化利用网络(DVLN),它由两个强耦合子网络组成,即数据集跨网络(DA-Net)和候选决策网络(CD-Net)。特别是,DA-Net利用不同数据集的不同特征和分布,而CD-Net根据DA-Net给出的候选假设做出最终决定,以利用某个数据集内的变化。广泛的评估表明,我们的方法具有实时性,并且在具有挑战性的300-W数据集上显著优于最先进的方法。
{"title":"Leveraging Intra and Inter-Dataset Variations for Robust Face Alignment","authors":"Wenyan Wu, Shuo Yang","doi":"10.1109/CVPRW.2017.261","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.261","url":null,"abstract":"Face alignment is a critical topic in the computer vision community. Numerous efforts have been made and various benchmark datasets have been released in recent decades. However, two significant issues remain in recent datasets, e.g., Intra-Dataset Variation and Inter-Dataset Variation. Inter-Dataset Variation refers to bias on expression, head pose, etc. inside one certain dataset, while Intra-Dataset Variation refers to different bias across different datasets. To address the mentioned problems, we proposed a novel Deep Variation Leveraging Network (DVLN), which consists of two strong coupling sub-networks, e.g., Dataset-Across Network (DA-Net) and Candidate-Decision Network (CD-Net). Extensive evaluations show that our approach demonstrates real-time performance and dramatically outperforms state-of-the-art methods on the challenging 300-W dataset.,,,,,, To address the mentioned problems, we proposed a novel Deep Variation Leveraging Network (DVLN), which consists of two strong coupling sub-networks, e.g., Dataset-Across Network (DA-Net) and Candidate-Decision Network (CD-Net). In particular, DA-Net takes advantage of different characteristics and distributions across different datasets, while CD-Net makes a final decision on candidate hypotheses given by DA-Net to leverage variations within one certain dataset. Extensive evaluations show that our approach demonstrates real-time performance and dramatically outperforms state-of-the-art methods on the challenging 300-W dataset.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"26 1","pages":"2096-2105"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81129729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 100
Real-Time Driver Drowsiness Detection for Embedded System Using Model Compression of Deep Neural Networks 基于深度神经网络模型压缩的嵌入式系统驾驶员困倦实时检测
B. Reddy, Ye-Hoon Kim, Sojung Yun, Chanwon Seo, Junik Jang
Driver’s status is crucial because one of the main reasons for motor vehicular accidents is related to driver’s inattention or drowsiness. Drowsiness detector on a car can reduce numerous accidents. Accidents occur because of a single moment of negligence, thus driver monitoring system which works in real-time is necessary. This detector should be deployable to an embedded device and perform at high accuracy. In this paper, a novel approach towards real-time drowsiness detection based on deep learning which can be implemented on a low cost embedded board and performs with a high accuracy is proposed. Main contribution of our paper is compression of heavy baseline model to a light weight model deployable to an embedded board. Moreover, minimized network structure was designed based on facial landmark input to recognize whether driver is drowsy or not. The proposed model achieved an accuracy of 89.5% on 3-class classification and speed of 14.9 frames per second (FPS) on Jetson TK1.
驾驶员的状态是至关重要的,因为机动车事故的主要原因之一与驾驶员的注意力不集中或困倦有关。汽车上的睡意检测器可以减少许多事故。事故的发生往往是由于一时的疏忽,因此需要实时工作的驾驶员监控系统。该检测器应可部署到嵌入式设备,并以高精度执行。本文提出了一种基于深度学习的实时困倦检测新方法,该方法可以在低成本的嵌入式电路板上实现,并且具有高精度。本文的主要贡献是将重型基线模型压缩为可部署到嵌入式板上的轻量级模型。此外,基于面部地标输入,设计了最小化网络结构来识别驾驶员是否昏昏欲睡。该模型在Jetson TK1上的3类分类准确率达到89.5%,速度达到14.9帧/秒。
{"title":"Real-Time Driver Drowsiness Detection for Embedded System Using Model Compression of Deep Neural Networks","authors":"B. Reddy, Ye-Hoon Kim, Sojung Yun, Chanwon Seo, Junik Jang","doi":"10.1109/CVPRW.2017.59","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.59","url":null,"abstract":"Driver’s status is crucial because one of the main reasons for motor vehicular accidents is related to driver’s inattention or drowsiness. Drowsiness detector on a car can reduce numerous accidents. Accidents occur because of a single moment of negligence, thus driver monitoring system which works in real-time is necessary. This detector should be deployable to an embedded device and perform at high accuracy. In this paper, a novel approach towards real-time drowsiness detection based on deep learning which can be implemented on a low cost embedded board and performs with a high accuracy is proposed. Main contribution of our paper is compression of heavy baseline model to a light weight model deployable to an embedded board. Moreover, minimized network structure was designed based on facial landmark input to recognize whether driver is drowsy or not. The proposed model achieved an accuracy of 89.5% on 3-class classification and speed of 14.9 frames per second (FPS) on Jetson TK1.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"16 1","pages":"438-445"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90194478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 158
FORMS-Locks: A Dataset for the Evaluation of Similarity Measures for Forensic Toolmark Images FORMS-Locks:用于评估法医工具标记图像相似性度量的数据集
M. Keglevic, Robert Sablatnig
We present a toolmark dataset created using lock cylinders seized during criminal investigations of break-ins. A total number of 197 cylinders from 48 linked criminal cases were photographed under a comparison microscope used by forensic experts for toolmark comparisons. In order to allow an assessment of the influence of different lighting conditions, all images were captured using a ring light with 11 different lighting settings. Further, matching image regions in the toolmark images were manually annotated. In addition to the annotated toolmark images and the annotation tool, extracted toolmark patches are provided for training and testing to allow a quantitative comparison of the performance of different similarity measures. Finally, results from an evaluation using a publicly available state-of-the-art image descriptor based on deep learning are presented to provide a baseline for future publications.
我们提出了一个工具标记数据集,该数据集使用在刑事调查中查获的锁圆柱体创建。来自48个相关刑事案件的197个圆柱体在法医专家用于工具标记比较的比较显微镜下被拍摄下来。为了评估不同照明条件的影响,所有图像都是使用具有11种不同照明设置的环形灯拍摄的。此外,对工具标记图像中的匹配图像区域进行手工标注。除了标注的工具标记图像和标注工具外,还提供了提取的工具标记补丁用于训练和测试,以允许对不同相似性度量的性能进行定量比较。最后,介绍了使用基于深度学习的公开可用的最先进图像描述符进行评估的结果,为未来的出版物提供基线。
{"title":"FORMS-Locks: A Dataset for the Evaluation of Similarity Measures for Forensic Toolmark Images","authors":"M. Keglevic, Robert Sablatnig","doi":"10.1109/CVPRW.2017.236","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.236","url":null,"abstract":"We present a toolmark dataset created using lock cylinders seized during criminal investigations of break-ins. A total number of 197 cylinders from 48 linked criminal cases were photographed under a comparison microscope used by forensic experts for toolmark comparisons. In order to allow an assessment of the influence of different lighting conditions, all images were captured using a ring light with 11 different lighting settings. Further, matching image regions in the toolmark images were manually annotated. In addition to the annotated toolmark images and the annotation tool, extracted toolmark patches are provided for training and testing to allow a quantitative comparison of the performance of different similarity measures. Finally, results from an evaluation using a publicly available state-of-the-art image descriptor based on deep learning are presented to provide a baseline for future publications.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"21 1","pages":"1890-1897"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87917495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Intel(R) RealSense(TM) Stereoscopic Depth Cameras Intel(R) RealSense(TM)立体深度相机
L. Keselman, J. Woodfill, A. Grunnet-Jepsen, A. Bhowmik
We present a comprehensive overview of the stereoscopic Intel RealSense RGBD imaging systems. We discuss these systems' mode-of-operation, functional behavior and include models of their expected performance, shortcomings, and limitations. We provide information about the systems' optical characteristics, their correlation algorithms, and how these properties can affect different applications, including 3D reconstruction and gesture recognition. Our discussion covers the Intel RealSense R200 and RS400.
我们提出了立体英特尔RealSense RGBD成像系统的全面概述。我们将讨论这些系统的操作模式、功能行为,并包括其预期性能、缺点和限制的模型。我们提供了有关系统光学特性的信息,它们的相关算法,以及这些特性如何影响不同的应用,包括3D重建和手势识别。我们的讨论涵盖了英特尔RealSense R200和RS400。
{"title":"Intel(R) RealSense(TM) Stereoscopic Depth Cameras","authors":"L. Keselman, J. Woodfill, A. Grunnet-Jepsen, A. Bhowmik","doi":"10.1109/CVPRW.2017.167","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.167","url":null,"abstract":"We present a comprehensive overview of the stereoscopic Intel RealSense RGBD imaging systems. We discuss these systems' mode-of-operation, functional behavior and include models of their expected performance, shortcomings, and limitations. We provide information about the systems' optical characteristics, their correlation algorithms, and how these properties can affect different applications, including 3D reconstruction and gesture recognition. Our discussion covers the Intel RealSense R200 and RS400.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"13 1","pages":"1267-1276"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90313840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 116
Balanced Two-Stage Residual Networks for Image Super-Resolution 图像超分辨率的平衡两级残差网络
Yuchen Fan, Humphrey Shi, Jiahui Yu, Ding Liu, Wei Han, Haichao Yu, Zhangyang Wang, Xinchao Wang, Thomas S. Huang
In this paper, balanced two-stage residual networks (BTSRN) are proposed for single image super-resolution. The deep residual design with constrained depth achieves the optimal balance between the accuracy and the speed for super-resolving images. The experiments show that the balanced two-stage structure, together with our lightweight two-layer PConv residual block design, achieves very promising results when considering both accuracy and speed. We evaluated our models on the New Trends in Image Restoration and Enhancement workshop and challenge on image super-resolution (NTIRE SR 2017). Our final model with only 10 residual blocks ranked among the best ones in terms of not only accuracy (6th among 20 final teams) but also speed (2nd among top 6 teams in terms of accuracy). The source code both for training and evaluation is available in https://github.com/ychfan/sr_ntire2017.
针对单幅图像的超分辨率问题,提出了平衡两级残差网络(BTSRN)。深度约束下的深度残差设计实现了超分辨图像精度和速度的最佳平衡。实验表明,平衡两级结构和轻量化两层PConv残块设计在精度和速度方面都取得了很好的效果。我们在图像恢复和增强的新趋势研讨会和图像超分辨率的挑战(NTIRE SR 2017)上评估了我们的模型。我们的最终模型只有10个剩余块,不仅在准确性(在20个最终团队中排名第6)和速度(在前6个团队中排名第2)方面都名列前茅。培训和评估的源代码可从https://github.com/ychfan/sr_ntire2017获得。
{"title":"Balanced Two-Stage Residual Networks for Image Super-Resolution","authors":"Yuchen Fan, Humphrey Shi, Jiahui Yu, Ding Liu, Wei Han, Haichao Yu, Zhangyang Wang, Xinchao Wang, Thomas S. Huang","doi":"10.1109/CVPRW.2017.154","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.154","url":null,"abstract":"In this paper, balanced two-stage residual networks (BTSRN) are proposed for single image super-resolution. The deep residual design with constrained depth achieves the optimal balance between the accuracy and the speed for super-resolving images. The experiments show that the balanced two-stage structure, together with our lightweight two-layer PConv residual block design, achieves very promising results when considering both accuracy and speed. We evaluated our models on the New Trends in Image Restoration and Enhancement workshop and challenge on image super-resolution (NTIRE SR 2017). Our final model with only 10 residual blocks ranked among the best ones in terms of not only accuracy (6th among 20 final teams) but also speed (2nd among top 6 teams in terms of accuracy). The source code both for training and evaluation is available in https://github.com/ychfan/sr_ntire2017.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"88 1","pages":"1157-1164"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78537194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 88
期刊
2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1