首页 > 最新文献

2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)最新文献

英文 中文
Ball 3D Trajectory Reconstruction without Preliminary Temporal and Geometrical Camera Calibration 无初步时间和几何摄像机标定的球三维轨迹重建
S. Miyata, H. Saito, Kosuke Takahashi, Dan Mikami, Mariko Isogawa, H. Kimata
This paper proposes a method for reconstructing 3D ball trajectories by using multiple temporally and geometrically uncalibrated cameras. To use cameras to measure the trajectory of a fast-moving object, such as a ball thrown by a pitcher, the cameras must be temporally synchronized and their position and orientation should be calibrated. In some cases, these conditions cannot be met, e.g., one cannot geometrically calibrate cameras when one cannot step into a baseball stadium. The basic idea of the proposed method is to use a ball captured by multiple cameras as a corresponding point. The method first detects a ball. Then, it estimates temporal difference between cameras. After that, the ball positions are used as corresponding points for geometrically calibrating the cameras. Experiments using actual pitching videos verify the effectiveness of our method.
本文提出了一种利用多个时间和几何未标定摄像机重建三维球轨迹的方法。为了使用摄像机来测量快速运动物体的轨迹,例如投手抛出的球,摄像机必须在时间上同步,并且应该校准它们的位置和方向。在某些情况下,这些条件无法满足,例如,当一个人无法进入棒球场时,就无法对相机进行几何校准。所提出的方法的基本思想是使用多个摄像机捕获的球作为对应点。该方法首先检测一个球。然后,估计相机之间的时间差。然后,将球的位置作为对应点,对摄像机进行几何标定。实际投球视频的实验验证了该方法的有效性。
{"title":"Ball 3D Trajectory Reconstruction without Preliminary Temporal and Geometrical Camera Calibration","authors":"S. Miyata, H. Saito, Kosuke Takahashi, Dan Mikami, Mariko Isogawa, H. Kimata","doi":"10.1109/CVPRW.2017.26","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.26","url":null,"abstract":"This paper proposes a method for reconstructing 3D ball trajectories by using multiple temporally and geometrically uncalibrated cameras. To use cameras to measure the trajectory of a fast-moving object, such as a ball thrown by a pitcher, the cameras must be temporally synchronized and their position and orientation should be calibrated. In some cases, these conditions cannot be met, e.g., one cannot geometrically calibrate cameras when one cannot step into a baseball stadium. The basic idea of the proposed method is to use a ball captured by multiple cameras as a corresponding point. The method first detects a ball. Then, it estimates temporal difference between cameras. After that, the ball positions are used as corresponding points for geometrically calibrating the cameras. Experiments using actual pitching videos verify the effectiveness of our method.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"1 1","pages":"164-169"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81705468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Aff-Wild: Valence and Arousal ‘In-the-Wild’ Challenge 野性:价态和唤醒“野性”挑战
S. Zafeiriou, D. Kollias, M. Nicolaou, A. Papaioannou, Guoying Zhao, I. Kotsia
The Affect-in-the-Wild (Aff-Wild) Challenge proposes a new comprehensive benchmark for assessing the performance of facial affect/behaviour analysis/understanding 'in-the-wild'. The Aff-wild benchmark contains about 300 videos (over 2,000 minutes of data) annotated with regards to valence and arousal, all captured 'in-the-wild' (the main source being Youtube videos). The paper presents the database description, the experimental set up, the baseline method used for the Challenge and finally the summary of the performance of the different methods submitted to the Affect-in-the-Wild Challenge for Valence and Arousal estimation. The challenge demonstrates that meticulously designed deep neural networks can achieve very good performance when trained with in-the-wild data.
“野外情感挑战”提出了一个新的综合基准,用于评估面部情感/行为分析/理解“野外”的表现。off -wild基准包含大约300个视频(超过2000分钟的数据),这些视频都是“in-the-wild”捕获的(主要来源是Youtube视频)。本文介绍了数据库描述、实验设置、挑战中使用的基线方法,最后总结了在价性和唤醒估计中提交的不同方法的性能。该挑战表明,精心设计的深度神经网络在使用野外数据进行训练时可以获得非常好的性能。
{"title":"Aff-Wild: Valence and Arousal ‘In-the-Wild’ Challenge","authors":"S. Zafeiriou, D. Kollias, M. Nicolaou, A. Papaioannou, Guoying Zhao, I. Kotsia","doi":"10.1109/CVPRW.2017.248","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.248","url":null,"abstract":"The Affect-in-the-Wild (Aff-Wild) Challenge proposes a new comprehensive benchmark for assessing the performance of facial affect/behaviour analysis/understanding 'in-the-wild'. The Aff-wild benchmark contains about 300 videos (over 2,000 minutes of data) annotated with regards to valence and arousal, all captured 'in-the-wild' (the main source being Youtube videos). The paper presents the database description, the experimental set up, the baseline method used for the Challenge and finally the summary of the performance of the different methods submitted to the Affect-in-the-Wild Challenge for Valence and Arousal estimation. The challenge demonstrates that meticulously designed deep neural networks can achieve very good performance when trained with in-the-wild data.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"120 1","pages":"1980-1987"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85705189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 275
A Taxonomy and Evaluation of Dense Light Field Depth Estimation Algorithms 密集光场深度估计算法的分类与评价
O. Johannsen, Katrin Honauer, Bastian Goldlücke, A. Alperovich, F. Battisti, Yunsu Bok, Michele Brizzi, M. Carli, Gyeongmin Choe, M. Diebold, M. Gutsche, Hae-Gon Jeon, In-So Kweon, Jaesik Park, Jinsun Park, H. Schilling, Hao Sheng, Lipeng Si, Michael Strecke, Antonin Sulc, Yu-Wing Tai, Qing Wang, Tingxian Wang, S. Wanner, Z. Xiong, Jingyi Yu, Shuo Zhang, Hao Zhu
This paper presents the results of the depth estimation challenge for dense light fields, which took place at the second workshop on Light Fields for Computer Vision (LF4CV) in conjunction with CVPR 2017. The challenge consisted of submission to a recent benchmark [7], which allows a thorough performance analysis. While individual results are readily available on the benchmark web page http://www.lightfield-analysis.net, we take this opportunity to give a detailed overview of the current participants. Based on the algorithms submitted to our challenge, we develop a taxonomy of light field disparity estimation algorithms and give a report on the current state-ofthe- art. In addition, we include more comparative metrics, and discuss the relative strengths and weaknesses of the algorithms. Thus, we obtain a snapshot of where light field algorithm development stands at the moment and identify aspects with potential for further improvement.
本文介绍了在与CVPR 2017联合举行的第二届计算机视觉光场研讨会(LF4CV)上进行的密集光场深度估计挑战的结果。挑战包括提交到最近的基准测试[7],它允许进行彻底的性能分析。虽然在基准测试网页http://www.lightfield-analysis.net上可以很容易地获得个人结果,但我们借此机会对当前参与者进行详细概述。基于提交的算法,我们发展了一种光场视差估计算法的分类,并对当前的技术状况进行了报告。此外,我们包括更多的比较指标,并讨论了算法的相对优势和劣势。因此,我们获得了光场算法发展现状的快照,并确定了有进一步改进潜力的方面。
{"title":"A Taxonomy and Evaluation of Dense Light Field Depth Estimation Algorithms","authors":"O. Johannsen, Katrin Honauer, Bastian Goldlücke, A. Alperovich, F. Battisti, Yunsu Bok, Michele Brizzi, M. Carli, Gyeongmin Choe, M. Diebold, M. Gutsche, Hae-Gon Jeon, In-So Kweon, Jaesik Park, Jinsun Park, H. Schilling, Hao Sheng, Lipeng Si, Michael Strecke, Antonin Sulc, Yu-Wing Tai, Qing Wang, Tingxian Wang, S. Wanner, Z. Xiong, Jingyi Yu, Shuo Zhang, Hao Zhu","doi":"10.1109/CVPRW.2017.226","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.226","url":null,"abstract":"This paper presents the results of the depth estimation challenge for dense light fields, which took place at the second workshop on Light Fields for Computer Vision (LF4CV) in conjunction with CVPR 2017. The challenge consisted of submission to a recent benchmark [7], which allows a thorough performance analysis. While individual results are readily available on the benchmark web page http://www.lightfield-analysis.net, we take this opportunity to give a detailed overview of the current participants. Based on the algorithms submitted to our challenge, we develop a taxonomy of light field disparity estimation algorithms and give a report on the current state-ofthe- art. In addition, we include more comparative metrics, and discuss the relative strengths and weaknesses of the algorithms. Thus, we obtain a snapshot of where light field algorithm development stands at the moment and identify aspects with potential for further improvement.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"70 1","pages":"1795-1812"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73890287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
Analysis, Comparison, and Assessment of Latent Fingerprint Image Preprocessing 潜在指纹图像预处理的分析、比较与评价
Haiying Guan, Paul Lee, A. Dienstfrey, M. Theofanos, C. Lamp, Brian C. Stanton, Matthew T. Schwarz
Latent fingerprints obtained from crime scenes are rarely immediately suitable for identification purposes. Instead, most latent fingerprint images must be preprocessed to enhance the fingerprint information held within the digital image, while suppressing interference arising from noise and otherwise unwanted image features. In the following we present results of our ongoing research to assess this critical step in the forensic workflow. Previously we discussed the creation of a new database of latent fingerprint images to support such research. The new contributions of this paper are twofold. First, we implement a study in which a group of trained Latent Print Examiners provide Extended Feature Set markups of all images. We discuss the experimental design of this study, and its execution. Next, we propose metrics for measuring the increase of fingerprint information provided by latent fingerprint image preprocessing, and we present preliminary analysis of these metrics when applied to the images in our database. We consider formally defined quality scales (Good, Bad, Ugly), and minutiae identifications of latent fingerprint images before and after preprocessing. All analyses show that latent fingerprint image preprocessing results in a statistically significant increase in fingerprint information and quality.
从犯罪现场获得的潜在指纹很少能立即用于身份识别。相反,大多数潜在的指纹图像必须进行预处理,以增强数字图像中保存的指纹信息,同时抑制噪声和其他不需要的图像特征引起的干扰。在下文中,我们介绍了我们正在进行的研究结果,以评估法医工作流程中的这一关键步骤。之前我们讨论了建立一个新的潜在指纹图像数据库来支持这样的研究。本文的新贡献是双重的。首先,我们实现了一项研究,在该研究中,一组训练有素的潜在打印审查员提供了所有图像的扩展特征集标记。我们讨论了本研究的实验设计和实施。接下来,我们提出了衡量潜在指纹图像预处理所提供的指纹信息增加的指标,并对这些指标应用于我们数据库中的图像进行了初步分析。我们考虑了正式定义的质量尺度(好,坏,丑),以及预处理前后潜在指纹图像的细节识别。分析结果表明,对潜在指纹图像进行预处理后,指纹信息和质量均有显著提高。
{"title":"Analysis, Comparison, and Assessment of Latent Fingerprint Image Preprocessing","authors":"Haiying Guan, Paul Lee, A. Dienstfrey, M. Theofanos, C. Lamp, Brian C. Stanton, Matthew T. Schwarz","doi":"10.1109/CVPRW.2017.91","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.91","url":null,"abstract":"Latent fingerprints obtained from crime scenes are rarely immediately suitable for identification purposes. Instead, most latent fingerprint images must be preprocessed to enhance the fingerprint information held within the digital image, while suppressing interference arising from noise and otherwise unwanted image features. In the following we present results of our ongoing research to assess this critical step in the forensic workflow. Previously we discussed the creation of a new database of latent fingerprint images to support such research. The new contributions of this paper are twofold. First, we implement a study in which a group of trained Latent Print Examiners provide Extended Feature Set markups of all images. We discuss the experimental design of this study, and its execution. Next, we propose metrics for measuring the increase of fingerprint information provided by latent fingerprint image preprocessing, and we present preliminary analysis of these metrics when applied to the images in our database. We consider formally defined quality scales (Good, Bad, Ugly), and minutiae identifications of latent fingerprint images before and after preprocessing. All analyses show that latent fingerprint image preprocessing results in a statistically significant increase in fingerprint information and quality.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"196 1","pages":"628-635"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77436937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Deep Convolutional Neural Network with Selection Units for Super-Resolution 具有超分辨率选择单元的深度卷积神经网络
Jae-Seok Choi, Munchurl Kim
Rectified linear units (ReLU) are known to be effective in many deep learning methods. Inspired by linear-mapping technique used in other super-resolution (SR) methods, we reinterpret ReLU into point-wise multiplication of an identity mapping and a switch, and finally present a novel nonlinear unit, called a selection unit (SU). While conventional ReLU has no direct control through which data is passed, the proposed SU optimizes this on-off switching control, and is therefore capable of better handling nonlinearity functionality than ReLU in a more flexible way. Our proposed deep network with SUs, called SelNet, was top-5th ranked in NTIRE2017 Challenge, which has a much lower computation complexity compared to the top-4 entries. Further experiment results show that our proposed SelNet outperforms our baseline only with ReLU (without SUs), and other state-of-the-art deep-learning-based SR methods.
整流线性单元(ReLU)在许多深度学习方法中都是有效的。受其他超分辨率(SR)方法中使用的线性映射技术的启发,我们将ReLU重新解释为单位映射和开关的逐点乘法,最后提出了一种新的非线性单元,称为选择单元(SU)。传统的ReLU没有数据传递的直接控制,而该SU优化了这种开关控制,因此能够以更灵活的方式比ReLU更好地处理非线性功能。我们提出的带有SUs的深度网络SelNet在NTIRE2017挑战赛中排名前5,与前4名的参赛作品相比,其计算复杂度要低得多。进一步的实验结果表明,我们提出的SelNet仅在使用ReLU(没有SUs)和其他最先进的基于深度学习的SR方法时才优于我们的基线。
{"title":"A Deep Convolutional Neural Network with Selection Units for Super-Resolution","authors":"Jae-Seok Choi, Munchurl Kim","doi":"10.1109/CVPRW.2017.153","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.153","url":null,"abstract":"Rectified linear units (ReLU) are known to be effective in many deep learning methods. Inspired by linear-mapping technique used in other super-resolution (SR) methods, we reinterpret ReLU into point-wise multiplication of an identity mapping and a switch, and finally present a novel nonlinear unit, called a selection unit (SU). While conventional ReLU has no direct control through which data is passed, the proposed SU optimizes this on-off switching control, and is therefore capable of better handling nonlinearity functionality than ReLU in a more flexible way. Our proposed deep network with SUs, called SelNet, was top-5th ranked in NTIRE2017 Challenge, which has a much lower computation complexity compared to the top-4 entries. Further experiment results show that our proposed SelNet outperforms our baseline only with ReLU (without SUs), and other state-of-the-art deep-learning-based SR methods.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"37 4 1","pages":"1150-1156"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80293065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 103
End-to-End Driving in a Realistic Racing Game with Deep Reinforcement Learning 端到端驾驶在现实的赛车游戏与深度强化学习
E. Perot, M. Jaritz, Marin Toromanoff, Raoul de Charette
We address the problem of autonomous race car driving. Using a recent rally game (WRC6) with realistic physics and graphics we train an Asynchronous Actor Critic (A3C) in an end-to-end fashion and propose an improved reward function to learn faster. The network is trained simultaneously on three very different tracks (snow, mountain, and coast) with various road structures, graphics and physics. Despite the more complex environments the trained agent learns significant features and exhibits good performance while driving in a more stable way than existing end-to-end approaches.
我们解决了自动驾驶赛车的问题。使用最近的拉力赛游戏(WRC6)与现实的物理和图形,我们训练异步演员评论家(A3C)在端到端方式,并提出改进的奖励功能,以更快地学习。该网络同时在三条非常不同的轨道(雪地、山地和海岸)上进行训练,这些轨道具有不同的道路结构、图形和物理特性。尽管环境更复杂,但经过训练的智能体在以比现有端到端方法更稳定的方式驾驶时,仍能学习到重要的特征,并表现出良好的性能。
{"title":"End-to-End Driving in a Realistic Racing Game with Deep Reinforcement Learning","authors":"E. Perot, M. Jaritz, Marin Toromanoff, Raoul de Charette","doi":"10.1109/CVPRW.2017.64","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.64","url":null,"abstract":"We address the problem of autonomous race car driving. Using a recent rally game (WRC6) with realistic physics and graphics we train an Asynchronous Actor Critic (A3C) in an end-to-end fashion and propose an improved reward function to learn faster. The network is trained simultaneously on three very different tracks (snow, mountain, and coast) with various road structures, graphics and physics. Despite the more complex environments the trained agent learns significant features and exhibits good performance while driving in a more stable way than existing end-to-end approaches.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"103 1","pages":"474-475"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89931752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
Infrared Variation Optimized Deep Convolutional Neural Network for Robust Automatic Ground Target Recognition 红外变化优化的深度卷积神经网络鲁棒自动地面目标识别
Sungho Kim, Woo‐Jin Song, Sohyeon Kim
Automatic infrared target recognition (ATR) is a traditionally unsolved problem in military applications because of the wide range of infrared (IR) image variations and limited number of training images, which is caused by various 3D target poses, non-cooperative weather conditions, and difficult target acquisition environments. Recently, deep convolutional neural network-based approaches in RGB images (RGB-CNN) showed breakthrough performance in computer vision problems, such as object detection and classification. The direct use of the RGB-CNN to IR ATR problem fails to work because of the IR database problems. This paper presents a novel infrared variation-optimized deep convolutional neural network (IVO-CNN) by considering database management, such as increasing the database by a thermal simulator, controlling the image contrast automatically and suppressing the thermal noise to reduce the effects of infrared image variations in deep convolutional neural network-based automatic ground target recognition. The experimental results on the synthesized infrared images generated by the thermal simulator (OKTAL-SE) validated the feasibility of IVO-CNN for military ATR applications.
红外目标自动识别(ATR)在军事应用中是一个传统的未解决的问题,因为红外图像变化范围大,训练图像数量有限,这是由各种三维目标姿态、非合作天气条件和困难的目标获取环境造成的。近年来,基于深度卷积神经网络的RGB图像方法(RGB- cnn)在目标检测和分类等计算机视觉问题上取得了突破性的进展。由于红外数据库问题,直接使用RGB-CNN来红外ATR问题无法工作。在基于深度卷积神经网络的地面目标自动识别中,通过热模拟器增加数据库、自动控制图像对比度和抑制热噪声等方法来降低红外图像变化的影响,提出了一种新的红外变化优化深度卷积神经网络(IVO-CNN)。在热模拟器(OKTAL-SE)生成的红外合成图像上的实验结果验证了IVO-CNN在军事ATR应用中的可行性。
{"title":"Infrared Variation Optimized Deep Convolutional Neural Network for Robust Automatic Ground Target Recognition","authors":"Sungho Kim, Woo‐Jin Song, Sohyeon Kim","doi":"10.1109/CVPRW.2017.30","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.30","url":null,"abstract":"Automatic infrared target recognition (ATR) is a traditionally unsolved problem in military applications because of the wide range of infrared (IR) image variations and limited number of training images, which is caused by various 3D target poses, non-cooperative weather conditions, and difficult target acquisition environments. Recently, deep convolutional neural network-based approaches in RGB images (RGB-CNN) showed breakthrough performance in computer vision problems, such as object detection and classification. The direct use of the RGB-CNN to IR ATR problem fails to work because of the IR database problems. This paper presents a novel infrared variation-optimized deep convolutional neural network (IVO-CNN) by considering database management, such as increasing the database by a thermal simulator, controlling the image contrast automatically and suppressing the thermal noise to reduce the effects of infrared image variations in deep convolutional neural network-based automatic ground target recognition. The experimental results on the synthesized infrared images generated by the thermal simulator (OKTAL-SE) validated the feasibility of IVO-CNN for military ATR applications.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"15 1","pages":"195-202"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87390410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
High-Magnification Multi-views Based Classification of Breast Fine Needle Aspiration Cytology Cell Samples Using Fusion of Decisions from Deep Convolutional Networks 基于深度卷积网络决策融合的高倍多视图乳腺细针穿刺细胞学样本分类
Hrushikesh Garud, S. Karri, D. Sheet, J. Chatterjee, M. Mahadevappa, A. Ray, Arindam Ghosh, A. Maity
Fine needle aspiration cytology is commonly used for diagnosis of breast cancer, with traditional practice being based on the subjective visual assessment of the breast cytopathology cell samples under a microscope to evaluate the state of various cytological features. Therefore, there are many challenges in maintaining consistency and reproducibility of findings. However, digital imaging and computational aid in diagnosis can improve the diagnostic accuracy and reduce the effective workload of pathologists. This paper presents a deep convolutional neural network (CNN) based classification approach for the diagnosis of the cell samples using their microscopic high-magnification multi-views. The proposed approach has been tested using GoogLeNet architecture of CNN on an image dataset of 37 breast cytopathology samples (24 benign and 13 malignant), where the network was trained using images of ~54% cell samples and tested on the rest, achieving 89.7% mean accuracy in 8 fold validation.
细针抽吸细胞学是常用的乳腺癌诊断方法,传统做法是在显微镜下对乳腺细胞病理学细胞样本进行主观视觉评估,评估各种细胞学特征的状态。因此,在保持研究结果的一致性和可重复性方面存在许多挑战。然而,数字成像和计算辅助诊断可以提高诊断准确性,减少病理学家的有效工作量。本文提出了一种基于深度卷积神经网络(CNN)的分类方法,用于细胞样本的显微高倍多视图诊断。我们使用CNN的GoogLeNet架构在37个乳腺细胞病理学样本(24个良性和13个恶性)的图像数据集上对所提出的方法进行了测试,其中网络使用约54%的细胞样本图像进行训练,并在其余细胞样本上进行测试,在8次验证中达到89.7%的平均准确率。
{"title":"High-Magnification Multi-views Based Classification of Breast Fine Needle Aspiration Cytology Cell Samples Using Fusion of Decisions from Deep Convolutional Networks","authors":"Hrushikesh Garud, S. Karri, D. Sheet, J. Chatterjee, M. Mahadevappa, A. Ray, Arindam Ghosh, A. Maity","doi":"10.1109/CVPRW.2017.115","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.115","url":null,"abstract":"Fine needle aspiration cytology is commonly used for diagnosis of breast cancer, with traditional practice being based on the subjective visual assessment of the breast cytopathology cell samples under a microscope to evaluate the state of various cytological features. Therefore, there are many challenges in maintaining consistency and reproducibility of findings. However, digital imaging and computational aid in diagnosis can improve the diagnostic accuracy and reduce the effective workload of pathologists. This paper presents a deep convolutional neural network (CNN) based classification approach for the diagnosis of the cell samples using their microscopic high-magnification multi-views. The proposed approach has been tested using GoogLeNet architecture of CNN on an image dataset of 37 breast cytopathology samples (24 benign and 13 malignant), where the network was trained using images of ~54% cell samples and tested on the rest, achieving 89.7% mean accuracy in 8 fold validation.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"7 1","pages":"828-833"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88358026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Finding Anomalies with Generative Adversarial Networks for a Patrolbot 基于生成对抗网络的巡逻机器人异常发现
W. Lawson, Esube Bekele, Keith Sullivan
We present an anomaly detection system based on an autonomous robot performing a patrol task. Using a generative adversarial network (GAN), we compare the robot's current view with a learned model of normality. Our preliminary experimental results show that the approach is well suited for anomaly detection, providing efficient results with a low false positive rate.
我们提出了一种基于自主机器人执行巡逻任务的异常检测系统。使用生成对抗网络(GAN),我们将机器人的当前视图与学习到的正态性模型进行比较。我们的初步实验结果表明,该方法非常适合于异常检测,提供了低误报率的高效结果。
{"title":"Finding Anomalies with Generative Adversarial Networks for a Patrolbot","authors":"W. Lawson, Esube Bekele, Keith Sullivan","doi":"10.1109/CVPRW.2017.68","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.68","url":null,"abstract":"We present an anomaly detection system based on an autonomous robot performing a patrol task. Using a generative adversarial network (GAN), we compare the robot's current view with a learned model of normality. Our preliminary experimental results show that the approach is well suited for anomaly detection, providing efficient results with a low false positive rate.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"1 1","pages":"484-485"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90439113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Court-Based Volleyball Video Summarization Focusing on Rally Scene 以拉力赛场景为重点的基于球场的排球视频总结
Takahiro Itazuri, Tsukasa Fukusato, Shugo Yamaguchi, S. Morishima
In this paper, we propose a video summarization system for volleyball videos. Our system automatically detects rally scenes as self-consumable video segments and evaluates rally-rank for each rally scene to decide priority. In the priority decision, features representing the contents of the game are necessary; however such features have not been considered in most previous methods. Although several visual features such as the position of a ball and players should be used, acquisition of such features is still non-robust and unreliable in low resolution or low frame rate volleyball videos. Instead, we utilize the court transition information caused by camera operation. Experimental results demonstrate the robustness of our rally scene detection and the effectiveness of our rally-rank to reflect viewers' preferences over previous methods.
本文提出了一种针对排球视频的视频摘要系统。我们的系统自动检测拉力赛场景作为自消费视频片段,并评估每个拉力赛场景的拉力赛排名来决定优先级。在优先级决策中,代表游戏内容的功能是必要的;然而,在大多数以前的方法中没有考虑到这些特征。虽然应该使用一些视觉特征,如球和球员的位置,但在低分辨率或低帧率排球视频中,这些特征的获取仍然是非鲁棒的和不可靠的。相反,我们利用摄像机操作产生的球场过渡信息。实验结果证明了我们的拉力赛场景检测的鲁棒性,以及我们的拉力赛排名比以前的方法更能反映观众的偏好。
{"title":"Court-Based Volleyball Video Summarization Focusing on Rally Scene","authors":"Takahiro Itazuri, Tsukasa Fukusato, Shugo Yamaguchi, S. Morishima","doi":"10.1109/CVPRW.2017.28","DOIUrl":"https://doi.org/10.1109/CVPRW.2017.28","url":null,"abstract":"In this paper, we propose a video summarization system for volleyball videos. Our system automatically detects rally scenes as self-consumable video segments and evaluates rally-rank for each rally scene to decide priority. In the priority decision, features representing the contents of the game are necessary; however such features have not been considered in most previous methods. Although several visual features such as the position of a ball and players should be used, acquisition of such features is still non-robust and unreliable in low resolution or low frame rate volleyball videos. Instead, we utilize the court transition information caused by camera operation. Experimental results demonstrate the robustness of our rally scene detection and the effectiveness of our rally-rank to reflect viewers' preferences over previous methods.","PeriodicalId":6668,"journal":{"name":"2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"70 1","pages":"179-186"},"PeriodicalIF":0.0,"publicationDate":"2017-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72979838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1