首页 > 最新文献

2017 14th Conference on Computer and Robot Vision (CRV)最新文献

英文 中文
Person Following Robot Using Selected Online Ada-Boosting with Stereo Camera 基于立体摄像头的人跟踪机器人选择在线数据增强
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.55
B. Chen, Raghavender Sahdev, John K. Tsotsos
Person following behavior is an important task for social robots. To enable robots to follow a person, we have to track the target in real-time without critical failures. There are many situations where the robot will potentially loose tracking in a dynamic environment, e.g., occlusion, illumination, pose-changes, etc. Often, people use a complex tracking algorithm to improve robustness. However, the trade-off is that their approaches may not able to run in real-time on mobile robots. In this paper, we present Selected Online Ada-Boosting (SOAB) technique, a modified Online Ada-Boosting (OAB) tracking algorithm with integrated scene depth information obtained from a stereo camera which runs in real-time on a mobile robot. We build and share our results on the performance of our technique on a new stereo dataset for the task of person following. The dataset covers different challenging situations like squatting, partial and complete occlusion of the target being tracked, people wearing similar clothes, appearance changes, walking facing the front and back side of the person to the robot, and normal walking.
人的行为跟踪是社交机器人的一项重要任务。为了使机器人能够跟随人,我们必须在没有严重故障的情况下实时跟踪目标。在许多情况下,机器人在动态环境中可能会失去跟踪,例如,遮挡,照明,姿势变化等。通常,人们使用复杂的跟踪算法来提高鲁棒性。然而,代价是他们的方法可能无法在移动机器人上实时运行。在本文中,我们提出了一种改进的在线Ada-Boosting (OAB)跟踪算法,该算法集成了实时运行在移动机器人上的立体摄像机获取的场景深度信息。我们建立并分享了我们的技术在一个新的立体数据集上的性能结果,用于人跟随任务。该数据集涵盖了不同的具有挑战性的情况,如蹲着、被跟踪目标的部分和完全遮挡、穿着相似衣服的人、外表变化、面向机器人的人的前后侧行走以及正常行走。
{"title":"Person Following Robot Using Selected Online Ada-Boosting with Stereo Camera","authors":"B. Chen, Raghavender Sahdev, John K. Tsotsos","doi":"10.1109/CRV.2017.55","DOIUrl":"https://doi.org/10.1109/CRV.2017.55","url":null,"abstract":"Person following behavior is an important task for social robots. To enable robots to follow a person, we have to track the target in real-time without critical failures. There are many situations where the robot will potentially loose tracking in a dynamic environment, e.g., occlusion, illumination, pose-changes, etc. Often, people use a complex tracking algorithm to improve robustness. However, the trade-off is that their approaches may not able to run in real-time on mobile robots. In this paper, we present Selected Online Ada-Boosting (SOAB) technique, a modified Online Ada-Boosting (OAB) tracking algorithm with integrated scene depth information obtained from a stereo camera which runs in real-time on a mobile robot. We build and share our results on the performance of our technique on a new stereo dataset for the task of person following. The dataset covers different challenging situations like squatting, partial and complete occlusion of the target being tracked, people wearing similar clothes, appearance changes, walking facing the front and back side of the person to the robot, and normal walking.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116854494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Towards an Improved Vision-Based Web Page Segmentation Algorithm 一种改进的基于视觉的网页分割算法
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.38
M. Cormier, R. Mann, Karyn Moffatt, R. Cohen
In this paper we introduce an edge-based segmentation algorithm designed for web pages. We consider each web page as an image and perform segmentation as the initial stage of a planned parsing system that will also include region classification. The motivation for our work is to enable improved online experiences for users with assistive needs (serving as the back-end process for such front-end tasks as zooming and decluttering the image being presented to those with visual or cognitive challenges, or producing less unwieldy output from screenreaders). Our focus is therefore on the interpretation of a class of man-made images (where web pages consist of one particular set of these images which have important constraints that assist in performing the processing). After clarifying some comparisons with an earlier model of ours, we show validation for our method. Following this, we briefly discuss the contribution for the field of computer vision, offering a contrast with current work in segmentation focused on the processing of natural images.
本文介绍了一种基于边缘的网页分割算法。我们将每个网页视为图像,并将分割作为计划解析系统的初始阶段,该系统还将包括区域分类。我们工作的动机是为有辅助需求的用户提供更好的在线体验(作为前端任务的后端流程,如缩放和整理呈现给有视觉或认知挑战的用户的图像,或从屏幕阅读器产生更少笨拙的输出)。因此,我们的重点是对一类人造图像的解释(其中网页由一组特定的图像组成,这些图像具有协助执行处理的重要约束)。在澄清了与我们早期模型的一些比较之后,我们证明了我们的方法的有效性。在此之后,我们简要讨论了计算机视觉领域的贡献,并与当前专注于自然图像处理的分割工作进行了对比。
{"title":"Towards an Improved Vision-Based Web Page Segmentation Algorithm","authors":"M. Cormier, R. Mann, Karyn Moffatt, R. Cohen","doi":"10.1109/CRV.2017.38","DOIUrl":"https://doi.org/10.1109/CRV.2017.38","url":null,"abstract":"In this paper we introduce an edge-based segmentation algorithm designed for web pages. We consider each web page as an image and perform segmentation as the initial stage of a planned parsing system that will also include region classification. The motivation for our work is to enable improved online experiences for users with assistive needs (serving as the back-end process for such front-end tasks as zooming and decluttering the image being presented to those with visual or cognitive challenges, or producing less unwieldy output from screenreaders). Our focus is therefore on the interpretation of a class of man-made images (where web pages consist of one particular set of these images which have important constraints that assist in performing the processing). After clarifying some comparisons with an earlier model of ours, we show validation for our method. Following this, we briefly discuss the contribution for the field of computer vision, offering a contrast with current work in segmentation focused on the processing of natural images.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132119729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Self-Organization of a Robot Swarm into Concentric Shapes 同心圆机器人群的自组织
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.58
Geoff Nagy, R. Vaughan
In this paper, we show how a swarm of differential-drive robots can self-organize into multiple nested layers of a given shape. A key component of our work is the reliance on inter-robot collisions to provide information on how the formation should grow. We describe a simple controller and experimentally evaluate how its performance scales as the number of robots in the swarm increases from tens to several hundred robots. The average quality of the formation is shown to be a linearly decreasing function of swarm size, although the steepness of this line depends on the complexity of the formation. We also show that the time for a swarm to form a given shape does not grow quickly even as the number of robots in the swarm increases by a large amount.
在本文中,我们展示了一群差动驱动机器人如何自组织成给定形状的多个嵌套层。我们工作的一个关键组成部分是依赖于机器人之间的碰撞来提供关于队形应该如何增长的信息。我们描述了一个简单的控制器,并通过实验评估了当群中的机器人数量从几十个增加到几百个时,它的性能是如何变化的。地层的平均质量显示为群大小的线性递减函数,尽管这条线的陡峭程度取决于地层的复杂性。我们还表明,即使群体中的机器人数量大量增加,群体形成给定形状的时间也不会迅速增长。
{"title":"Self-Organization of a Robot Swarm into Concentric Shapes","authors":"Geoff Nagy, R. Vaughan","doi":"10.1109/CRV.2017.58","DOIUrl":"https://doi.org/10.1109/CRV.2017.58","url":null,"abstract":"In this paper, we show how a swarm of differential-drive robots can self-organize into multiple nested layers of a given shape. A key component of our work is the reliance on inter-robot collisions to provide information on how the formation should grow. We describe a simple controller and experimentally evaluate how its performance scales as the number of robots in the swarm increases from tens to several hundred robots. The average quality of the formation is shown to be a linearly decreasing function of swarm size, although the steepness of this line depends on the complexity of the formation. We also show that the time for a swarm to form a given shape does not grow quickly even as the number of robots in the swarm increases by a large amount.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129646195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Collaborative Sampling Using Heterogeneous Marine Robots Driven by Visual Cues 基于视觉线索驱动的异构海洋机器人协同采样
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.49
Sandeep Manjanna, Johanna Hansen, Alberto Quattrini Li, Ioannis M. Rekleitis, G. Dudek
This paper addresses distributed data sampling in marine environments using robotic devices. We present a method to strategically sample locally observable features using two classes of sensor platforms. Our system consists of a sophisticated autonomous surface vehicle (ASV) which strategically samples based on information provided by a team of inexpensive sensor nodes. The sensor nodes effectively extend the observational capabilities of the vehicle by capturing georeferenced samples from disparate and moving points across the region. The ASV uses this information, along with its own observations, to plan a path so as to sample points which it expects to be particularly informative. We compare our approach to a traditional exhaustive survey approach and show that we are able to effectively represent a region with less energy expenditure. We validate our approach through simulations and test the system on real robots in field.
本文讨论了利用机器人设备在海洋环境中进行分布式数据采样。我们提出了一种使用两类传感器平台对局部可观察特征进行策略性采样的方法。我们的系统由一个复杂的自动水面车辆(ASV)组成,它根据一组廉价的传感器节点提供的信息进行策略性采样。传感器节点通过从整个区域的不同和移动点捕获地理参考样本,有效地扩展了车辆的观测能力。ASV使用这些信息,连同它自己的观察,来规划一条路径,以便对它期望特别有用的点进行采样。我们将我们的方法与传统的详尽调查方法进行比较,并表明我们能够有效地代表一个能源消耗较少的地区。我们通过仿真验证了我们的方法,并在真实的机器人上进行了现场测试。
{"title":"Collaborative Sampling Using Heterogeneous Marine Robots Driven by Visual Cues","authors":"Sandeep Manjanna, Johanna Hansen, Alberto Quattrini Li, Ioannis M. Rekleitis, G. Dudek","doi":"10.1109/CRV.2017.49","DOIUrl":"https://doi.org/10.1109/CRV.2017.49","url":null,"abstract":"This paper addresses distributed data sampling in marine environments using robotic devices. We present a method to strategically sample locally observable features using two classes of sensor platforms. Our system consists of a sophisticated autonomous surface vehicle (ASV) which strategically samples based on information provided by a team of inexpensive sensor nodes. The sensor nodes effectively extend the observational capabilities of the vehicle by capturing georeferenced samples from disparate and moving points across the region. The ASV uses this information, along with its own observations, to plan a path so as to sample points which it expects to be particularly informative. We compare our approach to a traditional exhaustive survey approach and show that we are able to effectively represent a region with less energy expenditure. We validate our approach through simulations and test the system on real robots in field.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"145 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123052958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Pitch and Roll Camera Orientation from a Single 2D Image Using Convolutional Neural Networks 使用卷积神经网络从单个2D图像中获得俯仰和滚动相机方向
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.53
Greg Olmschenk, Hao Tang, Zhigang Zhu
In this paper, we propose using convolutional neural networks (CNNs) to automatically determine the pitch and roll of a camera using a single, scene agnostic, 2D image. We compared a linear regressor, a two-layer neural network, and two CNNs. We show the CNNs produce high levels of accuracy in estimating the ground truth orientations which can be used in various computer vision tasks where calculating the camera orientation is necessary or useful. By utilizing accelerometer data in an existing image dataset, we were able to provide the large camera orientation ground truth dataset needed to train such a network with approximately correct values. The trained network is then fine-tuned to smaller datasets with exact camera orientation labels. Additionally, the network is fine-tuned to a dataset with different intrinsic camera parameters to demonstrate the transferability of the network.
在本文中,我们建议使用卷积神经网络(cnn)来自动确定相机的俯仰和滚动,使用单个场景不可知的2D图像。我们比较了线性回归器、双层神经网络和两个cnn。我们展示了cnn在估计地面真值方向方面具有很高的精度,这可以用于计算相机方向是必要或有用的各种计算机视觉任务。通过利用现有图像数据集中的加速度计数据,我们能够提供训练这种网络所需的大型相机方向地面真值数据集,并具有近似正确的值。然后,经过训练的网络被微调到具有精确相机方向标签的更小的数据集。此外,该网络还对具有不同内在相机参数的数据集进行了微调,以证明网络的可移植性。
{"title":"Pitch and Roll Camera Orientation from a Single 2D Image Using Convolutional Neural Networks","authors":"Greg Olmschenk, Hao Tang, Zhigang Zhu","doi":"10.1109/CRV.2017.53","DOIUrl":"https://doi.org/10.1109/CRV.2017.53","url":null,"abstract":"In this paper, we propose using convolutional neural networks (CNNs) to automatically determine the pitch and roll of a camera using a single, scene agnostic, 2D image. We compared a linear regressor, a two-layer neural network, and two CNNs. We show the CNNs produce high levels of accuracy in estimating the ground truth orientations which can be used in various computer vision tasks where calculating the camera orientation is necessary or useful. By utilizing accelerometer data in an existing image dataset, we were able to provide the large camera orientation ground truth dataset needed to train such a network with approximately correct values. The trained network is then fine-tuned to smaller datasets with exact camera orientation labels. Additionally, the network is fine-tuned to a dataset with different intrinsic camera parameters to demonstrate the transferability of the network.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124556812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Effect of Denoising Algorithms on Video Stabilization 去噪算法对视频稳定的影响
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.34
Abdelrahman Ahmed, M. Shehata
Various denoising algorithms exist in the literature, however, no studies have ever been made to measure the impact of denoising algorithms on the quality of the video produced by a stabilization algorithm. In this paper, the impact of state of the art denoising algorithms on a feature-based video stabilization is measured and evaluated. Also, a quantitative measure is proposed which can give more insight on the impact of the chosen denoising algorithm on stabilization. The results show that the denoising algorithm can drastically affect the quality of stabilization results and choosing the latest denoising algorithm does not always guarantee the best stabilization results.
文献中有各种各样的去噪算法,但是没有研究衡量去噪算法对稳定化算法产生的视频质量的影响。在本文中,测量和评估了当前最先进的去噪算法对基于特征的视频防抖的影响。此外,本文还提出了一种定量度量方法,可以更深入地了解所选择的去噪算法对稳定化的影响。结果表明,去噪算法会极大地影响稳定结果的质量,选择最新的去噪算法并不一定能保证最佳的稳定结果。
{"title":"Effect of Denoising Algorithms on Video Stabilization","authors":"Abdelrahman Ahmed, M. Shehata","doi":"10.1109/CRV.2017.34","DOIUrl":"https://doi.org/10.1109/CRV.2017.34","url":null,"abstract":"Various denoising algorithms exist in the literature, however, no studies have ever been made to measure the impact of denoising algorithms on the quality of the video produced by a stabilization algorithm. In this paper, the impact of state of the art denoising algorithms on a feature-based video stabilization is measured and evaluated. Also, a quantitative measure is proposed which can give more insight on the impact of the chosen denoising algorithm on stabilization. The results show that the denoising algorithm can drastically affect the quality of stabilization results and choosing the latest denoising algorithm does not always guarantee the best stabilization results.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"65 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132678890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Saliency of an Object Using Genetic Algorithm 利用遗传算法增强目标的显著性
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.33
R. Pal, Dipanjan Roy
It is often required to emphasize an object in an image. Artists, illustrators, cinematographers and photographers have long used the principles of contrast and composition to guide visual attention. In order to achieve this, a novel perceptually-driven approach is put forth which leads to the enhancement of visual saliency of target object without destroying the naturalness of the contents of the image. The proposed approach computes new feature values for the intended object by maximizing the feature dissimilarity (which is weighted by positional proximity) with other objects. Too much change in feature values in the target segment may destroy naturality of the image. This poses as the constraint in the proposed maximization problem. Genetic algorithm has been used, in this context, to find the feature values which maximize the saliency of the target object. Experimental validation through objective evaluation metrics using saliency maps, as well as analysis of eye-tracking data, establish the success of the proposed method.
通常需要在图像中强调一个对象。长期以来,艺术家、插画家、电影摄影师和摄影师一直使用对比和构图的原则来引导视觉注意力。为了实现这一目标,提出了一种新的感知驱动方法,在不破坏图像内容的自然性的情况下增强目标物体的视觉显著性。该方法通过最大化目标与其他目标的特征不相似性(通过位置接近度加权)来计算目标的新特征值。目标段的特征值变化太大可能会破坏图像的自然性。这是所提最大化问题的约束条件。在这种情况下,遗传算法被用来寻找使目标对象的显著性最大化的特征值。通过使用显著性图进行客观评价指标的实验验证,以及对眼动追踪数据的分析,验证了所提方法的有效性。
{"title":"Enhancing Saliency of an Object Using Genetic Algorithm","authors":"R. Pal, Dipanjan Roy","doi":"10.1109/CRV.2017.33","DOIUrl":"https://doi.org/10.1109/CRV.2017.33","url":null,"abstract":"It is often required to emphasize an object in an image. Artists, illustrators, cinematographers and photographers have long used the principles of contrast and composition to guide visual attention. In order to achieve this, a novel perceptually-driven approach is put forth which leads to the enhancement of visual saliency of target object without destroying the naturalness of the contents of the image. The proposed approach computes new feature values for the intended object by maximizing the feature dissimilarity (which is weighted by positional proximity) with other objects. Too much change in feature values in the target segment may destroy naturality of the image. This poses as the constraint in the proposed maximization problem. Genetic algorithm has been used, in this context, to find the feature values which maximize the saliency of the target object. Experimental validation through objective evaluation metrics using saliency maps, as well as analysis of eye-tracking data, establish the success of the proposed method.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126365675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Fast and Accurate Tracking of Highly Deformable Heart Valves with Locally Constrained Level Sets 基于局部约束水平集的高度可变形心脏瓣膜快速准确跟踪
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.13
A. Burden, Melissa Cote, A. Albu
This paper focuses on the automatic quantitative performance analysis of bioprosthetic heart valves from video footage acquired during in vitro testing. Bioprosthetic heart valves, mimicking the shape and functionality of a human heart valve, are routinely used in valve replacement procedures to substitute defective native valves. Their reliability in both functionality and durability is crucial to the patients' well-being, as such, valve designs must be rigorously tested before deployment. A key quality metric of a heart valve design is the cyclical temporal evolution of the valve's area. This metric is typically computed manually from input video data, a time-consuming and error-prone task. We propose a novel, cost-effective approach for the automatic tracking and segmentation of valve orifices that integrates a probabilistic motion boundary model into a distance regularized level set evolution formulation. The proposed method constrains the level set evolution domain using data about characteristic motion patterns of heart valves. Experiments including comparisons with two other methods demonstrate the value of the proposed approach on three levels: an improved segmented orifice shape accuracy, a greater computational efficiency, and a better ability to identify video frames with orifice area content (open valve).
本文主要研究了体外测试过程中获得的视频片段对生物假体心脏瓣膜的自动定量性能分析。生物人工心脏瓣膜,模仿人类心脏瓣膜的形状和功能,通常用于瓣膜置换手术,以替代有缺陷的天然瓣膜。它们在功能和耐用性方面的可靠性对患者的健康至关重要,因此,瓣膜的设计必须在部署之前经过严格的测试。心脏瓣膜设计的一个关键质量指标是瓣膜面积的周期性时间演变。该指标通常是根据输入的视频数据手动计算的,这是一项耗时且容易出错的任务。我们提出了一种新颖的、经济有效的方法来自动跟踪和分割阀口,该方法将概率运动边界模型集成到距离正则化水平集进化公式中。该方法利用心脏瓣膜特征运动模式数据对水平集演化域进行约束。实验包括与其他两种方法的比较,证明了该方法在三个层面上的价值:改进的分段孔板形状精度,更高的计算效率,以及更好的识别具有孔板面积内容(打开阀)的视频帧的能力。
{"title":"Fast and Accurate Tracking of Highly Deformable Heart Valves with Locally Constrained Level Sets","authors":"A. Burden, Melissa Cote, A. Albu","doi":"10.1109/CRV.2017.13","DOIUrl":"https://doi.org/10.1109/CRV.2017.13","url":null,"abstract":"This paper focuses on the automatic quantitative performance analysis of bioprosthetic heart valves from video footage acquired during in vitro testing. Bioprosthetic heart valves, mimicking the shape and functionality of a human heart valve, are routinely used in valve replacement procedures to substitute defective native valves. Their reliability in both functionality and durability is crucial to the patients' well-being, as such, valve designs must be rigorously tested before deployment. A key quality metric of a heart valve design is the cyclical temporal evolution of the valve's area. This metric is typically computed manually from input video data, a time-consuming and error-prone task. We propose a novel, cost-effective approach for the automatic tracking and segmentation of valve orifices that integrates a probabilistic motion boundary model into a distance regularized level set evolution formulation. The proposed method constrains the level set evolution domain using data about characteristic motion patterns of heart valves. Experiments including comparisons with two other methods demonstrate the value of the proposed approach on three levels: an improved segmented orifice shape accuracy, a greater computational efficiency, and a better ability to identify video frames with orifice area content (open valve).","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126655143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Automatic Photo Orientation Detection with Convolutional Neural Networks 基于卷积神经网络的照片方向自动检测
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.59
Ujash Joshi, Michael Guerzhoy
We apply convolutional neural networks (CNN) to the problem of image orientation detection in the context of determining the correct orientation (from 0, 90, 180, and 270 degrees) of a consumer photo. The problem is especially important for digitazing analog photographs. We substantially improve on the published state of the art in terms of the performance on one of the standard datasets, and test our system on a more difficult large dataset of consumer photos. We use Guided Backpropagation to obtain insights into how our CNN detects photo orientation, and to explain its mistakes.
我们将卷积神经网络(CNN)应用于确定消费者照片的正确方向(0度、90度、180度和270度)的图像方向检测问题。这个问题对于模拟照片的数字化尤其重要。我们在一个标准数据集的性能方面大大改进了已发布的最新技术,并在一个更困难的消费者照片大型数据集上测试了我们的系统。我们使用引导反向传播来深入了解我们的CNN如何检测照片方向,并解释其错误。
{"title":"Automatic Photo Orientation Detection with Convolutional Neural Networks","authors":"Ujash Joshi, Michael Guerzhoy","doi":"10.1109/CRV.2017.59","DOIUrl":"https://doi.org/10.1109/CRV.2017.59","url":null,"abstract":"We apply convolutional neural networks (CNN) to the problem of image orientation detection in the context of determining the correct orientation (from 0, 90, 180, and 270 degrees) of a consumer photo. The problem is especially important for digitazing analog photographs. We substantially improve on the published state of the art in terms of the performance on one of the standard datasets, and test our system on a more difficult large dataset of consumer photos. We use Guided Backpropagation to obtain insights into how our CNN detects photo orientation, and to explain its mistakes.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"187 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132576402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Fully Automatic, Real-Time Vehicle Tracking for Surveillance Video 全自动,实时车辆跟踪监控视频
Pub Date : 2017-05-16 DOI: 10.1109/CRV.2017.43
Yanzi Jin, Jakob Eriksson
We present an object tracking framework which fuses multiple unstable video-based methods and supports automatic tracker initialization and termination. To evaluate our system, we collected a large dataset of hand-annotated 5-minute traffic surveillance videos, which we are releasing to the community. To the best of our knowledge, this is the first publicly available dataset of such long videos, providing a diverse range of real-world object variation, scale change, interaction, different resolutions and illumination conditions. In our comprehensive evaluation using this dataset, we show that our automatic object tracking system often outperforms state-of-the-art trackers, even when these are provided with proper manual initialization. We also demonstrate tracking throughput improvements of 5× or more vs. the competition.
提出了一种融合了多种不稳定视频跟踪方法的目标跟踪框架,并支持自动跟踪初始化和终止。为了评估我们的系统,我们收集了一个大型数据集,其中包括手动注释的5分钟交通监控视频,我们将向社区发布这些视频。据我们所知,这是第一个公开的长视频数据集,提供了各种现实世界的物体变化、尺度变化、交互、不同的分辨率和照明条件。在我们使用该数据集的综合评估中,我们表明,我们的自动对象跟踪系统通常优于最先进的跟踪器,即使这些跟踪器提供了适当的手动初始化。我们还演示了与竞争对手相比,跟踪吞吐量提高了5倍或更多。
{"title":"Fully Automatic, Real-Time Vehicle Tracking for Surveillance Video","authors":"Yanzi Jin, Jakob Eriksson","doi":"10.1109/CRV.2017.43","DOIUrl":"https://doi.org/10.1109/CRV.2017.43","url":null,"abstract":"We present an object tracking framework which fuses multiple unstable video-based methods and supports automatic tracker initialization and termination. To evaluate our system, we collected a large dataset of hand-annotated 5-minute traffic surveillance videos, which we are releasing to the community. To the best of our knowledge, this is the first publicly available dataset of such long videos, providing a diverse range of real-world object variation, scale change, interaction, different resolutions and illumination conditions. In our comprehensive evaluation using this dataset, we show that our automatic object tracking system often outperforms state-of-the-art trackers, even when these are provided with proper manual initialization. We also demonstrate tracking throughput improvements of 5× or more vs. the competition.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134062051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2017 14th Conference on Computer and Robot Vision (CRV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1