首页 > 最新文献

2017 14th Conference on Computer and Robot Vision (CRV)最新文献

英文 中文
Developing a Cubature Multi-state Constraint Kalman Filter for Visual-Inertial Navigation System 一种用于视觉惯性导航系统的多状态约束卡尔曼滤波
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.19
Trung Nguyen, G. Mann, A. Vardy, R. Gosine
The objective of this paper is to develop a cubature Multi-State Constraint Kalman Filter (MSCKF) for a VisualInertial Navigation System (VINS). MSCKF is a tightly-coupled EKF-based filter operating over a sliding window of multiple sequent states. In order to decrease the complexity and the computational cost of the original EKF-based measurement, the measurement model is built on the Trifocal Tensor Geometry (TTG). The predicted measurement does not need to reconstruct the 3D position of the visual landmarks. In order to employ that nonlinear TTG-based measurement model, this paper will implement cubature approach (i.e. popularly associated with Cubature Kalman Filter (CKF)). Compared to other advanced nonlinear filter, specifically Unscented Kalman Filter (UKF), the CKF has removed the positive-definite condition of the covariance matrix computation, which may halt or fail the filter operation. The proposed filter is validated with three KITTI datasets [1] of residential area to evaluate its performance.
本文的目的是开发一种用于视觉惯性导航系统(VINS)的多状态约束卡尔曼滤波器(MSCKF)。MSCKF是一种紧耦合的基于ekf的滤波器,在多个连续状态的滑动窗口上运行。为了降低原来基于ekf的测量的复杂性和计算成本,在三焦张量几何(TTG)上建立了测量模型。预测的测量不需要重建视觉地标的三维位置。为了采用基于非线性ttg的测量模型,本文将实现cubature方法(即通常与cubature Kalman Filter (CKF)相关联)。与其他先进的非线性滤波器,特别是Unscented卡尔曼滤波器(UKF)相比,CKF去除了协方差矩阵计算的正定条件,这可能会导致滤波操作停止或失败。用3个居民区的KITTI数据集[1]对该滤波器进行了验证,以评价其性能。
{"title":"Developing a Cubature Multi-state Constraint Kalman Filter for Visual-Inertial Navigation System","authors":"Trung Nguyen, G. Mann, A. Vardy, R. Gosine","doi":"10.1109/CRV.2017.19","DOIUrl":"https://doi.org/10.1109/CRV.2017.19","url":null,"abstract":"The objective of this paper is to develop a cubature Multi-State Constraint Kalman Filter (MSCKF) for a VisualInertial Navigation System (VINS). MSCKF is a tightly-coupled EKF-based filter operating over a sliding window of multiple sequent states. In order to decrease the complexity and the computational cost of the original EKF-based measurement, the measurement model is built on the Trifocal Tensor Geometry (TTG). The predicted measurement does not need to reconstruct the 3D position of the visual landmarks. In order to employ that nonlinear TTG-based measurement model, this paper will implement cubature approach (i.e. popularly associated with Cubature Kalman Filter (CKF)). Compared to other advanced nonlinear filter, specifically Unscented Kalman Filter (UKF), the CKF has removed the positive-definite condition of the covariance matrix computation, which may halt or fail the filter operation. The proposed filter is validated with three KITTI datasets [1] of residential area to evaluate its performance.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129352101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Bifurcation Localization in 3D Images via Evolutionary Geometric Deformable Templates 基于进化几何变形模板的三维图像分岔定位
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.12
Mengliu Zhao, G. Hamarneh
Given the importance of studying bifurcations in 3D anatomical trees (e.g. vasculature and airway), we propose a bifurcation detector that operates by fitting a parametric geometric deformable model to 3D medical images. A fitness function is designed to integrate features along the model skeletons, surfaces and internal areas. To overcome local optima while detecting multiple bifurcations in a single image, we adopt genetic algorithm with a tribes niching technique. Results on both VascuSynth data and clinical CT data demonstrate not only high bifurcation detection accuracy and stability, but the ability to locate parent and children branch directions and vessel wall locations simultaneously.
考虑到研究三维解剖树(如脉管系统和气道)分支的重要性,我们提出了一种分支检测器,该检测器通过拟合参数化几何可变形模型来操作三维医学图像。适应度函数被设计用来整合模型骨架、表面和内部区域的特征。为了在单幅图像中检测多个分支时克服局部最优,我们采用了带有部落小生境技术的遗传算法。结果表明,VascuSynth数据和临床CT数据不仅具有较高的分支检测准确性和稳定性,而且能够同时定位父母和孩子的分支方向和血管壁位置。
{"title":"Bifurcation Localization in 3D Images via Evolutionary Geometric Deformable Templates","authors":"Mengliu Zhao, G. Hamarneh","doi":"10.1109/CRV.2017.12","DOIUrl":"https://doi.org/10.1109/CRV.2017.12","url":null,"abstract":"Given the importance of studying bifurcations in 3D anatomical trees (e.g. vasculature and airway), we propose a bifurcation detector that operates by fitting a parametric geometric deformable model to 3D medical images. A fitness function is designed to integrate features along the model skeletons, surfaces and internal areas. To overcome local optima while detecting multiple bifurcations in a single image, we adopt genetic algorithm with a tribes niching technique. Results on both VascuSynth data and clinical CT data demonstrate not only high bifurcation detection accuracy and stability, but the ability to locate parent and children branch directions and vessel wall locations simultaneously.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134635485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Index Structure for Fast Range Search in Hamming Space 一种用于Hamming空间快速范围搜索的索引结构
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.37
E. M. Reina, K. Pu, F. Qureshi
This paper addresses the problem of indexing and querying very large databases of binary vectors. Such databases of binary vectors are a common occurrence in domains such as information retrieval and computer vision. We propose an indexing structure consisting of a compressed bitwise trie and a hash table for supporting range queries in Hamming space. The index structure, which can be updated incrementally, is able to solve the range queries for any radius. Our approach significantly outperforms state-of-the-art approaches.
本文解决了索引和查询超大型二进制向量数据库的问题。这种二值向量数据库在信息检索和计算机视觉等领域中很常见。我们提出了一个由压缩的按位树和哈希表组成的索引结构,用于支持汉明空间中的范围查询。可以增量更新的索引结构能够解决任何半径的范围查询。我们的方法明显优于最先进的方法。
{"title":"An Index Structure for Fast Range Search in Hamming Space","authors":"E. M. Reina, K. Pu, F. Qureshi","doi":"10.1109/CRV.2017.37","DOIUrl":"https://doi.org/10.1109/CRV.2017.37","url":null,"abstract":"This paper addresses the problem of indexing and querying very large databases of binary vectors. Such databases of binary vectors are a common occurrence in domains such as information retrieval and computer vision. We propose an indexing structure consisting of a compressed bitwise trie and a hash table for supporting range queries in Hamming space. The index structure, which can be updated incrementally, is able to solve the range queries for any radius. Our approach significantly outperforms state-of-the-art approaches.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115304876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Convolutional Residual Network for Grasp Localization 基于卷积残差网络的抓握定位
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.14
Ludovic Trottier, P. Giguère, B. Chaib-draa
Object grasping is an important ability for carrying out complex manipulation tasks with autonomous robotic systems. The grasp localization module plays an essential role in the success of the grasp maneuver. Generally viewed as a vision perception problem, its goal is determining regions of high graspability by interpreting light and depth information. Over the past few years, several works in Deep Learning (DL) have shown the high potential of Convolutional Neural Networks (CNNs) for solving vision-related problems. Advances in residual networks have further facilitated neural network training by improving convergence time and generalization performances with identity skip connections and residual mappings. In this paper, we investigate the use of residual networks for grasp localization. A standard residual CNN for object recognition uses a global average pooling layer prior to the fully-connected layers. Our experiments have shown that this pooling layer removes the spatial correlation in the back-propagated error signal, and this prevents the network from correctly localizing good grasp regions. We propose an architecture modification that removes this limitation. Our experiments on the Cornell task have shown that our network obtained state-of-the-art performances of 10.85% and 11.86% rectangle metric error on image-wise and object-wise splits respectively. We did not use pre-training but rather opted for on-line data augmentation for managing overfitting. In comparison to previous approach that employed off-line data augmentation, our network used 15x fewer observations, which significantly reduced training time.
物体抓取是自主机器人完成复杂操作任务的一项重要能力。抓握定位模块对抓握机动的成功与否起着至关重要的作用。它通常被视为一个视觉感知问题,其目标是通过解释光和深度信息来确定高可抓性的区域。在过去的几年里,深度学习(DL)领域的几项研究都显示了卷积神经网络(cnn)在解决视觉相关问题方面的巨大潜力。残差网络的进步通过提高收敛时间和使用身份跳过连接和残差映射的泛化性能,进一步促进了神经网络的训练。在本文中,我们研究了残差网络在抓取定位中的应用。用于目标识别的标准残差CNN在完全连接层之前使用全局平均池化层。我们的实验表明,池化层消除了反向传播误差信号中的空间相关性,这阻碍了网络正确定位好的抓取区域。我们建议对架构进行修改,以消除这一限制。我们在康奈尔任务上的实验表明,我们的网络在图像和对象分割上分别获得了10.85%和11.86%的矩形度量误差。我们没有使用预训练,而是选择在线数据增强来管理过拟合。与之前使用离线数据增强的方法相比,我们的网络使用的观察值减少了15倍,这大大减少了训练时间。
{"title":"Convolutional Residual Network for Grasp Localization","authors":"Ludovic Trottier, P. Giguère, B. Chaib-draa","doi":"10.1109/CRV.2017.14","DOIUrl":"https://doi.org/10.1109/CRV.2017.14","url":null,"abstract":"Object grasping is an important ability for carrying out complex manipulation tasks with autonomous robotic systems. The grasp localization module plays an essential role in the success of the grasp maneuver. Generally viewed as a vision perception problem, its goal is determining regions of high graspability by interpreting light and depth information. Over the past few years, several works in Deep Learning (DL) have shown the high potential of Convolutional Neural Networks (CNNs) for solving vision-related problems. Advances in residual networks have further facilitated neural network training by improving convergence time and generalization performances with identity skip connections and residual mappings. In this paper, we investigate the use of residual networks for grasp localization. A standard residual CNN for object recognition uses a global average pooling layer prior to the fully-connected layers. Our experiments have shown that this pooling layer removes the spatial correlation in the back-propagated error signal, and this prevents the network from correctly localizing good grasp regions. We propose an architecture modification that removes this limitation. Our experiments on the Cornell task have shown that our network obtained state-of-the-art performances of 10.85% and 11.86% rectangle metric error on image-wise and object-wise splits respectively. We did not use pre-training but rather opted for on-line data augmentation for managing overfitting. In comparison to previous approach that employed off-line data augmentation, our network used 15x fewer observations, which significantly reduced training time.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133659191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Depth Estimation of Semi-submerged Objects Using a Light-Field Camera 基于光场相机的半水下目标深度估计
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.44
Juehui Fan, Herbert Yang
We present an algorithm to estimate depth of realworld scenes containing an object semi-submerged in water using a light field camera. Existing hand-held consumer light field cameras are well-suited for automated refocusing, depth detection in out-door environment. However, when it comes to surveying marine environment and near water macro photography, all depth estimation algorithms based on traditional perspective camera model will fail because of the refracted rays. In this paper, we present a new method that explicitly accommodates the effect of refraction and resolves correct depths of underwater scene points. A semi-submerged object with opaque Lambertian surface with repeating textures is assumed. After removing the effect of refraction, the reconstructed underwater part of the semi-submerged object has consistent depth and shape with that of the above-water part.
我们提出了一种使用光场相机估计包含半淹没在水中的物体的真实世界场景深度的算法。现有的手持消费型光场相机非常适合在户外环境中进行自动调焦和深度检测。然而,在测量海洋环境和近水微距摄影时,所有基于传统透视相机模型的深度估计算法都会因为光线的折射而失效。在本文中,我们提出了一种新的方法,明确地适应折射的影响,并正确地解决水下场景点的深度。假设一个具有重复纹理的不透明朗伯曲面的半淹没物体。去除折射影响后,重建的半水下物体水下部分与水面部分深度和形状一致。
{"title":"Depth Estimation of Semi-submerged Objects Using a Light-Field Camera","authors":"Juehui Fan, Herbert Yang","doi":"10.1109/CRV.2017.44","DOIUrl":"https://doi.org/10.1109/CRV.2017.44","url":null,"abstract":"We present an algorithm to estimate depth of realworld scenes containing an object semi-submerged in water using a light field camera. Existing hand-held consumer light field cameras are well-suited for automated refocusing, depth detection in out-door environment. However, when it comes to surveying marine environment and near water macro photography, all depth estimation algorithms based on traditional perspective camera model will fail because of the refracted rays. In this paper, we present a new method that explicitly accommodates the effect of refraction and resolves correct depths of underwater scene points. A semi-submerged object with opaque Lambertian surface with repeating textures is assumed. After removing the effect of refraction, the reconstructed underwater part of the semi-submerged object has consistent depth and shape with that of the above-water part.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128455812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Fast Estimation of Large Displacement Optical Flow Using Dominant Motion Patterns & Sub-Volume PatchMatch Filtering 基于优势运动模式和亚体积补丁匹配滤波的大位移光流快速估计
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.40
M. Helala, F. Qureshi
This paper presents a new method for efficiently computing large-displacement optical flow. The method uses dominant motion patterns to identify a sparse set of sub-volumes within the cost volume and restricts subsequent Edge-Aware Filtering (EAF) to these sub-volumes. The method uses an extension of PatchMatch to filter these sub-volumes. The fact that our method only applies EAF to a small fraction of the entire cost volume boosts runtime performance. We also show that computational complexity is linear in the size of the images and does not depend upon the size of the label space. We evaluate the proposed technique on MPI Sintel, Middlebury and KITTI benchmarks and show that our method achieves accuracy comparable to those of several recent state-of-the-art methods, while posting significantly faster runtimes.
本文提出了一种有效计算大位移光流的新方法。该方法利用优势运动模式识别成本体积内的稀疏子体积集,并将后续边缘感知滤波(EAF)限制在这些子体积上。该方法使用PatchMatch的扩展来过滤这些子卷。我们的方法只将EAF应用于整个成本量的一小部分,这一事实提高了运行时性能。我们还表明,计算复杂度在图像的大小上是线性的,并且不依赖于标签空间的大小。我们在MPI sinintel、Middlebury和KITTI基准测试中评估了所提出的技术,并表明我们的方法达到了与最近几种最先进方法相当的准确性,同时运行时间明显更快。
{"title":"Fast Estimation of Large Displacement Optical Flow Using Dominant Motion Patterns & Sub-Volume PatchMatch Filtering","authors":"M. Helala, F. Qureshi","doi":"10.1109/CRV.2017.40","DOIUrl":"https://doi.org/10.1109/CRV.2017.40","url":null,"abstract":"This paper presents a new method for efficiently computing large-displacement optical flow. The method uses dominant motion patterns to identify a sparse set of sub-volumes within the cost volume and restricts subsequent Edge-Aware Filtering (EAF) to these sub-volumes. The method uses an extension of PatchMatch to filter these sub-volumes. The fact that our method only applies EAF to a small fraction of the entire cost volume boosts runtime performance. We also show that computational complexity is linear in the size of the images and does not depend upon the size of the label space. We evaluate the proposed technique on MPI Sintel, Middlebury and KITTI benchmarks and show that our method achieves accuracy comparable to those of several recent state-of-the-art methods, while posting significantly faster runtimes.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130923687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fully Automated Road Defect Detection Using Street View Images 使用街景图像的全自动道路缺陷检测
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.50
David Abou Chacra, J. Zelek
Road quality assessment is a crucial part in municipalities' work to maintain their infrastructure, plan upgrades, and manage their budgets. Properly maintaining this infrastructure relies heavily on consistently monitoring its condition and deterioration over time. This can be a challenge, especially in larger towns and cities where there is a lot of city property to keep an eye on. We review road quality assessment methods currently employed, and then describe our novel system, which integrates a collection of existing algorithms, aimed at identifying distressed road regions from street view images and pinpointing cracks within them. We predict distressed regions by computing Fisher vectors on local SIFT descriptors and classifying them with an SVM trained to distinguish between road qualities. We follow this step with a comparison to a weighed contour map within these distressed regions to identify exact crack and defect locations, and use the contour weights to predict the crack severity. Promising results are obtained on our manually annotated dataset, which indicate the viability of using this cost-effective system to perform road quality assessment at the municipal level.
道路质量评估是市政当局维护基础设施、规划升级和管理预算工作的重要组成部分。正确维护此基础设施在很大程度上依赖于持续监测其状况和随时间的恶化情况。这可能是一个挑战,尤其是在较大的城镇和城市,那里有很多城市房产需要关注。我们回顾了目前使用的道路质量评估方法,然后描述了我们的新系统,该系统集成了一系列现有算法,旨在从街景图像中识别受损道路区域并精确定位其中的裂缝。我们通过在局部SIFT描述符上计算Fisher向量并使用经过训练以区分道路质量的SVM对其进行分类来预测受损区域。我们遵循这一步骤,与这些受损区域内的加权等高线图进行比较,以确定准确的裂纹和缺陷位置,并使用等高线权重来预测裂纹的严重程度。在我们手工标注的数据集上获得了有希望的结果,这表明使用这种具有成本效益的系统在市级进行道路质量评估的可行性。
{"title":"Fully Automated Road Defect Detection Using Street View Images","authors":"David Abou Chacra, J. Zelek","doi":"10.1109/CRV.2017.50","DOIUrl":"https://doi.org/10.1109/CRV.2017.50","url":null,"abstract":"Road quality assessment is a crucial part in municipalities' work to maintain their infrastructure, plan upgrades, and manage their budgets. Properly maintaining this infrastructure relies heavily on consistently monitoring its condition and deterioration over time. This can be a challenge, especially in larger towns and cities where there is a lot of city property to keep an eye on. We review road quality assessment methods currently employed, and then describe our novel system, which integrates a collection of existing algorithms, aimed at identifying distressed road regions from street view images and pinpointing cracks within them. We predict distressed regions by computing Fisher vectors on local SIFT descriptors and classifying them with an SVM trained to distinguish between road qualities. We follow this step with a comparison to a weighed contour map within these distressed regions to identify exact crack and defect locations, and use the contour weights to predict the crack severity. Promising results are obtained on our manually annotated dataset, which indicate the viability of using this cost-effective system to perform road quality assessment at the municipal level.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125173817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A Structured Deep-Learning Based Approach for the Automated Segmentation of Human Leg Muscle from 3D MRI 基于结构化深度学习的三维MRI人体腿部肌肉自动分割方法
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.32
Shrimanti Ghosh, Nilanjan Ray, P. Boulanger
In this paper, we present an automated algorithm for segmenting human leg muscles from 3D MRI data using deep convolutional neural network (CNN). Using a generalized cylinder model the human leg muscle can be represented by two smooth 2D parametric images representing the contour of the muscle in the MRI image. The proposed CNN algorithm can predict these two parametrized images from raw 3D voxels. We use a pre-trained AlexNet as our baseline and further fine-tune the network that is suitable for this problem. In this scheme, AlexNet predicts a compressed vector obtained by applying principal component analysis, which is then back-projected into two parametric 2D images representing the leg muscle contours. We show that the proposed CNN with a structured regression model can out-perform conventional model-based segmentation approach such as the Active Appearance Model (AAM). The average Dice score between the ground truth segmentation and the obtained segmentation image is 0.87 using the proposed CNN model, whereas for AAM score is 0.68. One of the greatest advantages of our proposed method is that no initialization is needed to predict the segmentation contour, unlike AAM.
在本文中,我们提出了一种使用深度卷积神经网络(CNN)从3D MRI数据中分割人体腿部肌肉的自动算法。利用广义圆柱体模型,人体腿部肌肉可以用两个光滑的二维参数图像来表示MRI图像中肌肉的轮廓。本文提出的CNN算法可以从原始三维体素中预测这两种参数化图像。我们使用预训练的AlexNet作为基线,并进一步微调适合此问题的网络。在该方案中,AlexNet通过应用主成分分析预测压缩向量,然后将其反投影到代表腿部肌肉轮廓的两个参数2D图像中。我们表明,采用结构化回归模型的CNN可以优于传统的基于模型的分割方法,如活动外观模型(AAM)。使用本文提出的CNN模型,ground truth segmentation与得到的分割图像之间的平均Dice得分为0.87,而AAM得分为0.68。我们提出的方法最大的优点之一是不需要初始化来预测分割轮廓,这与AAM不同。
{"title":"A Structured Deep-Learning Based Approach for the Automated Segmentation of Human Leg Muscle from 3D MRI","authors":"Shrimanti Ghosh, Nilanjan Ray, P. Boulanger","doi":"10.1109/CRV.2017.32","DOIUrl":"https://doi.org/10.1109/CRV.2017.32","url":null,"abstract":"In this paper, we present an automated algorithm for segmenting human leg muscles from 3D MRI data using deep convolutional neural network (CNN). Using a generalized cylinder model the human leg muscle can be represented by two smooth 2D parametric images representing the contour of the muscle in the MRI image. The proposed CNN algorithm can predict these two parametrized images from raw 3D voxels. We use a pre-trained AlexNet as our baseline and further fine-tune the network that is suitable for this problem. In this scheme, AlexNet predicts a compressed vector obtained by applying principal component analysis, which is then back-projected into two parametric 2D images representing the leg muscle contours. We show that the proposed CNN with a structured regression model can out-perform conventional model-based segmentation approach such as the Active Appearance Model (AAM). The average Dice score between the ground truth segmentation and the obtained segmentation image is 0.87 using the proposed CNN model, whereas for AAM score is 0.68. One of the greatest advantages of our proposed method is that no initialization is needed to predict the segmentation contour, unlike AAM.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114502657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A Window to Your Smartphone: Exploring Interaction and Communication in Immersive VR with Augmented Virtuality 智能手机的窗口:探索沉浸式虚拟现实与增强虚拟的互动和交流
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.16
Amit P. Desai, Lourdes Peña Castillo, Oscar E. Meruvia Pastor
A major drawback of most Head Mounted Displays (HMDs) used in immersive Virtual Reality (VR) is the visual and social isolation of users from their real-world surroundings while wearing these headsets. This partial isolation of users from the real-world might hinder social interactions with friends and family. To address this issue, we present a new method to allow people wearing VR HMDs to use their smartphones or tablets without removing their HMDs. To do this, we augment the scene inside the VR HMD with a view of the user's device so that the user can interact with the device without removing the headset. The idea involves the use of additional cameras, such as the Leap Motion device or a high-resolution RGB camera to capture the user's real-world surrounding and augment the virtual world with the content displayed on the smartphone screen. This setup allows VR users to have a window to their smartphone from within the virtual world and afford all of the functionality provided by their smartphones, with the potential to reduce some of the undesirable isolation users may experience when using immersive VR systems.
沉浸式虚拟现实(VR)中使用的大多数头戴式显示器(hmd)的一个主要缺点是,当用户戴着这些头戴式显示器时,用户与现实世界的视觉和社交隔离。用户与现实世界的这种部分隔离可能会阻碍与朋友和家人的社交互动。为了解决这个问题,我们提出了一种新方法,让人们戴着VR头戴式设备使用他们的智能手机或平板电脑,而无需移除他们的头戴式设备。为了做到这一点,我们通过用户设备的视图来增强VR HMD内部的场景,这样用户就可以在不摘下耳机的情况下与设备进行交互。这个想法涉及到使用额外的摄像头,比如Leap Motion设备或高分辨率RGB摄像头来捕捉用户的真实世界环境,并通过智能手机屏幕上显示的内容来增强虚拟世界。这种设置允许VR用户在虚拟世界中有一个窗口到他们的智能手机,并负担得起智能手机提供的所有功能,有可能减少用户在使用沉浸式VR系统时可能遇到的一些不受欢迎的隔离。
{"title":"A Window to Your Smartphone: Exploring Interaction and Communication in Immersive VR with Augmented Virtuality","authors":"Amit P. Desai, Lourdes Peña Castillo, Oscar E. Meruvia Pastor","doi":"10.1109/CRV.2017.16","DOIUrl":"https://doi.org/10.1109/CRV.2017.16","url":null,"abstract":"A major drawback of most Head Mounted Displays (HMDs) used in immersive Virtual Reality (VR) is the visual and social isolation of users from their real-world surroundings while wearing these headsets. This partial isolation of users from the real-world might hinder social interactions with friends and family. To address this issue, we present a new method to allow people wearing VR HMDs to use their smartphones or tablets without removing their HMDs. To do this, we augment the scene inside the VR HMD with a view of the user's device so that the user can interact with the device without removing the headset. The idea involves the use of additional cameras, such as the Leap Motion device or a high-resolution RGB camera to capture the user's real-world surrounding and augment the virtual world with the content displayed on the smartphone screen. This setup allows VR users to have a window to their smartphone from within the virtual world and afford all of the functionality provided by their smartphones, with the potential to reduce some of the undesirable isolation users may experience when using immersive VR systems.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131647276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
LOST Highway: A Multiple-Lane Ant-Trail Algorithm to Reduce Congestion in Large-Population Multi-robot Systems LOST Highway:一种多车道反跟踪算法以减少大人口多机器人系统中的拥堵
Pub Date : 2017-05-01 DOI: 10.1109/CRV.2017.24
A. Abdelaal, Maram Sakr, R. Vaughan
We propose a modification of a well-known ant-inspired trail-following algorithm to reduce congestion in multi-robot systems. Our method results in robots moving in multiple lanes towards their goal location. Our algorithm is inspired by the idea of building multiple-lane highways to mitigate traffic congestion in traffic engineering. We consider the resource transportation task where autonomous robots repeatedly transport goods between a food source and a nest in an initially unknown environment. To evaluate our algorithm, we perform simulation experiments in several environments with and without obstacles. Compared with the baseline SO-LOST algorithm, we find that our modified method increases the system throughput by up to 3.9 times by supporting a larger productive robot population.
为了减少多机器人系统中的拥塞,我们提出了一种著名的蚁启发路径跟踪算法的改进。我们的方法使机器人在多个车道上朝着目标位置移动。该算法的灵感来自于交通工程中建设多车道高速公路以缓解交通拥堵的思想。我们考虑资源运输任务,其中自主机器人在最初未知的环境中反复在食物源和巢穴之间运输货物。为了评估我们的算法,我们在几个有障碍和没有障碍的环境中进行了模拟实验。与基线SO-LOST算法相比,我们发现改进的方法通过支持更大的生产机器人种群,将系统吞吐量提高了3.9倍。
{"title":"LOST Highway: A Multiple-Lane Ant-Trail Algorithm to Reduce Congestion in Large-Population Multi-robot Systems","authors":"A. Abdelaal, Maram Sakr, R. Vaughan","doi":"10.1109/CRV.2017.24","DOIUrl":"https://doi.org/10.1109/CRV.2017.24","url":null,"abstract":"We propose a modification of a well-known ant-inspired trail-following algorithm to reduce congestion in multi-robot systems. Our method results in robots moving in multiple lanes towards their goal location. Our algorithm is inspired by the idea of building multiple-lane highways to mitigate traffic congestion in traffic engineering. We consider the resource transportation task where autonomous robots repeatedly transport goods between a food source and a nest in an initially unknown environment. To evaluate our algorithm, we perform simulation experiments in several environments with and without obstacles. Compared with the baseline SO-LOST algorithm, we find that our modified method increases the system throughput by up to 3.9 times by supporting a larger productive robot population.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134365785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2017 14th Conference on Computer and Robot Vision (CRV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1