首页 > 最新文献

2017 IEEE International Conference on Computer Vision Workshops (ICCVW)最新文献

英文 中文
Coral-Segmentation: Training Dense Labeling Models with Sparse Ground Truth 珊瑚分割:训练稀疏地面真值的密集标记模型
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.339
Iñigo Alonso, Ana B. Cambra, A. Muñoz, T. Treibitz, A. C. Murillo
Biological datasets, such as our case of study, coral segmentation, often present scarce and sparse annotated image labels. Transfer learning techniques allow us to adapt existing deep learning models to new domains, even with small amounts of training data. Therefore, one of the main challenges to train dense segmentation models is to obtain the required dense labeled training data. This work presents a novel pipeline to address this pitfall and demonstrates the advantages of applying it to coral imagery segmentation. We fine tune state-of-the-art encoder-decoder CNN models for semantic segmentation thanks to a new proposed augmented labeling strategy. Our experiments run on a recent coral dataset [4], proving that this augmented ground truth allows us to effectively learn coral segmentation, as well as provide a relevant score of the segmentation quality based on it. Our approach provides a segmentation of comparable or better quality than the baseline presented with the dataset and a more flexible end-to-end pipeline.
生物数据集,例如我们的研究案例,珊瑚分割,经常呈现稀缺和稀疏的注释图像标签。迁移学习技术使我们能够使现有的深度学习模型适应新的领域,即使只有少量的训练数据。因此,训练密集分割模型的主要挑战之一是获得所需的密集标记训练数据。这项工作提出了一种新的管道来解决这个陷阱,并展示了将其应用于珊瑚图像分割的优势。我们微调最先进的编码器-解码器CNN模型的语义分割感谢一个新的提出的增强标签策略。我们的实验运行在最近的珊瑚数据集[4]上,证明了这种增强的地面真值使我们能够有效地学习珊瑚分割,并提供基于它的分割质量的相关分数。我们的方法提供了与数据集提供的基线相当或更好质量的分割,以及更灵活的端到端管道。
{"title":"Coral-Segmentation: Training Dense Labeling Models with Sparse Ground Truth","authors":"Iñigo Alonso, Ana B. Cambra, A. Muñoz, T. Treibitz, A. C. Murillo","doi":"10.1109/ICCVW.2017.339","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.339","url":null,"abstract":"Biological datasets, such as our case of study, coral segmentation, often present scarce and sparse annotated image labels. Transfer learning techniques allow us to adapt existing deep learning models to new domains, even with small amounts of training data. Therefore, one of the main challenges to train dense segmentation models is to obtain the required dense labeled training data. This work presents a novel pipeline to address this pitfall and demonstrates the advantages of applying it to coral imagery segmentation. We fine tune state-of-the-art encoder-decoder CNN models for semantic segmentation thanks to a new proposed augmented labeling strategy. Our experiments run on a recent coral dataset [4], proving that this augmented ground truth allows us to effectively learn coral segmentation, as well as provide a relevant score of the segmentation quality based on it. Our approach provides a segmentation of comparable or better quality than the baseline presented with the dataset and a more flexible end-to-end pipeline.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122250447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Near-Duplicate Video Retrieval with Deep Metric Learning 基于深度度量学习的近重复视频检索
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.49
Giorgos Kordopatis-Zilos, S. Papadopoulos, I. Patras, Y. Kompatsiaris
This work addresses the problem of Near-Duplicate Video Retrieval (NDVR). We propose an effective video-level NDVR scheme based on deep metric learning that leverages Convolutional Neural Network (CNN) features from intermediate layers to generate discriminative global video representations in tandem with a Deep Metric Learning (DML) framework with two fusion variations, trained to approximate an embedding function for accurate distance calculation between two near-duplicate videos. In contrast to most state-of-the-art methods, which exploit information deriving from the same source of data for both development and evaluation (which usually results to dataset-specific solutions), the proposed model is fed during training with sampled triplets generated from an independent dataset and is thoroughly tested on the widely used CC_WEB_VIDEO dataset, using two popular deep CNN architectures (AlexNet, GoogleNet). We demonstrate that the proposed approach achieves outstanding performance against the state-of-the-art, either with or without access to the evaluation dataset.
这项工作解决了近重复视频检索(NDVR)的问题。我们提出了一种有效的基于深度度量学习的视频级NDVR方案,该方案利用来自中间层的卷积神经网络(CNN)特征,与具有两个融合变化的深度度量学习(DML)框架一起生成判别性全局视频表示,该框架被训练成近似嵌入函数,用于精确计算两个近重复视频之间的距离。与大多数最先进的方法相反,这些方法利用来自同一数据源的信息进行开发和评估(通常会产生特定于数据集的解决方案),所提出的模型在训练期间使用从独立数据集生成的采样三元组进行输入,并使用两种流行的深度CNN架构(AlexNet, GoogleNet)在广泛使用的CC_WEB_VIDEO数据集上进行彻底测试。我们证明,无论是否访问评估数据集,所提出的方法都能在最先进的情况下取得出色的性能。
{"title":"Near-Duplicate Video Retrieval with Deep Metric Learning","authors":"Giorgos Kordopatis-Zilos, S. Papadopoulos, I. Patras, Y. Kompatsiaris","doi":"10.1109/ICCVW.2017.49","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.49","url":null,"abstract":"This work addresses the problem of Near-Duplicate Video Retrieval (NDVR). We propose an effective video-level NDVR scheme based on deep metric learning that leverages Convolutional Neural Network (CNN) features from intermediate layers to generate discriminative global video representations in tandem with a Deep Metric Learning (DML) framework with two fusion variations, trained to approximate an embedding function for accurate distance calculation between two near-duplicate videos. In contrast to most state-of-the-art methods, which exploit information deriving from the same source of data for both development and evaluation (which usually results to dataset-specific solutions), the proposed model is fed during training with sampled triplets generated from an independent dataset and is thoroughly tested on the widely used CC_WEB_VIDEO dataset, using two popular deep CNN architectures (AlexNet, GoogleNet). We demonstrate that the proposed approach achieves outstanding performance against the state-of-the-art, either with or without access to the evaluation dataset.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"2002 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129572568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Fast Approximate Karhunen-Loève Transform for Three-Way Array Data 三向阵列数据的快速近似karhunen - lo<e:1>变换
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.216
Hayato Itoh, A. Imiya, T. Sakai
Organs, cells and microstructures in cells dealt with in biomedical image analysis are volumetric data. We are required to process and analyse these data as volumetric data without embedding into higher-dimensional vector space from the viewpoints of object oriented data analysis. Sampled values of volumetric data are expressed as three-way array data. Therefore, principal component analysis of multi-way data is an essential technique for subspace-based pattern recognition, data retrievals and data compression of volumetric data. For one-way array (the vector form) problem the discrete cosine transform matrix is a good relaxed solution of the eigenmatrix for principal component analysis. This algebraic property of principal component analysis, derives an approximate fast algorithm for PCA of three-way data arrays.
在生物医学图像分析中处理的器官、细胞和细胞中的微结构是体积数据。从面向对象数据分析的角度来看,我们需要将这些数据作为体数据来处理和分析,而不是嵌入到高维向量空间中。体积数据的采样值表示为三向阵列数据。因此,多向数据的主成分分析是基于子空间的模式识别、数据检索和体积数据压缩的关键技术。对于单向阵列(矢量形式)问题,离散余弦变换矩阵是主成分分析中特征矩阵的一种很好的松弛解。利用主成分分析的代数性质,导出了三向数据阵列主成分分析的近似快速算法。
{"title":"Fast Approximate Karhunen-Loève Transform for Three-Way Array Data","authors":"Hayato Itoh, A. Imiya, T. Sakai","doi":"10.1109/ICCVW.2017.216","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.216","url":null,"abstract":"Organs, cells and microstructures in cells dealt with in biomedical image analysis are volumetric data. We are required to process and analyse these data as volumetric data without embedding into higher-dimensional vector space from the viewpoints of object oriented data analysis. Sampled values of volumetric data are expressed as three-way array data. Therefore, principal component analysis of multi-way data is an essential technique for subspace-based pattern recognition, data retrievals and data compression of volumetric data. For one-way array (the vector form) problem the discrete cosine transform matrix is a good relaxed solution of the eigenmatrix for principal component analysis. This algebraic property of principal component analysis, derives an approximate fast algorithm for PCA of three-way data arrays.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128240831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Color Consistency Correction Based on Remapping Optimization for Image Stitching 基于重映射优化的图像拼接颜色一致性校正
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.351
Menghan Xia, Jian Yao, Renping Xie, Mi Zhang, Jinsheng Xiao
Color consistency correction is a challenging problem in image stitching, because it matters several factors, including tone, contrast and fidelity, to present a natural appearance. In this paper, we propose an effective color correction method which is feasible to optimize the color consistency across images and guarantee the imaging quality of individual image meanwhile. Our method first apply well-directed alteration detection algorithms to find coherent-content regions in inter-image overlaps where reliable color correspondences are extracted. Then, we parameterize the color remapping curve as transform model, and express the constraints of color consistency, contrast and gradient in an uniform energy function. It can be formulated as a convex quadratic programming problem which provides the global optimal solution efficiently. Our method has a good performance in color consistency and suffers no pixel saturation or tonal dimming. Experimental results of representative datasets demonstrate the superiority of our method over state-of-the-art algorithms.
色彩一致性校正是图像拼接中的一个具有挑战性的问题,因为它关系到色调、对比度和保真度等几个因素,以呈现自然的外观。本文提出了一种有效的色彩校正方法,既能优化图像间的色彩一致性,又能保证单个图像的成像质量。我们的方法首先应用定向改变检测算法在图像间重叠中找到相干内容区域,提取可靠的颜色对应。然后,将颜色重映射曲线参数化为变换模型,将颜色一致性、对比度和梯度约束表示为均匀能量函数;它可以被表述为一个凸二次规划问题,并能有效地给出全局最优解。该方法具有良好的色彩一致性,不受像素饱和和色调调光的影响。代表性数据集的实验结果表明,我们的方法优于最先进的算法。
{"title":"Color Consistency Correction Based on Remapping Optimization for Image Stitching","authors":"Menghan Xia, Jian Yao, Renping Xie, Mi Zhang, Jinsheng Xiao","doi":"10.1109/ICCVW.2017.351","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.351","url":null,"abstract":"Color consistency correction is a challenging problem in image stitching, because it matters several factors, including tone, contrast and fidelity, to present a natural appearance. In this paper, we propose an effective color correction method which is feasible to optimize the color consistency across images and guarantee the imaging quality of individual image meanwhile. Our method first apply well-directed alteration detection algorithms to find coherent-content regions in inter-image overlaps where reliable color correspondences are extracted. Then, we parameterize the color remapping curve as transform model, and express the constraints of color consistency, contrast and gradient in an uniform energy function. It can be formulated as a convex quadratic programming problem which provides the global optimal solution efficiently. Our method has a good performance in color consistency and suffers no pixel saturation or tonal dimming. Experimental results of representative datasets demonstrate the superiority of our method over state-of-the-art algorithms.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129180135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Darwintrees for Action Recognition 动作识别的达尔文树
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.375
Albert Clapés, T. Tuytelaars, Sergio Escalera
We propose a novel mid-level representation for action/activity recognition on RGB videos. We model the evolution of improved dense trajectory features not only for the entire video sequence, but also on subparts of the video. Subparts are obtained using a spectral divisive clustering that yields an unordered binary tree decomposing the entire cloud of trajectories of a sequence. We then compute video-darwin on video subparts, exploiting more finegrained temporal information and reducing the sensitivity of the standard time varying mean strategy of videodarwin. After decomposition, we model the evolution of features through both frames of subparts and descending/ascending paths in tree branches. We refer to these mid-level representations as node-darwintree and branch-darwintree respectively. For the final classification, we construct a kernel representation for both mid-level and holistic videodarwin representations. Our approach achieves better performance than standard videodarwin and defines the current state-of-the-art on UCF-Sports and Highfive action recognition datasets.
我们提出了一种新的用于RGB视频动作/活动识别的中级表示。我们不仅对整个视频序列,而且还对视频的子部分进行了改进的密集轨迹特征的演化建模。子部分是通过光谱分裂聚类获得的,该聚类产生一棵无序二叉树,分解序列的整个轨迹云。然后,我们在视频子部分上计算视频达尔文,利用更细粒度的时间信息,降低视频达尔文标准时变均值策略的灵敏度。分解后,我们通过子部件框架和树分支中的下降/上升路径来建模特征的演化。我们将这些中层表示分别称为节点-达尔文树和分支-达尔文树。对于最后的分类,我们为中级和整体视频达尔文表示构建了一个核表示。我们的方法实现了比标准videodarwin更好的性能,并在UCF-Sports和Highfive动作识别数据集上定义了当前最先进的技术。
{"title":"Darwintrees for Action Recognition","authors":"Albert Clapés, T. Tuytelaars, Sergio Escalera","doi":"10.1109/ICCVW.2017.375","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.375","url":null,"abstract":"We propose a novel mid-level representation for action/activity recognition on RGB videos. We model the evolution of improved dense trajectory features not only for the entire video sequence, but also on subparts of the video. Subparts are obtained using a spectral divisive clustering that yields an unordered binary tree decomposing the entire cloud of trajectories of a sequence. We then compute video-darwin on video subparts, exploiting more finegrained temporal information and reducing the sensitivity of the standard time varying mean strategy of videodarwin. After decomposition, we model the evolution of features through both frames of subparts and descending/ascending paths in tree branches. We refer to these mid-level representations as node-darwintree and branch-darwintree respectively. For the final classification, we construct a kernel representation for both mid-level and holistic videodarwin representations. Our approach achieves better performance than standard videodarwin and defines the current state-of-the-art on UCF-Sports and Highfive action recognition datasets.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130664906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Particle Tracking Accuracy Measurement Based on Comparison of Linear Oriented Forests 基于线性导向森林对比的粒子跟踪精度测量
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.8
M. Maška, P. Matula
Particle tracking is of fundamental importance in diverse quantitative analyses of dynamic intracellular processes using time-lapse microscopy. Due to frequent impracticability of tracking particles manually, a number of fully automated algorithms have been developed over past decades, carrying out the tracking task in two subsequent phases: (1) particle detection and (2) particle linking. An objective benchmark for assessing the performance of such algorithms was recently established by the Particle Tracking Challenge. Because its performance evaluation protocol finds correspondences between a reference and algorithm-generated tracking result at the level of individual tracks, the performance assessment strongly depends on the algorithm linking capabilities. In this paper, we propose a novel performance evaluation protocol based on a simplified version of the tracking accuracy measure employed in the Cell Tracking Challenge, which establishes the correspondences at the level of individual particle detections, thus allowing one to evaluate the performance of each of the two phases in an isolated, unbiased manner By analyzing the tracking results of all 14 algorithms competing in the Particle Tracking Challenge using the proposed evaluation protocol, we reveal substantial changes in their detection and linking performance, yielding rankings different from those reported previously.
粒子跟踪在使用延时显微镜对动态细胞内过程进行各种定量分析中具有重要意义。由于人工跟踪粒子往往不可行,在过去的几十年里,许多全自动算法被开发出来,在随后的两个阶段完成跟踪任务:(1)粒子检测和(2)粒子连接。最近,粒子跟踪挑战赛(Particle Tracking Challenge)建立了一个评估这些算法性能的客观基准。由于其性能评估协议在单个轨迹级别上发现参考和算法生成的跟踪结果之间的对应关系,因此性能评估在很大程度上依赖于算法链接能力。在本文中,我们提出了一种新的性能评估协议,该协议基于细胞跟踪挑战中采用的跟踪精度测量的简化版本,该协议建立了单个粒子检测级别的对应关系,从而允许人们以孤立、无偏的方式评估两个阶段中每个阶段的性能。我们揭示了它们在检测和链接性能方面的重大变化,产生了与之前报道的不同的排名。
{"title":"Particle Tracking Accuracy Measurement Based on Comparison of Linear Oriented Forests","authors":"M. Maška, P. Matula","doi":"10.1109/ICCVW.2017.8","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.8","url":null,"abstract":"Particle tracking is of fundamental importance in diverse quantitative analyses of dynamic intracellular processes using time-lapse microscopy. Due to frequent impracticability of tracking particles manually, a number of fully automated algorithms have been developed over past decades, carrying out the tracking task in two subsequent phases: (1) particle detection and (2) particle linking. An objective benchmark for assessing the performance of such algorithms was recently established by the Particle Tracking Challenge. Because its performance evaluation protocol finds correspondences between a reference and algorithm-generated tracking result at the level of individual tracks, the performance assessment strongly depends on the algorithm linking capabilities. In this paper, we propose a novel performance evaluation protocol based on a simplified version of the tracking accuracy measure employed in the Cell Tracking Challenge, which establishes the correspondences at the level of individual particle detections, thus allowing one to evaluate the performance of each of the two phases in an isolated, unbiased manner By analyzing the tracking results of all 14 algorithms competing in the Particle Tracking Challenge using the proposed evaluation protocol, we reveal substantial changes in their detection and linking performance, yielding rankings different from those reported previously.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124214110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
HSCNN: CNN-Based Hyperspectral Image Recovery from Spectrally Undersampled Projections HSCNN:基于cnn的光谱欠采样投影高光谱图像恢复
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.68
Zhiwei Xiong, Zhan Shi, Huiqun Li, Lizhi Wang, Dong Liu, Feng Wu
This paper presents a unified deep learning framework to recover hyperspectral images from spectrally undersampled projections. Specifically, we investigate two kinds of representative projections, RGB and compressive sensing (CS) measurements. These measurements are first upsampled in the spectral dimension through simple interpolation or CS reconstruction, and the proposed method learns an end-to-end mapping from a large number of up-sampled/groundtruth hyperspectral image pairs. The mapping is represented as a deep convolutional neural network (CNN) that takes the spectrally upsampled image as input and outputs the enhanced hyperspetral one. We explore different network configurations to achieve high reconstruction fidelity. Experimental results on a variety of test images demonstrate significantly improved performance of the proposed method over the state-of-the-arts.
本文提出了一个统一的深度学习框架,用于从光谱欠采样投影中恢复高光谱图像。具体来说,我们研究了两种具有代表性的投影,RGB和压缩感知(CS)测量。这些测量首先通过简单的插值或CS重建在光谱维度上采样,并且该方法从大量上采样/底真高光谱图像对中学习端到端映射。该映射被表示为一个深度卷积神经网络(CNN),该网络将光谱上采样图像作为输入,输出增强的高光谱图像。我们探索了不同的网络配置来实现高重建保真度。在各种测试图像上的实验结果表明,所提出的方法的性能明显优于最先进的方法。
{"title":"HSCNN: CNN-Based Hyperspectral Image Recovery from Spectrally Undersampled Projections","authors":"Zhiwei Xiong, Zhan Shi, Huiqun Li, Lizhi Wang, Dong Liu, Feng Wu","doi":"10.1109/ICCVW.2017.68","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.68","url":null,"abstract":"This paper presents a unified deep learning framework to recover hyperspectral images from spectrally undersampled projections. Specifically, we investigate two kinds of representative projections, RGB and compressive sensing (CS) measurements. These measurements are first upsampled in the spectral dimension through simple interpolation or CS reconstruction, and the proposed method learns an end-to-end mapping from a large number of up-sampled/groundtruth hyperspectral image pairs. The mapping is represented as a deep convolutional neural network (CNN) that takes the spectrally upsampled image as input and outputs the enhanced hyperspetral one. We explore different network configurations to achieve high reconstruction fidelity. Experimental results on a variety of test images demonstrate significantly improved performance of the proposed method over the state-of-the-arts.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121209780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 125
Finding Mirror Symmetry via Registration and Optimal Symmetric Pairwise Assignment of Curves: Algorithm and Results 通过配准和最优对称曲线成对分配寻找镜像对称性:算法和结果
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.207
Marcelo Cicconet, David Grant Colburn Hildebrand, H. Elliott
We demonstrate that the problem of fitting a plane of mirror symmetry to data in any Euclidian space can be reduced to the problem of registering two datasets, and that the exactness of the solution depends entirely on the registration accuracy. This new Mirror Symmetry via Registration (MSR) framework involves (1) data reflection with respect to an arbitrary plane, (2) registration of original and reflected datasets, and (3) calculation of the eigenvector of eigenvalue -1 for the transformation matrix representing the reflection and registration mappings. To support MSR, we also introduce a novel 2D registration method based on random sample consensus of an ensemble of normalized cross-correlation matches. We further demonstrate the generality of MSR by testing it on a database of 3D shapes with an iterative closest point registration back-end.
我们证明了任意欧几里德空间中数据的镜面对称拟合问题可以简化为两个数据集的配准问题,并且解的准确性完全取决于配准精度。这个新的镜面对称配准(MSR)框架涉及(1)相对于任意平面的数据反射,(2)原始和反射数据集的配准,以及(3)计算表示反射和配准映射的变换矩阵的特征值-1的特征向量。为了支持MSR,我们还引入了一种新的基于归一化互相关匹配集合的随机样本一致性的二维配准方法。我们通过在三维形状数据库上使用迭代最近点配准后端进行测试,进一步证明了MSR的通用性。
{"title":"Finding Mirror Symmetry via Registration and Optimal Symmetric Pairwise Assignment of Curves: Algorithm and Results","authors":"Marcelo Cicconet, David Grant Colburn Hildebrand, H. Elliott","doi":"10.1109/ICCVW.2017.207","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.207","url":null,"abstract":"We demonstrate that the problem of fitting a plane of mirror symmetry to data in any Euclidian space can be reduced to the problem of registering two datasets, and that the exactness of the solution depends entirely on the registration accuracy. This new Mirror Symmetry via Registration (MSR) framework involves (1) data reflection with respect to an arbitrary plane, (2) registration of original and reflected datasets, and (3) calculation of the eigenvector of eigenvalue -1 for the transformation matrix representing the reflection and registration mappings. To support MSR, we also introduce a novel 2D registration method based on random sample consensus of an ensemble of normalized cross-correlation matches. We further demonstrate the generality of MSR by testing it on a database of 3D shapes with an iterative closest point registration back-end.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114637471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Deep Modality Invariant Adversarial Network for Shared Representation Learning 面向共享表示学习的深度模态不变对抗网络
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.311
T. Harada, Kuniaki Saito, Yusuke Mukuta, Y. Ushiku
In this work, we propose a novel method to learn the mapping to the common space wherein different modalities have the same information for shared representation learning. Our goal is to correctly classify the target modality with a classifier trained on source modality samples and their labels in common representations. We call these representations modality-invariant representations. Our proposed method has the major advantage of not needing any labels for the target samples in order to learn representations. For example, we obtain modality-invariant representations from pairs of images and texts. Then, we train the text classifier on the modality-invariant space. Although we do not give any explicit relationship between images and labels, we can expect that images can be classified correctly in that space. Our method draws upon the theory of domain adaptation and we propose to learn modality-invariant representations by utilizing adversarial training. We call our method the Deep Modality Invariant Adversarial Network (DeMIAN). We demonstrate the effectiveness of our method in experiments.
在这项工作中,我们提出了一种新的方法来学习映射到公共空间,其中不同的模态具有相同的信息用于共享表示学习。我们的目标是使用基于源模态样本和它们在共同表示中的标签训练的分类器来正确分类目标模态。我们称这些表示为模态不变表示。我们提出的方法的主要优点是不需要对目标样本进行任何标记来学习表征。例如,我们从图像和文本对中获得模态不变表示。然后,我们在模态不变空间上训练文本分类器。虽然我们没有给出图像和标签之间的任何明确的关系,但我们可以期望在该空间中图像可以被正确分类。我们的方法借鉴了领域适应理论,我们建议通过使用对抗性训练来学习模态不变表示。我们将这种方法称为深度模态不变对抗网络(DeMIAN)。我们在实验中证明了我们方法的有效性。
{"title":"Deep Modality Invariant Adversarial Network for Shared Representation Learning","authors":"T. Harada, Kuniaki Saito, Yusuke Mukuta, Y. Ushiku","doi":"10.1109/ICCVW.2017.311","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.311","url":null,"abstract":"In this work, we propose a novel method to learn the mapping to the common space wherein different modalities have the same information for shared representation learning. Our goal is to correctly classify the target modality with a classifier trained on source modality samples and their labels in common representations. We call these representations modality-invariant representations. Our proposed method has the major advantage of not needing any labels for the target samples in order to learn representations. For example, we obtain modality-invariant representations from pairs of images and texts. Then, we train the text classifier on the modality-invariant space. Although we do not give any explicit relationship between images and labels, we can expect that images can be classified correctly in that space. Our method draws upon the theory of domain adaptation and we propose to learn modality-invariant representations by utilizing adversarial training. We call our method the Deep Modality Invariant Adversarial Network (DeMIAN). We demonstrate the effectiveness of our method in experiments.","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121705137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
SkiMap++: Real-Time Mapping and Object Recognition for Robotics skimap++:机器人的实时映射和对象识别
Pub Date : 2017-10-01 DOI: 10.1109/ICCVW.2017.84
Daniele De Gregorio, Tommaso Cavallari, L. D. Stefano
We introduce SkiMap++, an extension to the recently proposed SkiMap mapping framework for robot navigation [1]. The extension deals with enriching the map with semantic information concerning the presence in the environment of certain objects that may be usefully recognized by the robot, e.g. for the sake of grasping them. More precisely, the map can accommodate information about the spatial locations of certain 3D object features, as determined by matching the visual features extracted from the incoming frames through a random forest learned off-line from a set of object models. Thereby, evidence about the presence of object features is gathered from multiple vantage points alongside with the standard geometric mapping task, so to enable recognizing the objects and estimating their 6 DOF poses. As a result, SkiMap++ can reconstruct the geometry of large scale environments as well as localize some relevant objects therein (Fig.1) in real-time on CPU. As an additional contribution, we present an RGB-D dataset featuring ground-truth camera and object poses, which may be deployed by researchers interested in pursuing SLAM alongside with object recognition, a topic often referred to as Semantic SLAM. 1
我们介绍了skimap++,这是最近提出的用于机器人导航[1]的SkiMap映射框架的扩展。扩展处理的是用语义信息来丰富地图,这些信息涉及机器人在环境中可能有效识别的某些物体的存在,例如为了抓取它们。更准确地说,地图可以容纳某些3D物体特征的空间位置信息,这是通过从一组物体模型中离线学习的随机森林来匹配从传入帧中提取的视觉特征来确定的。因此,关于物体特征存在的证据是与标准几何映射任务一起从多个有利位置收集的,因此能够识别物体并估计其6自由度姿势。因此,skimap++可以在CPU上实时重建大规模环境的几何结构,并对其中的一些相关物体进行定位(图1)。作为额外的贡献,我们提出了一个RGB-D数据集,该数据集具有地面真实相机和物体姿势,可以由有兴趣在物体识别的同时追求SLAM的研究人员部署,这个主题通常被称为语义SLAM。1
{"title":"SkiMap++: Real-Time Mapping and Object Recognition for Robotics","authors":"Daniele De Gregorio, Tommaso Cavallari, L. D. Stefano","doi":"10.1109/ICCVW.2017.84","DOIUrl":"https://doi.org/10.1109/ICCVW.2017.84","url":null,"abstract":"We introduce SkiMap++, an extension to the recently proposed SkiMap mapping framework for robot navigation [1]. The extension deals with enriching the map with semantic information concerning the presence in the environment of certain objects that may be usefully recognized by the robot, e.g. for the sake of grasping them. More precisely, the map can accommodate information about the spatial locations of certain 3D object features, as determined by matching the visual features extracted from the incoming frames through a random forest learned off-line from a set of object models. Thereby, evidence about the presence of object features is gathered from multiple vantage points alongside with the standard geometric mapping task, so to enable recognizing the objects and estimating their 6 DOF poses. As a result, SkiMap++ can reconstruct the geometry of large scale environments as well as localize some relevant objects therein (Fig.1) in real-time on CPU. As an additional contribution, we present an RGB-D dataset featuring ground-truth camera and object poses, which may be deployed by researchers interested in pursuing SLAM alongside with object recognition, a topic often referred to as Semantic SLAM. 1","PeriodicalId":149766,"journal":{"name":"2017 IEEE International Conference on Computer Vision Workshops (ICCVW)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114766056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2017 IEEE International Conference on Computer Vision Workshops (ICCVW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1