首页 > 最新文献

2016 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Shape matching using a self similar affine invariant descriptor 使用自相似仿射不变描述子的形状匹配
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532803
Joonsoo Kim, He Li, Jiaju Yue, E. Delp
In this paper we introduce a shape descriptor known as Self Similar Affine Invariant (SSAI) descriptor for shape retrieval. The SSAI descriptor is based on the property that two sets of points are transformed by an affine transform, then subsets of each set of points are also related by the same affine transformation. Also, the SSAI descriptor is insensitive to local shape distortions. We use multiple SSAI descriptors based on different sets of neighbor points to improve shape recognition accuracy. We also describe an efficient image matching method for the multiple SSAI descriptors. Experimental results show that our approach achieves very good performance on two publicly available shape datasets.
本文引入了一种用于形状检索的自相似仿射不变量(SSAI)形状描述符。SSAI描述符基于这样的性质:两个点的集合通过一个仿射变换进行变换,那么每个点的子集也通过同一个仿射变换进行关联。此外,SSAI描述符对局部形状畸变不敏感。我们使用基于不同相邻点集的多个SSAI描述符来提高形状识别的精度。我们还描述了一种针对多个SSAI描述符的高效图像匹配方法。实验结果表明,我们的方法在两个公开的形状数据集上取得了很好的性能。
{"title":"Shape matching using a self similar affine invariant descriptor","authors":"Joonsoo Kim, He Li, Jiaju Yue, E. Delp","doi":"10.1109/ICIP.2016.7532803","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532803","url":null,"abstract":"In this paper we introduce a shape descriptor known as Self Similar Affine Invariant (SSAI) descriptor for shape retrieval. The SSAI descriptor is based on the property that two sets of points are transformed by an affine transform, then subsets of each set of points are also related by the same affine transformation. Also, the SSAI descriptor is insensitive to local shape distortions. We use multiple SSAI descriptors based on different sets of neighbor points to improve shape recognition accuracy. We also describe an efficient image matching method for the multiple SSAI descriptors. Experimental results show that our approach achieves very good performance on two publicly available shape datasets.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"46 1","pages":"2470-2474"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86885975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Variational EM approach for high resolution hyper-spectral imaging based on probabilistic matrix factorization 基于概率矩阵分解的高分辨率高光谱成像变分EM方法
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532663
Baihong Lin, Xiaoming Tao, Linhao Dong, Jianhua Lu
High resolution hyper-spectral imaging works as a scheme to obtain images with high spatial and spectral resolutions by merging a low spatial resolution hyper-spectral image (HSI) with a high spatial resolution multi-spectral image (MSI). In this paper, we propose a novel method based on probabilistic matrix factorization under Bayesian framework: First, Gaussian priors, as observations' distributions, are given upon two HSI-MSI-pair-based images, in which two variances share the same hyper-parameter to ensure fair and effective constraints on two observations. Second, to avoid the manual tuning process and learn a better setting automatically, hyper-priors are adopted for all hyper-parameters. To that end, a variational expectation-maximization (EM) approach is devised to figure out the result expectation for its simplicity and effectiveness. Exhaustive experiments of two different cases prove that our algorithm outperforms many state-of-the-art methods.
高分辨率高光谱成像是一种将低空间分辨率高光谱图像(HSI)与高空间分辨率多光谱图像(MSI)合并,获得高空间分辨率和光谱分辨率图像的方案。本文提出了一种基于贝叶斯框架下概率矩阵分解的新方法:首先,在两幅基于hsi - msi对的图像上给出高斯先验作为观测值的分布,其中两个方差共享相同的超参数,以确保对两个观测值的公平有效约束;其次,为了避免人工调优过程,自动学习更好的设置,所有超参数都采用超先验。为此,设计了一种简单有效的变分期望最大化(EM)方法来计算结果期望。两种不同情况的详尽实验证明,我们的算法优于许多最先进的方法。
{"title":"Variational EM approach for high resolution hyper-spectral imaging based on probabilistic matrix factorization","authors":"Baihong Lin, Xiaoming Tao, Linhao Dong, Jianhua Lu","doi":"10.1109/ICIP.2016.7532663","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532663","url":null,"abstract":"High resolution hyper-spectral imaging works as a scheme to obtain images with high spatial and spectral resolutions by merging a low spatial resolution hyper-spectral image (HSI) with a high spatial resolution multi-spectral image (MSI). In this paper, we propose a novel method based on probabilistic matrix factorization under Bayesian framework: First, Gaussian priors, as observations' distributions, are given upon two HSI-MSI-pair-based images, in which two variances share the same hyper-parameter to ensure fair and effective constraints on two observations. Second, to avoid the manual tuning process and learn a better setting automatically, hyper-priors are adopted for all hyper-parameters. To that end, a variational expectation-maximization (EM) approach is devised to figure out the result expectation for its simplicity and effectiveness. Exhaustive experiments of two different cases prove that our algorithm outperforms many state-of-the-art methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"31 1","pages":"1774-1778"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86258858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Person re-identification via adaboost ranking ensemble 通过adaboost排名集合重新识别人员
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533165
Zhaoju Li, Zhenjun Han, Qixiang Ye
Matching specific persons across scenes, known as person re-identification, is an important yet unsolved computer vision problem. Feature representation and metric learning are two fundamental factors in person re-identification. However, current person re-identification methods, which use single handcrafted feature with corresponding metric, could be not powerful enough when facing illumination, viewpoint and pose variations. Thus it inevitably produces suboptimal ranking lists. In this paper, we propose incorporating multiple features with metrics to build weak learners, and aggregate the base ranking lists by AdaBoost Ranking. Experiments on two commonly used datasets, VIPeR and CUHK01, show that our proposed approach greatly improves recognition rates over the state-of-the-art methods.
在场景中匹配特定的人,即人的再识别,是一个重要但尚未解决的计算机视觉问题。特征表征和度量学习是人再识别的两个基本因素。然而,现有的人脸再识别方法在面对光照、视点和姿态变化时,使用单个手工特征和相应的度量,可能不够强大。因此,它不可避免地产生次优排名列表。在本文中,我们提出将多个特征与度量相结合来构建弱学习器,并通过AdaBoost排名来汇总基本排名列表。在两个常用的数据集(VIPeR和CUHK01)上进行的实验表明,我们提出的方法比最先进的方法大大提高了识别率。
{"title":"Person re-identification via adaboost ranking ensemble","authors":"Zhaoju Li, Zhenjun Han, Qixiang Ye","doi":"10.1109/ICIP.2016.7533165","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533165","url":null,"abstract":"Matching specific persons across scenes, known as person re-identification, is an important yet unsolved computer vision problem. Feature representation and metric learning are two fundamental factors in person re-identification. However, current person re-identification methods, which use single handcrafted feature with corresponding metric, could be not powerful enough when facing illumination, viewpoint and pose variations. Thus it inevitably produces suboptimal ranking lists. In this paper, we propose incorporating multiple features with metrics to build weak learners, and aggregate the base ranking lists by AdaBoost Ranking. Experiments on two commonly used datasets, VIPeR and CUHK01, show that our proposed approach greatly improves recognition rates over the state-of-the-art methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"21 6 1","pages":"4269-4273"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83658683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Saliency-context two-stream convnets for action recognition 用于动作识别的显著性-上下文双流卷积
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532925
Quan-Qi Chen, Feng Liu, Xue Li, Baodi Liu, Yujin Zhang
Recently, very deep two-stream ConvNets have achieved great discriminative power for video classification, which is especially the case for the temporal ConvNets when trained on multi-frame optical flow. However, action recognition in videos often fall prey to the wild camera motion, which poses challenges on the extraction of reliable optical flow for human body. In light of this, we propose a novel method to remove the global camera motion, which explicitly calculates a homography between two consecutive frames without human detection. Given the estimated homography due to camera motion, background motion can be canceled out from the warped optical flow. We take this a step further and design a new architecture called Saliency-Context two-stream ConvNets, where the context two-stream ConvNets are employed to recognize the entire scene in video frames, whilst the saliency streams are trained on salient human motion regions that are detected from the warped optical flow. Finally, the Saliency-Context two-stream ConvNets allow us to capture complementary information and achieve state-of-the-art performance on UCF101 dataset.
近年来,深度双流卷积神经网络在视频分类中取得了较好的判别能力,特别是在多帧光流训练下的时域卷积神经网络。然而,视频中的动作识别往往受到摄像机剧烈运动的影响,这对提取可靠的人体光流提出了挑战。鉴于此,我们提出了一种新的消除全局摄像机运动的方法,该方法在不需要人工检测的情况下显式计算两个连续帧之间的单应性。给定由相机运动引起的估计单应性,背景运动可以从扭曲的光流中抵消。我们进一步设计了一种名为“显著性-上下文两流卷积神经网络”的新架构,其中上下文两流卷积神经网络用于识别视频帧中的整个场景,而显著性流则在从扭曲光流检测到的显著人体运动区域上进行训练。最后,显著性-上下文两流卷积神经网络允许我们捕获互补信息,并在UCF101数据集上实现最先进的性能。
{"title":"Saliency-context two-stream convnets for action recognition","authors":"Quan-Qi Chen, Feng Liu, Xue Li, Baodi Liu, Yujin Zhang","doi":"10.1109/ICIP.2016.7532925","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532925","url":null,"abstract":"Recently, very deep two-stream ConvNets have achieved great discriminative power for video classification, which is especially the case for the temporal ConvNets when trained on multi-frame optical flow. However, action recognition in videos often fall prey to the wild camera motion, which poses challenges on the extraction of reliable optical flow for human body. In light of this, we propose a novel method to remove the global camera motion, which explicitly calculates a homography between two consecutive frames without human detection. Given the estimated homography due to camera motion, background motion can be canceled out from the warped optical flow. We take this a step further and design a new architecture called Saliency-Context two-stream ConvNets, where the context two-stream ConvNets are employed to recognize the entire scene in video frames, whilst the saliency streams are trained on salient human motion regions that are detected from the warped optical flow. Finally, the Saliency-Context two-stream ConvNets allow us to capture complementary information and achieve state-of-the-art performance on UCF101 dataset.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"32 1","pages":"3076-3080"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82963847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
The visibility of motion artifacts and their effect on motion quality 运动伪影的可见性及其对运动质量的影响
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532796
Alex Mackin, K. Noland, D. Bull
The visibility of motion artifacts in a video sequence e.g. motion blur and temporal aliasing, affects perceived motion quality. The frame rate required to render these motion artifacts imperceptible is far higher than is currently feasible or specified in current video formats. This paper investigates the perception of temporal aliasing and its associated artifacts below this frame rate, along with their influence on motion quality, with the aim of making suitable frame rate recommendations for future formats. Results show impairment in motion quality due to temporal aliasing can be tolerated to a degree, and that it may be acceptable to sample at frame rates 50% lower than those needed to eliminate perceptible temporal aliasing.
视频序列中运动伪影的可见性,如运动模糊和时间混叠,会影响感知的运动质量。使这些运动伪影难以察觉所需的帧率远远高于当前可行的或当前视频格式中指定的帧率。本文研究了时间混叠的感知及其相关的工件低于这个帧率,以及它们对运动质量的影响,目的是为未来的格式提出合适的帧率建议。结果表明,由于时间混叠造成的运动质量损害在一定程度上是可以容忍的,并且以比消除可感知的时间混叠所需的帧率低50%的帧率进行采样是可以接受的。
{"title":"The visibility of motion artifacts and their effect on motion quality","authors":"Alex Mackin, K. Noland, D. Bull","doi":"10.1109/ICIP.2016.7532796","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532796","url":null,"abstract":"The visibility of motion artifacts in a video sequence e.g. motion blur and temporal aliasing, affects perceived motion quality. The frame rate required to render these motion artifacts imperceptible is far higher than is currently feasible or specified in current video formats. This paper investigates the perception of temporal aliasing and its associated artifacts below this frame rate, along with their influence on motion quality, with the aim of making suitable frame rate recommendations for future formats. Results show impairment in motion quality due to temporal aliasing can be tolerated to a degree, and that it may be acceptable to sample at frame rates 50% lower than those needed to eliminate perceptible temporal aliasing.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"17 1","pages":"2435-2439"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88873894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Real-time temporally coherent local HDR tone mapping 实时时间相干本地HDR色调映射
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532485
S. Croci, T. Aydin, N. Stefanoski, M. Gross, A. Smolic
Subjective studies showed that most HDR video tone mapping operators either produce disturbing temporal artifacts, or are limited in their local contrast reproduction capability. Recently, both these issues have been addressed by a novel temporally coherent local HDR tone mapping method, which has been shown, both qualitatively and through a subjective study, to be advantageous compared to previous methods. However, this method's high-quality results came at the cost of a computationally expensive workflow that could only be executed offline. In this paper, we present a modified algorithm which builds upon the previous work by redesigning key components to achieve real-time performance. We accomplish this by replacing the optical flow based per-pixel temporal coherency with a tone-curve-space alternative. This way we eliminate the main computational burden of the original method with little sacrifice in visual quality.
主观研究表明,大多数HDR视频色调映射算子要么产生干扰的时间伪影,要么局部对比度再现能力有限。最近,一种新的时间相干局部HDR色调映射方法解决了这两个问题,该方法在定性和主观研究中都显示出与以前的方法相比具有优势。然而,这种方法的高质量结果是以计算成本昂贵的工作流为代价的,而工作流只能离线执行。在本文中,我们提出了一种改进的算法,该算法建立在以前的工作基础上,通过重新设计关键组件来实现实时性能。我们通过用音调曲线空间替代基于每像素时间相干的光流来实现这一目标。这种方法消除了原始方法的主要计算负担,并且在视觉质量上几乎没有牺牲。
{"title":"Real-time temporally coherent local HDR tone mapping","authors":"S. Croci, T. Aydin, N. Stefanoski, M. Gross, A. Smolic","doi":"10.1109/ICIP.2016.7532485","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532485","url":null,"abstract":"Subjective studies showed that most HDR video tone mapping operators either produce disturbing temporal artifacts, or are limited in their local contrast reproduction capability. Recently, both these issues have been addressed by a novel temporally coherent local HDR tone mapping method, which has been shown, both qualitatively and through a subjective study, to be advantageous compared to previous methods. However, this method's high-quality results came at the cost of a computationally expensive workflow that could only be executed offline. In this paper, we present a modified algorithm which builds upon the previous work by redesigning key components to achieve real-time performance. We accomplish this by replacing the optical flow based per-pixel temporal coherency with a tone-curve-space alternative. This way we eliminate the main computational burden of the original method with little sacrifice in visual quality.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"30 1","pages":"889-893"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83720014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Pursuing face identity from view-specific representation to view-invariant representation 从特定视点表征到视点不变表征对人脸身份的追求
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532959
Ting Zhang, Qiulei Dong, Zhanyi Hu
How to learn view-invariant facial representations is an important task for view-invariant face recognition. The recent work [1] discovered that the brain of the macaque monkey has a face-processing network, where some neurons are view-specific. Motivated by this discovery, this paper proposes a deep convolutional learning model for face recognition, which explicitly enforces this view-specific mechanism for learning view-invariant facial representations. The proposed model consists of two concatenated modules: the first one is a convolutional neural network (CNN) for learning the corresponding viewing pose to the input face image; the second one consists of multiple CNNs, each of which learns the corresponding frontal image of an image under a specific viewing pose. This method is of low computational cost, and it can be well trained with a relatively small number of samples. The experimental results on the MultiPIE dataset demonstrate the effectiveness of our proposed convolutional model in contrast to three state-of-the-art works.
如何学习视觉不变的人脸表征是视觉不变人脸识别的一个重要课题。最近的研究[1]发现,猕猴的大脑有一个面部处理网络,其中一些神经元是特定于视觉的。基于这一发现,本文提出了一种用于人脸识别的深度卷积学习模型,该模型明确地执行了这种特定于视图的机制来学习视图不变的面部表征。该模型由两个串联模块组成:第一个模块是卷积神经网络(CNN),用于学习输入人脸图像的相应观看姿态;第二种方法由多个cnn组成,每个cnn学习特定观看姿势下图像对应的正面图像。该方法计算成本低,并且可以用相对较少的样本进行很好的训练。在MultiPIE数据集上的实验结果证明了我们提出的卷积模型与三个最新研究成果的有效性。
{"title":"Pursuing face identity from view-specific representation to view-invariant representation","authors":"Ting Zhang, Qiulei Dong, Zhanyi Hu","doi":"10.1109/ICIP.2016.7532959","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532959","url":null,"abstract":"How to learn view-invariant facial representations is an important task for view-invariant face recognition. The recent work [1] discovered that the brain of the macaque monkey has a face-processing network, where some neurons are view-specific. Motivated by this discovery, this paper proposes a deep convolutional learning model for face recognition, which explicitly enforces this view-specific mechanism for learning view-invariant facial representations. The proposed model consists of two concatenated modules: the first one is a convolutional neural network (CNN) for learning the corresponding viewing pose to the input face image; the second one consists of multiple CNNs, each of which learns the corresponding frontal image of an image under a specific viewing pose. This method is of low computational cost, and it can be well trained with a relatively small number of samples. The experimental results on the MultiPIE dataset demonstrate the effectiveness of our proposed convolutional model in contrast to three state-of-the-art works.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"24 1","pages":"3244-3248"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83224578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Video object segmentation by Multi-Scale Pyramidal Multi-Dimensional LSTM with generated depth context 基于生成深度上下文的多尺度锥体多维LSTM视频目标分割
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532363
Qiurui Wang, C. Yuan
Existing deep neural networks, such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs), typically treat volumetric video data as several single images and deal with one frame at one time, thus the relevance to frames can hardly be fully exploited. Besides, depth context plays the unique role in motion scenes for primates, but is seldom used in no depth label situations. In this paper, we use a more suitable architecture Multi-Scale Pyramidal Multi-Dimensional Long Short Term Memory (MSPMD-LSTM) to reveal the strong relevance within video frames. Furthermore, depth context is extracted and refined to enhance the performance of the model. Experiments demonstrate that our models yield competitive results on Youtube-Objects dataset and Segtrack v2 dataset.
现有的深度神经网络,如卷积神经网络(cnn)和递归神经网络(rnn),通常将体积视频数据视为几张单个图像,一次处理一帧,因此很难充分利用与帧的相关性。此外,深度上下文在灵长类动物的运动场景中发挥着独特的作用,但在没有深度标签的情况下很少使用。在本文中,我们使用了一个更合适的多尺度金字塔多维长短期记忆(MSPMD-LSTM)架构来揭示视频帧内的强相关性。进一步,对深度上下文进行提取和细化,以提高模型的性能。实验表明,我们的模型在Youtube-Objects数据集和Segtrack v2数据集上产生了具有竞争力的结果。
{"title":"Video object segmentation by Multi-Scale Pyramidal Multi-Dimensional LSTM with generated depth context","authors":"Qiurui Wang, C. Yuan","doi":"10.1109/ICIP.2016.7532363","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532363","url":null,"abstract":"Existing deep neural networks, such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs), typically treat volumetric video data as several single images and deal with one frame at one time, thus the relevance to frames can hardly be fully exploited. Besides, depth context plays the unique role in motion scenes for primates, but is seldom used in no depth label situations. In this paper, we use a more suitable architecture Multi-Scale Pyramidal Multi-Dimensional Long Short Term Memory (MSPMD-LSTM) to reveal the strong relevance within video frames. Furthermore, depth context is extracted and refined to enhance the performance of the model. Experiments demonstrate that our models yield competitive results on Youtube-Objects dataset and Segtrack v2 dataset.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"115 1","pages":"281-285"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83530185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A novel covariance function for predicting vegetation biochemistry from hyperspectral imagery with Gaussian processes 基于高斯过程的高光谱影像植被生物化学预测新协方差函数
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532752
Utsav B. Gewali, S. Monteiro
Remotely extracting information about the biochemical properties of the materials in an environment from airborne- or satellite-based hyperspectral sensor has a variety of applications in forestry, agriculture, mining, environmental monitoring and space exploration. In this paper, we propose a new non-stationary covariance function, called exponential spectral angle mapper (ESAM) for predicting the biochemistry of vegetation from hyperspectral imagery using Gaussian processes. The proposed covariance function is based on the angle between the spectra, which is known to be a better measure of similarity for hyperspectral data due to its robustness to illumination variations. We demonstrate the efficacy of the proposed method with experiments on a real-world hy-perspectral dataset.
利用机载或卫星高光谱传感器远程提取环境中材料的生化特性信息,在林业、农业、矿业、环境监测和空间探索等领域有着广泛的应用。在本文中,我们提出了一个新的非平稳协方差函数,称为指数光谱角映射(ESAM),用于利用高斯过程预测高光谱图像中的植被生物化学。所提出的协方差函数基于光谱之间的角度,由于其对光照变化的鲁棒性,它被认为是高光谱数据相似性的更好度量。我们通过实际高光谱数据集的实验证明了所提出方法的有效性。
{"title":"A novel covariance function for predicting vegetation biochemistry from hyperspectral imagery with Gaussian processes","authors":"Utsav B. Gewali, S. Monteiro","doi":"10.1109/ICIP.2016.7532752","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532752","url":null,"abstract":"Remotely extracting information about the biochemical properties of the materials in an environment from airborne- or satellite-based hyperspectral sensor has a variety of applications in forestry, agriculture, mining, environmental monitoring and space exploration. In this paper, we propose a new non-stationary covariance function, called exponential spectral angle mapper (ESAM) for predicting the biochemistry of vegetation from hyperspectral imagery using Gaussian processes. The proposed covariance function is based on the angle between the spectra, which is known to be a better measure of similarity for hyperspectral data due to its robustness to illumination variations. We demonstrate the efficacy of the proposed method with experiments on a real-world hy-perspectral dataset.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"58 1","pages":"2216-2220"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90559051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Person re-identification using a patch-based appearance model 使用基于补丁的外观模型重新识别人员
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532460
Khalid Tahboub, Blanca Delgado, E. Delp
Person re-identification is the process of recognizing a person across a network of cameras with non-overlapping fields of view. In this paper we present an unsupervised multi-shot approach based on a patch-based dynamic appearance model. We use deformable graph matching for person re-identification using histograms of color and texture as features of nodes. Each graph model spans multiple images and each node is a local patch in the shape of a rectangle. We evaluate our proposed method on publicly available PRID 2011 and iLIDS-VID databases.
人的再识别是通过具有非重叠视场的摄像机网络识别一个人的过程。本文提出了一种基于补丁动态外观模型的无监督多镜头方法。我们利用颜色直方图和纹理直方图作为节点特征,使用可变形图匹配进行人物再识别。每个图模型跨越多个图像,每个节点是矩形形状的局部补丁。我们在公开的PRID 2011和iLIDS-VID数据库上对我们提出的方法进行了评估。
{"title":"Person re-identification using a patch-based appearance model","authors":"Khalid Tahboub, Blanca Delgado, E. Delp","doi":"10.1109/ICIP.2016.7532460","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532460","url":null,"abstract":"Person re-identification is the process of recognizing a person across a network of cameras with non-overlapping fields of view. In this paper we present an unsupervised multi-shot approach based on a patch-based dynamic appearance model. We use deformable graph matching for person re-identification using histograms of color and texture as features of nodes. Each graph model spans multiple images and each node is a local patch in the shape of a rectangle. We evaluate our proposed method on publicly available PRID 2011 and iLIDS-VID databases.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"29 1","pages":"764-768"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86659930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2016 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1