首页 > 最新文献

2016 IEEE International Conference on Image Processing (ICIP)最新文献

英文 中文
Adaptive reduced-set matching pursuit for compressed sensing recovery 压缩感知恢复的自适应约简集匹配追踪
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532809
Michael M. Abdel-Sayed, Ahmed K. F. Khattab, Mohamed Fathy Abu Elyazeed
Compressed sensing enables the acquisition of sparse signals at a rate that is much lower than the Nyquist rate. Various greedy recovery algorithms have been proposed to achieve a lower computational complexity compared to the optimal ℓ1 minimization, while maintaining a good reconstruction accuracy. We propose a new greedy recovery algorithm for compressed sensing, called the Adaptive Reduced-set Matching Pursuit (ARMP). Our algorithm achieves higher reconstruction accuracy at a significantly low computational complexity compared to existing greedy recovery algorithms. It is even superior to ℓ1 minimization in terms of the normalized time-error product, a metric that we introduced to measure the trade-off between the reconstruction time and error.
压缩感知能够以比奈奎斯特速率低得多的速率获取稀疏信号。各种贪婪恢复算法已经提出,以实现较低的计算复杂度相比,最优的最小化,同时保持良好的重建精度。我们提出了一种新的贪婪恢复算法,称为自适应约简集匹配追踪(ARMP)。与现有的贪婪恢复算法相比,我们的算法在较低的计算复杂度下实现了更高的重建精度。在标准化的时间误差积方面,它甚至优于l1最小化,我们引入了一个度量来衡量重建时间和误差之间的权衡。
{"title":"Adaptive reduced-set matching pursuit for compressed sensing recovery","authors":"Michael M. Abdel-Sayed, Ahmed K. F. Khattab, Mohamed Fathy Abu Elyazeed","doi":"10.1109/ICIP.2016.7532809","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532809","url":null,"abstract":"Compressed sensing enables the acquisition of sparse signals at a rate that is much lower than the Nyquist rate. Various greedy recovery algorithms have been proposed to achieve a lower computational complexity compared to the optimal ℓ1 minimization, while maintaining a good reconstruction accuracy. We propose a new greedy recovery algorithm for compressed sensing, called the Adaptive Reduced-set Matching Pursuit (ARMP). Our algorithm achieves higher reconstruction accuracy at a significantly low computational complexity compared to existing greedy recovery algorithms. It is even superior to ℓ1 minimization in terms of the normalized time-error product, a metric that we introduced to measure the trade-off between the reconstruction time and error.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"4 1","pages":"2499-2503"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86935242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Pedestrian detection in crowded scenes via scale and occlusion analysis 基于尺度和遮挡分析的拥挤场景行人检测
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532550
Lu Wang, Lisheng Xu, Ming-Hsuan Yang
Despite significant progress in pedestrian detection has been made in recent years, detecting pedestrians in crowded scenes remains a challenging problem. In this paper, we propose to use visual contexts based on scale and occlusion cues from detections at proximity to better detect pedestrians for surveillance applications. Specifically, we first apply detectors based on full body and parts to generate initial detections. Scale prior at each image location is estimated using the cues provided by neighboring detections, and the confidence score of each detection is refined according to its consistency with the estimated scale prior. Local occlusion analysis is exploited in refining detection confidence scores which facilitates the final detection cluster based Non-Maximum Suppression. Experimental results on benchmark data sets show that the proposed algorithm performs favorably against the state-of-the-art methods.
尽管近年来行人检测取得了重大进展,但在拥挤场景中检测行人仍然是一个具有挑战性的问题。在本文中,我们建议使用基于近距离检测的尺度和遮挡线索的视觉上下文来更好地检测行人的监视应用。具体来说,我们首先应用基于全身和部位的检测器来生成初始检测。利用相邻检测提供的线索估计每个图像位置的尺度先验,并根据其与估计的尺度先验的一致性对每个检测的置信度评分进行细化。利用局部遮挡分析来改进检测置信度分数,从而促进基于非最大抑制的最终检测聚类。在基准数据集上的实验结果表明,该算法与现有方法相比具有良好的性能。
{"title":"Pedestrian detection in crowded scenes via scale and occlusion analysis","authors":"Lu Wang, Lisheng Xu, Ming-Hsuan Yang","doi":"10.1109/ICIP.2016.7532550","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532550","url":null,"abstract":"Despite significant progress in pedestrian detection has been made in recent years, detecting pedestrians in crowded scenes remains a challenging problem. In this paper, we propose to use visual contexts based on scale and occlusion cues from detections at proximity to better detect pedestrians for surveillance applications. Specifically, we first apply detectors based on full body and parts to generate initial detections. Scale prior at each image location is estimated using the cues provided by neighboring detections, and the confidence score of each detection is refined according to its consistency with the estimated scale prior. Local occlusion analysis is exploited in refining detection confidence scores which facilitates the final detection cluster based Non-Maximum Suppression. Experimental results on benchmark data sets show that the proposed algorithm performs favorably against the state-of-the-art methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"45 1","pages":"1210-1214"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90738825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Sign language recognition with long short-term memory 具有长短期记忆的手语识别
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532884
Tao Liu, Wen-gang Zhou, Houqiang Li
Sign Language Recognition (SLR) aims at translating the Sign Language (SL) into speech or text, so as to facilitate the communication between hearing-impaired people and the normal people. This problem has broad social impact, however it is challenging due to the variation for different people and the complexity in sign words. Traditional methods for SLR generally use handcrafted feature and Hidden Markov Models (HMMs) modeling temporal information. But reliable handcrafted features are difficult to design and not able to adapt to the large variations of sign words. To approach this problem, considering that Long Short-Term memory (LSTM) can model the contextual information of temporal sequence well, we propose an end-to-end method for SLR based on LSTM. Our system takes the moving trajectories of 4 skeleton joints as inputs without any prior knowledge and is free of explicit feature design. To evaluate our proposed model, we built a large isolated Chinese sign language vocabulary with Kinect 2.0. Experimental results demonstrate the effectiveness of our approach compared with traditional HMM based methods.
手语识别(Sign Language Recognition, SLR)旨在将手语(Sign Language, SL)翻译成语音或文字,以方便听障人士与正常人之间的交流。这个问题具有广泛的社会影响,但由于不同人的差异和手语的复杂性,它具有挑战性。传统的单反方法一般使用手工特征和隐马尔可夫模型(hmm)来建模时间信息。但是,可靠的手工特征很难设计,并且不能适应大量变化的标志文字。为了解决这一问题,考虑到长短期记忆(LSTM)可以很好地模拟时间序列的上下文信息,我们提出了一种基于LSTM的端到端SLR方法。我们的系统以4个骨骼关节的运动轨迹作为输入,没有任何先验知识,也没有明确的特征设计。为了评估我们提出的模型,我们用Kinect 2.0建立了一个大型孤立的中文手语词汇表。实验结果表明,与传统的HMM方法相比,该方法是有效的。
{"title":"Sign language recognition with long short-term memory","authors":"Tao Liu, Wen-gang Zhou, Houqiang Li","doi":"10.1109/ICIP.2016.7532884","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532884","url":null,"abstract":"Sign Language Recognition (SLR) aims at translating the Sign Language (SL) into speech or text, so as to facilitate the communication between hearing-impaired people and the normal people. This problem has broad social impact, however it is challenging due to the variation for different people and the complexity in sign words. Traditional methods for SLR generally use handcrafted feature and Hidden Markov Models (HMMs) modeling temporal information. But reliable handcrafted features are difficult to design and not able to adapt to the large variations of sign words. To approach this problem, considering that Long Short-Term memory (LSTM) can model the contextual information of temporal sequence well, we propose an end-to-end method for SLR based on LSTM. Our system takes the moving trajectories of 4 skeleton joints as inputs without any prior knowledge and is free of explicit feature design. To evaluate our proposed model, we built a large isolated Chinese sign language vocabulary with Kinect 2.0. Experimental results demonstrate the effectiveness of our approach compared with traditional HMM based methods.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"85 1","pages":"2871-2875"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90989885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Subtle consumer-photo quality evaluation 微妙的消费者照片质量评估
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7533066
Michele A. Saad, David G. Nicholas, Patrick McKnight, Jake Quartuccio, Ramesh Jaladi, P. Corriveau
Subjective test methodologies are morphing to enable researchers to answer questions relevant to rapidly evolving technologies in an efficient and reliable manner. This paper is an exploration of how subjective testing that employs crowdsourcing can be refined to drive stability and reliability in subjective results. We investigate how various design decisions can lead to disparate subjective responses; motivated by the need for efficient acquisition of large volumes of data, and the need to understand and mitigate pitfalls in online tests, when the stimuli are complex and subtle as is the case in popular consumer scenarios.
主观测试方法正在演变,使研究人员能够以有效和可靠的方式回答与快速发展的技术相关的问题。本文探讨了如何改进采用众包的主观测试,以提高主观结果的稳定性和可靠性。我们研究了不同的设计决策如何导致不同的主观反应;由于需要有效地获取大量数据,并且需要理解和减轻在线测试中的陷阱,当刺激复杂而微妙时,就像在流行的消费者场景中一样。
{"title":"Subtle consumer-photo quality evaluation","authors":"Michele A. Saad, David G. Nicholas, Patrick McKnight, Jake Quartuccio, Ramesh Jaladi, P. Corriveau","doi":"10.1109/ICIP.2016.7533066","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7533066","url":null,"abstract":"Subjective test methodologies are morphing to enable researchers to answer questions relevant to rapidly evolving technologies in an efficient and reliable manner. This paper is an exploration of how subjective testing that employs crowdsourcing can be refined to drive stability and reliability in subjective results. We investigate how various design decisions can lead to disparate subjective responses; motivated by the need for efficient acquisition of large volumes of data, and the need to understand and mitigate pitfalls in online tests, when the stimuli are complex and subtle as is the case in popular consumer scenarios.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"67 1","pages":"3778-3782"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91264093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A new multi-criteria fusion model for color textured image segmentation 一种新的彩色纹理图像分割多准则融合模型
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532825
Lazhar Khelifi, M. Mignotte
Fusion of image segmentations using consensus clustering and based on the optimization of a single criterion (commonly called the median partition based approach) may bias and limit the performance of an image segmentation model. To address this issue, we propose, in this paper, a new fusion model of image segmentation based on multi-objective optimization which aims to avoid the bias caused by a single criterion and to achieve a final improved segmentation. The proposed fusion model combines two conflicting and complementary segmentation criteria, namely; the region-based variation of information (VoI) criterion and the contour-based F-Measure (precision-recall) criterion with an entropy-based confidence weighting factor. To optimize our energy-based model we use an optimization procedure derived from the iterative conditional modes (ICM) algorithm. The experimental results on the Berkeley database with manual ground truth segmentations clearly show the effectiveness and the robustness of our multi-objective median partition based approach.
使用共识聚类和基于单一标准优化(通常称为基于中位数划分的方法)的图像分割融合可能会影响图像分割模型的性能。针对这一问题,本文提出了一种新的基于多目标优化的图像分割融合模型,以避免单一准则带来的偏差,最终达到改进的分割效果。所提出的融合模型结合了两个相互冲突又互补的分割准则,即;基于区域的信息变异(VoI)准则和基于轮廓的F-Measure(精确召回率)准则,并带有基于熵的置信度加权因子。为了优化基于能量的模型,我们使用了从迭代条件模式(ICM)算法派生的优化程序。在Berkeley数据库上进行人工地真值分割的实验结果表明了多目标中值分割方法的有效性和鲁棒性。
{"title":"A new multi-criteria fusion model for color textured image segmentation","authors":"Lazhar Khelifi, M. Mignotte","doi":"10.1109/ICIP.2016.7532825","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532825","url":null,"abstract":"Fusion of image segmentations using consensus clustering and based on the optimization of a single criterion (commonly called the median partition based approach) may bias and limit the performance of an image segmentation model. To address this issue, we propose, in this paper, a new fusion model of image segmentation based on multi-objective optimization which aims to avoid the bias caused by a single criterion and to achieve a final improved segmentation. The proposed fusion model combines two conflicting and complementary segmentation criteria, namely; the region-based variation of information (VoI) criterion and the contour-based F-Measure (precision-recall) criterion with an entropy-based confidence weighting factor. To optimize our energy-based model we use an optimization procedure derived from the iterative conditional modes (ICM) algorithm. The experimental results on the Berkeley database with manual ground truth segmentations clearly show the effectiveness and the robustness of our multi-objective median partition based approach.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"345 1","pages":"2579-2583"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76573999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Shearlet-based reduced reference image quality assessment 基于shearlet的简化参考图像质量评估
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532719
S. Bosse, Qiaobo Chen, Mischa Siekmann, W. Samek, T. Wiegand
This paper proposes a reduced reference image quality assessment method using only a low number of features. It involves a shearlet decomposition, directional pooling of the obtained coefficient and extracts the scalewise statistical location parameter as a feature. The proposed method is tested and compared to similar approaches on the LIVE image database. On this database it outperforms the compared methods on five of seven distortion types and on the full testset with a linear correlation of = 0.89.
本文提出了一种仅使用少量特征的简化参考图像质量评估方法。它涉及剪切波分解,对得到的系数进行定向池化,并提取按比例的统计位置参数作为特征。在LIVE图像数据库上对该方法进行了测试,并与类似方法进行了比较。在这个数据库中,它在7种失真类型中的5种和完整测试集上的性能优于比较方法,线性相关性为0.89。
{"title":"Shearlet-based reduced reference image quality assessment","authors":"S. Bosse, Qiaobo Chen, Mischa Siekmann, W. Samek, T. Wiegand","doi":"10.1109/ICIP.2016.7532719","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532719","url":null,"abstract":"This paper proposes a reduced reference image quality assessment method using only a low number of features. It involves a shearlet decomposition, directional pooling of the obtained coefficient and extracts the scalewise statistical location parameter as a feature. The proposed method is tested and compared to similar approaches on the LIVE image database. On this database it outperforms the compared methods on five of seven distortion types and on the full testset with a linear correlation of = 0.89.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"27 1","pages":"2052-2056"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78514716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
SIASM: Sparsity-based image alignment and stitching method for robust image mosaicking 基于稀疏性的图像对齐和拼接方法,用于鲁棒图像拼接
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532674
Yuelong Li, V. Monga
Image alignment and stitching continue to be the topics of great interest. Image mosaicking is a key application that involves both alignment and stitching of multiple images. Despite significant previous effort, existing methods have limited robustness in dealing with occlusions and local object motion in different captures. To address this issue, we investigate the potential of applying sparsity-based methods to the task of image alignment and stitching. We formulate the alignment problem as a low-rank and sparse matrix decomposition problem under incomplete observations (multiple parts of a scene), and the stitching problem as a multiple labeling problem which utilizes the sparse components. Additionally we develop efficient algorithms for solving them. Unlike typical pairwise alignment manners in classical image alignment algorithms, our algorithm is capable of simultaneously aligning multiple images, making full use of inter-frame relationships among all images. Experimental results demonstrate that the proposed algorithm is capable of generating artifact-free stitched image mosaics that are robust against occlusions and object motion.
图像对齐和拼接仍然是人们非常感兴趣的话题。图像拼接是一项关键的应用,涉及多个图像的对齐和拼接。尽管之前做了大量的工作,但现有方法在处理不同捕获的遮挡和局部目标运动方面的鲁棒性有限。为了解决这个问题,我们研究了将基于稀疏性的方法应用于图像对齐和拼接任务的潜力。我们将对齐问题表述为不完全观测(场景的多个部分)下的低秩稀疏矩阵分解问题,将拼接问题表述为利用稀疏分量的多重标记问题。此外,我们开发了有效的算法来解决它们。与经典图像对齐算法中典型的两两对齐方式不同,我们的算法能够同时对齐多幅图像,充分利用了所有图像之间的帧间关系。实验结果表明,该算法能够生成无伪影的拼接图像,对遮挡和目标运动具有较强的鲁棒性。
{"title":"SIASM: Sparsity-based image alignment and stitching method for robust image mosaicking","authors":"Yuelong Li, V. Monga","doi":"10.1109/ICIP.2016.7532674","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532674","url":null,"abstract":"Image alignment and stitching continue to be the topics of great interest. Image mosaicking is a key application that involves both alignment and stitching of multiple images. Despite significant previous effort, existing methods have limited robustness in dealing with occlusions and local object motion in different captures. To address this issue, we investigate the potential of applying sparsity-based methods to the task of image alignment and stitching. We formulate the alignment problem as a low-rank and sparse matrix decomposition problem under incomplete observations (multiple parts of a scene), and the stitching problem as a multiple labeling problem which utilizes the sparse components. Additionally we develop efficient algorithms for solving them. Unlike typical pairwise alignment manners in classical image alignment algorithms, our algorithm is capable of simultaneously aligning multiple images, making full use of inter-frame relationships among all images. Experimental results demonstrate that the proposed algorithm is capable of generating artifact-free stitched image mosaics that are robust against occlusions and object motion.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"99 1","pages":"1828-1832"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78440811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Self-restraint object recognition by model based CNN learning 基于模型CNN学习的自我约束目标识别
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532438
Yida Wang, Weihong Deng
CNN has shown excellent performance on object recognition based on huge amount of real images. For training with synthetic data rendered from 3D models alone to reduce the workload of collecting real images, we propose a concatenated self-restraint learning structure lead by a triplet and softmax jointed loss function for object recognition. Locally connected auto encoder trained from rendered images with and without background used for object reconstruction against environment variables produces an additional channel automatically concatenated to RGB channels as input of classification network. This structure makes it possible training a softmax classifier directly from CNN based on synthetic data with our rendering strategy. Our structure halves the gap between training based on real photos and 3D model in both PASCAL and ImageNet database compared to GoogleNet.
CNN在基于大量真实图像的目标识别方面表现出了优异的性能。为了减少采集真实图像的工作量,我们提出了一种由triplet和softmax联合损失函数引导的连接自我约束学习结构,用于物体识别。局部连接的自动编码器从有背景和没有背景的渲染图像中训练,用于根据环境变量进行对象重建,产生一个额外的通道,自动连接到RGB通道作为分类网络的输入。这种结构使得使用我们的渲染策略直接从CNN训练一个基于合成数据的softmax分类器成为可能。与GoogleNet相比,我们的结构将PASCAL和ImageNet数据库中基于真实照片和3D模型的训练差距缩小了一半。
{"title":"Self-restraint object recognition by model based CNN learning","authors":"Yida Wang, Weihong Deng","doi":"10.1109/ICIP.2016.7532438","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532438","url":null,"abstract":"CNN has shown excellent performance on object recognition based on huge amount of real images. For training with synthetic data rendered from 3D models alone to reduce the workload of collecting real images, we propose a concatenated self-restraint learning structure lead by a triplet and softmax jointed loss function for object recognition. Locally connected auto encoder trained from rendered images with and without background used for object reconstruction against environment variables produces an additional channel automatically concatenated to RGB channels as input of classification network. This structure makes it possible training a softmax classifier directly from CNN based on synthetic data with our rendering strategy. Our structure halves the gap between training based on real photos and 3D model in both PASCAL and ImageNet database compared to GoogleNet.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"14 1","pages":"654-658"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73425893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A novel structural variation detection strategy for image quality assessment 一种新的图像质量结构变化检测策略
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532723
Yibing Zhan, Rong Zhang
Structural information is critical in image quality assessment (IQA). Although existing objective IQA methods have achieved high consistency with subjective perception, detecting structural variation remains a difficult task. In this paper, we propose a novel structural variation detection strategy that is based on binary logic and inspired by the bag-of-words model. The proposed strategy detects structural variation by comparing the occurrences of structural features within the original and distorted images. In order to show the effectiveness of this strategy, this paper also proposes a novel and simple IQA method based on this strategy. The proposed method evaluates the image quality from two aspects: the structure distortion and the luminance distortion. The experimental results from four public databases show that the proposed method is highly congruous with subjective evaluation. The results also prove that the detection strategy is useful.
结构信息是图像质量评估(IQA)的关键。虽然现有的客观IQA方法与主观感知的一致性很高,但检测结构变化仍然是一项艰巨的任务。在本文中,我们提出了一种基于二元逻辑并受词袋模型启发的结构变异检测策略。该策略通过比较原始图像和扭曲图像中结构特征的出现情况来检测结构变化。为了证明该策略的有效性,本文还提出了一种基于该策略的新颖而简单的IQA方法。该方法从结构畸变和亮度畸变两方面对图像质量进行评价。四个公共数据库的实验结果表明,该方法与主观评价高度一致。结果也证明了该检测策略的有效性。
{"title":"A novel structural variation detection strategy for image quality assessment","authors":"Yibing Zhan, Rong Zhang","doi":"10.1109/ICIP.2016.7532723","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532723","url":null,"abstract":"Structural information is critical in image quality assessment (IQA). Although existing objective IQA methods have achieved high consistency with subjective perception, detecting structural variation remains a difficult task. In this paper, we propose a novel structural variation detection strategy that is based on binary logic and inspired by the bag-of-words model. The proposed strategy detects structural variation by comparing the occurrences of structural features within the original and distorted images. In order to show the effectiveness of this strategy, this paper also proposes a novel and simple IQA method based on this strategy. The proposed method evaluates the image quality from two aspects: the structure distortion and the luminance distortion. The experimental results from four public databases show that the proposed method is highly congruous with subjective evaluation. The results also prove that the detection strategy is useful.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"147 1","pages":"2072-2076"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73722827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Image retargeting based on spatially varying defocus blur map 基于空间变化散焦模糊图的图像重定位
Pub Date : 2016-09-01 DOI: 10.1109/ICIP.2016.7532848
Ali Karaali, C. Jung
This paper presents a new image retargeting method that explores blur information. Given the input image, we compute the blur map and estimate in-focus regions. For retargeting, we first try to crop image boundaries as much as possible (preserving in-focus regions). If cropping is not enough, we use seam carving exploring a novel blur-aware energy function that concentrates the seams in blurred regions of the image. Experimental results show that the proposed blur-aware retargeting scheme works better at preserving in-focus objects than other competitive retargeting algorithms.
本文提出了一种利用模糊信息进行图像重定位的新方法。给定输入图像,我们计算模糊映射并估计焦点区域。为了重新定位,我们首先尝试尽可能地裁剪图像边界(保留焦点区域)。如果裁剪不够,我们使用接缝雕刻探索一种新的模糊感知能量函数,将图像模糊区域的接缝集中在一起。实验结果表明,所提出的模糊感知重定向方案比其他竞争性重定向算法能更好地保留焦点内的目标。
{"title":"Image retargeting based on spatially varying defocus blur map","authors":"Ali Karaali, C. Jung","doi":"10.1109/ICIP.2016.7532848","DOIUrl":"https://doi.org/10.1109/ICIP.2016.7532848","url":null,"abstract":"This paper presents a new image retargeting method that explores blur information. Given the input image, we compute the blur map and estimate in-focus regions. For retargeting, we first try to crop image boundaries as much as possible (preserving in-focus regions). If cropping is not enough, we use seam carving exploring a novel blur-aware energy function that concentrates the seams in blurred regions of the image. Experimental results show that the proposed blur-aware retargeting scheme works better at preserving in-focus objects than other competitive retargeting algorithms.","PeriodicalId":6521,"journal":{"name":"2016 IEEE International Conference on Image Processing (ICIP)","volume":"84 1","pages":"2693-2697"},"PeriodicalIF":0.0,"publicationDate":"2016-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79339218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2016 IEEE International Conference on Image Processing (ICIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1