首页 > 最新文献

2007 IEEE International Conference on Image Processing最新文献

英文 中文
Semantics-Based Video Indexing using a Stochastic Modeling Approach 基于语义的随机建模视频索引
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4380017
Yong Wei, S. Bhandarkar, Kang Li
Semantic video indexing is the first step towards automatic video retrieval and personalization. We propose a data-driven stochastic modeling approach to perform both video segmentation and video indexing in a single pass. Compared with the existing hidden Markov model (HMM)-based video segmentation and indexing techniques, the advantages of the proposed approach are as follows: (1) the probabilistic grammar defining the video program is generated entirely from the training data allowing the proposed approach to handle various kinds of videos without having to manually redefine the program model; (2) the proposed use of the Tamura features improves the accuracy of temporal segmentation and indexing; (3) the need to use an HMM to model the video edit effects is obviated thus simplifying the processing and collection of training data and ensuring that all video segments in the database are labeled with concepts that have clear semantic meanings in order to facilitate semantics-based video retrieval. Experimental results on broadcast news video are presented.
语义视频索引是实现视频自动检索和个性化的第一步。我们提出了一个数据驱动的随机建模方法来执行视频分割和视频索引在一个单一的通道。与现有的基于隐马尔可夫模型(HMM)的视频分割和索引技术相比,该方法具有以下优点:(1)定义视频节目的概率语法完全由训练数据生成,使得该方法无需手动重新定义节目模型即可处理各种类型的视频;(2) Tamura特征的使用提高了时间分割和索引的准确性;(3)消除了使用HMM对视频编辑效果建模的需要,从而简化了训练数据的处理和收集,并确保数据库中的所有视频片段都标注了具有明确语义的概念,以便于基于语义的视频检索。给出了广播新闻视频的实验结果。
{"title":"Semantics-Based Video Indexing using a Stochastic Modeling Approach","authors":"Yong Wei, S. Bhandarkar, Kang Li","doi":"10.1109/ICIP.2007.4380017","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4380017","url":null,"abstract":"Semantic video indexing is the first step towards automatic video retrieval and personalization. We propose a data-driven stochastic modeling approach to perform both video segmentation and video indexing in a single pass. Compared with the existing hidden Markov model (HMM)-based video segmentation and indexing techniques, the advantages of the proposed approach are as follows: (1) the probabilistic grammar defining the video program is generated entirely from the training data allowing the proposed approach to handle various kinds of videos without having to manually redefine the program model; (2) the proposed use of the Tamura features improves the accuracy of temporal segmentation and indexing; (3) the need to use an HMM to model the video edit effects is obviated thus simplifying the processing and collection of training data and ensuring that all video segments in the database are labeled with concepts that have clear semantic meanings in order to facilitate semantics-based video retrieval. Experimental results on broadcast news video are presented.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124498671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Common Spatial Pattern Discovery by Efficient Candidate Pruning 基于高效候选剪枝的公共空间模式发现
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4378917
Junsong Yuan, Zhu Li, Yun Fu, Ying Wu, Thomas S. Huang
Automatically discovering common visual patterns in images is very challenging due to the uncertainties in the visual appearances of such spatial patterns and the enormous computational cost involved in exploring the huge solution space. Instead of performing exhaustive search on all possible candidates of such spatial patterns at various locations and scales, this paper presents a novel and very efficient algorithm for discovering common visual patterns by designing a provably correct and computationally efficient pruning procedure that has a quadratic complexity. This new approach is able to efficiently search a set of images for unknown visual patterns that exhibit large appearance variations because of rotation, scale changes, slight view changes, color variations and partial occlusions.
自动发现图像中常见的视觉模式是非常具有挑战性的,因为这种空间模式的视觉外观具有不确定性,并且在探索巨大的解空间时涉及巨大的计算成本。本文提出了一种新的、非常有效的算法,通过设计一个可证明正确的、计算效率高的二次复杂度剪枝过程,来发现常见的视觉模式,而不是在不同位置和尺度上对所有可能的候选空间模式进行穷尽搜索。这种新方法能够有效地搜索一组图像,寻找由于旋转、比例变化、轻微视图变化、颜色变化和部分遮挡而表现出巨大外观变化的未知视觉模式。
{"title":"Common Spatial Pattern Discovery by Efficient Candidate Pruning","authors":"Junsong Yuan, Zhu Li, Yun Fu, Ying Wu, Thomas S. Huang","doi":"10.1109/ICIP.2007.4378917","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4378917","url":null,"abstract":"Automatically discovering common visual patterns in images is very challenging due to the uncertainties in the visual appearances of such spatial patterns and the enormous computational cost involved in exploring the huge solution space. Instead of performing exhaustive search on all possible candidates of such spatial patterns at various locations and scales, this paper presents a novel and very efficient algorithm for discovering common visual patterns by designing a provably correct and computationally efficient pruning procedure that has a quadratic complexity. This new approach is able to efficiently search a set of images for unknown visual patterns that exhibit large appearance variations because of rotation, scale changes, slight view changes, color variations and partial occlusions.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126243001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
A Wavelet-Based Noise-Aware Method for Fusing Noisy Imagery 基于小波的噪声感知图像融合方法
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379602
Xiaohui Yuan, B. Buckles
Fusion of images in the presence of noise is a challenging problem. Conventional fusion methods focus on aggregating prominent image features, which usually result in noise enhancement. To address this problem, we developed a wavelet-based, noise-aware fusion method that distinguishes signal and noise coefficients on-the-fly and fuses them with weighted averaging and majority voting respectively. Our method retains coefficients that reconstruct salient features, whereas noise components are discarded. The performance is evaluated in terms of noise removal and feature retention. The comparisons with five state-of-the-art fusion methods and a combination with denoising method demonstrated that our method significantly outperformed the existing techniques with noisy inputs.
存在噪声的图像融合是一个具有挑战性的问题。传统的融合方法主要集中在聚集突出的图像特征,这通常会导致噪声增强。为了解决这个问题,我们开发了一种基于小波的噪声感知融合方法,该方法可以实时区分信号和噪声系数,并分别使用加权平均和多数投票进行融合。我们的方法保留了重建显著特征的系数,而丢弃了噪声成分。性能评估方面的噪声去除和特征保留。通过与五种最先进的融合方法的比较以及与去噪方法的结合表明,我们的方法明显优于现有的带有噪声输入的技术。
{"title":"A Wavelet-Based Noise-Aware Method for Fusing Noisy Imagery","authors":"Xiaohui Yuan, B. Buckles","doi":"10.1109/ICIP.2007.4379602","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379602","url":null,"abstract":"Fusion of images in the presence of noise is a challenging problem. Conventional fusion methods focus on aggregating prominent image features, which usually result in noise enhancement. To address this problem, we developed a wavelet-based, noise-aware fusion method that distinguishes signal and noise coefficients on-the-fly and fuses them with weighted averaging and majority voting respectively. Our method retains coefficients that reconstruct salient features, whereas noise components are discarded. The performance is evaluated in terms of noise removal and feature retention. The comparisons with five state-of-the-art fusion methods and a combination with denoising method demonstrated that our method significantly outperformed the existing techniques with noisy inputs.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126267990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Optimised Compression Strategy in Wavelet-Based Video Coding using Improved Context Models 基于改进上下文模型的小波视频编码优化压缩策略
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379331
Toni Zgaljic, M. Mrak, E. Izquierdo
Accurate probability estimation is a key to efficient compression in entropy coding phase of state-of-the-art video coding systems. Probability estimation can be enhanced if contexts in which symbols occur are used during the probability estimation phase. However, these contexts have to be carefully designed in order to avoid negative effects. Methods that use tree structures to model contexts of various syntax elements have been proven efficient in image and video coding. In this paper we use such structure to build optimised contexts for application in scalable wavelet-based video coding. With the proposed approach context are designed separately for intra-coded frames and motion-compensated frames considering varying statistics across different spatio-temporal subbands. Moreover, contexts are separately designed for different bit-planes. Comparison with compression using fixed contexts from embedded ZeroBlock coding (EZBC) has been performed showing improvements when context modelling on tree structures is applied.
在目前最先进的视频编码系统中,准确的概率估计是保证熵编码阶段有效压缩的关键。如果在概率估计阶段使用出现符号的上下文,则可以增强概率估计。但是,必须仔细设计这些环境,以避免负面影响。在图像和视频编码中,使用树形结构对各种语法元素的上下文建模的方法已被证明是有效的。在本文中,我们使用这种结构来构建优化的上下文,用于可扩展的基于小波的视频编码。考虑到不同时空子带的不同统计信息,该方法分别为编码内帧和运动补偿帧设计上下文。此外,上下文是针对不同的位平面单独设计的。与使用来自嵌入式零块编码(EZBC)的固定上下文的压缩进行了比较,显示了在树形结构上应用上下文建模时的改进。
{"title":"Optimised Compression Strategy in Wavelet-Based Video Coding using Improved Context Models","authors":"Toni Zgaljic, M. Mrak, E. Izquierdo","doi":"10.1109/ICIP.2007.4379331","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379331","url":null,"abstract":"Accurate probability estimation is a key to efficient compression in entropy coding phase of state-of-the-art video coding systems. Probability estimation can be enhanced if contexts in which symbols occur are used during the probability estimation phase. However, these contexts have to be carefully designed in order to avoid negative effects. Methods that use tree structures to model contexts of various syntax elements have been proven efficient in image and video coding. In this paper we use such structure to build optimised contexts for application in scalable wavelet-based video coding. With the proposed approach context are designed separately for intra-coded frames and motion-compensated frames considering varying statistics across different spatio-temporal subbands. Moreover, contexts are separately designed for different bit-planes. Comparison with compression using fixed contexts from embedded ZeroBlock coding (EZBC) has been performed showing improvements when context modelling on tree structures is applied.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126298439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Subspace Extension to Phase Correlation Approach for Fast Image Registration 快速图像配准的相位相关子空间扩展方法
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4378996
Jinchang Ren, T. Vlachos, Jianmin Jiang
A novel extension of phase correlation to subspace correlation is proposed, in which 2-D translation is decomposed into two 1-D motions thus only 1-D Fourier transform is used to estimate the corresponding motion. In each subspace, the first two highest peaks from 1-D correlation are linearly interpolated for subpixel accuracy. Experimental results have shown both the robustness and accuracy of our method.
提出了一种将相位相关扩展到子空间相关的新方法,将二维平移分解为两个一维运动,从而只用一维傅里叶变换来估计相应的运动。在每个子空间中,为了达到亚像素精度,对一维相关的前两个峰值进行线性插值。实验结果表明了该方法的鲁棒性和准确性。
{"title":"Subspace Extension to Phase Correlation Approach for Fast Image Registration","authors":"Jinchang Ren, T. Vlachos, Jianmin Jiang","doi":"10.1109/ICIP.2007.4378996","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4378996","url":null,"abstract":"A novel extension of phase correlation to subspace correlation is proposed, in which 2-D translation is decomposed into two 1-D motions thus only 1-D Fourier transform is used to estimate the corresponding motion. In each subspace, the first two highest peaks from 1-D correlation are linearly interpolated for subpixel accuracy. Experimental results have shown both the robustness and accuracy of our method.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126459225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Detection and Recovery of Film Dirt for Archive Restoration Applications 胶片污垢的检测与修复在档案修复中的应用
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379943
Jinchang Ren, T. Vlachos
A novel spatio-temporal method is proposed for film dirt detection and recovery. Firstly, a more reliable confidence measurement of dirt is extracted for color films. False alarms caused by motion are filtered using consistency checks among several measurements. Then, candidate dirt is detected by filtering and thresholding this confidence measurement. Finally, bi-directional local motion compensation and ML3Dex filtering are taken for the recovery of dirt pixels. Experiments on real data demonstrate the efficiency and effectiveness of our method in terms of both detection and recovery of dirt.
提出了一种新的薄膜污垢检测与恢复的时空方法。首先,对彩色胶片提取更可靠的污物置信度。由运动引起的假警报通过几个测量之间的一致性检查来过滤。然后,通过对该置信度测量值进行滤波和阈值处理来检测候选污垢。最后,采用双向局部运动补偿和ML3Dex滤波对污物像素进行恢复。在实际数据上的实验证明了该方法在污垢检测和回收方面的效率和有效性。
{"title":"Detection and Recovery of Film Dirt for Archive Restoration Applications","authors":"Jinchang Ren, T. Vlachos","doi":"10.1109/ICIP.2007.4379943","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379943","url":null,"abstract":"A novel spatio-temporal method is proposed for film dirt detection and recovery. Firstly, a more reliable confidence measurement of dirt is extracted for color films. False alarms caused by motion are filtered using consistency checks among several measurements. Then, candidate dirt is detected by filtering and thresholding this confidence measurement. Finally, bi-directional local motion compensation and ML3Dex filtering are taken for the recovery of dirt pixels. Experiments on real data demonstrate the efficiency and effectiveness of our method in terms of both detection and recovery of dirt.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128151155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Temporally Consistent Gaussian Random Field for Video Semantic Analysis 视频语义分析的时间一致高斯随机场
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4380070
Jinhui Tang, Xiansheng Hua, Tao Mei, Guo-Jun Qi, Shipeng Li, Xiuqing Wu
As a major family of semi-supervised learning, graph based semi-supervised learning methods have attracted lots of interests in the machine learning community as well as many application areas recently. However, for the application of video semantic annotation, these methods only consider the relations among samples in the feature space and neglect an intrinsic property of video data: the temporally adjacent video segments (e.g., shots) usually have similar semantic concept. In this paper, we adapt this temporal consistency property of video data into graph based semi-supervised learning and propose a novel method named temporally consistent Gaussian random field (TCGRF) to improve the annotation results. Experiments conducted on the TREC VID data set have demonstrated its effectiveness.
基于图的半监督学习方法作为半监督学习的一个主要分支,近年来引起了机器学习界和许多应用领域的广泛关注。然而,对于视频语义注释的应用,这些方法只考虑了特征空间中样本之间的关系,而忽略了视频数据的一个内在属性:在时间上相邻的视频片段(如镜头)通常具有相似的语义概念。本文将视频数据的这种时间一致性特性应用到基于图的半监督学习中,提出了一种名为时间一致高斯随机场(TCGRF)的新方法来改善标注结果。在TREC VID数据集上进行的实验证明了该方法的有效性。
{"title":"Temporally Consistent Gaussian Random Field for Video Semantic Analysis","authors":"Jinhui Tang, Xiansheng Hua, Tao Mei, Guo-Jun Qi, Shipeng Li, Xiuqing Wu","doi":"10.1109/ICIP.2007.4380070","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4380070","url":null,"abstract":"As a major family of semi-supervised learning, graph based semi-supervised learning methods have attracted lots of interests in the machine learning community as well as many application areas recently. However, for the application of video semantic annotation, these methods only consider the relations among samples in the feature space and neglect an intrinsic property of video data: the temporally adjacent video segments (e.g., shots) usually have similar semantic concept. In this paper, we adapt this temporal consistency property of video data into graph based semi-supervised learning and propose a novel method named temporally consistent Gaussian random field (TCGRF) to improve the annotation results. Experiments conducted on the TREC VID data set have demonstrated its effectiveness.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125766516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image Resolution Enhancement using Wavelet Domain Hidden Markov Tree and Coefficient Sign Estimation 基于小波域隐马尔可夫树和系数符号估计的图像分辨率增强
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379845
A. Temi̇zel
Image resolution enhancement using wavelets is a relatively new subject and many new algorithms have been proposed recently. These algorithms assume that the low resolution image is the approximation subband of a higher resolution image and attempts to estimate the unknown detail coefficients to reconstruct a high resolution image. A subset of these recent approaches utilized probabilistic models to estimate these unknown coefficients. Particularly, hidden Markov tree (HMT) based methods using Gaussian mixture models have been shown to produce promising results. However, one drawback of these methods is that, as the Gaussian is symmetrical around zero, signs of the coefficients generated using this distribution function are inherently random, adversely affecting the resulting image quality. In this paper, we demonstrate that, sign information is an important element affecting the results and propose a method to estimate signs of these coefficients more accurately.
利用小波增强图像分辨率是一门较新的学科,近年来提出了许多新的算法。这些算法假设低分辨率图像是高分辨率图像的近似子带,并尝试估计未知细节系数以重建高分辨率图像。这些最新方法的一个子集利用概率模型来估计这些未知系数。特别是,基于隐马尔可夫树(HMT)的方法使用高斯混合模型已被证明产生有希望的结果。然而,这些方法的一个缺点是,由于高斯分布在零附近是对称的,使用该分布函数生成的系数的符号本质上是随机的,从而对生成的图像质量产生不利影响。在本文中,我们证明了符号信息是影响结果的重要因素,并提出了一种更准确地估计这些系数的符号的方法。
{"title":"Image Resolution Enhancement using Wavelet Domain Hidden Markov Tree and Coefficient Sign Estimation","authors":"A. Temi̇zel","doi":"10.1109/ICIP.2007.4379845","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379845","url":null,"abstract":"Image resolution enhancement using wavelets is a relatively new subject and many new algorithms have been proposed recently. These algorithms assume that the low resolution image is the approximation subband of a higher resolution image and attempts to estimate the unknown detail coefficients to reconstruct a high resolution image. A subset of these recent approaches utilized probabilistic models to estimate these unknown coefficients. Particularly, hidden Markov tree (HMT) based methods using Gaussian mixture models have been shown to produce promising results. However, one drawback of these methods is that, as the Gaussian is symmetrical around zero, signs of the coefficients generated using this distribution function are inherently random, adversely affecting the resulting image quality. In this paper, we demonstrate that, sign information is an important element affecting the results and propose a method to estimate signs of these coefficients more accurately.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121616622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 93
Design and Implementation of a Real-Time Global Tone Mapping Processor for High Dynamic Range Video 高动态范围视频实时全局色调映射处理器的设计与实现
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379558
Tsun-Hsien Wang, Wei-Su Wong, F. Chen, C. Chiu
As the development in high dynamic range (HDR) video capture technologies, the bit-depth video encoding and decoding has become an interesting topic. In this paper, we show that the real-time HDR video display is possible. A tone mapping based HDR video architecture pipelined with a video CODEC is presented. The HDR video is compressed by the tone mapping processor. The compressed HDR video can be encoded and decoded by the video standards, such as MPEG2, MPEG4 or H.264 for transmission and display. We propose and implement a modified photographic tone mapping algorithm for the tone mapping processor. The required luminance wordlength in the processor is analyzed and the quantization error is estimated. We also develop the digit-by-digit exponent and logarithm hardware architecture for the tone mapping processor. The synthesized results show that our real-time tone mapping processor can process a NTSC video with 720*480 resolution at 30 frames per second.
随着高动态范围(HDR)视频捕获技术的发展,位深度视频编解码成为人们关注的话题。在本文中,我们证明了实时HDR视频显示是可能的。提出了一种基于音色映射的HDR视频架构,并将其与视频编解码器流水线化。HDR视频通过色调映射处理器进行压缩。压缩后的HDR视频可以按照MPEG2、MPEG4或H.264等视频标准进行编码和解码,进行传输和显示。针对色调映射处理器,提出并实现了一种改进的摄影色调映射算法。分析了处理器所需的亮度字长,估计了量化误差。我们还开发了音调映射处理器的逐位指数和对数硬件架构。合成结果表明,我们的实时色调映射处理器能够以30帧/秒的速度处理720*480分辨率的NTSC视频。
{"title":"Design and Implementation of a Real-Time Global Tone Mapping Processor for High Dynamic Range Video","authors":"Tsun-Hsien Wang, Wei-Su Wong, F. Chen, C. Chiu","doi":"10.1109/ICIP.2007.4379558","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379558","url":null,"abstract":"As the development in high dynamic range (HDR) video capture technologies, the bit-depth video encoding and decoding has become an interesting topic. In this paper, we show that the real-time HDR video display is possible. A tone mapping based HDR video architecture pipelined with a video CODEC is presented. The HDR video is compressed by the tone mapping processor. The compressed HDR video can be encoded and decoded by the video standards, such as MPEG2, MPEG4 or H.264 for transmission and display. We propose and implement a modified photographic tone mapping algorithm for the tone mapping processor. The required luminance wordlength in the processor is analyzed and the quantization error is estimated. We also develop the digit-by-digit exponent and logarithm hardware architecture for the tone mapping processor. The synthesized results show that our real-time tone mapping processor can process a NTSC video with 720*480 resolution at 30 frames per second.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121725521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Kernels on Bags of Fuzzy Regions for Fast Object retrieval 基于模糊区域袋的快速目标检索核算法
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4378920
P. Gosselin, M. Cord, S. Philipp-Foliguet
We propose in this paper a general kernel framework to deal with database object retrieval embedded in images with heterogeneous background. We use local features computed on fuzzy regions for image representation summarized in bags, and we propose original kernel functions to deal with sets of features and spatial constraints. Combined with SVMs classification and online learning scheme, the resulting algorithm satisfies the robustness requirements for representation and classification of objects. Experiments on a specific database having objects with heterogeneous backgrounds show the performance of our object retrieval technique.
本文提出了一个通用的内核框架来处理嵌入在异构背景图像中的数据库对象检索问题。我们使用在模糊区域上计算的局部特征进行图像表示,并提出原始核函数来处理特征集和空间约束。结合支持向量机分类和在线学习方案,得到的算法满足对象表示和分类的鲁棒性要求。在一个具有异构背景对象的数据库上进行了实验,验证了本文方法的有效性。
{"title":"Kernels on Bags of Fuzzy Regions for Fast Object retrieval","authors":"P. Gosselin, M. Cord, S. Philipp-Foliguet","doi":"10.1109/ICIP.2007.4378920","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4378920","url":null,"abstract":"We propose in this paper a general kernel framework to deal with database object retrieval embedded in images with heterogeneous background. We use local features computed on fuzzy regions for image representation summarized in bags, and we propose original kernel functions to deal with sets of features and spatial constraints. Combined with SVMs classification and online learning scheme, the resulting algorithm satisfies the robustness requirements for representation and classification of objects. Experiments on a specific database having objects with heterogeneous backgrounds show the performance of our object retrieval technique.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115917418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
期刊
2007 IEEE International Conference on Image Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1