首页 > 最新文献

2007 IEEE International Conference on Image Processing最新文献

英文 中文
Time-Varying Linear Autoregressive Models for Segmentation 时变线性自回归分割模型
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379003
Charles Florin, N. Paragios, G. Funka-Lea, James P. Williams
Tracking highly deforming structures in space and time arises in numerous applications in computer vision. Static Models are often referred to as linear combinations of a mean model and modes of variation learned from training examples. In Dynamic Modeling, the shape is represented as a function of shapes at previous time steps. In this paper, we introduce a novel technique that uses the spatial and the temporal information on the object deformation. We reformulate tracking as a high order time series prediction mechanism that adapts itself on-line to the newest results. Samples (toward dimensionality reduction) are represented in an orthogonal basis, and are introduced in an auto-regressive model that is determined through an optimization process in appropriate metric spaces. Toward capturing evolving deformations as well as cases that have not been part of the learning stage, a process that updates on-line both the orthogonal basis decomposition and the parameters of the autoregressive model is proposed. Experimental results with a nonstationary dynamic system prove adaptive AR models give better results than both stationary models and models learned over the whole sequence.
在空间和时间上跟踪高度变形的结构在计算机视觉中有许多应用。静态模型通常被称为均值模型和从训练样本中学习到的变化模式的线性组合。在动态建模中,形状表示为前一个时间步的形状的函数。在本文中,我们介绍了一种利用物体变形的空间和时间信息的新技术。我们将跟踪重新定义为一种高阶时间序列预测机制,它可以在线适应最新的结果。样本(向降维方向)以正交基表示,并引入通过适当度量空间中的优化过程确定的自回归模型。为了捕获演化变形以及未参与学习阶段的情况,提出了一种在线更新正交基分解和自回归模型参数的过程。对非平稳动态系统的实验结果表明,自适应增强现实模型比平稳模型和全序列学习模型具有更好的效果。
{"title":"Time-Varying Linear Autoregressive Models for Segmentation","authors":"Charles Florin, N. Paragios, G. Funka-Lea, James P. Williams","doi":"10.1109/ICIP.2007.4379003","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379003","url":null,"abstract":"Tracking highly deforming structures in space and time arises in numerous applications in computer vision. Static Models are often referred to as linear combinations of a mean model and modes of variation learned from training examples. In Dynamic Modeling, the shape is represented as a function of shapes at previous time steps. In this paper, we introduce a novel technique that uses the spatial and the temporal information on the object deformation. We reformulate tracking as a high order time series prediction mechanism that adapts itself on-line to the newest results. Samples (toward dimensionality reduction) are represented in an orthogonal basis, and are introduced in an auto-regressive model that is determined through an optimization process in appropriate metric spaces. Toward capturing evolving deformations as well as cases that have not been part of the learning stage, a process that updates on-line both the orthogonal basis decomposition and the parameters of the autoregressive model is proposed. Experimental results with a nonstationary dynamic system prove adaptive AR models give better results than both stationary models and models learned over the whole sequence.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133630352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Rate Control and Motion Estimation for H.264 Encoder 改进的H.264编码器速率控制和运动估计
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379827
Loren Merritt, R. Vanam
In this paper, we describe rate control and motion estimation in x264, an open source H.264/AVC encoder. We compare the rate control methods of x264 with the JM reference encoder and show that our approach performs well in both PSNR and bitrate. In motion estimation, we describe our implementation of initialization and show that it improves PSNR. We also propose an early termination for simplified uneven cross multi hexagon grid search (UMH) in x264 and show that it improves the speed by a factor of 1.5. Finally, we show that x264 performs 50 times faster and provides bitrates within 5% of the JM reference encoder for the same PSNR.
本文描述了开源的H.264/AVC编码器x264中的速率控制和运动估计。我们将x264的速率控制方法与JM参考编码器进行了比较,结果表明我们的方法在PSNR和比特率方面都有很好的表现。在运动估计中,我们描述了初始化的实现,并表明它提高了PSNR。我们还提出了x264中简化的不均匀交叉多六边形网格搜索(UMH)的早期终止,并表明它将速度提高了1.5倍。最后,我们证明了x264的执行速度快50倍,并且在相同的PSNR下提供的比特率在JM参考编码器的5%以内。
{"title":"Improved Rate Control and Motion Estimation for H.264 Encoder","authors":"Loren Merritt, R. Vanam","doi":"10.1109/ICIP.2007.4379827","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379827","url":null,"abstract":"In this paper, we describe rate control and motion estimation in x264, an open source H.264/AVC encoder. We compare the rate control methods of x264 with the JM reference encoder and show that our approach performs well in both PSNR and bitrate. In motion estimation, we describe our implementation of initialization and show that it improves PSNR. We also propose an early termination for simplified uneven cross multi hexagon grid search (UMH) in x264 and show that it improves the speed by a factor of 1.5. Finally, we show that x264 performs 50 times faster and provides bitrates within 5% of the JM reference encoder for the same PSNR.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"483 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132266478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
Peak Transform - A Nonlinear Transform for Efficient Image Representation and Coding 峰值变换-一种用于高效图像表示和编码的非线性变换
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379275
Zhihai He
In this work, we introduce a nonlinear geometric transform, called peak transform, for efficient image representation and coding. Coupled with wavelet transform and subband decomposition, the peak transform is able to significantly reduce signal energy in high-frequency subbands and achieve a significant transform coding gain. This has important applications in efficient data representation and compression. Based on peak transform (PT), we design an image encoder, called PT encoder, for efficient image compression. Our extensive experimental results demonstrate that, in wavelet-based subband decomposition, the signal energy in high-frequency subbands can be reduced by up to 60% if a peak transform is applied. The PT image encoder outperforms state-of-the-art JPEG2000 and H.264 (INTRA) encoders by up to 2-3 dB in PSNR (peak signal-to-noise ratio), especially for images with a significant amount of high-frequency components.
在这项工作中,我们引入了一种非线性几何变换,称为峰值变换,用于有效的图像表示和编码。与小波变换和子带分解相结合,峰值变换能够显著降低高频子带的信号能量,实现显著的变换编码增益。这在有效的数据表示和压缩方面有重要的应用。为了实现高效的图像压缩,我们设计了一种基于峰值变换的图像编码器,称为峰值变换编码器。我们大量的实验结果表明,在基于小波的子带分解中,如果应用峰值变换,高频子带中的信号能量可以降低高达60%。PT图像编码器在PSNR(峰值信噪比)方面优于最先进的JPEG2000和H.264 (INTRA)编码器高达2-3 dB,特别是对于具有大量高频成分的图像。
{"title":"Peak Transform - A Nonlinear Transform for Efficient Image Representation and Coding","authors":"Zhihai He","doi":"10.1109/ICIP.2007.4379275","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379275","url":null,"abstract":"In this work, we introduce a nonlinear geometric transform, called peak transform, for efficient image representation and coding. Coupled with wavelet transform and subband decomposition, the peak transform is able to significantly reduce signal energy in high-frequency subbands and achieve a significant transform coding gain. This has important applications in efficient data representation and compression. Based on peak transform (PT), we design an image encoder, called PT encoder, for efficient image compression. Our extensive experimental results demonstrate that, in wavelet-based subband decomposition, the signal energy in high-frequency subbands can be reduced by up to 60% if a peak transform is applied. The PT image encoder outperforms state-of-the-art JPEG2000 and H.264 (INTRA) encoders by up to 2-3 dB in PSNR (peak signal-to-noise ratio), especially for images with a significant amount of high-frequency components.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132579637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image Recognition for Mobile Applications 移动应用的图像识别
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379550
J. Lee, K. Yow
Our paper presents a system for efficient recognition of landmarks taken from camera phones. Information such as tutorial rooms within the captured landmarks is returned to user within seconds. The system uses a database of multiple viewpoint's images for matching. Various navigational aids and sensors are used to optimize accuracy and retrieval time by providing complementary information about relative position and viewpoint of each query image. This makes our system less sensitive to orientation, scale and perspective distortion. Multi-scale approach and a reliability score model are proposed in this application. Our system is validated by several experiments in the campus, with images taken from different resolution's camera phones, positions and times of day.
我们的论文提出了一种有效识别从照相手机拍摄的地标的系统。在捕捉到的地标内的教程房间等信息会在几秒钟内返回给用户。该系统使用多视点图像数据库进行匹配。各种导航辅助设备和传感器通过提供每个查询图像的相对位置和视点的补充信息来优化精度和检索时间。这使得我们的系统对方向、比例和透视失真不那么敏感。在此应用中提出了多尺度方法和可靠性评分模型。我们的系统在校园里进行了几次实验,从不同分辨率的相机手机,位置和时间拍摄的图像进行了验证。
{"title":"Image Recognition for Mobile Applications","authors":"J. Lee, K. Yow","doi":"10.1109/ICIP.2007.4379550","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379550","url":null,"abstract":"Our paper presents a system for efficient recognition of landmarks taken from camera phones. Information such as tutorial rooms within the captured landmarks is returned to user within seconds. The system uses a database of multiple viewpoint's images for matching. Various navigational aids and sensors are used to optimize accuracy and retrieval time by providing complementary information about relative position and viewpoint of each query image. This makes our system less sensitive to orientation, scale and perspective distortion. Multi-scale approach and a reliability score model are proposed in this application. Our system is validated by several experiments in the campus, with images taken from different resolution's camera phones, positions and times of day.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128817442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Topological-Stabilization Based Threshold Quantization for Robust Change Detection 基于拓扑稳定的阈值量化鲁棒变化检测
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379318
Chang Su, A. Amer
A threshold quantization algorithm for robust change detection is proposed in this paper. According to the threshold distribution of difference frames, a 4-level Lloyd-Max quantizer is designed, and then, based on the topological stabilization of video frames, the Lloyd-Max quantizer is refined by a linear adjusting function to form the proposed threshold quantizer. Objective and subjective experiments show that the proposed quantizer greatly improves the robustness of the thresholding methods for change detection thus significantly improves the quality of change masks without increasing computation loads.
提出了一种用于鲁棒变化检测的阈值量化算法。根据差分帧的阈值分布,设计了一个4级Lloyd-Max量化器,然后基于视频帧的拓扑稳定性,通过线性调节函数对Lloyd-Max量化器进行细化,形成所提出的阈值量化器。客观和主观实验表明,该量化器在不增加计算量的情况下,大大提高了阈值检测方法的鲁棒性,从而显著提高了变化掩码的质量。
{"title":"Topological-Stabilization Based Threshold Quantization for Robust Change Detection","authors":"Chang Su, A. Amer","doi":"10.1109/ICIP.2007.4379318","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379318","url":null,"abstract":"A threshold quantization algorithm for robust change detection is proposed in this paper. According to the threshold distribution of difference frames, a 4-level Lloyd-Max quantizer is designed, and then, based on the topological stabilization of video frames, the Lloyd-Max quantizer is refined by a linear adjusting function to form the proposed threshold quantizer. Objective and subjective experiments show that the proposed quantizer greatly improves the robustness of the thresholding methods for change detection thus significantly improves the quality of change masks without increasing computation loads.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128841237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weighted Adaptive Lifting-Basedwavelet Transform 基于加权自适应提升的小波变换
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379278
Yu Liu, K. Ngan
In this paper, we propose a new weighted adaptive lifting (WAL)-based wavelet transform that is designed to solve the problems existing in the previous adaptive directional lifting (ADL) approach. The proposed approach uses the weighted function to make sure that the prediction and update stages are consistent, the directional interpolation to improve the orientation property of interpolated image, and adaptive interpolation filter to adjust to statistical property of each image. Experimental results show that the proposed WAL-based wavelet transform for image coding outperforms the conventional lifting-based wavelet transform up to 3.02 dB in PSNR and significant improvement in subjective quality is also observed. Compared with the ADL approach, up to 1.18 dB improvement in PSNR is reported.
针对自适应定向提升(ADL)方法存在的问题,提出了一种新的加权自适应提升(WAL)小波变换。该方法利用加权函数保证预测和更新阶段的一致性,利用定向插值提高插值图像的方向性,利用自适应插值滤波器根据图像的统计特性进行调整。实验结果表明,基于小波变换的图像编码比传统的提升小波变换的PSNR提高了3.02 dB,主观质量也有了明显改善。与ADL方法相比,PSNR提高了1.18 dB。
{"title":"Weighted Adaptive Lifting-Basedwavelet Transform","authors":"Yu Liu, K. Ngan","doi":"10.1109/ICIP.2007.4379278","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379278","url":null,"abstract":"In this paper, we propose a new weighted adaptive lifting (WAL)-based wavelet transform that is designed to solve the problems existing in the previous adaptive directional lifting (ADL) approach. The proposed approach uses the weighted function to make sure that the prediction and update stages are consistent, the directional interpolation to improve the orientation property of interpolated image, and adaptive interpolation filter to adjust to statistical property of each image. Experimental results show that the proposed WAL-based wavelet transform for image coding outperforms the conventional lifting-based wavelet transform up to 3.02 dB in PSNR and significant improvement in subjective quality is also observed. Compared with the ADL approach, up to 1.18 dB improvement in PSNR is reported.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131785582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Robust Object Tracking using Local Kernels and Background Information 基于局部核和背景信息的鲁棒目标跟踪
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379762
Jaideep Jeyakar, R. Venkatesh Babu, K. Ramakrishnan
The mean shift algorithm has been proved to be efficient for tracking 2D blobs through a video sequence. Even so, this algorithm has certain inherent disadvantages. In this paper, we propose a robust tracking algorithm which overcomes the drawbacks of global color histogram based tracking. We incorporate tracking based only on reliable colors by separating the object from its background. A fast yet robust model update is employed to overcome illumination changes. This algorithm is computationally simple enough to be executed real time and was tested on several complex video sequences. The proposed technique could be easily extended to other tracking algorithms too.
均值移位算法已被证明可以有效地跟踪视频序列中的二维斑点。即便如此,这种算法也有一些固有的缺点。本文提出了一种鲁棒跟踪算法,克服了基于全局颜色直方图跟踪的缺点。我们通过将物体从背景中分离出来,结合基于可靠颜色的跟踪。采用快速而稳健的模型更新来克服光照变化。该算法计算简单,可以实时执行,并在多个复杂的视频序列上进行了测试。所提出的技术也可以很容易地扩展到其他跟踪算法。
{"title":"Robust Object Tracking using Local Kernels and Background Information","authors":"Jaideep Jeyakar, R. Venkatesh Babu, K. Ramakrishnan","doi":"10.1109/ICIP.2007.4379762","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379762","url":null,"abstract":"The mean shift algorithm has been proved to be efficient for tracking 2D blobs through a video sequence. Even so, this algorithm has certain inherent disadvantages. In this paper, we propose a robust tracking algorithm which overcomes the drawbacks of global color histogram based tracking. We incorporate tracking based only on reliable colors by separating the object from its background. A fast yet robust model update is employed to overcome illumination changes. This algorithm is computationally simple enough to be executed real time and was tested on several complex video sequences. The proposed technique could be easily extended to other tracking algorithms too.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131903795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Some Techniques for Wow Effect Reduction 减少哇噢声效果的一些技巧
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379946
A. Czyżewski, P. Maziewski
Wow distortion reduction has not attracted an adequate scientific attention so far. Only few papers on the subject are available, concerning mostly archive gramophone records, wax cylinders, and magnetic tapes affected by wow. This paper outlines researched wow reduction algorithms concerning archive movie soundtracks, or more generally audio recordings accompanying archival visual contents. The methods presented here are based on the pilot tone tracking, on the spectral analysis of genuine audio components, and on non-uniform resampling. The paper provides only a short overview of the concepts founding those methods; other studied approaches to the wow processing, as well as a more detailed description of the presented ones, can be found in referenced papers.
迄今为止,减少哇音失真尚未引起足够的科学关注。关于这一主题的论文寥寥无几,主要涉及受 "哇 "声影响的档案留声机唱片、蜡盘和磁带。本文概述了有关档案电影原声带,或更广泛地说是伴随档案视觉内容的音频记录的 "啸叫 "衰减算法研究。本文介绍的方法基于先导音跟踪、真实音频成分的频谱分析和非均匀重采样。本文仅简要概述了这些方法的基本概念;其他已研究过的哇声处理方法,以及对本文所介绍方法的更详细描述,可参阅参考文献。
{"title":"Some Techniques for Wow Effect Reduction","authors":"A. Czyżewski, P. Maziewski","doi":"10.1109/ICIP.2007.4379946","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379946","url":null,"abstract":"Wow distortion reduction has not attracted an adequate scientific attention so far. Only few papers on the subject are available, concerning mostly archive gramophone records, wax cylinders, and magnetic tapes affected by wow. This paper outlines researched wow reduction algorithms concerning archive movie soundtracks, or more generally audio recordings accompanying archival visual contents. The methods presented here are based on the pilot tone tracking, on the spectral analysis of genuine audio components, and on non-uniform resampling. The paper provides only a short overview of the concepts founding those methods; other studied approaches to the wow processing, as well as a more detailed description of the presented ones, can be found in referenced papers.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127432556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Epipolar Spaces and Optimal Sampling Strategies 极空间与最优采样策略
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379642
J. Monaco, A. Bovik, L. Cormack
If precise calibration information is unavailable, as is often the case for active binocular vision systems, the determination of epipolar lines becomes untenable. Yet, even without instantaneous knowledge of the geometry, the search for corresponding points can be restricted to areas called epipolar spaces. For each point in one image, we define the corresponding epipolar space in the other image as the union of all associated epipolar lines over all possible system geometries. Epipolar spaces eliminate the need for calibration at the cost of an increased search region. One approach to mitigate this increase is the application of a space variant sampling or foveation strategy. While the application of such strategies to stereo vision tasks is not new, only rarely has a foveation scheme been specifically tailored for a stereo vision task. In this paper we derive a foundation of theorems that provide a means for obtaining optimal sampling schemes for a given set of epipolar spaces. An optimal sampling scheme is defined as a strategy that minimizes the average area per epipolar space.
如果没有精确的校准信息,就像主动双目视觉系统经常出现的情况一样,极线的确定就站不住脚了。然而,即使没有即时的几何知识,对相应点的搜索也可以限制在称为极外空间的区域。对于一幅图像中的每个点,我们将另一幅图像中相应的极空间定义为所有可能的系统几何上所有相关的极线的并。极外空间消除了校准的需要,但代价是增加了搜索区域。缓解这种增加的一种方法是应用空间变采样或注视点策略。虽然这种策略在立体视觉任务中的应用并不新鲜,但很少有专门为立体视觉任务量身定制的注视方案。在本文中,我们推导了一些定理的基础,这些定理提供了一种方法来获得给定的一组极空间的最优抽样格式。最优采样方案被定义为使每极空间的平均面积最小的策略。
{"title":"Epipolar Spaces and Optimal Sampling Strategies","authors":"J. Monaco, A. Bovik, L. Cormack","doi":"10.1109/ICIP.2007.4379642","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379642","url":null,"abstract":"If precise calibration information is unavailable, as is often the case for active binocular vision systems, the determination of epipolar lines becomes untenable. Yet, even without instantaneous knowledge of the geometry, the search for corresponding points can be restricted to areas called epipolar spaces. For each point in one image, we define the corresponding epipolar space in the other image as the union of all associated epipolar lines over all possible system geometries. Epipolar spaces eliminate the need for calibration at the cost of an increased search region. One approach to mitigate this increase is the application of a space variant sampling or foveation strategy. While the application of such strategies to stereo vision tasks is not new, only rarely has a foveation scheme been specifically tailored for a stereo vision task. In this paper we derive a foundation of theorems that provide a means for obtaining optimal sampling schemes for a given set of epipolar spaces. An optimal sampling scheme is defined as a strategy that minimizes the average area per epipolar space.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115342417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Inter Frame Coding with Template Matching Averaging 基于模板匹配平均的帧间编码
Pub Date : 2007-11-12 DOI: 10.1109/ICIP.2007.4379333
Yoshinori Suzuki, C. Boon, T. K. Tan
A template matching prediction based on a group of reconstructed pixels surrounding a target block enables prediction of pixels in the target block without motion information. The predictor of a target block is produced by minimizing the matching error of the template. Due to the freedom possessed by the template, the residuals of a target block may become large in flat regions. Our previous paper proposed to predictively encode the decimated version of a target block in flat regions to suppress the prediction errors. In this paper, the performance of template matching prediction is further improved. Multiple candidates are created by template matching at decoder. An average of the multiple candidates then forms the final predictor, which can reduce coding noise residing in the reference frames. Simulation results show that the proposed scheme improves coding efficiency of H.264 up to 7.9%.
一种基于目标块周围的一组重构像素的模板匹配预测能够在没有运动信息的情况下预测目标块中的像素。目标块的预测器是通过最小化模板的匹配误差产生的。由于模板所具有的自由度,目标块的残差在平坦区域可能会变大。我们之前的论文提出在平面区域对目标块的抽取版本进行预测编码,以抑制预测误差。本文进一步提高了模板匹配预测的性能。在解码器处通过模板匹配创建多个候选对象。然后,多个候选的平均值形成最终的预测器,这可以减少驻留在参考帧中的编码噪声。仿真结果表明,该方案可将H.264的编码效率提高7.9%。
{"title":"Inter Frame Coding with Template Matching Averaging","authors":"Yoshinori Suzuki, C. Boon, T. K. Tan","doi":"10.1109/ICIP.2007.4379333","DOIUrl":"https://doi.org/10.1109/ICIP.2007.4379333","url":null,"abstract":"A template matching prediction based on a group of reconstructed pixels surrounding a target block enables prediction of pixels in the target block without motion information. The predictor of a target block is produced by minimizing the matching error of the template. Due to the freedom possessed by the template, the residuals of a target block may become large in flat regions. Our previous paper proposed to predictively encode the decimated version of a target block in flat regions to suppress the prediction errors. In this paper, the performance of template matching prediction is further improved. Multiple candidates are created by template matching at decoder. An average of the multiple candidates then forms the final predictor, which can reduce coding noise residing in the reference frames. Simulation results show that the proposed scheme improves coding efficiency of H.264 up to 7.9%.","PeriodicalId":131177,"journal":{"name":"2007 IEEE International Conference on Image Processing","volume":"241 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115586422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 56
期刊
2007 IEEE International Conference on Image Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1