首页 > 最新文献

2008 IEEE 10th Workshop on Multimedia Signal Processing最新文献

英文 中文
Rich representation and ranking for photographic image retrieval in ImageCLEF 2007 ImageCLEF 2007中图像检索的丰富表示和排序
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665139
Sheng Gao, J. Chevallet, Joo-Hwee Lim
The task of ad hoc photographic image retrieval in ImageCLEF 2007 international benchmark is to retrieve relevant images in the database to the user query formulated as keywords and image examples. This paper presents rich representation and indexing technologies exploited in our system that participated in ImageCLEF 2007. It uses diverse visual content representation, text representation, pseudo-relevance feedback and fusion, which make our system, with mean average precision 0.2833, in the 4th place among 457 automatic runs submitted from 20 participants to photographic ImageCLEF 2007 and in the 2nd place in terms of participants. Our systematic analysis in the paper demonstrates that 1) combing diverse low-level visual features and ranking technologies significantly improves the content-based image retrieval (CBIR) system; 2) cross-modality pseudo-relevance feedback improves the system performance; and 3) fusion of CBIR and TBIR outperforms individual modality based system.
ImageCLEF 2007国际基准中的临时摄影图像检索任务是将数据库中的相关图像检索到以关键字和图像示例形式表述的用户查询中。本文介绍了参与ImageCLEF 2007的系统所采用的丰富的表示和索引技术。它使用了多种视觉内容表示、文本表示、伪相关反馈和融合,使我们的系统以0.2833的平均精度在20名参与者提交给2007年摄影ImageCLEF的457次自动运行中排名第4,在参与者方面排名第2。本文的系统分析表明:1)结合多种低层次视觉特征和排序技术显著改善了基于内容的图像检索(CBIR)系统;2)跨模态伪相关反馈提高了系统性能;3) CBIR和tir的融合优于基于个体模态的系统。
{"title":"Rich representation and ranking for photographic image retrieval in ImageCLEF 2007","authors":"Sheng Gao, J. Chevallet, Joo-Hwee Lim","doi":"10.1109/MMSP.2008.4665139","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665139","url":null,"abstract":"The task of ad hoc photographic image retrieval in ImageCLEF 2007 international benchmark is to retrieve relevant images in the database to the user query formulated as keywords and image examples. This paper presents rich representation and indexing technologies exploited in our system that participated in ImageCLEF 2007. It uses diverse visual content representation, text representation, pseudo-relevance feedback and fusion, which make our system, with mean average precision 0.2833, in the 4th place among 457 automatic runs submitted from 20 participants to photographic ImageCLEF 2007 and in the 2nd place in terms of participants. Our systematic analysis in the paper demonstrates that 1) combing diverse low-level visual features and ranking technologies significantly improves the content-based image retrieval (CBIR) system; 2) cross-modality pseudo-relevance feedback improves the system performance; and 3) fusion of CBIR and TBIR outperforms individual modality based system.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126228477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Generalized framework for reduced precision global motion estimation between digital images 数字图像间低精度全局运动估计的广义框架
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665052
K. Yang, M. Frater, E. Huntington, M. Pickering, J. Arnold
The efficiency of real-time digital image processing operations has an important impact on the cost and realizability of complex algorithms. Global motion estimation is an example of such a complex algorithm. Most digital image processing is carried out with a precision of 8 bits per pixel, however there has always been interest in low-complexity algorithms. One way of achieving low complexity is through low precision, such as might be achieved by quantization of each pixel to a single bit. Previous approaches to one-bit motion estimation have achieved quantization through a combination of spatial filtering/averaging and threshold setting. In this paper we present a generalized framework for precision reduction. Motivated by this framework, we show that bit-plane selection provides higher performance, with lower complexity, than conventional approaches to quantization.
实时数字图像处理操作的效率对复杂算法的成本和可实现性有着重要的影响。全局运动估计就是这种复杂算法的一个例子。大多数数字图像处理以每像素8位的精度进行,但是人们一直对低复杂度算法感兴趣。实现低复杂度的一种方法是通过低精度,例如可以通过将每个像素量化为单个比特来实现。以前的一比特运动估计方法是通过空间滤波/平均和阈值设置的组合来实现量化的。在本文中,我们提出了一个广义的精度约简框架。在这个框架的激励下,我们证明了比特平面选择比传统的量化方法提供了更高的性能和更低的复杂性。
{"title":"Generalized framework for reduced precision global motion estimation between digital images","authors":"K. Yang, M. Frater, E. Huntington, M. Pickering, J. Arnold","doi":"10.1109/MMSP.2008.4665052","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665052","url":null,"abstract":"The efficiency of real-time digital image processing operations has an important impact on the cost and realizability of complex algorithms. Global motion estimation is an example of such a complex algorithm. Most digital image processing is carried out with a precision of 8 bits per pixel, however there has always been interest in low-complexity algorithms. One way of achieving low complexity is through low precision, such as might be achieved by quantization of each pixel to a single bit. Previous approaches to one-bit motion estimation have achieved quantization through a combination of spatial filtering/averaging and threshold setting. In this paper we present a generalized framework for precision reduction. Motivated by this framework, we show that bit-plane selection provides higher performance, with lower complexity, than conventional approaches to quantization.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125621470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Automatic video object segmentation using depth information and an active contour model 基于深度信息和主动轮廓模型的自动视频目标分割
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665204
Yingdong Ma, S. Worrall, A. Kondoz
Automatic video object segmentation based on spatial-temporal information has been a research topic for many years. Existing approaches can achieve good results in some cases, such as where there is a simple background. However, in the case of cluttered backgrounds or low quality video input, automatic video object segmentation is still a problem without a general solution. A novel approach is introduced in this work, to deal with this problem by using depth information in the algorithm. The proposed approach obtains the initial object masks based on depth map and on motion detection. The object boundaries are obtained by updating object masks using a simultaneous combination of multiple cues, including spatial location, intensity, and edge, within an active contour model. The experimental result shows that this method is effective and has good output, even with cluttered backgrounds. It is also robust when the quality of input depth and video is low.
基于时空信息的视频目标自动分割是一个研究多年的课题。现有的方法可以在某些情况下取得良好的效果,例如在背景简单的情况下。然而,在背景杂乱或视频输入质量低的情况下,视频对象的自动分割仍然是一个没有通用解决方案的问题。本文提出了一种新的方法,在算法中使用深度信息来处理这一问题。该方法基于深度图和运动检测获得初始目标掩模。目标边界是通过在活动轮廓模型中使用多个线索(包括空间位置、强度和边缘)同时组合更新目标掩模来获得的。实验结果表明,即使在杂乱的背景下,该方法也具有良好的输出效果。当输入深度和视频质量较低时,它也具有鲁棒性。
{"title":"Automatic video object segmentation using depth information and an active contour model","authors":"Yingdong Ma, S. Worrall, A. Kondoz","doi":"10.1109/MMSP.2008.4665204","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665204","url":null,"abstract":"Automatic video object segmentation based on spatial-temporal information has been a research topic for many years. Existing approaches can achieve good results in some cases, such as where there is a simple background. However, in the case of cluttered backgrounds or low quality video input, automatic video object segmentation is still a problem without a general solution. A novel approach is introduced in this work, to deal with this problem by using depth information in the algorithm. The proposed approach obtains the initial object masks based on depth map and on motion detection. The object boundaries are obtained by updating object masks using a simultaneous combination of multiple cues, including spatial location, intensity, and edge, within an active contour model. The experimental result shows that this method is effective and has good output, even with cluttered backgrounds. It is also robust when the quality of input depth and video is low.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132240773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Computationally efficient interference detection in videokeratoscopy images 视频角膜炎图像中计算效率高的干扰检测
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665127
D. Alonso-Caneiro, D. R. Iskander, M. Collins
An optimal videokeratoscopic image presents a strong well-oriented pattern over the majority of the measured corneal surface. In the presence of interference, arising from reflections from eyelashes or tear film instability, the patternpsilas flow is disturbed and the local orientation of the area of interference is no longer coherent with the global flow. Detecting and analysing videokeratoscopic pattern interference is important when assessing tear film surface quality, break-up time and location as well as designing tools that provide a more accurate static measurement of corneal topography. In this paper a set of algorithms for detecting interference patterns in videokeratoscopic images is presented. First a frequency approach is used to subtract the background information from the oriented structure and then a gradient-based analysis is used to obtain the patternpsilas orientation and coherence. The proposed techniques are compared to a previously reported method based on statistical block normalisation and Gabor filtering. The results indicate that the proposed technique leads, in most cases: to a better videokeratoscopic interference detection system, that for a given probability of the useful signal detection (99.7%) has a significantly lower probability of false alarm, and at the same time is computationally much more efficient than the previously reported method.
一个最佳的视频透视图像呈现了一个强大的定向模式在大多数测量角膜表面。在干扰存在的情况下,由于睫毛的反射或泪膜的不稳定性,图案流受到干扰,干扰区域的局部方向不再与全局流一致。在评估泪膜表面质量、破裂时间和位置以及设计提供更准确的角膜地形图静态测量的工具时,检测和分析视频角化镜模式干扰非常重要。本文提出了一套检测视频透视图像中干涉图案的算法。首先采用频率法从取向结构中去除背景信息,然后采用梯度分析方法获得图案的取向和相干性。将提出的技术与先前报道的基于统计块归一化和Gabor滤波的方法进行了比较。结果表明,在大多数情况下,所提出的技术导致了一个更好的视频角化干扰检测系统,对于给定的有用信号检测概率(99.7%),虚警概率显着降低,同时计算效率比先前报道的方法高得多。
{"title":"Computationally efficient interference detection in videokeratoscopy images","authors":"D. Alonso-Caneiro, D. R. Iskander, M. Collins","doi":"10.1109/MMSP.2008.4665127","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665127","url":null,"abstract":"An optimal videokeratoscopic image presents a strong well-oriented pattern over the majority of the measured corneal surface. In the presence of interference, arising from reflections from eyelashes or tear film instability, the patternpsilas flow is disturbed and the local orientation of the area of interference is no longer coherent with the global flow. Detecting and analysing videokeratoscopic pattern interference is important when assessing tear film surface quality, break-up time and location as well as designing tools that provide a more accurate static measurement of corneal topography. In this paper a set of algorithms for detecting interference patterns in videokeratoscopic images is presented. First a frequency approach is used to subtract the background information from the oriented structure and then a gradient-based analysis is used to obtain the patternpsilas orientation and coherence. The proposed techniques are compared to a previously reported method based on statistical block normalisation and Gabor filtering. The results indicate that the proposed technique leads, in most cases: to a better videokeratoscopic interference detection system, that for a given probability of the useful signal detection (99.7%) has a significantly lower probability of false alarm, and at the same time is computationally much more efficient than the previously reported method.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130172255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
SDI: New metric for quantification of speckle noise in ultrasound imaging 超声成像中散斑噪声量化的新度量
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665060
K. Joshi, R. Kamathe
In ultrasound images a special type of acoustic noise, technically known as speckle noise, is the major factor of image quality degradation. In order to improve the image quality by means of speckle suppression and thus to increase the diagnostic potential of medical ultrasound, it is important to quantify the speckle. This paper describes, quality metrics for speckle in coherent imaging and their limitations. It also describes a new metric-SDI, its uniqueness in quantifying the speckle and comparison of performance with existing metrics. Empirical verification of SDI with a set of Test Images proves its speckle quantification in ultrasound. A subjective criterion is also taken into account to support the results.
在超声图像中,一种特殊类型的噪声,技术上称为斑点噪声,是图像质量下降的主要因素。为了通过抑制散斑来改善图像质量,从而提高医学超声的诊断潜力,对散斑进行量化是很重要的。本文介绍了相干成像中散斑的质量度量及其局限性。它还描述了一个新的指标- sdi,它在量化散斑和与现有指标的性能比较方面的独特性。用一组测试图像对SDI进行了实证验证,证明了其在超声中的散斑量化。还考虑了一个主观标准来支持结果。
{"title":"SDI: New metric for quantification of speckle noise in ultrasound imaging","authors":"K. Joshi, R. Kamathe","doi":"10.1109/MMSP.2008.4665060","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665060","url":null,"abstract":"In ultrasound images a special type of acoustic noise, technically known as speckle noise, is the major factor of image quality degradation. In order to improve the image quality by means of speckle suppression and thus to increase the diagnostic potential of medical ultrasound, it is important to quantify the speckle. This paper describes, quality metrics for speckle in coherent imaging and their limitations. It also describes a new metric-SDI, its uniqueness in quantifying the speckle and comparison of performance with existing metrics. Empirical verification of SDI with a set of Test Images proves its speckle quantification in ultrasound. A subjective criterion is also taken into account to support the results.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134131616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Tennis video enrichment with content layer separation and real-time rendering in sprite plane 网球视频丰富与内容层分离和实时渲染在精灵平面
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665160
Jui-Hsin Lai, Shao-Yi Chien
Sport video enrichment can provide viewers more interaction and user experiences. In this paper, with tennis sport video as an example, two techniques are proposed for video enrichment: content layer separation and real-time rendering. The video content is decomposed into different layers, like field, players and ball, and the enriched video is rendered by re-integrated these layers information. They are both executed in sprite plane to avoid complex 3D model construction and rendering. Experiments shows that it can generate nature and seamless edited video by viewerspsila requests, and the real-time processing speed of 30 720times480 frames per second can be achieved on a 3 GHz CPU.
体育视频的丰富可以为观众提供更多的互动和用户体验。本文以网球运动视频为例,提出了两种视频丰富技术:内容层分离和实时渲染。将视频内容分解为场地、球员、球等不同的图层,将这些图层信息重新整合后呈现丰富的视频。它们都是在精灵平面中执行,以避免复杂的3D模型构建和渲染。实验表明,该方法可以根据观众的要求生成自然、无缝编辑的视频,在3 GHz的CPU上可实现30 720 × 480帧/秒的实时处理速度。
{"title":"Tennis video enrichment with content layer separation and real-time rendering in sprite plane","authors":"Jui-Hsin Lai, Shao-Yi Chien","doi":"10.1109/MMSP.2008.4665160","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665160","url":null,"abstract":"Sport video enrichment can provide viewers more interaction and user experiences. In this paper, with tennis sport video as an example, two techniques are proposed for video enrichment: content layer separation and real-time rendering. The video content is decomposed into different layers, like field, players and ball, and the enriched video is rendered by re-integrated these layers information. They are both executed in sprite plane to avoid complex 3D model construction and rendering. Experiments shows that it can generate nature and seamless edited video by viewerspsila requests, and the real-time processing speed of 30 720times480 frames per second can be achieved on a 3 GHz CPU.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131512091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A structural method for quality evaluation of desynchronization attacks in image watermarking 一种图像水印中去同步攻击质量评价的结构化方法
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665175
Angela D'Angelo, M. Barni
Geometric transformations are known to be one of the most serious threats against any digital watermarking scheme. The goal of this work is to design an objective measurement scheme for geometric distortions in order to investigate the perceptual quality impact of geometric attacks on the watermarked images. The proposed approach is a full-reference image quality metric focusing on the problem of local geometric attacks and it is based on the use of Gabor filters. The novelty of the proposed metric is that it considers both the displacement field describing the distortion and the structure of the image.The experimental results show the good performances of the metric.
已知几何变换是对任何数字水印方案最严重的威胁之一。为了研究几何攻击对水印图像感知质量的影响,本文设计了一种客观的几何失真测量方案。该方法是一种基于Gabor滤波器的全参考图像质量度量,主要关注局部几何攻击问题。该度量的新颖之处在于它同时考虑了描述图像畸变和图像结构的位移场。实验结果表明,该度量方法具有良好的性能。
{"title":"A structural method for quality evaluation of desynchronization attacks in image watermarking","authors":"Angela D'Angelo, M. Barni","doi":"10.1109/MMSP.2008.4665175","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665175","url":null,"abstract":"Geometric transformations are known to be one of the most serious threats against any digital watermarking scheme. The goal of this work is to design an objective measurement scheme for geometric distortions in order to investigate the perceptual quality impact of geometric attacks on the watermarked images. The proposed approach is a full-reference image quality metric focusing on the problem of local geometric attacks and it is based on the use of Gabor filters. The novelty of the proposed metric is that it considers both the displacement field describing the distortion and the structure of the image.The experimental results show the good performances of the metric.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131632641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Intra-frame video coding using an open-loop predictive coding approach 帧内视频编码采用开环预测编码方法
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665088
Frederik Verbist, A. Munteanu, J. Cornelis, P. Schelkens
A novel coding approach, applying open-loop coding principles in predictive coding systems is proposed in this paper. The proposed approach is instantiated with an intra-frame video codec employing the transform and spatial prediction modes from H.264. Additionally, a novel rate-distortion model for open-loop predictive coding is proposed and experimentally validated. Optimally allocating rate based on the proposed model provides significant gains in comparison to a straightforward rate allocation not accounting for drift. Furthermore, the proposed open-loop predictive codec provides gains of up to 2.3 dB in comparison to an equivalent closed-loop intra-frame video codec employing the transform, prediction modes and rate-allocation from H.264. This indicates that, with appropriate drift compensation, open-loop predictive coding offers the possibility for further improving the compression performance in predictive coding systems.
提出了一种新的编码方法,即在预测编码系统中应用开环编码原理。采用H.264的变换和空间预测模式对帧内视频编解码器进行了实例化。此外,提出了一种新的开环预测编码的率失真模型,并进行了实验验证。与不考虑漂移的直接速率分配相比,基于所提模型的最佳速率分配提供了显着的增益。此外,与采用H.264的变换、预测模式和速率分配的等效闭环帧内视频编解码器相比,所提出的开环预测编解码器提供高达2.3 dB的增益。这表明,通过适当的漂移补偿,开环预测编码为进一步提高预测编码系统的压缩性能提供了可能。
{"title":"Intra-frame video coding using an open-loop predictive coding approach","authors":"Frederik Verbist, A. Munteanu, J. Cornelis, P. Schelkens","doi":"10.1109/MMSP.2008.4665088","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665088","url":null,"abstract":"A novel coding approach, applying open-loop coding principles in predictive coding systems is proposed in this paper. The proposed approach is instantiated with an intra-frame video codec employing the transform and spatial prediction modes from H.264. Additionally, a novel rate-distortion model for open-loop predictive coding is proposed and experimentally validated. Optimally allocating rate based on the proposed model provides significant gains in comparison to a straightforward rate allocation not accounting for drift. Furthermore, the proposed open-loop predictive codec provides gains of up to 2.3 dB in comparison to an equivalent closed-loop intra-frame video codec employing the transform, prediction modes and rate-allocation from H.264. This indicates that, with appropriate drift compensation, open-loop predictive coding offers the possibility for further improving the compression performance in predictive coding systems.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132580196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
3-D mesh representation and retrieval using Isomap manifold 基于Isomap流形的三维网格表示与检索
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665137
Jung-Shiong Chang, A. C. Shih, Hsueh-Yi Sean Lin, Hai-Feng Kao, H. Liao, Wen-Hsien Fang
We propose a compact 3D object representation scheme that can greatly assist the search/retrieval process in a network environment. A 3D mesh-based object is transformed into a new coordinate frame by using the Isomap (isometric feature mapping) method. During the transformation process, not only the structure of the salient parts of an object will be kept, but also the geometrical relationships will be preserved. From the viewpoint of cognitive psychology, the data distributed on the Isomap manifold can be regarded as a set of significant features of a 3D mesh-based object. To perform efficient matching, we project the Isomap domain 3D object onto two different 2D maps, and the two 2D feature descriptors are used as the basis to measure the degree of similarity between two 3D mesh-based objects. Experiments demonstrate that the proposed method in retrieving similar 3D models is very effective. Most importantly, the proposed 3D mesh retrieval scheme is still valid even if a 3D mesh undergoes a mesh simplification process.
我们提出了一种紧凑的三维对象表示方案,可以极大地辅助网络环境中的搜索/检索过程。采用等距特征映射(Isomap)方法,将基于三维网格的物体转换为新的坐标框架。在变换过程中,不仅保留了物体突出部分的结构,而且保留了几何关系。从认知心理学的角度来看,分布在Isomap流形上的数据可以看作是基于三维网格的对象的一组重要特征。为了实现高效匹配,我们将Isomap域三维对象投影到两个不同的二维地图上,并将两个二维特征描述符作为度量两个三维网格对象之间相似度的基础。实验结果表明,该方法对相似三维模型的检索是非常有效的。最重要的是,所提出的三维网格检索方案即使在网格简化过程中仍然有效。
{"title":"3-D mesh representation and retrieval using Isomap manifold","authors":"Jung-Shiong Chang, A. C. Shih, Hsueh-Yi Sean Lin, Hai-Feng Kao, H. Liao, Wen-Hsien Fang","doi":"10.1109/MMSP.2008.4665137","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665137","url":null,"abstract":"We propose a compact 3D object representation scheme that can greatly assist the search/retrieval process in a network environment. A 3D mesh-based object is transformed into a new coordinate frame by using the Isomap (isometric feature mapping) method. During the transformation process, not only the structure of the salient parts of an object will be kept, but also the geometrical relationships will be preserved. From the viewpoint of cognitive psychology, the data distributed on the Isomap manifold can be regarded as a set of significant features of a 3D mesh-based object. To perform efficient matching, we project the Isomap domain 3D object onto two different 2D maps, and the two 2D feature descriptors are used as the basis to measure the degree of similarity between two 3D mesh-based objects. Experiments demonstrate that the proposed method in retrieving similar 3D models is very effective. Most importantly, the proposed 3D mesh retrieval scheme is still valid even if a 3D mesh undergoes a mesh simplification process.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114308770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast encoding algorithms for video coding with adaptive interpolation filters 自适应插值滤波器视频编码的快速编码算法
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665096
D. Rusanovskyy, K. Ugur, M. Gabbouj
In order to compensate for the temporally changing effect of aliasing and improve the coding efficiency of video coders, adaptive interpolation filtering schemes have been recently proposed. In such schemes, encoder computes the interpolation filter coefficients for each frame and then re-encodes the frame with the new adaptive filter. However, the coding efficiency benefit comes with the expense of increased encoding complexity due to this additional encoding pass. In this paper, we present two novel algorithms to reduce the encoding complexity of adaptive interpolation filtering schemes. First algorithm reduces the complexity of the second encoding pass by using a very lightweight motion estimation algorithm that reuses the data already computed in the first encoding pass. Second algorithm eliminates the second coding pass and re-uses the filter coefficients already computed for previous frames. Experimental results show that the proposed methods achieve between 1.5 to 2 times encoding complexity reduction with practically negligible penalty on coding efficiency.
为了补偿混叠的时变效应,提高视频编码器的编码效率,近年来提出了自适应插值滤波方案。在这种方案中,编码器计算每帧的插值滤波系数,然后用新的自适应滤波器对帧进行重新编码。然而,编码效率的提高是以增加编码复杂性为代价的,因为这是额外的编码通道。本文提出了两种新的算法来降低自适应插值滤波方案的编码复杂度。第一种算法通过使用一种非常轻量级的运动估计算法来减少第二次编码传递的复杂性,该算法重用了在第一次编码传递中已经计算的数据。第二种算法消除了第二个编码通道,并重新使用已经为前一帧计算过的滤波器系数。实验结果表明,该方法可以将编码复杂度降低1.5 ~ 2倍,而编码效率几乎可以忽略不计。
{"title":"Fast encoding algorithms for video coding with adaptive interpolation filters","authors":"D. Rusanovskyy, K. Ugur, M. Gabbouj","doi":"10.1109/MMSP.2008.4665096","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665096","url":null,"abstract":"In order to compensate for the temporally changing effect of aliasing and improve the coding efficiency of video coders, adaptive interpolation filtering schemes have been recently proposed. In such schemes, encoder computes the interpolation filter coefficients for each frame and then re-encodes the frame with the new adaptive filter. However, the coding efficiency benefit comes with the expense of increased encoding complexity due to this additional encoding pass. In this paper, we present two novel algorithms to reduce the encoding complexity of adaptive interpolation filtering schemes. First algorithm reduces the complexity of the second encoding pass by using a very lightweight motion estimation algorithm that reuses the data already computed in the first encoding pass. Second algorithm eliminates the second coding pass and re-uses the filter coefficients already computed for previous frames. Experimental results show that the proposed methods achieve between 1.5 to 2 times encoding complexity reduction with practically negligible penalty on coding efficiency.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"166 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114371466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2008 IEEE 10th Workshop on Multimedia Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1