首页 > 最新文献

2006 IEEE International Conference on Multimedia and Expo最新文献

英文 中文
Non-Bandlimited Resampling of Images 图像的无带宽限制重采样
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262591
Beilei Huang, E. Lai
The resampling of discrete-time signals where the underlying analog signal is non-bandlimited is considered in this paper. We extend the generalized sampling theory developed based on the principle of consistency to resampling. Realizing the resampling system has both discrete input and output, the performance of the resampling filter is considered in l2 instead of the traditionally used L2 . We show that the performance of the resampling system depends on the resampling rate instead of the actual interpolating kernels. The theory can be applied to image processing applications like zooming to provide better response to high frequency components. Since the resampling process is discrete in nature, our filter designed to optimize resampling in l2 is shown to outperform other techniques designed in L2
本文研究了底层模拟信号为无带宽限制的离散时间信号的重采样问题。我们将基于一致性原理的广义抽样理论推广到重抽样。实现重采样系统的输入和输出都是离散的,重采样滤波器的性能在l2中考虑,而不是传统使用的l2。我们证明了重采样系统的性能取决于重采样率而不是实际的插值核。该理论可以应用于图像处理应用,如变焦,以提供更好的响应高频成分。由于重采样过程本质上是离散的,我们设计用于优化l2重采样的滤波器被证明优于l2中设计的其他技术
{"title":"Non-Bandlimited Resampling of Images","authors":"Beilei Huang, E. Lai","doi":"10.1109/ICME.2006.262591","DOIUrl":"https://doi.org/10.1109/ICME.2006.262591","url":null,"abstract":"The resampling of discrete-time signals where the underlying analog signal is non-bandlimited is considered in this paper. We extend the generalized sampling theory developed based on the principle of consistency to resampling. Realizing the resampling system has both discrete input and output, the performance of the resampling filter is considered in l2 instead of the traditionally used L2 . We show that the performance of the resampling system depends on the resampling rate instead of the actual interpolating kernels. The theory can be applied to image processing applications like zooming to provide better response to high frequency components. Since the resampling process is discrete in nature, our filter designed to optimize resampling in l2 is shown to outperform other techniques designed in L2","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"138 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116357153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Combined Bayesshrinkwavelet-Ridgelet Technique for Image Denoising 联合贝叶斯收缩小波-脊波图像去噪技术
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262931
N. Nezamoddini-Kachouie, P. Fieguth
In this paper a combined Bayesshrink wavelet-ridgelet de-noising method is presented. In our previous work we have showed that Bayesshrink ridgelet performs better than Visushrink ridgelet and Visushrink wavelet. Although our Bayesshrink ridgelet technique performs somewhat poorer in comparison with Bayesshrink wavelet, based on SNR, visually it produces smoother results, especially for images with straight lines. In the proposed method Bayesshrink wavelet is combined with Bayesshrink ridgelet denoising method which performs better than each filter individually. The proposed combined denoising method gains the advantage of each filter in its specific domain, i.e., wavelet for natural and ridgelet for straight regions, and produces better and smoother results, both visually and in terms of SNR
本文提出了一种贝叶斯收缩小波-脊波联合去噪方法。在我们之前的工作中,我们已经证明Bayesshrink脊波比Visushrink脊波和Visushrink小波具有更好的性能。尽管我们的Bayesshrink脊波技术与Bayesshrink小波相比表现稍差,但基于信噪比,它在视觉上产生了更平滑的结果,特别是对于具有直线的图像。在该方法中,贝叶斯收缩小波与贝叶斯收缩脊波相结合,其降噪效果优于单个滤波器。所提出的组合去噪方法利用了每个滤波器在其特定领域的优势,即对自然区域使用小波,对直线区域使用脊波,在视觉上和信噪比上都得到了更好、更平滑的结果
{"title":"A Combined Bayesshrinkwavelet-Ridgelet Technique for Image Denoising","authors":"N. Nezamoddini-Kachouie, P. Fieguth","doi":"10.1109/ICME.2006.262931","DOIUrl":"https://doi.org/10.1109/ICME.2006.262931","url":null,"abstract":"In this paper a combined Bayesshrink wavelet-ridgelet de-noising method is presented. In our previous work we have showed that Bayesshrink ridgelet performs better than Visushrink ridgelet and Visushrink wavelet. Although our Bayesshrink ridgelet technique performs somewhat poorer in comparison with Bayesshrink wavelet, based on SNR, visually it produces smoother results, especially for images with straight lines. In the proposed method Bayesshrink wavelet is combined with Bayesshrink ridgelet denoising method which performs better than each filter individually. The proposed combined denoising method gains the advantage of each filter in its specific domain, i.e., wavelet for natural and ridgelet for straight regions, and produces better and smoother results, both visually and in terms of SNR","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114823920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Towards an Efficient Integration, Structure and Exploration of Landscape Architecture Project Information 景观设计项目信息的高效整合、构建与探索
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262520
Franck Favetta, R. Laurini
Landscape architecture projects have many specific requirements such as particular multimedia and geographic data integration and structure, information preview, user-friendly interface, and means of multi-actor participation. This article presents a solution for an efficient, quick and user-friendly integration, structure, exploration and management of landscape information. Our proposal extends different existing solutions and introduces useful preview abilities. A recently developed prototype implements the solution
景观设计项目对多媒体和地理数据的集成和结构、信息预览、用户友好界面、多主体参与方式等有许多特殊的要求。本文提出了一种高效、快速、用户友好的景观信息集成、构建、挖掘和管理解决方案。我们的建议扩展了不同的现有解决方案,并引入了有用的预览功能。最近开发的一个原型实现了该解决方案
{"title":"Towards an Efficient Integration, Structure and Exploration of Landscape Architecture Project Information","authors":"Franck Favetta, R. Laurini","doi":"10.1109/ICME.2006.262520","DOIUrl":"https://doi.org/10.1109/ICME.2006.262520","url":null,"abstract":"Landscape architecture projects have many specific requirements such as particular multimedia and geographic data integration and structure, information preview, user-friendly interface, and means of multi-actor participation. This article presents a solution for an efficient, quick and user-friendly integration, structure, exploration and management of landscape information. Our proposal extends different existing solutions and introduces useful preview abilities. A recently developed prototype implements the solution","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"134 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124511677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Extraction of Outcrop Points from Visual Hulls for Motion Estimation 用于运动估计的视觉船体露头点提取
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262421
M. Toyoura, M. Iiyama, K. Kakusho, M. Minoh
In this article, we discuss 3D shape reconstruction of an object in a rigid motion with the volume intersection method. When the object moves rigidly, the cameras change their relative positions to the object at every moment. To estimate the motion correctly, we propose new feature points called outcrop points on the reconstructed 3D shape. These points are guaranteed to be located on the real surface of the object. If the rigid motion of the object can be correctly estimated, cameras at different moments serve as the cameras in different positions virtually. With these cameras in time sequences, we can increase accuracy of the reconstructed 3D shape without increasing the number of cameras. Based on this idea, we reconstruct an accurate shape of the object in motion from images obtained by limited number of cameras. As the result, we can acquire an accurate shape from images in time sequences
本文讨论了用体交法重建刚体运动物体的三维形状。当物体作刚性运动时,摄像机会随时改变与物体的相对位置。为了正确估计运动,我们在重建的三维形状上提出了新的特征点,称为露头点。这些点保证位于物体的真实表面上。如果能正确估计物体的刚体运动,则不同时刻的摄像机实际上充当了不同位置的摄像机。在不增加相机数量的情况下,利用这些时间序列相机可以提高重建三维形状的精度。基于这一思想,我们从有限数量的相机获得的图像中重建运动物体的精确形状。因此,我们可以从时间序列的图像中获得精确的形状
{"title":"Extraction of Outcrop Points from Visual Hulls for Motion Estimation","authors":"M. Toyoura, M. Iiyama, K. Kakusho, M. Minoh","doi":"10.1109/ICME.2006.262421","DOIUrl":"https://doi.org/10.1109/ICME.2006.262421","url":null,"abstract":"In this article, we discuss 3D shape reconstruction of an object in a rigid motion with the volume intersection method. When the object moves rigidly, the cameras change their relative positions to the object at every moment. To estimate the motion correctly, we propose new feature points called outcrop points on the reconstructed 3D shape. These points are guaranteed to be located on the real surface of the object. If the rigid motion of the object can be correctly estimated, cameras at different moments serve as the cameras in different positions virtually. With these cameras in time sequences, we can increase accuracy of the reconstructed 3D shape without increasing the number of cameras. Based on this idea, we reconstruct an accurate shape of the object in motion from images obtained by limited number of cameras. As the result, we can acquire an accurate shape from images in time sequences","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124004179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Fingerprinting System for Musical Content 音乐内容的指纹识别系统
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262949
L. Ghouti, A. Bouridane, M. K. Ibrahim
Driven by the recent advances in digital entertainment technologies, digital multimedia content (such as music and movies) is becoming a major part of the average computer user experience. Through daily interaction with digital multimedia content, large digital collections of music, audio and sound effects have emerged. Furthermore, these collections are produced/consumed by different groups of users such as the entertainment, music, movie and animation industries. Therefore, the need for identification and management of such content grows proportionally to the increasing widespread availability of such media virtually "any time and any where" over the internet. In this paper, we propose a novel algorithm for robust perceptual hashing of musical content using balanced multiwavelets (BMW). The procedure for generating robust perceptual hash values (or fingerprints) is described in details. The generated hash values are used for identifying, searching, and retrieving musical content from large musical databases. Furthermore, we illustrate, through extensive computer simulation, the robustness of the proposed framework to efficiently represent audio content and withstand several signal processing attacks and manipulations
在数字娱乐技术最新进展的推动下,数字多媒体内容(如音乐和电影)正在成为普通计算机用户体验的主要部分。通过与数字多媒体内容的日常互动,大量的音乐、音频和音效的数字集合已经出现。此外,这些集合由不同的用户群体生产/消费,如娱乐、音乐、电影和动画行业。因此,识别和管理这类内容的需求随着这类媒体几乎“随时随地”在互联网上日益广泛的可用性而增长。在本文中,我们提出了一种利用平衡多小波对音乐内容进行鲁棒感知哈希的新算法。详细描述了生成鲁棒感知哈希值(或指纹)的过程。生成的散列值用于从大型音乐数据库中识别、搜索和检索音乐内容。此外,通过广泛的计算机模拟,我们说明了所提出的框架的鲁棒性,可以有效地表示音频内容,并承受多种信号处理攻击和操纵
{"title":"A Fingerprinting System for Musical Content","authors":"L. Ghouti, A. Bouridane, M. K. Ibrahim","doi":"10.1109/ICME.2006.262949","DOIUrl":"https://doi.org/10.1109/ICME.2006.262949","url":null,"abstract":"Driven by the recent advances in digital entertainment technologies, digital multimedia content (such as music and movies) is becoming a major part of the average computer user experience. Through daily interaction with digital multimedia content, large digital collections of music, audio and sound effects have emerged. Furthermore, these collections are produced/consumed by different groups of users such as the entertainment, music, movie and animation industries. Therefore, the need for identification and management of such content grows proportionally to the increasing widespread availability of such media virtually \"any time and any where\" over the internet. In this paper, we propose a novel algorithm for robust perceptual hashing of musical content using balanced multiwavelets (BMW). The procedure for generating robust perceptual hash values (or fingerprints) is described in details. The generated hash values are used for identifying, searching, and retrieving musical content from large musical databases. Furthermore, we illustrate, through extensive computer simulation, the robustness of the proposed framework to efficiently represent audio content and withstand several signal processing attacks and manipulations","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127645870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
An New Coefficients Transform Matrix for the Transform Domain MPEG-2 TO H.264/AVC Transcoding 变换域MPEG-2到H.264/AVC转码的新系数变换矩阵
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262463
Gao Chen, Shouxun Lin, Yongdong Zhang, Gang Cao
In this paper, a fast transform method is proposed to convert MPEG-2 8-tap discrete cosine transform (DCT) coefficients to H.264/AVC 4-tap integer transform coefficients directly in the transform domain. The proposed transform method saves 16 operations for each 8times8 DCT block by utilizing a novel transform kernel matrix and a fast computing method for multiplication of this new matrix. The simulation results show that the proposed method causes only a very little quality degradation, which is completely negligible in practice with the maximum value lower than 8times10-33dB, as compared with Jun Xin' s method. Hence, it can be efficiently used in the transform-domain MPEG-2 to H.264 transcoding
提出了一种直接在变换域中将MPEG-2 8分线离散余弦变换(DCT)系数转换为H.264/AVC 4分线整数变换系数的快速变换方法。该变换方法利用一种新的变换核矩阵和一种快速的矩阵乘法计算方法,为每个8times8 DCT块节省16次运算。仿真结果表明,与Jun Xin的方法相比,所提方法的质量衰减很小,在实际应用中完全可以忽略不计,其最大值小于8倍10- 33db。因此,它可以有效地用于转换域MPEG-2到H.264的转码
{"title":"An New Coefficients Transform Matrix for the Transform Domain MPEG-2 TO H.264/AVC Transcoding","authors":"Gao Chen, Shouxun Lin, Yongdong Zhang, Gang Cao","doi":"10.1109/ICME.2006.262463","DOIUrl":"https://doi.org/10.1109/ICME.2006.262463","url":null,"abstract":"In this paper, a fast transform method is proposed to convert MPEG-2 8-tap discrete cosine transform (DCT) coefficients to H.264/AVC 4-tap integer transform coefficients directly in the transform domain. The proposed transform method saves 16 operations for each 8times8 DCT block by utilizing a novel transform kernel matrix and a fast computing method for multiplication of this new matrix. The simulation results show that the proposed method causes only a very little quality degradation, which is completely negligible in practice with the maximum value lower than 8times10-33dB, as compared with Jun Xin' s method. Hence, it can be efficiently used in the transform-domain MPEG-2 to H.264 transcoding","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"148 Pt 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126319412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Computing a Multimedia Representation for Documents Given Time and Display Constraints 计算给定时间和显示约束的文档的多媒体表示
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262657
B. Erol, K. Berkner, S. Joshi, J. Hull
It is difficult to view multipage, high resolution documents on devices with small displays. As a solution, we introduce a multimedia thumbnail representation, which can be seen as a multimedia clip that provides an automated guided tour through a document. Multimedia thumbnails are automatically generated by taking a document image as input and first performing visual and audible information analysis on the document to determine salient document elements. Next, the time and information attributes for each document element are computed by taking into account the display and application constraints. An optimization routine, given a time constraint, selects elements to be included in the multimedia thumbnail. Last, the selected elements are synthesized into animated images and audio to create the final multimedia representation
在屏幕小的设备上查看多页高分辨率文档是很困难的。作为一种解决方案,我们引入了一个多媒体缩略图表示,它可以被看作是一个多媒体剪辑,它提供了通过文档的自动导览。多媒体缩略图是通过将文档图像作为输入并首先对文档执行视觉和听觉信息分析以确定突出的文档元素来自动生成的。接下来,通过考虑显示和应用程序约束来计算每个文档元素的时间和信息属性。在给定时间限制的情况下,优化例程选择要包含在多媒体缩略图中的元素。最后,将选定的元素合成为动画图像和音频,以创建最终的多媒体表示
{"title":"Computing a Multimedia Representation for Documents Given Time and Display Constraints","authors":"B. Erol, K. Berkner, S. Joshi, J. Hull","doi":"10.1109/ICME.2006.262657","DOIUrl":"https://doi.org/10.1109/ICME.2006.262657","url":null,"abstract":"It is difficult to view multipage, high resolution documents on devices with small displays. As a solution, we introduce a multimedia thumbnail representation, which can be seen as a multimedia clip that provides an automated guided tour through a document. Multimedia thumbnails are automatically generated by taking a document image as input and first performing visual and audible information analysis on the document to determine salient document elements. Next, the time and information attributes for each document element are computed by taking into account the display and application constraints. An optimization routine, given a time constraint, selects elements to be included in the multimedia thumbnail. Last, the selected elements are synthesized into animated images and audio to create the final multimedia representation","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126208512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Complexity-Distortion Optimized Motion Estimation Algorithm with Fine-Granular Scalable Complexity 具有细粒度可扩展复杂度的复杂度-畸变优化运动估计算法
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262815
Li Zhang, Wen Gao
Video encoding now is being implemented in various computing platforms with different computing capability, the requirement on the encoding complexity is also different according to different applications. As the most computation-intensive part of video encoding, the ME (motion estimation) should have a scalable complexity. This paper proposes a ME algorithm with fine-granular scalable complexity, a more important feature of the proposed algorithm is that it seeks for the complexity-distortion optimization. The given computation budget will be allocated to each MB (macroblock) in one frame. Each MB will consume its allocated computation by a hybrid search pattern. Experimental results show that the proposed algorithm can get a better computation-distortion performance than the existing ME algorithms
视频编码目前在各种计算平台上实现,计算能力不同,不同的应用对编码复杂度的要求也不同。运动估计作为视频编码中计算量最大的部分,其复杂度应具有可扩展性。本文提出了一种具有细粒度可扩展复杂度的ME算法,该算法的一个更重要的特点是寻求复杂度失真优化。给定的计算预算将在一帧中分配给每个MB(宏块)。每个MB将通过混合搜索模式消耗其分配的计算。实验结果表明,该算法比现有的ME算法具有更好的计算失真性能
{"title":"Complexity-Distortion Optimized Motion Estimation Algorithm with Fine-Granular Scalable Complexity","authors":"Li Zhang, Wen Gao","doi":"10.1109/ICME.2006.262815","DOIUrl":"https://doi.org/10.1109/ICME.2006.262815","url":null,"abstract":"Video encoding now is being implemented in various computing platforms with different computing capability, the requirement on the encoding complexity is also different according to different applications. As the most computation-intensive part of video encoding, the ME (motion estimation) should have a scalable complexity. This paper proposes a ME algorithm with fine-granular scalable complexity, a more important feature of the proposed algorithm is that it seeks for the complexity-distortion optimization. The given computation budget will be allocated to each MB (macroblock) in one frame. Each MB will consume its allocated computation by a hybrid search pattern. Experimental results show that the proposed algorithm can get a better computation-distortion performance than the existing ME algorithms","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128153587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Comparison of Three 3-D Facial Reconstruction Approaches 三种三维人脸重建方法的比较
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262619
A. Woodward, Da An, G. Gimel'farb, P. Delmas
We compare three computer vision approaches to 3-D reconstruction, namely passive binocular stereo and active structured lighting and photometric stereo, in application to human face reconstruction for modelling virtual humans. An integrated lab environment was set up to simultaneously acquire images for 3-D reconstruction and corresponding data from a 3-D scanner. This allowed us to quantitatively compare reconstruction results to accurate ground truth. Our goal was to determine whether any current computer vision approach is accurate enough for practically useful 3-D facial surface reconstruction. Comparative experiments show the combination of structured lighting with symmetric dynamic programming based binocular stereo has good prospects due to reasonable processing time and sufficient accuracy
我们比较了三种计算机视觉的三维重建方法,即被动双目立体和主动结构照明和光度立体,在人脸重建中的应用,以模拟虚拟人。建立了一个集成的实验室环境,同时获取用于三维重建的图像和三维扫描仪的相应数据。这使我们能够定量地将重建结果与准确的地面真实情况进行比较。我们的目标是确定当前的计算机视觉方法是否足够精确,可以用于实际有用的3-D面部表面重建。对比实验表明,结构照明与对称动态规划双目立体相结合,处理时间合理,精度足够,具有良好的应用前景
{"title":"A Comparison of Three 3-D Facial Reconstruction Approaches","authors":"A. Woodward, Da An, G. Gimel'farb, P. Delmas","doi":"10.1109/ICME.2006.262619","DOIUrl":"https://doi.org/10.1109/ICME.2006.262619","url":null,"abstract":"We compare three computer vision approaches to 3-D reconstruction, namely passive binocular stereo and active structured lighting and photometric stereo, in application to human face reconstruction for modelling virtual humans. An integrated lab environment was set up to simultaneously acquire images for 3-D reconstruction and corresponding data from a 3-D scanner. This allowed us to quantitatively compare reconstruction results to accurate ground truth. Our goal was to determine whether any current computer vision approach is accurate enough for practically useful 3-D facial surface reconstruction. Comparative experiments show the combination of structured lighting with symmetric dynamic programming based binocular stereo has good prospects due to reasonable processing time and sufficient accuracy","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125983929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
On-Demand Partial Schema Delivery for Multimedia Metadata 多媒体元数据的按需部分模式交付
Pub Date : 2006-07-09 DOI: 10.1109/ICME.2006.262830
S. Davis, I. Burnett
XML is a popular approach to interoperable exchange of multimedia metadata between a wide range of devices. This paper explores extending the use of the remote XML exchange protocol (previously proposed by the authors) as a mechanism to provide efficient interaction with complex multimedia XML documents and their associated schemas. This is particularly applicable to users with limited application complexity devices and/or limited bandwidth connections. Many XML documents do not fully utilize all the information present in a given schema; thus, users download substantial redundant information for the current application. This paper introduces the use of RXEP for the transmission of small, relevant schema sections. The paper investigates the advantages of schema retrieval using RXEP in terms of the bandwidth saved
XML是一种在各种设备之间进行多媒体元数据互操作交换的流行方法。本文探讨了扩展远程XML交换协议(作者先前提出的)的使用,将其作为一种机制来提供与复杂多媒体XML文档及其相关模式的有效交互。这特别适用于应用程序复杂性有限的设备和/或带宽有限的连接的用户。许多XML文档没有充分利用给定模式中存在的所有信息;因此,用户为当前应用程序下载大量冗余信息。本文介绍了使用RXEP来传输小的、相关的模式段。本文从节省带宽的角度研究了使用RXEP进行模式检索的优势
{"title":"On-Demand Partial Schema Delivery for Multimedia Metadata","authors":"S. Davis, I. Burnett","doi":"10.1109/ICME.2006.262830","DOIUrl":"https://doi.org/10.1109/ICME.2006.262830","url":null,"abstract":"XML is a popular approach to interoperable exchange of multimedia metadata between a wide range of devices. This paper explores extending the use of the remote XML exchange protocol (previously proposed by the authors) as a mechanism to provide efficient interaction with complex multimedia XML documents and their associated schemas. This is particularly applicable to users with limited application complexity devices and/or limited bandwidth connections. Many XML documents do not fully utilize all the information present in a given schema; thus, users download substantial redundant information for the current application. This paper introduces the use of RXEP for the transmission of small, relevant schema sections. The paper investigates the advantages of schema retrieval using RXEP in terms of the bandwidth saved","PeriodicalId":339258,"journal":{"name":"2006 IEEE International Conference on Multimedia and Expo","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121922153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2006 IEEE International Conference on Multimedia and Expo
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1