首页 > 最新文献

2011 IEEE 13th International Workshop on Multimedia Signal Processing最新文献

英文 中文
Automatic ridge network detection in crumpled paper based on graph density 基于图密度的皱纸脊网自动检测
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093840
Marvin Huang, Chiou-Ting Hsu, Kazuyuki Tanaka
Crumpled sheets of paper tend to exhibit specific and complex structure, which is usually described as ridge network by physicists. Existing literature has showed that it is difficult to automate ridge network detection in crumpled paper because of its complex structure. In this paper, we attempt to develop an automatic detection process in terms of our proposed density criterion. We model the ridge network as a weighted graph, where the nodes indicate the intersections of ridges and the edges are the straightened ridges detected in crumpled paper. We construct the weighted graph by first detecting the nodes and then determining the edge weight using the ridge responses. Next, we formulate a graph density criterion to evaluate the detected ridge network. Finally, we propose an edge linking method to construct the graph by maximizing the proposed density criterion. Our experimental results show that, with the density criterion, our proposed node detection together with the edge line linking method could effectively automate the ridge network detection.
皱巴巴的纸张往往呈现出特定而复杂的结构,物理学家通常将其描述为脊网。现有文献表明,由于皱褶纸结构复杂,很难实现皱褶纸脊网的自动检测。在本文中,我们试图根据我们提出的密度标准开发一个自动检测过程。我们将脊网络建模为加权图,其中节点表示脊的交叉点,边缘是在皱巴巴的纸上检测到的拉直的脊。我们首先通过检测节点,然后利用脊响应确定边缘权重来构建加权图。接下来,我们制定了一个图密度准则来评估检测到的脊网。最后,我们提出了一种边连接方法,通过最大化所提出的密度准则来构造图。实验结果表明,在密度准则下,我们提出的节点检测与边缘线连接方法可以有效地实现脊网检测的自动化。
{"title":"Automatic ridge network detection in crumpled paper based on graph density","authors":"Marvin Huang, Chiou-Ting Hsu, Kazuyuki Tanaka","doi":"10.1109/MMSP.2011.6093840","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093840","url":null,"abstract":"Crumpled sheets of paper tend to exhibit specific and complex structure, which is usually described as ridge network by physicists. Existing literature has showed that it is difficult to automate ridge network detection in crumpled paper because of its complex structure. In this paper, we attempt to develop an automatic detection process in terms of our proposed density criterion. We model the ridge network as a weighted graph, where the nodes indicate the intersections of ridges and the edges are the straightened ridges detected in crumpled paper. We construct the weighted graph by first detecting the nodes and then determining the edge weight using the ridge responses. Next, we formulate a graph density criterion to evaluate the detected ridge network. Finally, we propose an edge linking method to construct the graph by maximizing the proposed density criterion. Our experimental results show that, with the density criterion, our proposed node detection together with the edge line linking method could effectively automate the ridge network detection.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128345434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Compression of navigable speech soundfield zones 可导航语音声场区域的压缩
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093795
Xiguang Zheng, C. Ritz
This paper presents a new coding architecture for the compression of navigable speech soundfield zones. The proposed coding scheme encodes multiple speech soundfields, each representing different spatial zones, into a mono or stereo sound-field mixture signal that can be compressed with an existing speech or audio coder. The resulting compressed signals can be decoded back to individual soundfield zones. Objective and subjective testing results show that the approach successfully compresses up to 3 speech soundfields (each consisting of 4 individual speakers) at a bit rate of 48 kbps whilst maintaining the perceptual quality of each decoded soundfield zone.
本文提出了一种新的可导航语音声场区域压缩编码结构。所提出的编码方案将多个语音声场(每个声场代表不同的空间区域)编码成单声道或立体声声场混合信号,该信号可以用现有的语音或音频编码器进行压缩。由此产生的压缩信号可以被解码回单个声场区域。客观和主观测试结果表明,该方法成功地以48 kbps的比特率压缩了多达3个语音声场(每个声场由4个单独的说话者组成),同时保持了每个解码声场区域的感知质量。
{"title":"Compression of navigable speech soundfield zones","authors":"Xiguang Zheng, C. Ritz","doi":"10.1109/MMSP.2011.6093795","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093795","url":null,"abstract":"This paper presents a new coding architecture for the compression of navigable speech soundfield zones. The proposed coding scheme encodes multiple speech soundfields, each representing different spatial zones, into a mono or stereo sound-field mixture signal that can be compressed with an existing speech or audio coder. The resulting compressed signals can be decoded back to individual soundfield zones. Objective and subjective testing results show that the approach successfully compresses up to 3 speech soundfields (each consisting of 4 individual speakers) at a bit rate of 48 kbps whilst maintaining the perceptual quality of each decoded soundfield zone.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128607928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Optimal resource allocation for multimedia cloud based on queuing model 基于排队模型的多媒体云资源优化分配
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093813
Xiaoming Nan, Yifeng He, L. Guan
Multimedia cloud, as a specific cloud paradigm, addresses how cloud can effectively process multimedia services and provide QoS provisioning for multimedia applications. There are two major challenges in multimedia cloud. The first challenge is the service response time in multimedia cloud, and the second challenge is the cost of cloud resources. In this paper, we optimize resource allocation for multimedia cloud based on queuing model. Specifically, we optimize the resource allocation in both single-class service case and multiple-class service case. In each case, we formulate and solve the response time minimization problem and resource cost minimization problem, respectively. Simulation results demonstrate that the proposed optimal allocation scheme can optimally utilize the cloud resources to achieve a minimal mean response time or a minimal resource cost.
多媒体云作为一种特定的云范式,解决了云如何有效地处理多媒体服务并为多媒体应用程序提供QoS配置的问题。多媒体云面临着两大挑战。第一个挑战是多媒体云中的服务响应时间,第二个挑战是云资源的成本。本文基于排队模型对多媒体云的资源分配进行了优化。具体来说,我们对单类服务和多类服务的资源分配进行了优化。在每种情况下,我们分别制定和解决响应时间最小化问题和资源成本最小化问题。仿真结果表明,所提出的最优分配方案能够最优地利用云资源,实现最小的平均响应时间或最小的资源成本。
{"title":"Optimal resource allocation for multimedia cloud based on queuing model","authors":"Xiaoming Nan, Yifeng He, L. Guan","doi":"10.1109/MMSP.2011.6093813","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093813","url":null,"abstract":"Multimedia cloud, as a specific cloud paradigm, addresses how cloud can effectively process multimedia services and provide QoS provisioning for multimedia applications. There are two major challenges in multimedia cloud. The first challenge is the service response time in multimedia cloud, and the second challenge is the cost of cloud resources. In this paper, we optimize resource allocation for multimedia cloud based on queuing model. Specifically, we optimize the resource allocation in both single-class service case and multiple-class service case. In each case, we formulate and solve the response time minimization problem and resource cost minimization problem, respectively. Simulation results demonstrate that the proposed optimal allocation scheme can optimally utilize the cloud resources to achieve a minimal mean response time or a minimal resource cost.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126569983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 114
Desynchronization resilient video fingerprinting via randomized, low-rank tensor approximations 去同步弹性视频指纹通过随机,低秩张量近似
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093778
Mu Li, V. Monga
The problem of summarizing videos by short fingerprints or hashes has garnered significant attention recently. While traditional applications of video hashing lie in database search and content authentication, the emergence of websites such as YouTube and DailyMotion poses a challenging problem of anti-piracy video search. That is, hashes or fingerprints of an original video (provided to YouTube by the content owner) must be matched against those uploaded to YouTube by users to identify instances of “illegal” or undesirable uploads. Because the uploaded videos invariably differ from the original in their digital representation (owing to incidental or malicious distortions), robust video hashes are desired. In this paper, we model videos as order-3 tensors and use multilinear subspace projections, such as a reduced rank parallel factor analysis (PARAFAC) to construct video hashes. We observe that unlike most standard descriptors of video content, tensor based subspace projections can offer excellent robustness while effectively capturing the spatio-temporal essence of the video for discriminability. We further randomize the construction of the hash by dividing the video into randomly selected overlapping sub-cubes to prevent against intentional guessing and forgery. The most significant gains are seen for the difficult attacks of spatial (e.g. geometric) as well as temporal (random frame dropping) desynchronization. Experimental validation is provided in the form of ROC curves and we further perform detection-theoretic analysis which closely mimics empirically observed probability of error.
最近,通过短指纹或哈希来总结视频的问题引起了人们的极大关注。传统的视频哈希算法应用于数据库搜索和内容认证,而YouTube和DailyMotion等网站的出现,给反盗版视频搜索带来了挑战。也就是说,原始视频(由内容所有者提供给YouTube)的哈希值或指纹必须与用户上传到YouTube的视频相匹配,以识别“非法”或不受欢迎的上传实例。由于上传的视频总是在数字表示上与原始视频不同(由于偶然或恶意扭曲),因此需要鲁棒的视频哈希。在本文中,我们将视频建模为3阶张量,并使用多线性子空间投影,如降阶并行因子分析(PARAFAC)来构建视频哈希。我们观察到,与大多数视频内容的标准描述符不同,基于张量的子空间投影可以提供出色的鲁棒性,同时有效地捕捉视频的时空本质以实现可判别性。我们通过将视频分成随机选择的重叠子立方体来进一步随机化哈希的构造,以防止故意猜测和伪造。最显著的增益见于空间(例如几何)和时间(随机帧丢失)去同步的困难攻击。实验验证以ROC曲线的形式提供,我们进一步进行检测理论分析,接近模拟经验观察到的误差概率。
{"title":"Desynchronization resilient video fingerprinting via randomized, low-rank tensor approximations","authors":"Mu Li, V. Monga","doi":"10.1109/MMSP.2011.6093778","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093778","url":null,"abstract":"The problem of summarizing videos by short fingerprints or hashes has garnered significant attention recently. While traditional applications of video hashing lie in database search and content authentication, the emergence of websites such as YouTube and DailyMotion poses a challenging problem of anti-piracy video search. That is, hashes or fingerprints of an original video (provided to YouTube by the content owner) must be matched against those uploaded to YouTube by users to identify instances of “illegal” or undesirable uploads. Because the uploaded videos invariably differ from the original in their digital representation (owing to incidental or malicious distortions), robust video hashes are desired. In this paper, we model videos as order-3 tensors and use multilinear subspace projections, such as a reduced rank parallel factor analysis (PARAFAC) to construct video hashes. We observe that unlike most standard descriptors of video content, tensor based subspace projections can offer excellent robustness while effectively capturing the spatio-temporal essence of the video for discriminability. We further randomize the construction of the hash by dividing the video into randomly selected overlapping sub-cubes to prevent against intentional guessing and forgery. The most significant gains are seen for the difficult attacks of spatial (e.g. geometric) as well as temporal (random frame dropping) desynchronization. Experimental validation is provided in the form of ROC curves and we further perform detection-theoretic analysis which closely mimics empirically observed probability of error.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121434638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
End-to-end distortion optimized error control for real-time wireless video streaming 实时无线视频流端到端失真优化误差控制
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093820
Guangchao Peng, Yanwei Liu, Yahui Hu, S. Ci, Hui Tang
In wireless video streaming, the packet loss often occurs and affects the end-user visual quality. To alleviate the transmission error effects, intra refresh coding is usually used to improve the streaming error resilience ability from the view of source coding. At the physical layer, the adaptive modulation and coding (AMC) is also used to promote the transmission reliability at the transporting level. Both the error control components have their own influences on the received video quality. To achieve the best video transmission performance, it is crucial to make an error control tradeoff between intra refresh coding and AMC. In this paper, we propose an end-to-end video distortion optimized cross-layer error control method which jointly considers the video quantization parameter (QP) and intra refresh rate at the application layer, and AMC at the physical layer for delay-constraint real-time video streaming. The experimental results show that the proposed cross-layer error control streaming method can achieve the superior objective and subjective performances to the layer-independent error control streaming methods with and without cross-layer optimization.
在无线视频流中,经常会出现丢包现象,影响终端用户的视觉质量。为了减轻传输误差的影响,通常采用帧内刷新编码,从源编码的角度提高流的容错能力。在物理层,自适应调制编码(AMC)也被用于提高传输层的传输可靠性。这两个误差控制组件对接收到的视频质量都有各自的影响。为了获得最佳的视频传输性能,在帧内刷新编码和帧内刷新编码之间进行错误控制的权衡是至关重要的。本文提出了一种端到端视频失真优化的跨层误差控制方法,该方法在应用层综合考虑视频量化参数(QP)和帧内刷新率,在物理层综合考虑AMC,用于时延约束的实时视频流。实验结果表明,所提出的跨层误差控制流方法在客观和主观性能上都优于具有和不具有跨层优化的层无关误差控制流方法。
{"title":"End-to-end distortion optimized error control for real-time wireless video streaming","authors":"Guangchao Peng, Yanwei Liu, Yahui Hu, S. Ci, Hui Tang","doi":"10.1109/MMSP.2011.6093820","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093820","url":null,"abstract":"In wireless video streaming, the packet loss often occurs and affects the end-user visual quality. To alleviate the transmission error effects, intra refresh coding is usually used to improve the streaming error resilience ability from the view of source coding. At the physical layer, the adaptive modulation and coding (AMC) is also used to promote the transmission reliability at the transporting level. Both the error control components have their own influences on the received video quality. To achieve the best video transmission performance, it is crucial to make an error control tradeoff between intra refresh coding and AMC. In this paper, we propose an end-to-end video distortion optimized cross-layer error control method which jointly considers the video quantization parameter (QP) and intra refresh rate at the application layer, and AMC at the physical layer for delay-constraint real-time video streaming. The experimental results show that the proposed cross-layer error control streaming method can achieve the superior objective and subjective performances to the layer-independent error control streaming methods with and without cross-layer optimization.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"374 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115969215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Low latency live video streaming using HTTP chunked encoding 使用HTTP分块编码的低延迟实时视频流
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093825
Viswanathan Swaminathan, Sheng Wei
Hypertext transfer protocol (HTTP) based streaming solutions for live video and video on demand (VOD) applications have become available recently. However, the existing HTTP streaming solutions cannot provide a low latency experience due to the fact that inherently in all of them, latency is tied to the duration of the media fragments that are individually requested and obtained over HTTP. We propose a low latency HTTP streaming approach using HTTP chunked encoding, which enables the server to transmit partial fragments before the entire video fragment is published. We develop an analytical model to quantify and compare the live latencies in three HTTP streaming approaches. Then, we present the details of our experimental setup and implementation. Both the analysis and experimental results show that the chunked encoding approach is capable of reducing the live latency to one to two chunk durations and that the resulting live latency is independent of the fragment duration.
基于超文本传输协议(HTTP)的实时视频和视频点播(VOD)应用的流媒体解决方案最近已经出现。然而,现有的HTTP流解决方案无法提供低延迟体验,因为在所有这些解决方案中,延迟都与通过HTTP单独请求和获得的媒体片段的持续时间有关。我们提出了一种使用HTTP分块编码的低延迟HTTP流方法,该方法使服务器能够在发布整个视频片段之前传输部分片段。我们开发了一个分析模型来量化和比较三种HTTP流方法中的实时延迟。然后,我们介绍了我们的实验设置和实现的细节。分析和实验结果都表明,分块编码方法能够将活延迟减少到一到两个块持续时间,并且产生的活延迟与片段持续时间无关。
{"title":"Low latency live video streaming using HTTP chunked encoding","authors":"Viswanathan Swaminathan, Sheng Wei","doi":"10.1109/MMSP.2011.6093825","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093825","url":null,"abstract":"Hypertext transfer protocol (HTTP) based streaming solutions for live video and video on demand (VOD) applications have become available recently. However, the existing HTTP streaming solutions cannot provide a low latency experience due to the fact that inherently in all of them, latency is tied to the duration of the media fragments that are individually requested and obtained over HTTP. We propose a low latency HTTP streaming approach using HTTP chunked encoding, which enables the server to transmit partial fragments before the entire video fragment is published. We develop an analytical model to quantify and compare the live latencies in three HTTP streaming approaches. Then, we present the details of our experimental setup and implementation. Both the analysis and experimental results show that the chunked encoding approach is capable of reducing the live latency to one to two chunk durations and that the resulting live latency is independent of the fragment duration.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130902825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
SKRWM based descriptor for pedestrian detection in thermal images 基于SKRWM的热图像行人检测描述符
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093800
Zelin Li, Qiang Wu, Jian Zhang, G. Geers
Pedestrian detection in a thermal image is a difficult task due to intrinsic challenges:1) low image resolution, 2) thermal noising, 3) polarity changes, 4) lack of color, texture or depth information. To address these challenges, we propose a novel mid-level feature descriptor for pedestrian detection in thermal domain, which combines pixel-level Steering Kernel Regression Weights Matrix (SKRWM) with their corresponding covariances. SKRWM can properly capture the local structure of pixels, while the covariance computation can further provide the correlation of low level feature. This mid-level feature descriptor not only captures the pixel-level data difference and spatial differences of local structure, but also explores the correlations among low-level features. In the case of human detection, the proposed mid-level feature descriptor can discriminatively distinguish pedestrian from complexity. For testing the performance of proposed feature descriptor, a popular classifier framework based on Principal Component Analysis (PCA) and Support Vector Machine (SVM) is also built. Overall, our experimental results show that proposed approach has overcome the problems caused by background subtraction in [1] while attains comparable detection accuracy compared to the state-of-the-arts.
由于固有的挑战,热图像中的行人检测是一项艰巨的任务:1)低图像分辨率,2)热噪声,3)极性变化,4)缺乏颜色,纹理或深度信息。为了解决这些挑战,我们提出了一种新的用于热域行人检测的中级特征描述符,该描述符将像素级转向核回归权重矩阵(SKRWM)与其相应的协方差相结合。SKRWM可以很好地捕捉像素的局部结构,而协方差计算可以进一步提供低层特征的相关性。该中级特征描述符不仅捕获了像素级数据差异和局部结构的空间差异,而且还探索了低级特征之间的相关性。在人类检测的情况下,所提出的中级特征描述符可以区分行人和复杂性。为了测试所提出的特征描述符的性能,还构建了一个基于主成分分析(PCA)和支持向量机(SVM)的流行分类器框架。总体而言,我们的实验结果表明,所提出的方法克服了[1]中背景减法引起的问题,同时与最先进的方法相比,获得了相当的检测精度。
{"title":"SKRWM based descriptor for pedestrian detection in thermal images","authors":"Zelin Li, Qiang Wu, Jian Zhang, G. Geers","doi":"10.1109/MMSP.2011.6093800","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093800","url":null,"abstract":"Pedestrian detection in a thermal image is a difficult task due to intrinsic challenges:1) low image resolution, 2) thermal noising, 3) polarity changes, 4) lack of color, texture or depth information. To address these challenges, we propose a novel mid-level feature descriptor for pedestrian detection in thermal domain, which combines pixel-level Steering Kernel Regression Weights Matrix (SKRWM) with their corresponding covariances. SKRWM can properly capture the local structure of pixels, while the covariance computation can further provide the correlation of low level feature. This mid-level feature descriptor not only captures the pixel-level data difference and spatial differences of local structure, but also explores the correlations among low-level features. In the case of human detection, the proposed mid-level feature descriptor can discriminatively distinguish pedestrian from complexity. For testing the performance of proposed feature descriptor, a popular classifier framework based on Principal Component Analysis (PCA) and Support Vector Machine (SVM) is also built. Overall, our experimental results show that proposed approach has overcome the problems caused by background subtraction in [1] while attains comparable detection accuracy compared to the state-of-the-arts.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114896930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
An error resilient multiple description video coder 一种抗错误多描述视频编码器
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093819
Yi-Jen Huang, Hsu-Feng Hsiao
The idea of multiple description video coding has been introduced to deal with the issues of bandwidth/path diversity and packet loss due to network congestion and/or error-prone channels which might cause serious quality degradation in video applications such as multimedia streaming and video conferencing services. In this paper, two approaches to description generation are proposed to produce multiple descriptions at higher coding efficiency. One of them is motivated by the multiple description scalar quantizer to reduce the distortion and the other is the coefficient partition in transform domain in order to balance the descriptions better. An estimation mechanism is further proposed to alleviate the drifting problem due to description fluctuation by synchronizing the reference frames at the encoder and the decoder as much as possible. The experiments show that the proposed methods offer substantial improvement at the event of description loss.
多描述视频编码的思想已经被引入来处理带宽/路径多样性和由于网络拥塞和/或容易出错的通道而导致的数据包丢失问题,这些问题可能会导致视频应用(如多媒体流媒体和视频会议服务)的质量严重下降。本文提出了两种描述生成方法,以提高编码效率生成多个描述。其中一种方法是利用多重描述标量量化器来减少描述失真,另一种方法是在变换域进行系数分割来更好地平衡描述。进一步提出了一种估计机制,通过尽可能同步编码器和解码器的参考帧来缓解由于描述波动引起的漂移问题。实验表明,在描述丢失的情况下,所提出的方法有很大的改进。
{"title":"An error resilient multiple description video coder","authors":"Yi-Jen Huang, Hsu-Feng Hsiao","doi":"10.1109/MMSP.2011.6093819","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093819","url":null,"abstract":"The idea of multiple description video coding has been introduced to deal with the issues of bandwidth/path diversity and packet loss due to network congestion and/or error-prone channels which might cause serious quality degradation in video applications such as multimedia streaming and video conferencing services. In this paper, two approaches to description generation are proposed to produce multiple descriptions at higher coding efficiency. One of them is motivated by the multiple description scalar quantizer to reduce the distortion and the other is the coefficient partition in transform domain in order to balance the descriptions better. An estimation mechanism is further proposed to alleviate the drifting problem due to description fluctuation by synchronizing the reference frames at the encoder and the decoder as much as possible. The experiments show that the proposed methods offer substantial improvement at the event of description loss.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116291910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motion vector coding techniques for HEVC HEVC的运动矢量编码技术
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093817
Jian-Liang Lin, Yi-Wen Chen, Yu-Pao Tsai, Yu-Wen Huang, S. Lei
High Efficiency Video Coding (HEVC) is a new international video coding standard that has been developed by the Joint Collaborative Team on Video Coding (JCT-VC). In this paper, an overview of the motion vector coding techniques for HEVC is presented. Our three proposed coding tools for the motion vector predictor (MVP) in the Inter, Skip and Merge modes of HEVC are also presented, which includes a new location of the temporal MVP, a priority-based derivation method of spatial MVPs, and a derivation method of temporal MVPs. A combination of these three tools can achieve on average 1.3%, 1.8%, 1.2% and 2.2% bit rate reductions for high efficiency random access, low complexity random access, high efficiency low delay, and low complexity low delay, respectively.
HEVC (High Efficiency Video Coding)是由JCT-VC (Joint Collaborative Team on Video Coding)开发的一种新的国际视频编码标准。本文对HEVC的运动矢量编码技术进行了综述。我们提出了三种用于HEVC的运动矢量预测器(MVP)的编码工具,包括时间矢量预测器的新位置、基于优先级的空间矢量预测器派生方法和时间矢量预测器派生方法。在高效随机访问、低复杂度随机访问、高效低延迟和低复杂度低延迟方面,这三种工具的组合可以分别实现平均1.3%、1.8%、1.2%和2.2%的比特率降低。
{"title":"Motion vector coding techniques for HEVC","authors":"Jian-Liang Lin, Yi-Wen Chen, Yu-Pao Tsai, Yu-Wen Huang, S. Lei","doi":"10.1109/MMSP.2011.6093817","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093817","url":null,"abstract":"High Efficiency Video Coding (HEVC) is a new international video coding standard that has been developed by the Joint Collaborative Team on Video Coding (JCT-VC). In this paper, an overview of the motion vector coding techniques for HEVC is presented. Our three proposed coding tools for the motion vector predictor (MVP) in the Inter, Skip and Merge modes of HEVC are also presented, which includes a new location of the temporal MVP, a priority-based derivation method of spatial MVPs, and a derivation method of temporal MVPs. A combination of these three tools can achieve on average 1.3%, 1.8%, 1.2% and 2.2% bit rate reductions for high efficiency random access, low complexity random access, high efficiency low delay, and low complexity low delay, respectively.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122370181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Example-based image contrast enhancement 基于示例的图像对比度增强
Pub Date : 2011-12-01 DOI: 10.1109/MMSP.2011.6093777
Zhiyuan Zhou, Xiaokang Yang, Li Chen, Guangtao Zhai, Wenjun Zhang
In this paper, a novel example-based contrast enhancement algorithm is proposed. The proposed approach enhances the contrast by learning some important informative priors from the histogram of the example image. The experimental results indicate that the proposed Example-based Dist-Stretched (ExDS) contrast enhancement algorithm can boost the image contrast effectively. And thanks to the example-based learning process, the output images from the ExDS algorithm have more natural looking than those of traditional histogram equalization based methods. The proposed ExDS algorithm can also be extended to the applications of contrast correction for old film restoration as well as tone mapping for image and video post-productions.
本文提出了一种新的基于实例的对比度增强算法。该方法通过从样本图像的直方图中学习一些重要的先验信息来增强对比度。实验结果表明,本文提出的基于实例的离散拉伸(ExDS)对比度增强算法可以有效地提高图像的对比度。由于基于示例的学习过程,ExDS算法的输出图像比传统的基于直方图均衡化的方法更自然。所提出的ExDS算法还可以扩展到旧电影修复中的对比度校正以及图像和视频后期制作中的色调映射。
{"title":"Example-based image contrast enhancement","authors":"Zhiyuan Zhou, Xiaokang Yang, Li Chen, Guangtao Zhai, Wenjun Zhang","doi":"10.1109/MMSP.2011.6093777","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093777","url":null,"abstract":"In this paper, a novel example-based contrast enhancement algorithm is proposed. The proposed approach enhances the contrast by learning some important informative priors from the histogram of the example image. The experimental results indicate that the proposed Example-based Dist-Stretched (ExDS) contrast enhancement algorithm can boost the image contrast effectively. And thanks to the example-based learning process, the output images from the ExDS algorithm have more natural looking than those of traditional histogram equalization based methods. The proposed ExDS algorithm can also be extended to the applications of contrast correction for old film restoration as well as tone mapping for image and video post-productions.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130928089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2011 IEEE 13th International Workshop on Multimedia Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1