Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093840
Marvin Huang, Chiou-Ting Hsu, Kazuyuki Tanaka
Crumpled sheets of paper tend to exhibit specific and complex structure, which is usually described as ridge network by physicists. Existing literature has showed that it is difficult to automate ridge network detection in crumpled paper because of its complex structure. In this paper, we attempt to develop an automatic detection process in terms of our proposed density criterion. We model the ridge network as a weighted graph, where the nodes indicate the intersections of ridges and the edges are the straightened ridges detected in crumpled paper. We construct the weighted graph by first detecting the nodes and then determining the edge weight using the ridge responses. Next, we formulate a graph density criterion to evaluate the detected ridge network. Finally, we propose an edge linking method to construct the graph by maximizing the proposed density criterion. Our experimental results show that, with the density criterion, our proposed node detection together with the edge line linking method could effectively automate the ridge network detection.
{"title":"Automatic ridge network detection in crumpled paper based on graph density","authors":"Marvin Huang, Chiou-Ting Hsu, Kazuyuki Tanaka","doi":"10.1109/MMSP.2011.6093840","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093840","url":null,"abstract":"Crumpled sheets of paper tend to exhibit specific and complex structure, which is usually described as ridge network by physicists. Existing literature has showed that it is difficult to automate ridge network detection in crumpled paper because of its complex structure. In this paper, we attempt to develop an automatic detection process in terms of our proposed density criterion. We model the ridge network as a weighted graph, where the nodes indicate the intersections of ridges and the edges are the straightened ridges detected in crumpled paper. We construct the weighted graph by first detecting the nodes and then determining the edge weight using the ridge responses. Next, we formulate a graph density criterion to evaluate the detected ridge network. Finally, we propose an edge linking method to construct the graph by maximizing the proposed density criterion. Our experimental results show that, with the density criterion, our proposed node detection together with the edge line linking method could effectively automate the ridge network detection.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128345434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093795
Xiguang Zheng, C. Ritz
This paper presents a new coding architecture for the compression of navigable speech soundfield zones. The proposed coding scheme encodes multiple speech soundfields, each representing different spatial zones, into a mono or stereo sound-field mixture signal that can be compressed with an existing speech or audio coder. The resulting compressed signals can be decoded back to individual soundfield zones. Objective and subjective testing results show that the approach successfully compresses up to 3 speech soundfields (each consisting of 4 individual speakers) at a bit rate of 48 kbps whilst maintaining the perceptual quality of each decoded soundfield zone.
{"title":"Compression of navigable speech soundfield zones","authors":"Xiguang Zheng, C. Ritz","doi":"10.1109/MMSP.2011.6093795","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093795","url":null,"abstract":"This paper presents a new coding architecture for the compression of navigable speech soundfield zones. The proposed coding scheme encodes multiple speech soundfields, each representing different spatial zones, into a mono or stereo sound-field mixture signal that can be compressed with an existing speech or audio coder. The resulting compressed signals can be decoded back to individual soundfield zones. Objective and subjective testing results show that the approach successfully compresses up to 3 speech soundfields (each consisting of 4 individual speakers) at a bit rate of 48 kbps whilst maintaining the perceptual quality of each decoded soundfield zone.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128607928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093813
Xiaoming Nan, Yifeng He, L. Guan
Multimedia cloud, as a specific cloud paradigm, addresses how cloud can effectively process multimedia services and provide QoS provisioning for multimedia applications. There are two major challenges in multimedia cloud. The first challenge is the service response time in multimedia cloud, and the second challenge is the cost of cloud resources. In this paper, we optimize resource allocation for multimedia cloud based on queuing model. Specifically, we optimize the resource allocation in both single-class service case and multiple-class service case. In each case, we formulate and solve the response time minimization problem and resource cost minimization problem, respectively. Simulation results demonstrate that the proposed optimal allocation scheme can optimally utilize the cloud resources to achieve a minimal mean response time or a minimal resource cost.
{"title":"Optimal resource allocation for multimedia cloud based on queuing model","authors":"Xiaoming Nan, Yifeng He, L. Guan","doi":"10.1109/MMSP.2011.6093813","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093813","url":null,"abstract":"Multimedia cloud, as a specific cloud paradigm, addresses how cloud can effectively process multimedia services and provide QoS provisioning for multimedia applications. There are two major challenges in multimedia cloud. The first challenge is the service response time in multimedia cloud, and the second challenge is the cost of cloud resources. In this paper, we optimize resource allocation for multimedia cloud based on queuing model. Specifically, we optimize the resource allocation in both single-class service case and multiple-class service case. In each case, we formulate and solve the response time minimization problem and resource cost minimization problem, respectively. Simulation results demonstrate that the proposed optimal allocation scheme can optimally utilize the cloud resources to achieve a minimal mean response time or a minimal resource cost.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126569983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093778
Mu Li, V. Monga
The problem of summarizing videos by short fingerprints or hashes has garnered significant attention recently. While traditional applications of video hashing lie in database search and content authentication, the emergence of websites such as YouTube and DailyMotion poses a challenging problem of anti-piracy video search. That is, hashes or fingerprints of an original video (provided to YouTube by the content owner) must be matched against those uploaded to YouTube by users to identify instances of “illegal” or undesirable uploads. Because the uploaded videos invariably differ from the original in their digital representation (owing to incidental or malicious distortions), robust video hashes are desired. In this paper, we model videos as order-3 tensors and use multilinear subspace projections, such as a reduced rank parallel factor analysis (PARAFAC) to construct video hashes. We observe that unlike most standard descriptors of video content, tensor based subspace projections can offer excellent robustness while effectively capturing the spatio-temporal essence of the video for discriminability. We further randomize the construction of the hash by dividing the video into randomly selected overlapping sub-cubes to prevent against intentional guessing and forgery. The most significant gains are seen for the difficult attacks of spatial (e.g. geometric) as well as temporal (random frame dropping) desynchronization. Experimental validation is provided in the form of ROC curves and we further perform detection-theoretic analysis which closely mimics empirically observed probability of error.
{"title":"Desynchronization resilient video fingerprinting via randomized, low-rank tensor approximations","authors":"Mu Li, V. Monga","doi":"10.1109/MMSP.2011.6093778","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093778","url":null,"abstract":"The problem of summarizing videos by short fingerprints or hashes has garnered significant attention recently. While traditional applications of video hashing lie in database search and content authentication, the emergence of websites such as YouTube and DailyMotion poses a challenging problem of anti-piracy video search. That is, hashes or fingerprints of an original video (provided to YouTube by the content owner) must be matched against those uploaded to YouTube by users to identify instances of “illegal” or undesirable uploads. Because the uploaded videos invariably differ from the original in their digital representation (owing to incidental or malicious distortions), robust video hashes are desired. In this paper, we model videos as order-3 tensors and use multilinear subspace projections, such as a reduced rank parallel factor analysis (PARAFAC) to construct video hashes. We observe that unlike most standard descriptors of video content, tensor based subspace projections can offer excellent robustness while effectively capturing the spatio-temporal essence of the video for discriminability. We further randomize the construction of the hash by dividing the video into randomly selected overlapping sub-cubes to prevent against intentional guessing and forgery. The most significant gains are seen for the difficult attacks of spatial (e.g. geometric) as well as temporal (random frame dropping) desynchronization. Experimental validation is provided in the form of ROC curves and we further perform detection-theoretic analysis which closely mimics empirically observed probability of error.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121434638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093820
Guangchao Peng, Yanwei Liu, Yahui Hu, S. Ci, Hui Tang
In wireless video streaming, the packet loss often occurs and affects the end-user visual quality. To alleviate the transmission error effects, intra refresh coding is usually used to improve the streaming error resilience ability from the view of source coding. At the physical layer, the adaptive modulation and coding (AMC) is also used to promote the transmission reliability at the transporting level. Both the error control components have their own influences on the received video quality. To achieve the best video transmission performance, it is crucial to make an error control tradeoff between intra refresh coding and AMC. In this paper, we propose an end-to-end video distortion optimized cross-layer error control method which jointly considers the video quantization parameter (QP) and intra refresh rate at the application layer, and AMC at the physical layer for delay-constraint real-time video streaming. The experimental results show that the proposed cross-layer error control streaming method can achieve the superior objective and subjective performances to the layer-independent error control streaming methods with and without cross-layer optimization.
{"title":"End-to-end distortion optimized error control for real-time wireless video streaming","authors":"Guangchao Peng, Yanwei Liu, Yahui Hu, S. Ci, Hui Tang","doi":"10.1109/MMSP.2011.6093820","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093820","url":null,"abstract":"In wireless video streaming, the packet loss often occurs and affects the end-user visual quality. To alleviate the transmission error effects, intra refresh coding is usually used to improve the streaming error resilience ability from the view of source coding. At the physical layer, the adaptive modulation and coding (AMC) is also used to promote the transmission reliability at the transporting level. Both the error control components have their own influences on the received video quality. To achieve the best video transmission performance, it is crucial to make an error control tradeoff between intra refresh coding and AMC. In this paper, we propose an end-to-end video distortion optimized cross-layer error control method which jointly considers the video quantization parameter (QP) and intra refresh rate at the application layer, and AMC at the physical layer for delay-constraint real-time video streaming. The experimental results show that the proposed cross-layer error control streaming method can achieve the superior objective and subjective performances to the layer-independent error control streaming methods with and without cross-layer optimization.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"374 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115969215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093825
Viswanathan Swaminathan, Sheng Wei
Hypertext transfer protocol (HTTP) based streaming solutions for live video and video on demand (VOD) applications have become available recently. However, the existing HTTP streaming solutions cannot provide a low latency experience due to the fact that inherently in all of them, latency is tied to the duration of the media fragments that are individually requested and obtained over HTTP. We propose a low latency HTTP streaming approach using HTTP chunked encoding, which enables the server to transmit partial fragments before the entire video fragment is published. We develop an analytical model to quantify and compare the live latencies in three HTTP streaming approaches. Then, we present the details of our experimental setup and implementation. Both the analysis and experimental results show that the chunked encoding approach is capable of reducing the live latency to one to two chunk durations and that the resulting live latency is independent of the fragment duration.
{"title":"Low latency live video streaming using HTTP chunked encoding","authors":"Viswanathan Swaminathan, Sheng Wei","doi":"10.1109/MMSP.2011.6093825","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093825","url":null,"abstract":"Hypertext transfer protocol (HTTP) based streaming solutions for live video and video on demand (VOD) applications have become available recently. However, the existing HTTP streaming solutions cannot provide a low latency experience due to the fact that inherently in all of them, latency is tied to the duration of the media fragments that are individually requested and obtained over HTTP. We propose a low latency HTTP streaming approach using HTTP chunked encoding, which enables the server to transmit partial fragments before the entire video fragment is published. We develop an analytical model to quantify and compare the live latencies in three HTTP streaming approaches. Then, we present the details of our experimental setup and implementation. Both the analysis and experimental results show that the chunked encoding approach is capable of reducing the live latency to one to two chunk durations and that the resulting live latency is independent of the fragment duration.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130902825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093800
Zelin Li, Qiang Wu, Jian Zhang, G. Geers
Pedestrian detection in a thermal image is a difficult task due to intrinsic challenges:1) low image resolution, 2) thermal noising, 3) polarity changes, 4) lack of color, texture or depth information. To address these challenges, we propose a novel mid-level feature descriptor for pedestrian detection in thermal domain, which combines pixel-level Steering Kernel Regression Weights Matrix (SKRWM) with their corresponding covariances. SKRWM can properly capture the local structure of pixels, while the covariance computation can further provide the correlation of low level feature. This mid-level feature descriptor not only captures the pixel-level data difference and spatial differences of local structure, but also explores the correlations among low-level features. In the case of human detection, the proposed mid-level feature descriptor can discriminatively distinguish pedestrian from complexity. For testing the performance of proposed feature descriptor, a popular classifier framework based on Principal Component Analysis (PCA) and Support Vector Machine (SVM) is also built. Overall, our experimental results show that proposed approach has overcome the problems caused by background subtraction in [1] while attains comparable detection accuracy compared to the state-of-the-arts.
{"title":"SKRWM based descriptor for pedestrian detection in thermal images","authors":"Zelin Li, Qiang Wu, Jian Zhang, G. Geers","doi":"10.1109/MMSP.2011.6093800","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093800","url":null,"abstract":"Pedestrian detection in a thermal image is a difficult task due to intrinsic challenges:1) low image resolution, 2) thermal noising, 3) polarity changes, 4) lack of color, texture or depth information. To address these challenges, we propose a novel mid-level feature descriptor for pedestrian detection in thermal domain, which combines pixel-level Steering Kernel Regression Weights Matrix (SKRWM) with their corresponding covariances. SKRWM can properly capture the local structure of pixels, while the covariance computation can further provide the correlation of low level feature. This mid-level feature descriptor not only captures the pixel-level data difference and spatial differences of local structure, but also explores the correlations among low-level features. In the case of human detection, the proposed mid-level feature descriptor can discriminatively distinguish pedestrian from complexity. For testing the performance of proposed feature descriptor, a popular classifier framework based on Principal Component Analysis (PCA) and Support Vector Machine (SVM) is also built. Overall, our experimental results show that proposed approach has overcome the problems caused by background subtraction in [1] while attains comparable detection accuracy compared to the state-of-the-arts.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114896930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093819
Yi-Jen Huang, Hsu-Feng Hsiao
The idea of multiple description video coding has been introduced to deal with the issues of bandwidth/path diversity and packet loss due to network congestion and/or error-prone channels which might cause serious quality degradation in video applications such as multimedia streaming and video conferencing services. In this paper, two approaches to description generation are proposed to produce multiple descriptions at higher coding efficiency. One of them is motivated by the multiple description scalar quantizer to reduce the distortion and the other is the coefficient partition in transform domain in order to balance the descriptions better. An estimation mechanism is further proposed to alleviate the drifting problem due to description fluctuation by synchronizing the reference frames at the encoder and the decoder as much as possible. The experiments show that the proposed methods offer substantial improvement at the event of description loss.
{"title":"An error resilient multiple description video coder","authors":"Yi-Jen Huang, Hsu-Feng Hsiao","doi":"10.1109/MMSP.2011.6093819","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093819","url":null,"abstract":"The idea of multiple description video coding has been introduced to deal with the issues of bandwidth/path diversity and packet loss due to network congestion and/or error-prone channels which might cause serious quality degradation in video applications such as multimedia streaming and video conferencing services. In this paper, two approaches to description generation are proposed to produce multiple descriptions at higher coding efficiency. One of them is motivated by the multiple description scalar quantizer to reduce the distortion and the other is the coefficient partition in transform domain in order to balance the descriptions better. An estimation mechanism is further proposed to alleviate the drifting problem due to description fluctuation by synchronizing the reference frames at the encoder and the decoder as much as possible. The experiments show that the proposed methods offer substantial improvement at the event of description loss.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116291910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093817
Jian-Liang Lin, Yi-Wen Chen, Yu-Pao Tsai, Yu-Wen Huang, S. Lei
High Efficiency Video Coding (HEVC) is a new international video coding standard that has been developed by the Joint Collaborative Team on Video Coding (JCT-VC). In this paper, an overview of the motion vector coding techniques for HEVC is presented. Our three proposed coding tools for the motion vector predictor (MVP) in the Inter, Skip and Merge modes of HEVC are also presented, which includes a new location of the temporal MVP, a priority-based derivation method of spatial MVPs, and a derivation method of temporal MVPs. A combination of these three tools can achieve on average 1.3%, 1.8%, 1.2% and 2.2% bit rate reductions for high efficiency random access, low complexity random access, high efficiency low delay, and low complexity low delay, respectively.
HEVC (High Efficiency Video Coding)是由JCT-VC (Joint Collaborative Team on Video Coding)开发的一种新的国际视频编码标准。本文对HEVC的运动矢量编码技术进行了综述。我们提出了三种用于HEVC的运动矢量预测器(MVP)的编码工具,包括时间矢量预测器的新位置、基于优先级的空间矢量预测器派生方法和时间矢量预测器派生方法。在高效随机访问、低复杂度随机访问、高效低延迟和低复杂度低延迟方面,这三种工具的组合可以分别实现平均1.3%、1.8%、1.2%和2.2%的比特率降低。
{"title":"Motion vector coding techniques for HEVC","authors":"Jian-Liang Lin, Yi-Wen Chen, Yu-Pao Tsai, Yu-Wen Huang, S. Lei","doi":"10.1109/MMSP.2011.6093817","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093817","url":null,"abstract":"High Efficiency Video Coding (HEVC) is a new international video coding standard that has been developed by the Joint Collaborative Team on Video Coding (JCT-VC). In this paper, an overview of the motion vector coding techniques for HEVC is presented. Our three proposed coding tools for the motion vector predictor (MVP) in the Inter, Skip and Merge modes of HEVC are also presented, which includes a new location of the temporal MVP, a priority-based derivation method of spatial MVPs, and a derivation method of temporal MVPs. A combination of these three tools can achieve on average 1.3%, 1.8%, 1.2% and 2.2% bit rate reductions for high efficiency random access, low complexity random access, high efficiency low delay, and low complexity low delay, respectively.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122370181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2011-12-01DOI: 10.1109/MMSP.2011.6093777
Zhiyuan Zhou, Xiaokang Yang, Li Chen, Guangtao Zhai, Wenjun Zhang
In this paper, a novel example-based contrast enhancement algorithm is proposed. The proposed approach enhances the contrast by learning some important informative priors from the histogram of the example image. The experimental results indicate that the proposed Example-based Dist-Stretched (ExDS) contrast enhancement algorithm can boost the image contrast effectively. And thanks to the example-based learning process, the output images from the ExDS algorithm have more natural looking than those of traditional histogram equalization based methods. The proposed ExDS algorithm can also be extended to the applications of contrast correction for old film restoration as well as tone mapping for image and video post-productions.
{"title":"Example-based image contrast enhancement","authors":"Zhiyuan Zhou, Xiaokang Yang, Li Chen, Guangtao Zhai, Wenjun Zhang","doi":"10.1109/MMSP.2011.6093777","DOIUrl":"https://doi.org/10.1109/MMSP.2011.6093777","url":null,"abstract":"In this paper, a novel example-based contrast enhancement algorithm is proposed. The proposed approach enhances the contrast by learning some important informative priors from the histogram of the example image. The experimental results indicate that the proposed Example-based Dist-Stretched (ExDS) contrast enhancement algorithm can boost the image contrast effectively. And thanks to the example-based learning process, the output images from the ExDS algorithm have more natural looking than those of traditional histogram equalization based methods. The proposed ExDS algorithm can also be extended to the applications of contrast correction for old film restoration as well as tone mapping for image and video post-productions.","PeriodicalId":214459,"journal":{"name":"2011 IEEE 13th International Workshop on Multimedia Signal Processing","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130928089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}