Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958799
Ligang Zhang, C. Lau, D. Tjondronegoro, V. Chandran
The proliferation of news reports published in online websites and news information sharing among social media users necessitates effective techniques for analysing the image, text and video data related to news topics. This paper presents the first study to classify affective facial images on emerging news topics. The proposed system dynamically monitors and selects the current hot (of great interest) news topics with strong affective interestingness using textual keywords in news articles and social media discussions. Images from the selected hot topics are extracted and classified into three categorized emotions, positive, neutral and negative, based on facial expressions of subjects in the images. Performance evaluations on two facial image datasets collected from realworld resources demonstrate the applicability and effectiveness of the proposed system in affective classification of facial images in news reports. Facial expression shows high consistency with the affective textual content in news reports for positive emotion, while only low correlation has been observed for neutral and negative. The system can be directly used for applications, such as assisting editors in choosing photos with a proper affective semantic for a certain topic during news report preparation.
{"title":"A pilot study on affective classification of facial images for emerging news topics","authors":"Ligang Zhang, C. Lau, D. Tjondronegoro, V. Chandran","doi":"10.1109/MMSP.2014.6958799","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958799","url":null,"abstract":"The proliferation of news reports published in online websites and news information sharing among social media users necessitates effective techniques for analysing the image, text and video data related to news topics. This paper presents the first study to classify affective facial images on emerging news topics. The proposed system dynamically monitors and selects the current hot (of great interest) news topics with strong affective interestingness using textual keywords in news articles and social media discussions. Images from the selected hot topics are extracted and classified into three categorized emotions, positive, neutral and negative, based on facial expressions of subjects in the images. Performance evaluations on two facial image datasets collected from realworld resources demonstrate the applicability and effectiveness of the proposed system in affective classification of facial images in news reports. Facial expression shows high consistency with the affective textual content in news reports for positive emotion, while only low correlation has been observed for neutral and negative. The system can be directly used for applications, such as assisting editors in choosing photos with a proper affective semantic for a certain topic during news report preparation.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126953609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958821
Xin Lu, G. Martin
An improved rate control algorithm for the Scalable Video Coding (SVC) extension of H.264/AVC is described. The rate control scheme applied to the Base Layer (BL) of SVC adopts the linear Mean Absolute Difference (MAD) prediction and quadratic Rate Distortion (RD) models inherited from H.264/AVC. A MAD prediction error always exists and cannot be avoided. However, some encoding results of the base layer can be used to inform the coding of the enhancement layers (ELs), thus benefitting from the bottom-up coding structure of SVC. This property forms the basis for the proposed rate control approach. Simulation results show that accurate rate control is achieved and, compared to the default rate control algorithm of SVC, namely the JVT-G012 rate control scheme, the average PSNR is increased by 0.27dB or the average bit rate is reduced by 4.81%.
{"title":"An improved rate control algorithm for SVC with optimised MAD prediction","authors":"Xin Lu, G. Martin","doi":"10.1109/MMSP.2014.6958821","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958821","url":null,"abstract":"An improved rate control algorithm for the Scalable Video Coding (SVC) extension of H.264/AVC is described. The rate control scheme applied to the Base Layer (BL) of SVC adopts the linear Mean Absolute Difference (MAD) prediction and quadratic Rate Distortion (RD) models inherited from H.264/AVC. A MAD prediction error always exists and cannot be avoided. However, some encoding results of the base layer can be used to inform the coding of the enhancement layers (ELs), thus benefitting from the bottom-up coding structure of SVC. This property forms the basis for the proposed rate control approach. Simulation results show that accurate rate control is achieved and, compared to the default rate control algorithm of SVC, namely the JVT-G012 rate control scheme, the average PSNR is increased by 0.27dB or the average bit rate is reduced by 4.81%.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128437079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958818
Keji Chen, Y. Duan, Jun Sun, Zongming Guo
High Efficiency Video Coding (HEVC) is the new generation video coding standard which achieves significant improvement in coding efficiency. Although HEVC is promising in many applications, the increased computational complexity is a serious problem, which makes parallelization necessary in HEVC encoding. To better understand the bottleneck of parallelization and improve the encoding speed, in this paper, we propose a Coding Tree Blocks (CTB) level parallelism analysis method as well as a novel Inter-Frame Wavefront (IFW) parallel encoding method. First, by establishing the relationship between parallelism and dependence, parallelism is precisely described by CTB-level dependence as a criterion to evaluate different parallel methods of HEVC. On this basis, by effectively decreasing the dependence based on Wavefront Parallel Processing (WPP), IFW method is developed. Finally, with the proposed parallelism analysis method, IFW is theoretically proved to be of higher parallelism compared with other HEVC representative parallel methods. Extensive experimental results show that, the proposed method and implementation can bring up to 17.81x, 14.34x and 24.40x speedup for HEVC encoding of WVGA, 720p and 1080p standard test sequences with the same ignorable coding performance degradation as WPP, thus showing a promising technology for future large-scale HEVC video application.
高效视频编码(High Efficiency Video Coding, HEVC)是新一代视频编码标准,大大提高了编码效率。尽管HEVC在许多应用中都很有前景,但计算复杂度的增加是一个严重的问题,这使得并行化成为HEVC编码的必要条件。为了更好地理解并行化的瓶颈并提高编码速度,本文提出了一种编码树块(CTB)级并行性分析方法和一种新的帧间波前(IFW)并行编码方法。首先,通过建立并行度与依赖度之间的关系,将并行度精确地描述为ctb级依赖度,作为评价HEVC不同并行方法的标准;在此基础上,通过有效降低波前并行处理(WPP)的依赖性,发展了IFW方法。最后,通过提出的并行性分析方法,从理论上证明了IFW与其他HEVC代表性并行方法相比具有更高的并行性。大量的实验结果表明,所提出的方法和实现对WVGA、720p和1080p标准测试序列的HEVC编码速度提高了17.81倍、14.34倍和24.40倍,编码性能的下降与WPP相同,可以忽略,为未来大规模HEVC视频应用展示了一种很有前景的技术。
{"title":"Towards efficient wavefront parallel encoding of HEVC: Parallelism analysis and improvement","authors":"Keji Chen, Y. Duan, Jun Sun, Zongming Guo","doi":"10.1109/MMSP.2014.6958818","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958818","url":null,"abstract":"High Efficiency Video Coding (HEVC) is the new generation video coding standard which achieves significant improvement in coding efficiency. Although HEVC is promising in many applications, the increased computational complexity is a serious problem, which makes parallelization necessary in HEVC encoding. To better understand the bottleneck of parallelization and improve the encoding speed, in this paper, we propose a Coding Tree Blocks (CTB) level parallelism analysis method as well as a novel Inter-Frame Wavefront (IFW) parallel encoding method. First, by establishing the relationship between parallelism and dependence, parallelism is precisely described by CTB-level dependence as a criterion to evaluate different parallel methods of HEVC. On this basis, by effectively decreasing the dependence based on Wavefront Parallel Processing (WPP), IFW method is developed. Finally, with the proposed parallelism analysis method, IFW is theoretically proved to be of higher parallelism compared with other HEVC representative parallel methods. Extensive experimental results show that, the proposed method and implementation can bring up to 17.81x, 14.34x and 24.40x speedup for HEVC encoding of WVGA, 720p and 1080p standard test sequences with the same ignorable coding performance degradation as WPP, thus showing a promising technology for future large-scale HEVC video application.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129632660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958813
Joseph Santarcangelo, Xiao-Ping Zhang
This paper categorizes children's videos according to an expertly assigned predefined positive or negative cognitive impact category. The method uses affective features to determine if a video belongs to an expertly assigned predefined positive or to a negative cognitive impact category. The work demonstrates that simple affective features outperform more complex systems in determining if content belongs to the positive or negative cognitive impact category. The work is tested on a set of videos that have been classified as having a short term or long term measurable negative or positive impact on cognition based on cited psychological literature. It found that affective analysis had superior performance using less features than state of the art video genre classification systems. It also found that arousal features performed better than valence features.
{"title":"Classifying harmful children's content using affective analysis","authors":"Joseph Santarcangelo, Xiao-Ping Zhang","doi":"10.1109/MMSP.2014.6958813","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958813","url":null,"abstract":"This paper categorizes children's videos according to an expertly assigned predefined positive or negative cognitive impact category. The method uses affective features to determine if a video belongs to an expertly assigned predefined positive or to a negative cognitive impact category. The work demonstrates that simple affective features outperform more complex systems in determining if content belongs to the positive or negative cognitive impact category. The work is tested on a set of videos that have been classified as having a short term or long term measurable negative or positive impact on cognition based on cited psychological literature. It found that affective analysis had superior performance using less features than state of the art video genre classification systems. It also found that arousal features performed better than valence features.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129226325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958824
Yunong Wei, Yuan Zhang, Jinyao Yan
The error resilience and low-complexity video encoding are two major requirements of real-time visual communications on mobile devices. To address the two requirements simultaneously, this paper presents a fast mode decision algorithm for the error resilient video coding in packet loss environment. The proposed algorithm is a two-step method: early skip mode decision and early intra mode decision. Different from the existing methods for early skip mode decision, the proposed method takes the error-propagation distortion into account in estimating the coding cost. Considering the intra blocks are frequently used to terminate the error propagations, we also propose a method to fast estimate the intra block coding cost, so that the intra mode can be early determined. Overall, the proposed method can significantly reduce the encoding time while keeping the coding efficiency similar to the rate-distortion optimized mode decision method.
{"title":"Fast mode decision for error resilient video coding","authors":"Yunong Wei, Yuan Zhang, Jinyao Yan","doi":"10.1109/MMSP.2014.6958824","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958824","url":null,"abstract":"The error resilience and low-complexity video encoding are two major requirements of real-time visual communications on mobile devices. To address the two requirements simultaneously, this paper presents a fast mode decision algorithm for the error resilient video coding in packet loss environment. The proposed algorithm is a two-step method: early skip mode decision and early intra mode decision. Different from the existing methods for early skip mode decision, the proposed method takes the error-propagation distortion into account in estimating the coding cost. Considering the intra blocks are frequently used to terminate the error propagations, we also propose a method to fast estimate the intra block coding cost, so that the intra mode can be early determined. Overall, the proposed method can significantly reduce the encoding time while keeping the coding efficiency similar to the rate-distortion optimized mode decision method.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"16 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116341957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958833
Claire Mantel, Stefan Catalin Ferchiu, Søren Forchhammer
In this paper a subjective test in which participants evaluate the quality of JPEG-XT compressed HDR images is presented. Results show that for the selected test images and display, the subjective quality reached its saturation point starting around 3bpp. Objective evaluations are obtained by applying a model of the display and providing the modeled images to three objective metrics dedicated to HDR content. Objective grades are compared with subjective data both in physical domain and using a gamma correction to approximate perceptually uniform luminance coding. The MRSE metric obtains the best performance with the limit that it does not capture the quality saturation. The usage of the gamma correction prior to applying metrics depends on the characteristics of each objective metric.
{"title":"Comparing subjective and objective quality assessment of HDR images compressed with JPEG-XT","authors":"Claire Mantel, Stefan Catalin Ferchiu, Søren Forchhammer","doi":"10.1109/MMSP.2014.6958833","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958833","url":null,"abstract":"In this paper a subjective test in which participants evaluate the quality of JPEG-XT compressed HDR images is presented. Results show that for the selected test images and display, the subjective quality reached its saturation point starting around 3bpp. Objective evaluations are obtained by applying a model of the display and providing the modeled images to three objective metrics dedicated to HDR content. Objective grades are compared with subjective data both in physical domain and using a gamma correction to approximate perceptually uniform luminance coding. The MRSE metric obtains the best performance with the limit that it does not capture the quality saturation. The usage of the gamma correction prior to applying metrics depends on the characteristics of each objective metric.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128060172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958791
Xueyang Fu, Yue Huang, Delu Zeng, Xiao-Ping Zhang, Xinghao Ding
In this paper, a novel image enhancing approach focuses on single sandstorm image is proposed. The degraded image has some problems, such as color distortion, low-visibility, fuzz and non-uniform luminance, due to the light is absorbed and scattered by particles in sandstorm. The proposed approach based on fusion principles aims to overcome the aforementioned limitations. First, the degraded image is color corrected by adopting a statistical strategy. Then two inputs, which represent different brightness, are derived only from the color corrected image by applying Gamma correction. Three weighted maps (sharpness, chromaticity and prominence), which contain important features to increase the quality of the degraded image, are computed from the derived inputs. Finally, the enhanced image is obtained by fusing the inputs with the weight maps. The proposed method is the first to adopt a fusion-based method for enhancing single sandstorm image. Experimental results show that enhanced results can be improved by color correction, well enhanced details and local contrast while promoted global brightness, increasing the visibility, naturalness preservation. Moreover, the proposed algorithm is mostly calculated by per-pixel operation, which is appropriate for real-time applications.
{"title":"A fusion-based enhancing approach for single sandstorm image","authors":"Xueyang Fu, Yue Huang, Delu Zeng, Xiao-Ping Zhang, Xinghao Ding","doi":"10.1109/MMSP.2014.6958791","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958791","url":null,"abstract":"In this paper, a novel image enhancing approach focuses on single sandstorm image is proposed. The degraded image has some problems, such as color distortion, low-visibility, fuzz and non-uniform luminance, due to the light is absorbed and scattered by particles in sandstorm. The proposed approach based on fusion principles aims to overcome the aforementioned limitations. First, the degraded image is color corrected by adopting a statistical strategy. Then two inputs, which represent different brightness, are derived only from the color corrected image by applying Gamma correction. Three weighted maps (sharpness, chromaticity and prominence), which contain important features to increase the quality of the degraded image, are computed from the derived inputs. Finally, the enhanced image is obtained by fusing the inputs with the weight maps. The proposed method is the first to adopt a fusion-based method for enhancing single sandstorm image. Experimental results show that enhanced results can be improved by color correction, well enhanced details and local contrast while promoted global brightness, increasing the visibility, naturalness preservation. Moreover, the proposed algorithm is mostly calculated by per-pixel operation, which is appropriate for real-time applications.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134239002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958825
Sylvia O. N’guessan, N. Ling, Zhouye Gu
The emergence of social networks combined with the prevalence of mobile technology has led to an increasing demand of high definition video transmission and storage. One of the challenges of video compression is the ability to reduce the video size without significant visual quality loss. In this paper, we propose a new method that achieves compression reduction levels ranging from 2.6% to 16.9% while maintaining or improving subjective quality. Precisely, our approach is a saliency-aware mechanism that predicts and classifies regions-of-interests (ROIs) of a typical human eye gaze according to the static attention model (SAM) from the human visual system (HVS). We coin the term contrast human attention regions of interest (Contrast-HAROIs) to refer to those identified regions. Finally, we reduce the data load of those non Contrast-HAROIs via a smoothing spatial filter. Experimental results carried on eight sequences show that our technique reduces the size of HD videos further than the standard H.264/AVC. Moreover, it is in average 30% times faster than another saliency and motion aware algorithm.
{"title":"Compression of HD videos by a contrast-based human attention algorithm","authors":"Sylvia O. N’guessan, N. Ling, Zhouye Gu","doi":"10.1109/MMSP.2014.6958825","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958825","url":null,"abstract":"The emergence of social networks combined with the prevalence of mobile technology has led to an increasing demand of high definition video transmission and storage. One of the challenges of video compression is the ability to reduce the video size without significant visual quality loss. In this paper, we propose a new method that achieves compression reduction levels ranging from 2.6% to 16.9% while maintaining or improving subjective quality. Precisely, our approach is a saliency-aware mechanism that predicts and classifies regions-of-interests (ROIs) of a typical human eye gaze according to the static attention model (SAM) from the human visual system (HVS). We coin the term contrast human attention regions of interest (Contrast-HAROIs) to refer to those identified regions. Finally, we reduce the data load of those non Contrast-HAROIs via a smoothing spatial filter. Experimental results carried on eight sequences show that our technique reduces the size of HD videos further than the standard H.264/AVC. Moreover, it is in average 30% times faster than another saliency and motion aware algorithm.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"92 16","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131436591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958796
Sheng Wei, Viswanathan Swaminathan
The Hypertext Transfer Protocol (HTTP) has been widely adopted and deployed as the key protocol for video streaming over the Internet. One of the consequences of leveraging traditional HTTP for video streaming is the significantly increased request overhead due to the segmentation of the video content into HTTP resources. The overhead becomes even more significant when non-multiplexed video and audio segments are deployed. In this paper, we investigate and address the request overhead problem by employing the server push technology in the new HTTP 2.0 protocol. In particular, we develop a set of push strategies that actively deliver video and audio content from the HTTP server without requiring a request for each individual segment. We evaluate our approach in a Dynamic Adaptive Streaming over HTTP (DASH) streaming system. We show that the request overhead can be significantly reduced by using our push strategies. Also, we validate that the server push based approach is compatible with the existing HTTP streaming features, such as adaptive bitrate switching.
超文本传输协议(Hypertext Transfer Protocol, HTTP)作为Internet上视频流传输的关键协议已被广泛采用和部署。利用传统HTTP进行视频流的后果之一是,由于将视频内容分割为HTTP资源,请求开销显著增加。当部署非多路视频和音频段时,开销变得更加显著。在本文中,我们通过在新的HTTP 2.0协议中使用服务器推送技术来研究和解决请求开销问题。特别是,我们开发了一套推送策略,可以主动地从HTTP服务器发送视频和音频内容,而不需要对每个单独的片段发出请求。我们在基于HTTP的动态自适应流(DASH)流系统中评估了我们的方法。我们展示了通过使用我们的推送策略可以显著降低请求开销。此外,我们验证了基于服务器推送的方法与现有的HTTP流特性兼容,例如自适应比特率交换。
{"title":"Cost effective video streaming using server push over HTTP 2.0","authors":"Sheng Wei, Viswanathan Swaminathan","doi":"10.1109/MMSP.2014.6958796","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958796","url":null,"abstract":"The Hypertext Transfer Protocol (HTTP) has been widely adopted and deployed as the key protocol for video streaming over the Internet. One of the consequences of leveraging traditional HTTP for video streaming is the significantly increased request overhead due to the segmentation of the video content into HTTP resources. The overhead becomes even more significant when non-multiplexed video and audio segments are deployed. In this paper, we investigate and address the request overhead problem by employing the server push technology in the new HTTP 2.0 protocol. In particular, we develop a set of push strategies that actively deliver video and audio content from the HTTP server without requiring a request for each individual segment. We evaluate our approach in a Dynamic Adaptive Streaming over HTTP (DASH) streaming system. We show that the request overhead can be significantly reduced by using our push strategies. Also, we validate that the server push based approach is compatible with the existing HTTP streaming features, such as adaptive bitrate switching.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117037074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2014-11-20DOI: 10.1109/MMSP.2014.6958797
N. Erdogmus, Matthias Vanoni, S. Marcel
With its wide range of applicability, gender classification is an important task in face image analysis and it has drawn a great interest from the pattern recognition community. In this paper, we aim to deal with this problem using Local Binary Pattern Histogram Sequences as feature vectors in general. Differently from what has been done in similar studies, the algorithm parameters used in cropping and feature extraction steps are selected after an extensive grid search using BANCA and MOBIO databases. The final system which is evaluated on FERET, MORPH-II and LFW with gender balanced and imbalanced training sets is shown to achieve commensurate and better results compared to other state-of-the-art performances on those databases. The system is additionally tested for cross-database training in order to assess its accuracy in real world conditions. For LFW and MORPH-II, BeFIT protocols are used.
{"title":"Within- and cross- database evaluations for face gender classification via befit protocols","authors":"N. Erdogmus, Matthias Vanoni, S. Marcel","doi":"10.1109/MMSP.2014.6958797","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958797","url":null,"abstract":"With its wide range of applicability, gender classification is an important task in face image analysis and it has drawn a great interest from the pattern recognition community. In this paper, we aim to deal with this problem using Local Binary Pattern Histogram Sequences as feature vectors in general. Differently from what has been done in similar studies, the algorithm parameters used in cropping and feature extraction steps are selected after an extensive grid search using BANCA and MOBIO databases. The final system which is evaluated on FERET, MORPH-II and LFW with gender balanced and imbalanced training sets is shown to achieve commensurate and better results compared to other state-of-the-art performances on those databases. The system is additionally tested for cross-database training in order to assess its accuracy in real world conditions. For LFW and MORPH-II, BeFIT protocols are used.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127271236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}