首页 > 最新文献

2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)最新文献

英文 中文
A pilot study on affective classification of facial images for emerging news topics 新兴新闻主题面部图像情感分类的初步研究
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958799
Ligang Zhang, C. Lau, D. Tjondronegoro, V. Chandran
The proliferation of news reports published in online websites and news information sharing among social media users necessitates effective techniques for analysing the image, text and video data related to news topics. This paper presents the first study to classify affective facial images on emerging news topics. The proposed system dynamically monitors and selects the current hot (of great interest) news topics with strong affective interestingness using textual keywords in news articles and social media discussions. Images from the selected hot topics are extracted and classified into three categorized emotions, positive, neutral and negative, based on facial expressions of subjects in the images. Performance evaluations on two facial image datasets collected from realworld resources demonstrate the applicability and effectiveness of the proposed system in affective classification of facial images in news reports. Facial expression shows high consistency with the affective textual content in news reports for positive emotion, while only low correlation has been observed for neutral and negative. The system can be directly used for applications, such as assisting editors in choosing photos with a proper affective semantic for a certain topic during news report preparation.
在线网站上发布的新闻报道的激增以及社交媒体用户之间的新闻信息共享需要有效的技术来分析与新闻主题相关的图像、文本和视频数据。本文首次对新兴新闻话题中的情感面部图像进行分类研究。该系统利用新闻文章和社交媒体讨论中的文本关键词,动态监测和选择当前热点(极大兴趣)新闻话题,并产生强烈的情感兴趣。从选定的热点话题中提取图像,并根据图像中受试者的面部表情将其分为积极、中性和消极三大类情绪。通过对两组真实资源的人脸图像数据集进行性能评价,证明了该系统在新闻报道中人脸图像情感分类中的适用性和有效性。在新闻报道中,面部表情与情感文本内容在积极情绪方面具有较高的一致性,而在中性情绪和消极情绪方面相关性较低。该系统可以直接用于应用,例如在新闻报道准备过程中,协助编辑选择适合某一主题的情感语义图片。
{"title":"A pilot study on affective classification of facial images for emerging news topics","authors":"Ligang Zhang, C. Lau, D. Tjondronegoro, V. Chandran","doi":"10.1109/MMSP.2014.6958799","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958799","url":null,"abstract":"The proliferation of news reports published in online websites and news information sharing among social media users necessitates effective techniques for analysing the image, text and video data related to news topics. This paper presents the first study to classify affective facial images on emerging news topics. The proposed system dynamically monitors and selects the current hot (of great interest) news topics with strong affective interestingness using textual keywords in news articles and social media discussions. Images from the selected hot topics are extracted and classified into three categorized emotions, positive, neutral and negative, based on facial expressions of subjects in the images. Performance evaluations on two facial image datasets collected from realworld resources demonstrate the applicability and effectiveness of the proposed system in affective classification of facial images in news reports. Facial expression shows high consistency with the affective textual content in news reports for positive emotion, while only low correlation has been observed for neutral and negative. The system can be directly used for applications, such as assisting editors in choosing photos with a proper affective semantic for a certain topic during news report preparation.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126953609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An improved rate control algorithm for SVC with optimised MAD prediction 基于优化MAD预测的SVC改进速率控制算法
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958821
Xin Lu, G. Martin
An improved rate control algorithm for the Scalable Video Coding (SVC) extension of H.264/AVC is described. The rate control scheme applied to the Base Layer (BL) of SVC adopts the linear Mean Absolute Difference (MAD) prediction and quadratic Rate Distortion (RD) models inherited from H.264/AVC. A MAD prediction error always exists and cannot be avoided. However, some encoding results of the base layer can be used to inform the coding of the enhancement layers (ELs), thus benefitting from the bottom-up coding structure of SVC. This property forms the basis for the proposed rate control approach. Simulation results show that accurate rate control is achieved and, compared to the default rate control algorithm of SVC, namely the JVT-G012 rate control scheme, the average PSNR is increased by 0.27dB or the average bit rate is reduced by 4.81%.
针对H.264/AVC的可扩展视频编码(SVC)扩展,提出了一种改进的速率控制算法。应用于SVC基础层(BL)的速率控制方案采用了继承H.264/AVC的线性平均绝对差(MAD)预测和二次率失真(RD)模型。MAD预测误差总是存在并且无法避免。然而,基础层的一些编码结果可以用来通知增强层的编码,从而受益于SVC自下而上的编码结构。这一特性构成了拟议的费率控制办法的基础。仿真结果表明,该方法实现了精确的速率控制,与SVC的默认速率控制算法即JVT-G012速率控制方案相比,平均PSNR提高0.27dB,平均比特率降低4.81%。
{"title":"An improved rate control algorithm for SVC with optimised MAD prediction","authors":"Xin Lu, G. Martin","doi":"10.1109/MMSP.2014.6958821","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958821","url":null,"abstract":"An improved rate control algorithm for the Scalable Video Coding (SVC) extension of H.264/AVC is described. The rate control scheme applied to the Base Layer (BL) of SVC adopts the linear Mean Absolute Difference (MAD) prediction and quadratic Rate Distortion (RD) models inherited from H.264/AVC. A MAD prediction error always exists and cannot be avoided. However, some encoding results of the base layer can be used to inform the coding of the enhancement layers (ELs), thus benefitting from the bottom-up coding structure of SVC. This property forms the basis for the proposed rate control approach. Simulation results show that accurate rate control is achieved and, compared to the default rate control algorithm of SVC, namely the JVT-G012 rate control scheme, the average PSNR is increased by 0.27dB or the average bit rate is reduced by 4.81%.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128437079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Towards efficient wavefront parallel encoding of HEVC: Parallelism analysis and improvement 高效的HEVC波前并行编码:并行性分析与改进
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958818
Keji Chen, Y. Duan, Jun Sun, Zongming Guo
High Efficiency Video Coding (HEVC) is the new generation video coding standard which achieves significant improvement in coding efficiency. Although HEVC is promising in many applications, the increased computational complexity is a serious problem, which makes parallelization necessary in HEVC encoding. To better understand the bottleneck of parallelization and improve the encoding speed, in this paper, we propose a Coding Tree Blocks (CTB) level parallelism analysis method as well as a novel Inter-Frame Wavefront (IFW) parallel encoding method. First, by establishing the relationship between parallelism and dependence, parallelism is precisely described by CTB-level dependence as a criterion to evaluate different parallel methods of HEVC. On this basis, by effectively decreasing the dependence based on Wavefront Parallel Processing (WPP), IFW method is developed. Finally, with the proposed parallelism analysis method, IFW is theoretically proved to be of higher parallelism compared with other HEVC representative parallel methods. Extensive experimental results show that, the proposed method and implementation can bring up to 17.81x, 14.34x and 24.40x speedup for HEVC encoding of WVGA, 720p and 1080p standard test sequences with the same ignorable coding performance degradation as WPP, thus showing a promising technology for future large-scale HEVC video application.
高效视频编码(High Efficiency Video Coding, HEVC)是新一代视频编码标准,大大提高了编码效率。尽管HEVC在许多应用中都很有前景,但计算复杂度的增加是一个严重的问题,这使得并行化成为HEVC编码的必要条件。为了更好地理解并行化的瓶颈并提高编码速度,本文提出了一种编码树块(CTB)级并行性分析方法和一种新的帧间波前(IFW)并行编码方法。首先,通过建立并行度与依赖度之间的关系,将并行度精确地描述为ctb级依赖度,作为评价HEVC不同并行方法的标准;在此基础上,通过有效降低波前并行处理(WPP)的依赖性,发展了IFW方法。最后,通过提出的并行性分析方法,从理论上证明了IFW与其他HEVC代表性并行方法相比具有更高的并行性。大量的实验结果表明,所提出的方法和实现对WVGA、720p和1080p标准测试序列的HEVC编码速度提高了17.81倍、14.34倍和24.40倍,编码性能的下降与WPP相同,可以忽略,为未来大规模HEVC视频应用展示了一种很有前景的技术。
{"title":"Towards efficient wavefront parallel encoding of HEVC: Parallelism analysis and improvement","authors":"Keji Chen, Y. Duan, Jun Sun, Zongming Guo","doi":"10.1109/MMSP.2014.6958818","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958818","url":null,"abstract":"High Efficiency Video Coding (HEVC) is the new generation video coding standard which achieves significant improvement in coding efficiency. Although HEVC is promising in many applications, the increased computational complexity is a serious problem, which makes parallelization necessary in HEVC encoding. To better understand the bottleneck of parallelization and improve the encoding speed, in this paper, we propose a Coding Tree Blocks (CTB) level parallelism analysis method as well as a novel Inter-Frame Wavefront (IFW) parallel encoding method. First, by establishing the relationship between parallelism and dependence, parallelism is precisely described by CTB-level dependence as a criterion to evaluate different parallel methods of HEVC. On this basis, by effectively decreasing the dependence based on Wavefront Parallel Processing (WPP), IFW method is developed. Finally, with the proposed parallelism analysis method, IFW is theoretically proved to be of higher parallelism compared with other HEVC representative parallel methods. Extensive experimental results show that, the proposed method and implementation can bring up to 17.81x, 14.34x and 24.40x speedup for HEVC encoding of WVGA, 720p and 1080p standard test sequences with the same ignorable coding performance degradation as WPP, thus showing a promising technology for future large-scale HEVC video application.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129632660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Classifying harmful children's content using affective analysis 运用情感分析法对儿童有害内容进行分类
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958813
Joseph Santarcangelo, Xiao-Ping Zhang
This paper categorizes children's videos according to an expertly assigned predefined positive or negative cognitive impact category. The method uses affective features to determine if a video belongs to an expertly assigned predefined positive or to a negative cognitive impact category. The work demonstrates that simple affective features outperform more complex systems in determining if content belongs to the positive or negative cognitive impact category. The work is tested on a set of videos that have been classified as having a short term or long term measurable negative or positive impact on cognition based on cited psychological literature. It found that affective analysis had superior performance using less features than state of the art video genre classification systems. It also found that arousal features performed better than valence features.
本文根据专家指定的预定义的积极或消极认知影响类别对儿童视频进行分类。该方法使用情感特征来确定视频是否属于专家指定的预定义的积极或消极的认知影响类别。研究表明,简单的情感特征在确定内容是否属于积极或消极的认知影响类别方面优于更复杂的系统。这项工作是在一组视频上进行测试的,这些视频根据引用的心理学文献被分类为对认知有短期或长期可测量的消极或积极影响。它发现情感分析使用更少的特征比最先进的视频类型分类系统有更好的表现。研究还发现,唤起特征比效价特征表现得更好。
{"title":"Classifying harmful children's content using affective analysis","authors":"Joseph Santarcangelo, Xiao-Ping Zhang","doi":"10.1109/MMSP.2014.6958813","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958813","url":null,"abstract":"This paper categorizes children's videos according to an expertly assigned predefined positive or negative cognitive impact category. The method uses affective features to determine if a video belongs to an expertly assigned predefined positive or to a negative cognitive impact category. The work demonstrates that simple affective features outperform more complex systems in determining if content belongs to the positive or negative cognitive impact category. The work is tested on a set of videos that have been classified as having a short term or long term measurable negative or positive impact on cognition based on cited psychological literature. It found that affective analysis had superior performance using less features than state of the art video genre classification systems. It also found that arousal features performed better than valence features.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129226325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fast mode decision for error resilient video coding 抗错误性视频编码的快速模式决策
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958824
Yunong Wei, Yuan Zhang, Jinyao Yan
The error resilience and low-complexity video encoding are two major requirements of real-time visual communications on mobile devices. To address the two requirements simultaneously, this paper presents a fast mode decision algorithm for the error resilient video coding in packet loss environment. The proposed algorithm is a two-step method: early skip mode decision and early intra mode decision. Different from the existing methods for early skip mode decision, the proposed method takes the error-propagation distortion into account in estimating the coding cost. Considering the intra blocks are frequently used to terminate the error propagations, we also propose a method to fast estimate the intra block coding cost, so that the intra mode can be early determined. Overall, the proposed method can significantly reduce the encoding time while keeping the coding efficiency similar to the rate-distortion optimized mode decision method.
抗干扰性和低复杂度视频编码是移动设备实时可视通信的两大要求。为了同时满足这两方面的要求,本文提出了一种丢包环境下的抗错误性视频编码快速模式判定算法。该算法采用两步法:早期跳模决策和早期模内决策。与现有的早期跳过模式决策方法不同,该方法在估计编码成本时考虑了错误传播失真。考虑到码内块经常被用来终止错误传播,我们还提出了一种快速估计码内块编码代价的方法,以便尽早确定码内模式。总体而言,该方法在保持编码效率与率失真优化模式决策方法相似的同时,显著缩短了编码时间。
{"title":"Fast mode decision for error resilient video coding","authors":"Yunong Wei, Yuan Zhang, Jinyao Yan","doi":"10.1109/MMSP.2014.6958824","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958824","url":null,"abstract":"The error resilience and low-complexity video encoding are two major requirements of real-time visual communications on mobile devices. To address the two requirements simultaneously, this paper presents a fast mode decision algorithm for the error resilient video coding in packet loss environment. The proposed algorithm is a two-step method: early skip mode decision and early intra mode decision. Different from the existing methods for early skip mode decision, the proposed method takes the error-propagation distortion into account in estimating the coding cost. Considering the intra blocks are frequently used to terminate the error propagations, we also propose a method to fast estimate the intra block coding cost, so that the intra mode can be early determined. Overall, the proposed method can significantly reduce the encoding time while keeping the coding efficiency similar to the rate-distortion optimized mode decision method.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"16 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116341957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Comparing subjective and objective quality assessment of HDR images compressed with JPEG-XT 比较JPEG-XT压缩HDR图像的主客观质量评价
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958833
Claire Mantel, Stefan Catalin Ferchiu, Søren Forchhammer
In this paper a subjective test in which participants evaluate the quality of JPEG-XT compressed HDR images is presented. Results show that for the selected test images and display, the subjective quality reached its saturation point starting around 3bpp. Objective evaluations are obtained by applying a model of the display and providing the modeled images to three objective metrics dedicated to HDR content. Objective grades are compared with subjective data both in physical domain and using a gamma correction to approximate perceptually uniform luminance coding. The MRSE metric obtains the best performance with the limit that it does not capture the quality saturation. The usage of the gamma correction prior to applying metrics depends on the characteristics of each objective metric.
在本文中,提出了一个主观测试,参与者评估JPEG-XT压缩HDR图像的质量。结果表明,对于所选的测试图像和显示,主观质量在3bpp左右达到饱和点。通过应用显示模型并将建模图像提供给专用于HDR内容的三个客观度量来获得客观评估。客观等级与主观数据在物理领域进行比较,并使用伽玛校正来近似感知均匀的亮度编码。MRSE度量在不捕获质量饱和的限制下获得最佳性能。在应用度量之前使用伽马校正取决于每个客观度量的特性。
{"title":"Comparing subjective and objective quality assessment of HDR images compressed with JPEG-XT","authors":"Claire Mantel, Stefan Catalin Ferchiu, Søren Forchhammer","doi":"10.1109/MMSP.2014.6958833","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958833","url":null,"abstract":"In this paper a subjective test in which participants evaluate the quality of JPEG-XT compressed HDR images is presented. Results show that for the selected test images and display, the subjective quality reached its saturation point starting around 3bpp. Objective evaluations are obtained by applying a model of the display and providing the modeled images to three objective metrics dedicated to HDR content. Objective grades are compared with subjective data both in physical domain and using a gamma correction to approximate perceptually uniform luminance coding. The MRSE metric obtains the best performance with the limit that it does not capture the quality saturation. The usage of the gamma correction prior to applying metrics depends on the characteristics of each objective metric.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128060172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
A fusion-based enhancing approach for single sandstorm image 一种基于融合的沙尘暴单幅图像增强方法
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958791
Xueyang Fu, Yue Huang, Delu Zeng, Xiao-Ping Zhang, Xinghao Ding
In this paper, a novel image enhancing approach focuses on single sandstorm image is proposed. The degraded image has some problems, such as color distortion, low-visibility, fuzz and non-uniform luminance, due to the light is absorbed and scattered by particles in sandstorm. The proposed approach based on fusion principles aims to overcome the aforementioned limitations. First, the degraded image is color corrected by adopting a statistical strategy. Then two inputs, which represent different brightness, are derived only from the color corrected image by applying Gamma correction. Three weighted maps (sharpness, chromaticity and prominence), which contain important features to increase the quality of the degraded image, are computed from the derived inputs. Finally, the enhanced image is obtained by fusing the inputs with the weight maps. The proposed method is the first to adopt a fusion-based method for enhancing single sandstorm image. Experimental results show that enhanced results can be improved by color correction, well enhanced details and local contrast while promoted global brightness, increasing the visibility, naturalness preservation. Moreover, the proposed algorithm is mostly calculated by per-pixel operation, which is appropriate for real-time applications.
本文提出了一种针对单个沙尘暴图像的图像增强方法。由于沙尘暴中粒子对光线的吸收和散射,图像存在颜色失真、能见度低、模糊和亮度不均匀等问题。所提出的基于融合原理的方法旨在克服上述局限性。首先,采用统计策略对退化图像进行色彩校正;然后,通过伽玛校正,仅从颜色校正后的图像中获得代表不同亮度的两个输入。三个加权地图(清晰度,色度和突出),其中包含重要的特征,以提高退化图像的质量,计算从派生的输入。最后,将输入与权值映射融合得到增强图像。该方法首次采用基于融合的方法对单个沙尘暴图像进行增强。实验结果表明,通过色彩校正,增强后的图像细节和局部对比度得到了很好的增强,同时提高了图像的全局亮度,提高了图像的可见度和自然度。此外,该算法主要采用逐像素运算,适合于实时应用。
{"title":"A fusion-based enhancing approach for single sandstorm image","authors":"Xueyang Fu, Yue Huang, Delu Zeng, Xiao-Ping Zhang, Xinghao Ding","doi":"10.1109/MMSP.2014.6958791","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958791","url":null,"abstract":"In this paper, a novel image enhancing approach focuses on single sandstorm image is proposed. The degraded image has some problems, such as color distortion, low-visibility, fuzz and non-uniform luminance, due to the light is absorbed and scattered by particles in sandstorm. The proposed approach based on fusion principles aims to overcome the aforementioned limitations. First, the degraded image is color corrected by adopting a statistical strategy. Then two inputs, which represent different brightness, are derived only from the color corrected image by applying Gamma correction. Three weighted maps (sharpness, chromaticity and prominence), which contain important features to increase the quality of the degraded image, are computed from the derived inputs. Finally, the enhanced image is obtained by fusing the inputs with the weight maps. The proposed method is the first to adopt a fusion-based method for enhancing single sandstorm image. Experimental results show that enhanced results can be improved by color correction, well enhanced details and local contrast while promoted global brightness, increasing the visibility, naturalness preservation. Moreover, the proposed algorithm is mostly calculated by per-pixel operation, which is appropriate for real-time applications.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134239002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
Compression of HD videos by a contrast-based human attention algorithm 基于对比度的人类注意力算法压缩高清视频
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958825
Sylvia O. N’guessan, N. Ling, Zhouye Gu
The emergence of social networks combined with the prevalence of mobile technology has led to an increasing demand of high definition video transmission and storage. One of the challenges of video compression is the ability to reduce the video size without significant visual quality loss. In this paper, we propose a new method that achieves compression reduction levels ranging from 2.6% to 16.9% while maintaining or improving subjective quality. Precisely, our approach is a saliency-aware mechanism that predicts and classifies regions-of-interests (ROIs) of a typical human eye gaze according to the static attention model (SAM) from the human visual system (HVS). We coin the term contrast human attention regions of interest (Contrast-HAROIs) to refer to those identified regions. Finally, we reduce the data load of those non Contrast-HAROIs via a smoothing spatial filter. Experimental results carried on eight sequences show that our technique reduces the size of HD videos further than the standard H.264/AVC. Moreover, it is in average 30% times faster than another saliency and motion aware algorithm.
随着社交网络的出现和移动技术的普及,对高清视频传输和存储的需求不断增加。视频压缩面临的挑战之一是如何在不造成明显视觉质量损失的情况下减小视频大小。在本文中,我们提出了一种新的方法,在保持或提高主观质量的同时,实现了2.6%至16.9%的压缩降低水平。准确地说,我们的方法是一种显著性感知机制,根据人类视觉系统(HVS)的静态注意模型(SAM)预测和分类典型人眼注视的兴趣区域(roi)。我们创造了“对比人类感兴趣的注意力区域”(contrast - harois)一词来指代那些已识别的区域。最后,我们通过平滑空间滤波来减少非对比haroi的数据负载。在8个序列上进行的实验结果表明,与标准的H.264/AVC相比,我们的技术进一步减小了高清视频的大小。此外,它比另一种显著性和运动感知算法平均快30%。
{"title":"Compression of HD videos by a contrast-based human attention algorithm","authors":"Sylvia O. N’guessan, N. Ling, Zhouye Gu","doi":"10.1109/MMSP.2014.6958825","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958825","url":null,"abstract":"The emergence of social networks combined with the prevalence of mobile technology has led to an increasing demand of high definition video transmission and storage. One of the challenges of video compression is the ability to reduce the video size without significant visual quality loss. In this paper, we propose a new method that achieves compression reduction levels ranging from 2.6% to 16.9% while maintaining or improving subjective quality. Precisely, our approach is a saliency-aware mechanism that predicts and classifies regions-of-interests (ROIs) of a typical human eye gaze according to the static attention model (SAM) from the human visual system (HVS). We coin the term contrast human attention regions of interest (Contrast-HAROIs) to refer to those identified regions. Finally, we reduce the data load of those non Contrast-HAROIs via a smoothing spatial filter. Experimental results carried on eight sequences show that our technique reduces the size of HD videos further than the standard H.264/AVC. Moreover, it is in average 30% times faster than another saliency and motion aware algorithm.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"92 16","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131436591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Cost effective video streaming using server push over HTTP 2.0 通过HTTP 2.0使用服务器推送的高效视频流
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958796
Sheng Wei, Viswanathan Swaminathan
The Hypertext Transfer Protocol (HTTP) has been widely adopted and deployed as the key protocol for video streaming over the Internet. One of the consequences of leveraging traditional HTTP for video streaming is the significantly increased request overhead due to the segmentation of the video content into HTTP resources. The overhead becomes even more significant when non-multiplexed video and audio segments are deployed. In this paper, we investigate and address the request overhead problem by employing the server push technology in the new HTTP 2.0 protocol. In particular, we develop a set of push strategies that actively deliver video and audio content from the HTTP server without requiring a request for each individual segment. We evaluate our approach in a Dynamic Adaptive Streaming over HTTP (DASH) streaming system. We show that the request overhead can be significantly reduced by using our push strategies. Also, we validate that the server push based approach is compatible with the existing HTTP streaming features, such as adaptive bitrate switching.
超文本传输协议(Hypertext Transfer Protocol, HTTP)作为Internet上视频流传输的关键协议已被广泛采用和部署。利用传统HTTP进行视频流的后果之一是,由于将视频内容分割为HTTP资源,请求开销显著增加。当部署非多路视频和音频段时,开销变得更加显著。在本文中,我们通过在新的HTTP 2.0协议中使用服务器推送技术来研究和解决请求开销问题。特别是,我们开发了一套推送策略,可以主动地从HTTP服务器发送视频和音频内容,而不需要对每个单独的片段发出请求。我们在基于HTTP的动态自适应流(DASH)流系统中评估了我们的方法。我们展示了通过使用我们的推送策略可以显著降低请求开销。此外,我们验证了基于服务器推送的方法与现有的HTTP流特性兼容,例如自适应比特率交换。
{"title":"Cost effective video streaming using server push over HTTP 2.0","authors":"Sheng Wei, Viswanathan Swaminathan","doi":"10.1109/MMSP.2014.6958796","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958796","url":null,"abstract":"The Hypertext Transfer Protocol (HTTP) has been widely adopted and deployed as the key protocol for video streaming over the Internet. One of the consequences of leveraging traditional HTTP for video streaming is the significantly increased request overhead due to the segmentation of the video content into HTTP resources. The overhead becomes even more significant when non-multiplexed video and audio segments are deployed. In this paper, we investigate and address the request overhead problem by employing the server push technology in the new HTTP 2.0 protocol. In particular, we develop a set of push strategies that actively deliver video and audio content from the HTTP server without requiring a request for each individual segment. We evaluate our approach in a Dynamic Adaptive Streaming over HTTP (DASH) streaming system. We show that the request overhead can be significantly reduced by using our push strategies. Also, we validate that the server push based approach is compatible with the existing HTTP streaming features, such as adaptive bitrate switching.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117037074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Within- and cross- database evaluations for face gender classification via befit protocols 通过匹配协议进行面部性别分类的数据库内和数据库间评估
Pub Date : 2014-11-20 DOI: 10.1109/MMSP.2014.6958797
N. Erdogmus, Matthias Vanoni, S. Marcel
With its wide range of applicability, gender classification is an important task in face image analysis and it has drawn a great interest from the pattern recognition community. In this paper, we aim to deal with this problem using Local Binary Pattern Histogram Sequences as feature vectors in general. Differently from what has been done in similar studies, the algorithm parameters used in cropping and feature extraction steps are selected after an extensive grid search using BANCA and MOBIO databases. The final system which is evaluated on FERET, MORPH-II and LFW with gender balanced and imbalanced training sets is shown to achieve commensurate and better results compared to other state-of-the-art performances on those databases. The system is additionally tested for cross-database training in order to assess its accuracy in real world conditions. For LFW and MORPH-II, BeFIT protocols are used.
性别分类是人脸图像分析中的一项重要任务,具有广泛的适用性,引起了模式识别界的极大兴趣。在本文中,我们的目标是一般使用局部二值模式直方图序列作为特征向量来处理这个问题。与同类研究不同的是,裁剪和特征提取步骤中使用的算法参数是在使用BANCA和MOBIO数据库进行广泛的网格搜索后选择的。最终的系统在FERET、morphi - ii和LFW上进行了性别平衡和不平衡训练集的评估,结果表明,与这些数据库上的其他最先进的性能相比,该系统取得了相称的更好的结果。该系统还进行了跨数据库训练测试,以评估其在现实世界条件下的准确性。对于LFW和morphi - ii,使用BeFIT协议。
{"title":"Within- and cross- database evaluations for face gender classification via befit protocols","authors":"N. Erdogmus, Matthias Vanoni, S. Marcel","doi":"10.1109/MMSP.2014.6958797","DOIUrl":"https://doi.org/10.1109/MMSP.2014.6958797","url":null,"abstract":"With its wide range of applicability, gender classification is an important task in face image analysis and it has drawn a great interest from the pattern recognition community. In this paper, we aim to deal with this problem using Local Binary Pattern Histogram Sequences as feature vectors in general. Differently from what has been done in similar studies, the algorithm parameters used in cropping and feature extraction steps are selected after an extensive grid search using BANCA and MOBIO databases. The final system which is evaluated on FERET, MORPH-II and LFW with gender balanced and imbalanced training sets is shown to achieve commensurate and better results compared to other state-of-the-art performances on those databases. The system is additionally tested for cross-database training in order to assess its accuracy in real world conditions. For LFW and MORPH-II, BeFIT protocols are used.","PeriodicalId":164858,"journal":{"name":"2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127271236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2014 IEEE 16th International Workshop on Multimedia Signal Processing (MMSP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1