Efficient Encoding of Interactive Personalized Views Extracted from Immersive Video Content

J. D. Praeter, Pieter Duchi, G. Wallendael, J. Macq, P. Lambert
{"title":"Efficient Encoding of Interactive Personalized Views Extracted from Immersive Video Content","authors":"J. D. Praeter, Pieter Duchi, G. Wallendael, J. Macq, P. Lambert","doi":"10.1145/2983298.2983301","DOIUrl":null,"url":null,"abstract":"Traditional television limits people to a single viewpoint. However, with new technologies such as virtual reality glasses, the way in which people experience video will change. Instead of being limited to a single viewpoint, people will demand a more immersive experience that gives them a sense of being present in a sports stadium, a concert hall, or at other events. To satisfy these users, video such as 360-degree or panoramic video needs to be transported to their homes. Since these videos have an extremely high resolution, sending the entire video requires a high bandwidth capacity and also results in a high decoding complexity at the viewer. The traditional approach to this problem is to split the original video into tiles and only send the required tiles to the viewer. However, this approach still has a large bit rate overhead compared to sending only the required view. Therefore, we propose to send only a personalized view to each user. Since this paper focuses on reducing the computational cost of such a system, we accelerate the encoding of each personalized view based on coding information obtained from a pre-analysis on the entire ultra-high-resolution video. By doing this using the High Efficiency Video Coding Test Model (HM), the complexity of each individual encode of a personalized view is reduced by more than 96.5% compared to a full encode of the view. This acceleration results in a bit rate overhead of at most 19.5%, which is smaller compared to the bit rate overhead of the tile-based method.","PeriodicalId":293626,"journal":{"name":"Proceedings of the 1st International Workshop on Multimedia Alternate Realities","volume":"126 36","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 1st International Workshop on Multimedia Alternate Realities","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2983298.2983301","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Traditional television limits people to a single viewpoint. However, with new technologies such as virtual reality glasses, the way in which people experience video will change. Instead of being limited to a single viewpoint, people will demand a more immersive experience that gives them a sense of being present in a sports stadium, a concert hall, or at other events. To satisfy these users, video such as 360-degree or panoramic video needs to be transported to their homes. Since these videos have an extremely high resolution, sending the entire video requires a high bandwidth capacity and also results in a high decoding complexity at the viewer. The traditional approach to this problem is to split the original video into tiles and only send the required tiles to the viewer. However, this approach still has a large bit rate overhead compared to sending only the required view. Therefore, we propose to send only a personalized view to each user. Since this paper focuses on reducing the computational cost of such a system, we accelerate the encoding of each personalized view based on coding information obtained from a pre-analysis on the entire ultra-high-resolution video. By doing this using the High Efficiency Video Coding Test Model (HM), the complexity of each individual encode of a personalized view is reduced by more than 96.5% compared to a full encode of the view. This acceleration results in a bit rate overhead of at most 19.5%, which is smaller compared to the bit rate overhead of the tile-based method.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
从沉浸式视频内容中提取的交互式个性化视图的高效编码
传统电视把人们限制在一个单一的观点上。然而,随着虚拟现实眼镜等新技术的出现,人们体验视频的方式将发生变化。人们将不再局限于单一的视角,而是需要一种更加身临其境的体验,让他们有置身于体育场、音乐厅或其他活动中的感觉。为了满足这些用户,需要将360度或全景视频等视频传输到他们的家中。由于这些视频具有极高的分辨率,因此发送整个视频需要高带宽容量,并且在观看者处也会导致高解码复杂性。解决这个问题的传统方法是将原始视频分割成小块,只将需要的小块发送给观看者。然而,与只发送所需视图相比,这种方法仍然有很大的比特率开销。因此,我们建议只向每个用户发送一个个性化的视图。由于本文着眼于降低系统的计算成本,我们基于对整个超高分辨率视频进行预分析获得的编码信息来加速每个个性化视图的编码。通过使用高效视频编码测试模型(HM),与视图的完整编码相比,个性化视图的每个单独编码的复杂性降低了96.5%以上。这种加速导致的比特率开销最多为19.5%,这比基于tile的方法的比特率开销要小。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Evaluating the User Experience of Interactive Digital Narrative Efficient Encoding of Interactive Personalized Views Extracted from Immersive Video Content Typeface Emotion Analysis for Communication on Mobile Messengers Session details: Creation and Design Locomotion by Natural Gestures for Immersive Virtual Environments
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1