首页 > 最新文献

Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video最新文献

英文 中文
Vibra 摆动
Gangqiang Zhou, Run Wu, Miao Hu, Yipeng Zhou, Tom Z. J. Fu, Di Wu
Variable Bitrate (VBR) video encoding can provide much high quality-to-bits ratio compared to the widely adopted Constant Bitrate (CBR) encoding, and thus receives significant attentions by content providers in recent years. However, it is challenging to design efficient adaptive bitrate algorithms for VBR-encoded videos due to the sharply fluctuating chunk size and the resulting bitrate burstiness. In this paper, we propose a neural adaptive streaming framework called Vibra for VBR-encoded videos, which can well accommodate the high fluctuation of video chunk sizes and improve the quality-of-experience (QoE) of end users significantly. Our framework takes the characteristics of VBR-encoded videos into account, and adopts the technique of deep reinforcement learning to train a model for bitrate adaptation. We also conduct extensive trace-driven experiments, and the results show that Vibra outperforms the state-of-the-art ABR algorithms with an improvement of 8.17% -- 29.21% in terms of the average QoE.
{"title":"Vibra","authors":"Gangqiang Zhou, Run Wu, Miao Hu, Yipeng Zhou, Tom Z. J. Fu, Di Wu","doi":"10.1145/3458306.3460993","DOIUrl":"https://doi.org/10.1145/3458306.3460993","url":null,"abstract":"Variable Bitrate (VBR) video encoding can provide much high quality-to-bits ratio compared to the widely adopted Constant Bitrate (CBR) encoding, and thus receives significant attentions by content providers in recent years. However, it is challenging to design efficient adaptive bitrate algorithms for VBR-encoded videos due to the sharply fluctuating chunk size and the resulting bitrate burstiness. In this paper, we propose a neural adaptive streaming framework called Vibra for VBR-encoded videos, which can well accommodate the high fluctuation of video chunk sizes and improve the quality-of-experience (QoE) of end users significantly. Our framework takes the characteristics of VBR-encoded videos into account, and adopts the technique of deep reinforcement learning to train a model for bitrate adaptation. We also conduct extensive trace-driven experiments, and the results show that Vibra outperforms the state-of-the-art ABR algorithms with an improvement of 8.17% -- 29.21% in terms of the average QoE.","PeriodicalId":429348,"journal":{"name":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","volume":"25 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114389823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
CrowdSR CrowdSR
Zhenxiao Luo, Zelong Wang, Jinyu Chen, Miao Hu, Yipeng Zhou, Tom Z. J. Fu, Di Wu
The prevalence of personal devices motivates the rapid development of crowdsourced livecast in recent years. However, there exists huge diversity of upstream bandwidth among amateur broadcasters. Moreover, the highest video quality that can be streamed is limited by the hardware configuration of broadcaster devices (e.g., 540p for low-end mobile devices). The above factors pose significant challenges to the ingestion of high-resolution live video streams, and result in poor quality-of-experience (QoE) for viewers. In this paper, we propose a novel live video ingest approach called CrowdSR for crowdsourced livecast. CrowdSR can transform a low-resolution video stream uploaded by weak devices into a high-resolution video stream via super-resolution, and then deliver the stream to viewers. CrowdSR can exploit crowdsourced high-resolution video patches from similar broadcasters to speedup model training. Different from previous work, our approach does not require any modification at the client side, and thus is more practical and easy to implement. Finally, we implement and evaluate CrowdSR by conducting a series of real-world experiments. The results show that CrowdSR significantly outperforms the baseline approaches by 0.42-1.09 dB in terms of PSNR and 0.006-0.014 in terms of SSIM.
{"title":"CrowdSR","authors":"Zhenxiao Luo, Zelong Wang, Jinyu Chen, Miao Hu, Yipeng Zhou, Tom Z. J. Fu, Di Wu","doi":"10.1145/3458306.3462170","DOIUrl":"https://doi.org/10.1145/3458306.3462170","url":null,"abstract":"The prevalence of personal devices motivates the rapid development of crowdsourced livecast in recent years. However, there exists huge diversity of upstream bandwidth among amateur broadcasters. Moreover, the highest video quality that can be streamed is limited by the hardware configuration of broadcaster devices (e.g., 540p for low-end mobile devices). The above factors pose significant challenges to the ingestion of high-resolution live video streams, and result in poor quality-of-experience (QoE) for viewers. In this paper, we propose a novel live video ingest approach called CrowdSR for crowdsourced livecast. CrowdSR can transform a low-resolution video stream uploaded by weak devices into a high-resolution video stream via super-resolution, and then deliver the stream to viewers. CrowdSR can exploit crowdsourced high-resolution video patches from similar broadcasters to speedup model training. Different from previous work, our approach does not require any modification at the client side, and thus is more practical and easy to implement. Finally, we implement and evaluate CrowdSR by conducting a series of real-world experiments. The results show that CrowdSR significantly outperforms the baseline approaches by 0.42-1.09 dB in terms of PSNR and 0.006-0.014 in terms of SSIM.","PeriodicalId":429348,"journal":{"name":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121749688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
PAAS PAAS
Chenglei Wu, Zhi Wang, Lifeng Sun
Conventional tile-based 360° video streaming methods, including deep reinforcement learning (DRL) based, ignore the interactive nature of 360° video streaming and download tiles following fixed sequential orders, thus failing to respond to the user's head motion changes. We show that these existing solutions suffer from either the prefetch accuracy or the playback stability drop. Furthermore, these methods are constrained to serve only one fixed streaming preference, causing extra training overhead and the lack of generalization on unseen preferences. In this paper, we propose a dual-queue streaming framework, with accuracy and stability purposes respectively, to enable the DRL agent to determine and change the tile download order without incurring overhead. We also design a preference-aware DRL algorithm to incentivize the agent to learn preference-dependent ABR decisions efficiently. Compared with state-of-the-art DRL baselines, our method not only significantly improves the streaming quality, e.g., increasing the average streaming quality by 13.6% on a public dataset, but also demonstrates better performance and generalization under dynamic preferences, e.g., an average quality improvement of 19.9% on unseen preferences.
{"title":"PAAS","authors":"Chenglei Wu, Zhi Wang, Lifeng Sun","doi":"10.1145/3458306.3460995","DOIUrl":"https://doi.org/10.1145/3458306.3460995","url":null,"abstract":"Conventional tile-based 360° video streaming methods, including deep reinforcement learning (DRL) based, ignore the interactive nature of 360° video streaming and download tiles following fixed sequential orders, thus failing to respond to the user's head motion changes. We show that these existing solutions suffer from either the prefetch accuracy or the playback stability drop. Furthermore, these methods are constrained to serve only one fixed streaming preference, causing extra training overhead and the lack of generalization on unseen preferences. In this paper, we propose a dual-queue streaming framework, with accuracy and stability purposes respectively, to enable the DRL agent to determine and change the tile download order without incurring overhead. We also design a preference-aware DRL algorithm to incentivize the agent to learn preference-dependent ABR decisions efficiently. Compared with state-of-the-art DRL baselines, our method not only significantly improves the streaming quality, e.g., increasing the average streaming quality by 13.6% on a public dataset, but also demonstrates better performance and generalization under dynamic preferences, e.g., an average quality improvement of 19.9% on unseen preferences.","PeriodicalId":429348,"journal":{"name":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117206315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
360NorVic
C. Kattadige, Aravindh Raman, Kanchana Thilakarathna, Andra Lutu, Diego Perino
Streaming 360° video demands high bandwidth and low latency, and poses significant challenges to Internet Service Providers (ISPs) and Mobile Network Operators (MNOs). The identification of 360° video traffic can therefore benefits fixed and mobile carriers to optimize their network and provide better Quality of Experience (QoE) to the user. However, end-to-end encryption of network traffic has obstructed identifying those 360° videos from regular videos. As a solution this paper presents 360NorVic, a near-realtime and offline Machine Learning (ML) classification engine to distinguish 360° videos from regular videos when streamed from mobile devices. We collect packet and flow level data for over 800 video traces from YouTube & Facebook accounting for 200 unique videos under varying streaming conditions. Our results show that for near-realtime and offline classification at packet level, average accuracy exceeds 95%, and that for flow level, 360NorVic achieves more than 92% average accuracy. Finally, we pilot our solution in the commercial network of a large MNO showing the feasibility and effectiveness of 360NorVic in production settings.
{"title":"360NorVic","authors":"C. Kattadige, Aravindh Raman, Kanchana Thilakarathna, Andra Lutu, Diego Perino","doi":"10.1145/3458306.3460998","DOIUrl":"https://doi.org/10.1145/3458306.3460998","url":null,"abstract":"Streaming 360° video demands high bandwidth and low latency, and poses significant challenges to Internet Service Providers (ISPs) and Mobile Network Operators (MNOs). The identification of 360° video traffic can therefore benefits fixed and mobile carriers to optimize their network and provide better Quality of Experience (QoE) to the user. However, end-to-end encryption of network traffic has obstructed identifying those 360° videos from regular videos. As a solution this paper presents 360NorVic, a near-realtime and offline Machine Learning (ML) classification engine to distinguish 360° videos from regular videos when streamed from mobile devices. We collect packet and flow level data for over 800 video traces from YouTube & Facebook accounting for 200 unique videos under varying streaming conditions. Our results show that for near-realtime and offline classification at packet level, average accuracy exceeds 95%, and that for flow level, 360NorVic achieves more than 92% average accuracy. Finally, we pilot our solution in the commercial network of a large MNO showing the feasibility and effectiveness of 360NorVic in production settings.","PeriodicalId":429348,"journal":{"name":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","volume":"207 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114408660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Viewport-aware dynamic 360° video segment categorization 视窗感知的动态360°视频片段分类
A. Dharmasiri, C. Kattadige, V. Zhang, Kanchana Thilakarathna
Unlike conventional videos, 360° videos give freedom to users to turn their heads, watch and interact with the content owing to its immersive spherical environment. Although these movements are arbitrary, similarities can be observed between viewport patterns of different users and different videos. Identifying such patterns can assist both content and network providers to enhance the 360° video streaming process, eventually increasing the end-user Quality of Experience (QoE). But a study on how viewport patterns display similarities across different video content, and their potential applications has not yet been done. In this paper, we present a comprehensive analysis of a dataset of 88 360° videos and propose a novel video categorization algorithm that is based on similarities of viewports. First, we propose a novel viewport clustering algorithm that outperforms the existing algorithms in terms of clustering viewports with similar positioning and speed. Next, we develop a novel and unique dynamic video segment categorization algorithm that shows notable improvement in similarity for viewport distributions within the clusters when compared to that of existing static video categorizations.
与传统视频不同,360°视频由于其沉浸式球形环境,用户可以自由地转头观看并与内容互动。尽管这些运动是任意的,但在不同用户和不同视频的视口模式之间可以观察到相似之处。识别这些模式可以帮助内容和网络提供商增强360°视频流过程,最终提高终端用户的体验质量(QoE)。但是关于视口模式如何在不同视频内容中显示相似性及其潜在应用的研究尚未完成。在本文中,我们对88个360°视频数据集进行了全面分析,并提出了一种基于视口相似性的视频分类算法。首先,我们提出了一种新的视口聚类算法,该算法在相似位置和速度的情况下优于现有的视口聚类算法。接下来,我们开发了一种新颖而独特的动态视频片段分类算法,与现有的静态视频分类相比,该算法在聚类内视口分布的相似性方面有显著提高。
{"title":"Viewport-aware dynamic 360° video segment categorization","authors":"A. Dharmasiri, C. Kattadige, V. Zhang, Kanchana Thilakarathna","doi":"10.1145/3458306.3461000","DOIUrl":"https://doi.org/10.1145/3458306.3461000","url":null,"abstract":"Unlike conventional videos, 360° videos give freedom to users to turn their heads, watch and interact with the content owing to its immersive spherical environment. Although these movements are arbitrary, similarities can be observed between viewport patterns of different users and different videos. Identifying such patterns can assist both content and network providers to enhance the 360° video streaming process, eventually increasing the end-user Quality of Experience (QoE). But a study on how viewport patterns display similarities across different video content, and their potential applications has not yet been done. In this paper, we present a comprehensive analysis of a dataset of 88 360° videos and propose a novel video categorization algorithm that is based on similarities of viewports. First, we propose a novel viewport clustering algorithm that outperforms the existing algorithms in terms of clustering viewports with similar positioning and speed. Next, we develop a novel and unique dynamic video segment categorization algorithm that shows notable improvement in similarity for viewport distributions within the clusters when compared to that of existing static video categorizations.","PeriodicalId":429348,"journal":{"name":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127105940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video 第31届美国计算机学会数字音频和视频网络和操作系统支持研讨会论文集
{"title":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","authors":"","doi":"10.1145/3458306","DOIUrl":"https://doi.org/10.1145/3458306","url":null,"abstract":"","PeriodicalId":429348,"journal":{"name":"Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133487782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 31st ACM Workshop on Network and Operating Systems Support for Digital Audio and Video
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1