Pub Date : 2024-06-28DOI: 10.1109/TBC.2024.3407482
Haojiang Li;Wenjun Zhang;Yin Xu;Dazhi He;Haoyang Li
With the arrival of the 6G era, wireless communication networks will face increased pressure due to diversified service traffic with ultra-large bandwidth, ultra-low latency, and massive connections, making it difficult to guarantee quality of service. However, broadcasting can realize wide-area coverage with lower physical transmission resource occupancy. Therefore, the convergence of broadcasting and 6G networks can promote the evolution and upgrade of traditional broadcasting services towards flexibility, dynamics, and personalization, and at the same time, can effectively alleviate the data congestion in mobile communication networks. In this paper, we firstly introduce the three typical application scenarios of broadcasting and 6G convergence in the future, and summarize the vital technologies and challenges in constructing the converged network. On this basis, we propose a broadcasting and 6G converged network architecture and a next-generation 6G broadcasting core network architecture, and finally introduce the typical collaboration modes of the converged network.
{"title":"Broadcasting and 6G Converged Network Architecture","authors":"Haojiang Li;Wenjun Zhang;Yin Xu;Dazhi He;Haoyang Li","doi":"10.1109/TBC.2024.3407482","DOIUrl":"10.1109/TBC.2024.3407482","url":null,"abstract":"With the arrival of the 6G era, wireless communication networks will face increased pressure due to diversified service traffic with ultra-large bandwidth, ultra-low latency, and massive connections, making it difficult to guarantee quality of service. However, broadcasting can realize wide-area coverage with lower physical transmission resource occupancy. Therefore, the convergence of broadcasting and 6G networks can promote the evolution and upgrade of traditional broadcasting services towards flexibility, dynamics, and personalization, and at the same time, can effectively alleviate the data congestion in mobile communication networks. In this paper, we firstly introduce the three typical application scenarios of broadcasting and 6G convergence in the future, and summarize the vital technologies and challenges in constructing the converged network. On this basis, we propose a broadcasting and 6G converged network architecture and a next-generation 6G broadcasting core network architecture, and finally introduce the typical collaboration modes of the converged network.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 3","pages":"971-979"},"PeriodicalIF":3.2,"publicationDate":"2024-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141510395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-27DOI: 10.1109/TBC.2024.3407596
Yang Liu;Jie Wang;Ruohan Cao;Yueming Lu;Yaojun Qiao;Yuanqing Xia;Daoqi Han
This paper presents a comprehensive investigation into the crucial aspect of security within 5G broadcasting environments, with a particular focus on content production centers. It dives into the unique challenges and vulnerabilities associated with 5G technology, specifically within the context of broadcasting media. The study provides an up-to-date survey of the current landscape in 5G network security, emphasizing the specific requirements and risks specific to broadcasting. In response to these challenges, we propose a set of robust security strategies and technologies specifically tailored for these environments. Through rigorous simulations and compelling case studies, we demonstrate the efficacy of these strategies within a 5G broadcasting context. Ultimately, this paper aims to offer invaluable insights for broadcasters, policymakers, and technologists, enabling them to enhance the security and integrity of 5G broadcasting networks through informed decision-making and implementation of best practices.
{"title":"Securing Content Production Centers in 5G Broadcasting: Strategies and Technologies for Mitigating Cybersecurity Risks","authors":"Yang Liu;Jie Wang;Ruohan Cao;Yueming Lu;Yaojun Qiao;Yuanqing Xia;Daoqi Han","doi":"10.1109/TBC.2024.3407596","DOIUrl":"10.1109/TBC.2024.3407596","url":null,"abstract":"This paper presents a comprehensive investigation into the crucial aspect of security within 5G broadcasting environments, with a particular focus on content production centers. It dives into the unique challenges and vulnerabilities associated with 5G technology, specifically within the context of broadcasting media. The study provides an up-to-date survey of the current landscape in 5G network security, emphasizing the specific requirements and risks specific to broadcasting. In response to these challenges, we propose a set of robust security strategies and technologies specifically tailored for these environments. Through rigorous simulations and compelling case studies, we demonstrate the efficacy of these strategies within a 5G broadcasting context. Ultimately, this paper aims to offer invaluable insights for broadcasters, policymakers, and technologists, enabling them to enhance the security and integrity of 5G broadcasting networks through informed decision-making and implementation of best practices.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 3","pages":"1008-1017"},"PeriodicalIF":3.2,"publicationDate":"2024-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141528459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-25DOI: 10.1109/TBC.2024.3408643
Jingbo He;Xiaohai He;Shuhua Xiong;Honggang Chen
Single image super-resolution (SISR) is a task of reconstructing high-resolution (HR) images from low-resolution (LR) images, which are obtained by some degradation process. Deep neural networks (DNNs) have greatly advanced the frontier of image super-resolution research and replaced traditional methods as the de facto standard approach. The attention mechanism enables the SR algorithms to achieve breakthrough performance after another. However, limited research has been conducted on the interaction and integration of attention mechanisms across different dimensions. To tackle this issue, in this paper, we propose a cross-dimensional attention fusion network (CAFN) to effectively achieve cross-dimensional inter-action with long-range dependencies. Specifically, the proposed approach involves the utilization of a cross-dimensional aggrega-tion module (CAM) to effectively capture contextual information by integrating both spatial and channel importance maps. The design of information fusion module (IFM) in CAM serves as a bridge for parallel dual-attention information fusion. In addition, a novel memory-adaptive multi-stage (MAMS) training method is proposed. We perform warm-start retraining with the same setting as the previous stage, without increasing memory consumption. If the memory is sufficient, we finetune the model with a larger patch size after the warm-start. The experimental results definitively demonstrate the superior performance of our cross-dimensional attention fusion network and training strategy compared to state-of-the-art (SOTA) methods, as evidenced by both quantitative and qualitative metrics.
单幅图像超分辨率(SISR)是一项从低分辨率(LR)图像重建高分辨率(HR)图像的任务,而低分辨率(LR)图像是通过一定的降解过程获得的。深度神经网络(DNN)极大地推动了图像超分辨率研究的前沿发展,并取代传统方法成为事实上的标准方法。注意力机制使 SR 算法取得了一个又一个突破性的性能。然而,关于注意力机制在不同维度上的交互与融合的研究还很有限。为解决这一问题,我们在本文中提出了一种跨维注意力融合网络(CAFN),以有效实现具有长程依赖性的跨维交互作用。具体来说,所提出的方法包括利用跨维聚合模块(CAM),通过整合空间和通道重要性图来有效捕捉上下文信息。CAM 中信息融合模块(IFM)的设计可作为并行双注意信息融合的桥梁。此外,我们还提出了一种新颖的记忆自适应多阶段(MAMS)训练方法。我们在不增加内存消耗的情况下,以与前一阶段相同的设置执行热启动再训练。如果内存充足,我们会在热启动后使用更大的补丁尺寸对模型进行微调。实验结果从定量和定性指标两方面明确证明,与最先进的(SOTA)方法相比,我们的跨维注意力融合网络和训练策略具有更优越的性能。
{"title":"Cross-Dimensional Attention Fusion Network for Simulated Single Image Super-Resolution","authors":"Jingbo He;Xiaohai He;Shuhua Xiong;Honggang Chen","doi":"10.1109/TBC.2024.3408643","DOIUrl":"10.1109/TBC.2024.3408643","url":null,"abstract":"Single image super-resolution (SISR) is a task of reconstructing high-resolution (HR) images from low-resolution (LR) images, which are obtained by some degradation process. Deep neural networks (DNNs) have greatly advanced the frontier of image super-resolution research and replaced traditional methods as the de facto standard approach. The attention mechanism enables the SR algorithms to achieve breakthrough performance after another. However, limited research has been conducted on the interaction and integration of attention mechanisms across different dimensions. To tackle this issue, in this paper, we propose a cross-dimensional attention fusion network (CAFN) to effectively achieve cross-dimensional inter-action with long-range dependencies. Specifically, the proposed approach involves the utilization of a cross-dimensional aggrega-tion module (CAM) to effectively capture contextual information by integrating both spatial and channel importance maps. The design of information fusion module (IFM) in CAM serves as a bridge for parallel dual-attention information fusion. In addition, a novel memory-adaptive multi-stage (MAMS) training method is proposed. We perform warm-start retraining with the same setting as the previous stage, without increasing memory consumption. If the memory is sufficient, we finetune the model with a larger patch size after the warm-start. The experimental results definitively demonstrate the superior performance of our cross-dimensional attention fusion network and training strategy compared to state-of-the-art (SOTA) methods, as evidenced by both quantitative and qualitative metrics.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 3","pages":"909-923"},"PeriodicalIF":3.2,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141510396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-19DOI: 10.1109/TBC.2024.3399479
Axel De Decker;Jan De Cock;Peter Lambert;Glenn Van Wallendael
As the demand for high-quality video content continues to rise, accurately assessing the visual quality of digital videos has become more crucial than ever before. However, evaluating the perceptual quality of an impaired video in the absence of the original reference signal remains a significant challenge. To address this problem, we propose a novel No-Reference (NR) video quality metric called NR-VMAF. Our method is designed to replicate the popular Full-Reference (FR) metric VMAF in scenarios where the reference signal is unavailable or impractical to obtain. Like its FR counterpart, NR-VMAF is tailored specifically for measuring video quality in the presence of compression and scaling artifacts. The proposed model utilizes a deep convolutional neural network to extract quality-aware features from the pixel information of the distorted video, thereby eliminating the need for manual feature engineering. By adopting a patch-based approach, we are able to process high-resolution video data without any information loss. While the current model is trained solely on H.265/HEVC videos, its performance is verified on subjective datasets containing mainly H.264/AVC content. We demonstrate that NR-VMAF outperforms current state-of-the-art NR metrics while achieving a prediction accuracy that is comparable to VMAF and other FR metrics. Based on this strong performance, we believe that NR-VMAF is a viable approach to efficient and reliable No-Reference video quality assessment.
{"title":"No-Reference VMAF: A Deep Neural Network-Based Approach to Blind Video Quality Assessment","authors":"Axel De Decker;Jan De Cock;Peter Lambert;Glenn Van Wallendael","doi":"10.1109/TBC.2024.3399479","DOIUrl":"10.1109/TBC.2024.3399479","url":null,"abstract":"As the demand for high-quality video content continues to rise, accurately assessing the visual quality of digital videos has become more crucial than ever before. However, evaluating the perceptual quality of an impaired video in the absence of the original reference signal remains a significant challenge. To address this problem, we propose a novel No-Reference (NR) video quality metric called NR-VMAF. Our method is designed to replicate the popular Full-Reference (FR) metric VMAF in scenarios where the reference signal is unavailable or impractical to obtain. Like its FR counterpart, NR-VMAF is tailored specifically for measuring video quality in the presence of compression and scaling artifacts. The proposed model utilizes a deep convolutional neural network to extract quality-aware features from the pixel information of the distorted video, thereby eliminating the need for manual feature engineering. By adopting a patch-based approach, we are able to process high-resolution video data without any information loss. While the current model is trained solely on H.265/HEVC videos, its performance is verified on subjective datasets containing mainly H.264/AVC content. We demonstrate that NR-VMAF outperforms current state-of-the-art NR metrics while achieving a prediction accuracy that is comparable to VMAF and other FR metrics. Based on this strong performance, we believe that NR-VMAF is a viable approach to efficient and reliable No-Reference video quality assessment.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 3","pages":"844-861"},"PeriodicalIF":3.2,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141940837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-13DOI: 10.1109/TBC.2024.3407517
Lu Chen;Mao Ye;Luping Ji;Shuai Li;Hongwei Guo
To save transmission bandwidth, there exists an approach to down-sample a video and then up-sample the compressed video to save bit rates. The existing Super Resolution (SR) methods generally design powerful networks to compensate the loss information introduced by down-sampling. But the information of entire video is not fully utilized and effectively fused, resulting in the learned context information that is not enough for the high quality reconstruction. We propose a multi high-quality frames Referenced Cross-scale compressed Video Super Resolution method (RCVSR) that wisely uses past-and-future information, to pursue higher compression efficiency. Specifically, a joint reference motion alignment module is proposed. Low resolution (LR) frame after up-sampling is separately aligned with past-and-future reference frames to preserve more spatial details; at the same time this LR frame is aligned with neighborhood frames to get continuous motion information and similar contents. Then, a reference based refinement module is applied to compensate motion and lost texture details by computing similarity matrix across channel dimensions. Finally, an attention guided dual-branch residual module is employed to enhance the reconstructed result concurrently. Compared with the HEVC anchor, the average gain of Bjontegaard Delta Rate (BD-Rate) under the Low-Delay-P (LDP) setting is 24.86%. In addition, an experimental comparison is made with the advanced SR methods and compressed video quality enhancement (VQE) methods, and the superior efficiency and generalization of the proposed algorithm are further reported.
为了节省传输带宽,有一种方法是对视频进行下采样,然后对压缩后的视频进行上采样,以节省比特率。现有的超分辨率(SR)方法一般会设计功能强大的网络来弥补下采样带来的信息损失。但是,整个视频的信息并没有得到充分利用和有效融合,导致学习到的上下文信息不足以进行高质量的重建。我们提出了一种多高质量帧参考跨尺度压缩视频超分辨率方法(RCVSR),该方法合理利用过去和未来信息,追求更高的压缩效率。具体来说,我们提出了一个联合参考运动对齐模块。将上采样后的低分辨率(LR)帧分别与过去和未来的参考帧对齐,以保留更多空间细节;同时将 LR 帧与邻近帧对齐,以获得连续的运动信息和相似的内容。然后,应用基于参考的细化模块,通过计算跨通道维度的相似性矩阵来补偿运动和丢失的纹理细节。最后,采用注意力引导的双分支残差模块来同时增强重建结果。与 HEVC 锚点相比,在低延迟 P(LDP)设置下,Bjontegaard Delta Rate(BD-Rate)的平均增益为 24.86%。此外,还与先进的 SR 方法和压缩视频质量增强(VQE)方法进行了实验比较,并进一步报告了所提算法的卓越效率和通用性。
{"title":"Multi-Reference-Based Cross-Scale Feature Fusion for Compressed Video Super Resolution","authors":"Lu Chen;Mao Ye;Luping Ji;Shuai Li;Hongwei Guo","doi":"10.1109/TBC.2024.3407517","DOIUrl":"10.1109/TBC.2024.3407517","url":null,"abstract":"To save transmission bandwidth, there exists an approach to down-sample a video and then up-sample the compressed video to save bit rates. The existing Super Resolution (SR) methods generally design powerful networks to compensate the loss information introduced by down-sampling. But the information of entire video is not fully utilized and effectively fused, resulting in the learned context information that is not enough for the high quality reconstruction. We propose a multi high-quality frames Referenced Cross-scale compressed Video Super Resolution method (RCVSR) that wisely uses past-and-future information, to pursue higher compression efficiency. Specifically, a joint reference motion alignment module is proposed. Low resolution (LR) frame after up-sampling is separately aligned with past-and-future reference frames to preserve more spatial details; at the same time this LR frame is aligned with neighborhood frames to get continuous motion information and similar contents. Then, a reference based refinement module is applied to compensate motion and lost texture details by computing similarity matrix across channel dimensions. Finally, an attention guided dual-branch residual module is employed to enhance the reconstructed result concurrently. Compared with the HEVC anchor, the average gain of Bjontegaard Delta Rate (BD-Rate) under the Low-Delay-P (LDP) setting is 24.86%. In addition, an experimental comparison is made with the advanced SR methods and compressed video quality enhancement (VQE) methods, and the superior efficiency and generalization of the proposed algorithm are further reported.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 3","pages":"895-908"},"PeriodicalIF":3.2,"publicationDate":"2024-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141940838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-12DOI: 10.1109/TBC.2024.3405346
Stephen DelMarco
In this paper we present a new approach to high-performance compander design to reduce the peak-to-average power ratio (PAPR) that typically occurs in orthogonal frequency division multiplexing (OFDM) systems. Whereas many current compander designs assume a parametric model for the form of the transformed Rayleigh amplitude distribution, we define a constrained optimization problem for the functional form of the transformed distribution. We determine an optimal distribution which minimally deviates from the Rayleigh distribution and use a Sobolev norm to quantify distance. Use of the Sobolev norm imposes smoothness constraints on the transformed distribution, which are associated with lower out-of-band interference levels. We incorporate Lagrange multipliers into the problem formulation to enforce the constant power and probability density function constraints. We solve the constrained optimization problem using techniques from the Calculus of Variations and discuss compander and decompander design. We investigate the effect of incorporating derivative information, into the optimization formulation, on compander performance. We demonstrate compander performance through numerical simulation and compare compander performance to performance from a state-of-the-art variational compander which does not use derivative information in the formulation. We demonstrate performance improvements in out-of-band power rejection using the new compander.
{"title":"A Sobolev Norm-Based Variational Approach to Companding for PAPR Reduction in OFDM Systems","authors":"Stephen DelMarco","doi":"10.1109/TBC.2024.3405346","DOIUrl":"10.1109/TBC.2024.3405346","url":null,"abstract":"In this paper we present a new approach to high-performance compander design to reduce the peak-to-average power ratio (PAPR) that typically occurs in orthogonal frequency division multiplexing (OFDM) systems. Whereas many current compander designs assume a parametric model for the form of the transformed Rayleigh amplitude distribution, we define a constrained optimization problem for the functional form of the transformed distribution. We determine an optimal distribution which minimally deviates from the Rayleigh distribution and use a Sobolev norm to quantify distance. Use of the Sobolev norm imposes smoothness constraints on the transformed distribution, which are associated with lower out-of-band interference levels. We incorporate Lagrange multipliers into the problem formulation to enforce the constant power and probability density function constraints. We solve the constrained optimization problem using techniques from the Calculus of Variations and discuss compander and decompander design. We investigate the effect of incorporating derivative information, into the optimization formulation, on compander performance. We demonstrate compander performance through numerical simulation and compare compander performance to performance from a state-of-the-art variational compander which does not use derivative information in the formulation. We demonstrate performance improvements in out-of-band power rejection using the new compander.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 3","pages":"955-962"},"PeriodicalIF":3.2,"publicationDate":"2024-06-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141940840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-07DOI: 10.1109/TBC.2024.3408433
{"title":"IEEE Transactions on Broadcasting Information for Authors","authors":"","doi":"10.1109/TBC.2024.3408433","DOIUrl":"https://doi.org/10.1109/TBC.2024.3408433","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 2","pages":"C3-C4"},"PeriodicalIF":4.5,"publicationDate":"2024-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10552064","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141294952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}