首页 > 最新文献

IEEE Transactions on Broadcasting最新文献

英文 中文
Multi-Level Perception Assessment for Underwater Image Enhancement 水下图像增强的多级感知评价
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-01-29 DOI: 10.1109/TBC.2025.3525972
Yiwen Xu;Yuxiang Lin;Nian He;Xuejin Wang;Tiesong Zhao
Due to the complex underwater imaging environment, existing Underwater Image Enhancement (UIE) techniques are unable to handle the increasing demand for high-quality underwater content in broadcasting systems. Thus, a robust quality assessment method is highly expected to effectively compare the quality of different enhanced underwater images. To this end, we propose a novel quality assessment method for enhanced underwater images by utilizing multiple levels of features at various stages of the network’s depth. We first select underwater images with different distortions to analyze the characteristics of different UIE results at various feature levels. We found that low-level features are more sensitive to color information, while mid-level features are more indicative of structural differences. Based on this, a Channel-Spatial-Pixel Attention Module (CSPAM) is designed for low-level perception to capture color characteristics, utilizing channel, spatial, and pixel dimensions. To capture structural variations, a Parallel Structural Perception Module (PSPM) with convolutional kernels of different scales is introduced for mid-level perception. For high-level perception, due to the accumulation of noise, an Adaptive Weighted Downsampling (AWD) layer is employed to restore the semantic information. Furthermore, a new top-down multi-level feature fusion method is designed. Information from different levels is integrated through a Selective Feature Fusion (SFF) mechanism, which produces semantically rich features and enhances the model’s feature representation capability. Experimental results demonstrate the superior performance of the proposed method over the competing image quality evaluation methods.
由于水下成像环境的复杂性,现有的水下图像增强(UIE)技术无法满足广播系统对高质量水下内容日益增长的需求。因此,需要一种鲁棒的质量评估方法来有效地比较不同增强后的水下图像的质量。为此,我们提出了一种利用网络深度不同阶段的多层次特征对增强水下图像进行质量评估的新方法。我们首先选取具有不同失真的水下图像,分析不同特征层次上不同UIE结果的特征。我们发现,低级特征对颜色信息更敏感,而中级特征更能表明结构差异。在此基础上,设计了通道-空间-像素注意模块(channel - spatial - pixel Attention Module, CSPAM),利用通道、空间和像素维度对颜色特征进行低层次感知。为了捕获结构变化,引入了具有不同尺度卷积核的并行结构感知模块(PSPM)用于中级感知。对于高级感知,由于噪声的积累,采用自适应加权降采样(AWD)层来恢复语义信息。在此基础上,设计了一种自顶向下的多层次特征融合方法。通过选择性特征融合(Selective Feature Fusion, SFF)机制整合不同层次的信息,生成语义丰富的特征,增强了模型的特征表示能力。实验结果表明,该方法具有较好的图像质量评价效果。
{"title":"Multi-Level Perception Assessment for Underwater Image Enhancement","authors":"Yiwen Xu;Yuxiang Lin;Nian He;Xuejin Wang;Tiesong Zhao","doi":"10.1109/TBC.2025.3525972","DOIUrl":"https://doi.org/10.1109/TBC.2025.3525972","url":null,"abstract":"Due to the complex underwater imaging environment, existing Underwater Image Enhancement (UIE) techniques are unable to handle the increasing demand for high-quality underwater content in broadcasting systems. Thus, a robust quality assessment method is highly expected to effectively compare the quality of different enhanced underwater images. To this end, we propose a novel quality assessment method for enhanced underwater images by utilizing multiple levels of features at various stages of the network’s depth. We first select underwater images with different distortions to analyze the characteristics of different UIE results at various feature levels. We found that low-level features are more sensitive to color information, while mid-level features are more indicative of structural differences. Based on this, a Channel-Spatial-Pixel Attention Module (CSPAM) is designed for low-level perception to capture color characteristics, utilizing channel, spatial, and pixel dimensions. To capture structural variations, a Parallel Structural Perception Module (PSPM) with convolutional kernels of different scales is introduced for mid-level perception. For high-level perception, due to the accumulation of noise, an Adaptive Weighted Downsampling (AWD) layer is employed to restore the semantic information. Furthermore, a new top-down multi-level feature fusion method is designed. Information from different levels is integrated through a Selective Feature Fusion (SFF) mechanism, which produces semantically rich features and enhances the model’s feature representation capability. Experimental results demonstrate the superior performance of the proposed method over the competing image quality evaluation methods.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 2","pages":"606-615"},"PeriodicalIF":3.2,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144243815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cross-Domain Feature Interaction Network for Stereo Image Quality Assessment Considering Difference Information Guiding Binocular Fusion 基于差分信息的双目融合立体图像质量评价跨域特征交互网络
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-01-22 DOI: 10.1109/TBC.2025.3525976
Yongli Chang;Guanghui Yue;Bo Zhao
Recently, convolutional neural network (CNN) based stereo image quality assessment (SIQA) has been extensively researched, achieving impressive performance. However, most SIQA methods tend to only mine features from distorted stereo image, neglecting the exploitation of valuable features present in other image domains. Moreover, some simple fusion strategies like addition and concatenation for binocular fusion further limit the network’s prediction performance. Therefore, we design a cross-domain feature interaction network (CDFINet) for SIQA in this paper, which considers the complementarity between different domain features and realizes binocular fusion between the left and right monocular features based on difference information. Specifically, to boost the prediction ability, we design a dual-branch network with image and gradient feature extraction branches, extracting hierarchical features from both domains. Moreover, to explore more proper binocular information, we propose a difference information guidance based binocular fusion (DIGBF) module to achieve the binocular fusion. Furthermore, to better achieve information compensation between image and gradient domain, binocular features obtained from image domain and gradient domain are fused in the proposed cross-domain feature fusion (CDFF) module. In addition, considering the feedback mechanism of the visual cortex, higher-level features are backpropagated to lower-level regions, and the proposed cross-layer feature interaction (CLFI) module realizes the guidance of higher-level features to lower-level features. Finally, to encourage a more effective way to get the perceptual quality, a hierarchical multi-score quality aggregation method is proposed. The experimental results on four SIQA databases show that our CDFINet outperforms the compared mainstream metrics.
近年来,基于卷积神经网络(CNN)的立体图像质量评估(SIQA)得到了广泛的研究,并取得了令人瞩目的成绩。然而,大多数SIQA方法倾向于只从扭曲的立体图像中挖掘特征,而忽略了对其他图像域中存在的有价值特征的开发。此外,双目融合的一些简单的融合策略,如加法和拼接,进一步限制了网络的预测性能。为此,本文设计了一种面向SIQA的跨域特征交互网络(CDFINet),该网络考虑了不同域特征之间的互补性,并基于差异信息实现了左右单目特征的双目融合。具体来说,为了提高预测能力,我们设计了一个带有图像和梯度特征提取分支的双分支网络,从两个领域提取层次特征。此外,为了探索更合适的双目信息,我们提出了基于差分信息制导的双目融合(DIGBF)模块来实现双目融合。此外,为了更好地实现图像和梯度域之间的信息补偿,本文提出的跨域特征融合(cross-domain feature fusion, CDFF)模块将图像域和梯度域获得的双目特征进行融合。此外,考虑到视觉皮层的反馈机制,将高阶特征反向传播到低阶区域,提出的跨层特征交互(cross-layer feature interaction, CLFI)模块实现了高阶特征对低阶特征的引导。最后,为了更有效地获取感知质量,提出了一种分层多分值质量聚合方法。在四个SIQA数据库上的实验结果表明,我们的CDFINet优于主流指标。
{"title":"Cross-Domain Feature Interaction Network for Stereo Image Quality Assessment Considering Difference Information Guiding Binocular Fusion","authors":"Yongli Chang;Guanghui Yue;Bo Zhao","doi":"10.1109/TBC.2025.3525976","DOIUrl":"https://doi.org/10.1109/TBC.2025.3525976","url":null,"abstract":"Recently, convolutional neural network (CNN) based stereo image quality assessment (SIQA) has been extensively researched, achieving impressive performance. However, most SIQA methods tend to only mine features from distorted stereo image, neglecting the exploitation of valuable features present in other image domains. Moreover, some simple fusion strategies like addition and concatenation for binocular fusion further limit the network’s prediction performance. Therefore, we design a cross-domain feature interaction network (CDFINet) for SIQA in this paper, which considers the complementarity between different domain features and realizes binocular fusion between the left and right monocular features based on difference information. Specifically, to boost the prediction ability, we design a dual-branch network with image and gradient feature extraction branches, extracting hierarchical features from both domains. Moreover, to explore more proper binocular information, we propose a difference information guidance based binocular fusion (DIGBF) module to achieve the binocular fusion. Furthermore, to better achieve information compensation between image and gradient domain, binocular features obtained from image domain and gradient domain are fused in the proposed cross-domain feature fusion (CDFF) module. In addition, considering the feedback mechanism of the visual cortex, higher-level features are backpropagated to lower-level regions, and the proposed cross-layer feature interaction (CLFI) module realizes the guidance of higher-level features to lower-level features. Finally, to encourage a more effective way to get the perceptual quality, a hierarchical multi-score quality aggregation method is proposed. The experimental results on four SIQA databases show that our CDFINet outperforms the compared mainstream metrics.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 2","pages":"593-605"},"PeriodicalIF":3.2,"publicationDate":"2025-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144243882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distortion Propagation Factor Estimation for VVC Low-Delay Hierarchical Coding VVC低延迟分层编码的失真传播因子估计
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-01-20 DOI: 10.1109/TBC.2024.3519909
Hongwei Guo;Ce Zhu;Junjie Chen;Lei Luo;Yongkai Huo;Yutian Liu
Previous studies have shown that temporally dependent rate-distortion optimization (RDO) methods can enhance the compression performance of video encoders. However, accurately quantifying temporal rate-distortion dependencies in the latest video coding standard, Versatile Video Coding (VVC), remains a significant challenge. To address this issue, this paper proposes a distortion propagation factor (DPF) estimation method tailored for VVC low-delay hierarchical coding, aiming to achieve temporally dependent RDO. Specifically, we first derive a formula for calculating the DPF based on coding distortion and motion-compensated prediction (MCP) errors. Building on this, we present several pre-encoding-based DPF estimation schemes designed for the VVC low-delay hierarchical coding structure. These schemes have very low computational complexity and do not require buffering subsequent unencoded frames for pre-analysis, thereby avoiding additional encoding delays. Finally, the estimated DPFs are used to adaptively adjust the Lagrange multipliers and quantization parameters of each coding tree unit, optimizing the allocation of coding bit resources. After integrating the proposed method into the VVC test model VTM-23.0, experimental results show that one of the proposed DPF estimation schemes achieves average bit rate savings of 4.25% for low-delay B slices and 4.12% for low-delay P slices, with only a 1% increase in computational complexity. The proposed method offers an effective solution for enhancing the compression performance of VVC encoders. Consequently, the proposed DPF estimation approaches have already been adopted by the Joint Video Experts Team (JVET) and officially integrated into the VVC reference software.
已有研究表明,时变率失真优化(RDO)方法可以提高视频编码器的压缩性能。然而,在最新的视频编码标准——通用视频编码(VVC)中,准确量化时间率失真依赖关系仍然是一个重大挑战。为了解决这一问题,本文提出了一种针对VVC低延迟分层编码的失真传播因子(DPF)估计方法,旨在实现时间相关的RDO。具体来说,我们首先推导了基于编码失真和运动补偿预测(MCP)误差的DPF计算公式。在此基础上,针对VVC低延迟分层编码结构,提出了几种基于预编码的DPF估计方案。这些方案具有非常低的计算复杂度,并且不需要缓冲后续的未编码帧进行预分析,从而避免了额外的编码延迟。最后,利用估计的dpf自适应调整每个编码树单元的拉格朗日乘子和量化参数,优化编码位资源的分配。将该方法与VVC测试模型VTM-23.0集成后,实验结果表明,其中一种DPF估计方案在低延迟B片和低延迟P片上平均比特率分别节省4.25%和4.12%,而计算复杂度仅增加1%。该方法为提高VVC编码器的压缩性能提供了一种有效的解决方案。因此,提出的DPF估计方法已被联合视频专家组(JVET)采用,并正式集成到VVC参考软件中。
{"title":"Distortion Propagation Factor Estimation for VVC Low-Delay Hierarchical Coding","authors":"Hongwei Guo;Ce Zhu;Junjie Chen;Lei Luo;Yongkai Huo;Yutian Liu","doi":"10.1109/TBC.2024.3519909","DOIUrl":"https://doi.org/10.1109/TBC.2024.3519909","url":null,"abstract":"Previous studies have shown that temporally dependent rate-distortion optimization (RDO) methods can enhance the compression performance of video encoders. However, accurately quantifying temporal rate-distortion dependencies in the latest video coding standard, Versatile Video Coding (VVC), remains a significant challenge. To address this issue, this paper proposes a distortion propagation factor (DPF) estimation method tailored for VVC low-delay hierarchical coding, aiming to achieve temporally dependent RDO. Specifically, we first derive a formula for calculating the DPF based on coding distortion and motion-compensated prediction (MCP) errors. Building on this, we present several pre-encoding-based DPF estimation schemes designed for the VVC low-delay hierarchical coding structure. These schemes have very low computational complexity and do not require buffering subsequent unencoded frames for pre-analysis, thereby avoiding additional encoding delays. Finally, the estimated DPFs are used to adaptively adjust the Lagrange multipliers and quantization parameters of each coding tree unit, optimizing the allocation of coding bit resources. After integrating the proposed method into the VVC test model VTM-23.0, experimental results show that one of the proposed DPF estimation schemes achieves average bit rate savings of 4.25% for low-delay B slices and 4.12% for low-delay P slices, with only a 1% increase in computational complexity. The proposed method offers an effective solution for enhancing the compression performance of VVC encoders. Consequently, the proposed DPF estimation approaches have already been adopted by the Joint Video Experts Team (JVET) and officially integrated into the VVC reference software.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 2","pages":"492-505"},"PeriodicalIF":3.2,"publicationDate":"2025-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144243634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TV 3.0: An Overview 电视3.0:概述
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-01-01 DOI: 10.1109/TBC.2024.3511928
Allan Seiti Sassaqui Chaubet;Rodrigo Admir Vaz;George Henrique Maranhão Garcia de Oliveira;Ricardo Seriacopi Rabaça;Isabela Coelho Dourado;Gustavo de Melo Valeira;Cristiano Akamine
A new Digital Terrestrial Television Broadcasting (DTTB) system, called Television (TV) 3.0, is being developed in Brazil and is expected to be on air by 2025 under the commercial name DTV+. It started with a Call for Proposals (CfP) for its systems components, for which organizations worldwide have submitted candidate technologies. After two testing and evaluation phases, the technologies for all layers were selected, the TV 3.0 architecture was completely defined, and the standards were written. It consists of modern Modulation and Code (MODCOD) techniques, mandatory transmission and reception in Multiple-Input Multiple-Output (MIMO) with cross-polarized antennas, an app-oriented interface, an Internet-based Transport Layer (TL), and state-of-the-art efficient coding for audio, video, and captions. This set of technologies will allow for several new use cases that change the user experience with TV, such as Geographically Segmented Broadcasting (GSB), targeted advertising, sensory effects, and interactivity. This paper reviews the phases already concluded for the TV 3.0 project and presents its potentialities and the current developments at its final stage.
巴西正在开发一种新的数字地面电视广播(DTTB)系统,称为电视(TV) 3.0,预计将于2025年以商业名称DTV+播出。它从系统组件的提案征集(CfP)开始,世界各地的组织已经提交了候选技术。经过两个阶段的测试和评估,选定了各层的技术,完整定义了TV 3.0架构,并编写了标准。它由现代调制和编码(MODCOD)技术、交叉极化天线的多输入多输出(MIMO)强制传输和接收、面向应用程序的接口、基于互联网的传输层(TL)以及最先进的音频、视频和字幕高效编码组成。这组技术将允许一些新的用例,改变用户的电视体验,如地理分段广播(GSB)、目标广告、感官效果和交互性。本文回顾了TV 3.0项目已经结束的阶段,并介绍了它的潜力和目前在最后阶段的发展。
{"title":"TV 3.0: An Overview","authors":"Allan Seiti Sassaqui Chaubet;Rodrigo Admir Vaz;George Henrique Maranhão Garcia de Oliveira;Ricardo Seriacopi Rabaça;Isabela Coelho Dourado;Gustavo de Melo Valeira;Cristiano Akamine","doi":"10.1109/TBC.2024.3511928","DOIUrl":"https://doi.org/10.1109/TBC.2024.3511928","url":null,"abstract":"A new Digital Terrestrial Television Broadcasting (DTTB) system, called Television (TV) 3.0, is being developed in Brazil and is expected to be on air by 2025 under the commercial name DTV+. It started with a Call for Proposals (CfP) for its systems components, for which organizations worldwide have submitted candidate technologies. After two testing and evaluation phases, the technologies for all layers were selected, the TV 3.0 architecture was completely defined, and the standards were written. It consists of modern Modulation and Code (MODCOD) techniques, mandatory transmission and reception in Multiple-Input Multiple-Output (MIMO) with cross-polarized antennas, an app-oriented interface, an Internet-based Transport Layer (TL), and state-of-the-art efficient coding for audio, video, and captions. This set of technologies will allow for several new use cases that change the user experience with TV, such as Geographically Segmented Broadcasting (GSB), targeted advertising, sensory effects, and interactivity. This paper reviews the phases already concluded for the TV 3.0 project and presents its potentialities and the current developments at its final stage.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"11-18"},"PeriodicalIF":3.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Digital Entity Management Methodology for Digital Twin Implementation: Concept, Definition, and Examples 数字孪生实现的数字实体管理方法:概念、定义和示例
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-01-01 DOI: 10.1109/TBC.2024.3517138
Yegi Lee;Myung-Sun Baek;Kyoungro Yoon
Many efforts to achieve cost savings through simulations have been ongoing in the cyber-physical system (CPS) industry and manufacturing field. Recently, the concept of digital twins has emerged as a promising solution for cost reduction in various fields, such as smart cities, factory optimization, architecture, and manufacturing. Digital twins offer enormous potential by continuously monitoring and updating data to study a wide range of issues and improve products and processes. However, the practical implementation of digital twins presents significant challenges. Additionally, while various studies have introduced the concepts and roles of digital twin systems and digital components, further research is needed to explore efficient operation and management strategies. This paper aims to present digital entity management methodology for the efficient implementation of digital twin systems. Our proposed class-level digital entity management methodology constructs complex and repetitively used digital entities into digital entity classes. This approach facilitates the abstraction, inheritance, and upcasting of digital entity classes. By leveraging class-level management and easily reusable and modifiable digital entities, the implementation of low-complexity digital twin systems becomes feasible. The proposed methodology aims to streamline the digital twin implementation process, addressing complex technical integration and practical implementation challenges.
在信息物理系统(CPS)行业和制造领域,通过模拟实现成本节约的努力一直在进行。最近,数字双胞胎的概念已经成为智能城市、工厂优化、建筑和制造等各个领域降低成本的有希望的解决方案。数字孪生通过持续监测和更新数据来研究广泛的问题并改进产品和流程,从而提供了巨大的潜力。然而,数字孪生的实际实施面临着重大挑战。此外,虽然各种研究已经介绍了数字孪生系统和数字组件的概念和作用,但需要进一步研究以探索有效的运营和管理策略。本文旨在提出有效实施数字孪生系统的数字实体管理方法。我们提出的类级数字实体管理方法将复杂和重复使用的数字实体构建到数字实体类中。这种方法促进了数字实体类的抽象、继承和上转换。通过利用类级管理和易于重用和修改的数字实体,低复杂性数字孪生系统的实现变得可行。提出的方法旨在简化数字孪生实施过程,解决复杂的技术集成和实际实施挑战。
{"title":"Digital Entity Management Methodology for Digital Twin Implementation: Concept, Definition, and Examples","authors":"Yegi Lee;Myung-Sun Baek;Kyoungro Yoon","doi":"10.1109/TBC.2024.3517138","DOIUrl":"https://doi.org/10.1109/TBC.2024.3517138","url":null,"abstract":"Many efforts to achieve cost savings through simulations have been ongoing in the cyber-physical system (CPS) industry and manufacturing field. Recently, the concept of digital twins has emerged as a promising solution for cost reduction in various fields, such as smart cities, factory optimization, architecture, and manufacturing. Digital twins offer enormous potential by continuously monitoring and updating data to study a wide range of issues and improve products and processes. However, the practical implementation of digital twins presents significant challenges. Additionally, while various studies have introduced the concepts and roles of digital twin systems and digital components, further research is needed to explore efficient operation and management strategies. This paper aims to present digital entity management methodology for the efficient implementation of digital twin systems. Our proposed class-level digital entity management methodology constructs complex and repetitively used digital entities into digital entity classes. This approach facilitates the abstraction, inheritance, and upcasting of digital entity classes. By leveraging class-level management and easily reusable and modifiable digital entities, the implementation of low-complexity digital twin systems becomes feasible. The proposed methodology aims to streamline the digital twin implementation process, addressing complex technical integration and practical implementation challenges.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"19-29"},"PeriodicalIF":3.2,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial Coupling Strategy and Improved BFGS-Based Advanced Rate Control for VVC 空间耦合策略及基于改进bfgs的VVC高级速率控制
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-31 DOI: 10.1109/TBC.2024.3517167
Jiahao Zhang;Shuhua Xiong;Xiaohai He;Zeming Zhao;Hongdong Qin
This paper presents an advanced rate control (ARC) algorithm for Versatile Video Coding (VVC). The proposed method is based on spatial coupling strategy and improved Broyden Fletcher Goldfarb Shanno (BFGS) algorithm to achieve a high performance rate control (RC). In this paper, we address the problem that the current coding block does not fully utilise the spatial information during the encoding process. Firstly, a parameter updating strategy at the coding tree unit (CTU) level is constructed based on spatial coupling strategy. The spatial coupling strategy established the relationship between video parameters and video texture, which enables the video parameters at the CTU level to be more closely aligned with the video content. Furthermore, in order to enhance the precision of RC, we have proposed an improved BFGS algorithm to update video parameters, which utilizes the optimal search direction of the different partial differentials and sets an adaptive speed control factor. The experimental results indicate that the proposed method offers better performance compared to the default RC in VVC Test Moder (VTM) 19.0, with Bjøntegaard Delta Rate (BD-Rate) savings of 6.35%, 5.09% and 5.43% under Low Delay P, Low Delay B and Random Access configurations, respectively. Moreover, the proposed method demonstrates superior performance compared to other state-of-the-art algorithms.
提出了一种用于通用视频编码(VVC)的先进速率控制(ARC)算法。该方法基于空间耦合策略和改进的BFGS (Broyden Fletcher Goldfarb Shanno)算法来实现高性能的速率控制。本文解决了当前编码块在编码过程中没有充分利用空间信息的问题。首先,基于空间耦合策略构造了编码树单元(CTU)级的参数更新策略;空间耦合策略建立了视频参数与视频纹理之间的关系,使CTU级别的视频参数与视频内容更加紧密地对齐。此外,为了提高RC的精度,我们提出了一种改进的BFGS算法来更新视频参数,该算法利用不同偏微分的最优搜索方向并设置自适应速度控制因子。实验结果表明,与VVC Test Moder (VTM) 19.0中的默认RC相比,该方法具有更好的性能,在低延迟P、低延迟B和随机访问配置下,Bjøntegaard Delta Rate (BD-Rate)分别节省了6.35%、5.09%和5.43%。此外,与其他最先进的算法相比,所提出的方法表现出优越的性能。
{"title":"Spatial Coupling Strategy and Improved BFGS-Based Advanced Rate Control for VVC","authors":"Jiahao Zhang;Shuhua Xiong;Xiaohai He;Zeming Zhao;Hongdong Qin","doi":"10.1109/TBC.2024.3517167","DOIUrl":"https://doi.org/10.1109/TBC.2024.3517167","url":null,"abstract":"This paper presents an advanced rate control (ARC) algorithm for Versatile Video Coding (VVC). The proposed method is based on spatial coupling strategy and improved Broyden Fletcher Goldfarb Shanno (BFGS) algorithm to achieve a high performance rate control (RC). In this paper, we address the problem that the current coding block does not fully utilise the spatial information during the encoding process. Firstly, a parameter updating strategy at the coding tree unit (CTU) level is constructed based on spatial coupling strategy. The spatial coupling strategy established the relationship between video parameters and video texture, which enables the video parameters at the CTU level to be more closely aligned with the video content. Furthermore, in order to enhance the precision of RC, we have proposed an improved BFGS algorithm to update video parameters, which utilizes the optimal search direction of the different partial differentials and sets an adaptive speed control factor. The experimental results indicate that the proposed method offers better performance compared to the default RC in VVC Test Moder (VTM) 19.0, with Bjøntegaard Delta Rate (BD-Rate) savings of 6.35%, 5.09% and 5.43% under Low Delay P, Low Delay B and Random Access configurations, respectively. Moreover, the proposed method demonstrates superior performance compared to other state-of-the-art algorithms.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"111-124"},"PeriodicalIF":3.2,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalizable Underwater Image Quality Assessment With Curriculum Learning-Inspired Domain Adaption 基于课程学习启发域自适应的广义水下图像质量评价
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-27 DOI: 10.1109/TBC.2024.3511962
Shihui Wu;Qiuping Jiang;Guanghui Yue;Shiqi Wang;Guangtao Zhai
The complex distortions suffered by real-world underwater images pose urgent demands on accurate underwater image quality assessment (UIQA) approaches that can predict underwater image quality consistently with human perception. Deep learning techniques have achieved great success in many applications, yet usually requiring a substantial amount of human-labeled data, which is time-consuming and labor-intensive. Developing a deep learning-based UIQA method that does not rely on any human labeled underwater images for model training poses a great challenge. In this work, we propose a novel UIQA method based on domain adaption (DA) from a curriculum learning perspective. The proposed method is called curriculum learning-inspired DA (CLIDA), aiming to learn an robust and generalizable UIQA model by conducting DA between the labeled natural images and unlabeled underwater images progressively, i.e., from easy to hard. The key is how to select easy samples from all underwater images in the target domain so that the difficulty of DA can be well-controlled at each stage. To this end, we propose a simple yet effective easy sample selection (ESS) scheme to form an easy sample set at each stage. Then, DA is performed between the entire natural image set in the source domain (with labels) and the selected easy sample set in the target domain (with pseudo labels) at each stage. As only those reliable easy examples are involved in DA at each stage, the difficulty of DA is well-controlled and the capability of the model is expected to be progressively enhanced. We conduct extensive experiments to verify the superiority of the proposed CLIDA method and also the effectiveness of each key component involved in our CLIDA framework. The source code will be made available at https://github.com/zzeu001/CLIDA.
真实世界的水下图像所遭受的复杂畸变,对准确的水下图像质量评估(UIQA)方法提出了迫切的要求,该方法可以预测与人类感知一致的水下图像质量。深度学习技术在许多应用中取得了巨大的成功,但通常需要大量的人工标记数据,这是耗时和劳动密集型的。开发一种基于深度学习的UIQA方法,不依赖于任何人类标记的水下图像进行模型训练,这是一个巨大的挑战。本文从课程学习的角度提出了一种基于领域自适应(DA)的UIQA方法。本文提出的方法被称为课程学习启发DA (curriculum learning-inspired DA, CLIDA),其目的是通过在有标记的自然图像和未标记的水下图像之间,由易到难,逐步进行DA,学习一个鲁棒的、可推广的UIQA模型。关键是如何从目标域的所有水下图像中选择简单的样本,使每个阶段的数据处理难度得到很好的控制。为此,我们提出了一种简单而有效的易样本选择(ESS)方案,在每个阶段形成一个易样本集。然后,在每个阶段对源域中的整个自然图像集(带标签)和目标域中选定的简单样本集(带伪标签)进行数据分析。由于每个阶段的数据分析只涉及那些可靠的简单示例,因此数据分析的难度得到了很好的控制,模型的能力有望逐步提高。我们进行了大量的实验来验证所提出的CLIDA方法的优越性以及我们的CLIDA框架中涉及的每个关键组件的有效性。源代码将在https://github.com/zzeu001/CLIDA上提供。
{"title":"Generalizable Underwater Image Quality Assessment With Curriculum Learning-Inspired Domain Adaption","authors":"Shihui Wu;Qiuping Jiang;Guanghui Yue;Shiqi Wang;Guangtao Zhai","doi":"10.1109/TBC.2024.3511962","DOIUrl":"https://doi.org/10.1109/TBC.2024.3511962","url":null,"abstract":"The complex distortions suffered by real-world underwater images pose urgent demands on accurate underwater image quality assessment (UIQA) approaches that can predict underwater image quality consistently with human perception. Deep learning techniques have achieved great success in many applications, yet usually requiring a substantial amount of human-labeled data, which is time-consuming and labor-intensive. Developing a deep learning-based UIQA method that does not rely on any human labeled underwater images for model training poses a great challenge. In this work, we propose a novel UIQA method based on domain adaption (DA) from a curriculum learning perspective. The proposed method is called curriculum learning-inspired DA (CLIDA), aiming to learn an robust and generalizable UIQA model by conducting DA between the labeled natural images and unlabeled underwater images progressively, i.e., from easy to hard. The key is how to select easy samples from all underwater images in the target domain so that the difficulty of DA can be well-controlled at each stage. To this end, we propose a simple yet effective easy sample selection (ESS) scheme to form an easy sample set at each stage. Then, DA is performed between the entire natural image set in the source domain (with labels) and the selected easy sample set in the target domain (with pseudo labels) at each stage. As only those reliable easy examples are involved in DA at each stage, the difficulty of DA is well-controlled and the capability of the model is expected to be progressively enhanced. We conduct extensive experiments to verify the superiority of the proposed CLIDA method and also the effectiveness of each key component involved in our CLIDA framework. The source code will be made available at <uri>https://github.com/zzeu001/CLIDA</uri>.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"252-263"},"PeriodicalIF":3.2,"publicationDate":"2024-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CLIPVQA: Video Quality Assessment via CLIP CLIPVQA:视频质量评估通过剪辑
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-27 DOI: 10.1109/TBC.2024.3511927
Fengchuang Xing;Mingjie Li;Yuan-Gen Wang;Guopu Zhu;Xiaochun Cao
In learning vision-language representations from Web-scale data, the contrastive language-image pre-training (CLIP) mechanism has demonstrated a remarkable performance in many vision tasks. However, its application to the widely studied video quality assessment (VQA) task is still an open issue. In this paper, we propose an efficient and effective CLIP-based Transformer method for the VQA problem (CLIPVQA). Specifically, we first design an effective video frame perception paradigm with the goal of extracting the rich spatiotemporal quality and content information among video frames. Then, the spatiotemporal quality features are adequately integrated together using a self-attention mechanism to yield video-level quality representation. To utilize the quality language descriptions of videos for supervision, we develop a CLIP-based encoder for language embedding, which is then fully aggregated with the generated content information via a cross-attention module for producing video-language representation. Finally, the video-level quality and video-language representations are fused together for final video quality prediction, where a vectorized regression loss is employed for efficient end-to-end optimization. Comprehensive experiments are conducted on eight in-the-wild video datasets with diverse resolutions to evaluate the performance of CLIPVQA. The experimental results show that the proposed CLIPVQA achieves new state-of-the-art VQA performance and up to 37% better generalizability than existing benchmark VQA methods. A series of ablation studies are also performed to validate the effectiveness of each module in CLIPVQA.
对比语言-图像预训练(CLIP)机制在网络尺度的视觉语言表征学习中,在许多视觉任务中表现出了显著的效果。然而,它在视频质量评估(VQA)任务中的应用仍是一个有待解决的问题。在本文中,我们提出了一种高效的基于CLIPVQA的VQA问题的变压器方法(CLIPVQA)。具体而言,我们首先设计了一种有效的视频帧感知范式,以提取视频帧之间丰富的时空质量和内容信息。然后,利用自关注机制将时空质量特征充分整合在一起,得到视频级质量表示。为了利用视频的高质量语言描述进行监督,我们开发了一个基于clip的语言嵌入编码器,然后通过交叉关注模块将生成的内容信息与生成的内容信息完全聚合,以产生视频语言表示。最后,将视频级质量和视频语言表示融合在一起进行最终的视频质量预测,其中使用向量化回归损失进行有效的端到端优化。在8个不同分辨率的野外视频数据集上进行了综合实验,以评估CLIPVQA的性能。实验结果表明,所提出的CLIPVQA达到了最新的VQA性能,与现有的基准VQA方法相比,泛化能力提高了37%。还进行了一系列消融研究,以验证CLIPVQA中每个模块的有效性。
{"title":"CLIPVQA: Video Quality Assessment via CLIP","authors":"Fengchuang Xing;Mingjie Li;Yuan-Gen Wang;Guopu Zhu;Xiaochun Cao","doi":"10.1109/TBC.2024.3511927","DOIUrl":"https://doi.org/10.1109/TBC.2024.3511927","url":null,"abstract":"In learning vision-language representations from Web-scale data, the contrastive language-image pre-training (CLIP) mechanism has demonstrated a remarkable performance in many vision tasks. However, its application to the widely studied video quality assessment (VQA) task is still an open issue. In this paper, we propose an efficient and effective CLIP-based Transformer method for the VQA problem (CLIPVQA). Specifically, we first design an effective video frame perception paradigm with the goal of extracting the rich spatiotemporal quality and content information among video frames. Then, the spatiotemporal quality features are adequately integrated together using a self-attention mechanism to yield video-level quality representation. To utilize the quality language descriptions of videos for supervision, we develop a CLIP-based encoder for language embedding, which is then fully aggregated with the generated content information via a cross-attention module for producing video-language representation. Finally, the video-level quality and video-language representations are fused together for final video quality prediction, where a vectorized regression loss is employed for efficient end-to-end optimization. Comprehensive experiments are conducted on eight in-the-wild video datasets with diverse resolutions to evaluate the performance of CLIPVQA. The experimental results show that the proposed CLIPVQA achieves new state-of-the-art VQA performance and up to 37% better generalizability than existing benchmark VQA methods. A series of ablation studies are also performed to validate the effectiveness of each module in CLIPVQA.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"291-306"},"PeriodicalIF":3.2,"publicationDate":"2024-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint Optimization of Beamforming and Subarray Assignment for Full-Duplex Arrays in Next Generation Broadcast Systems 下一代广播系统中全双工阵列波束形成与子阵列分配的联合优化
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-25 DOI: 10.1109/TBC.2024.3517141
Lang Lin;Wensheng Pan;Hongzhi Zhao;Shengfeng Zhang;Shihai Shao;Youxi Tang
The full-duplex (FD) technique provides spectrum-efficient transmission service by supporting uplink and downlink transmission at the same time over the same frequency band. Combined FD with multi-input multi-output (MIMO) antenna systems can improve the capability and efficiency of next-generation broadcasting to provide various services for multiple users. However, the strong self-interference (SI) coupled between the transmit and receive arrays is a significant challenge for simultaneous transmission and reception. This study considers multi-user oriented broadcast systems where the transmit array with subarray division allows simultaneous multi-beam transmission. Our objective is to mitigate SI from two aspects: 1) subarray assignment; 2) adaptive transmit beamforming (TxBF). We propose a min-SI TxBF design jointly with subarray assignment, which determines the best subarray assignment pattern and provides a closed-form solution of the optimal TxBF weight. Theoretical analysis and simulations show our design has a low algorithm complexity. Based on the measured SI channel data collected from a hardware prototype testbed, simulation results verify that the min-SI TxBF design with subarray assignment can effectively enhance SI cancelation.
全双工(full-duplex, FD)技术通过支持在同一频段内同时上行和下行传输,提供频谱高效的传输服务。FD与多输入多输出(MIMO)天线系统相结合,可以提高下一代广播的能力和效率,为多用户提供各种服务。然而,发射和接收阵列之间耦合的强自干扰(SI)是同时发送和接收的重大挑战。本研究考虑面向多用户的广播系统,其中带子阵分割的发射阵列允许同时发送多波束。我们的目标是从两个方面减轻SI: 1)子阵列分配;2)自适应发射波束形成(TxBF)。提出了一种结合子阵分配的最小si TxBF设计,确定了最佳子阵分配模式,并给出了最优TxBF权值的封闭解。理论分析和仿真结果表明,本设计具有较低的算法复杂度。基于硬件样机试验台采集的SI通道实测数据,仿真结果验证了采用子阵列分配的最小SI TxBF设计能有效增强SI对消。
{"title":"Joint Optimization of Beamforming and Subarray Assignment for Full-Duplex Arrays in Next Generation Broadcast Systems","authors":"Lang Lin;Wensheng Pan;Hongzhi Zhao;Shengfeng Zhang;Shihai Shao;Youxi Tang","doi":"10.1109/TBC.2024.3517141","DOIUrl":"https://doi.org/10.1109/TBC.2024.3517141","url":null,"abstract":"The full-duplex (FD) technique provides spectrum-efficient transmission service by supporting uplink and downlink transmission at the same time over the same frequency band. Combined FD with multi-input multi-output (MIMO) antenna systems can improve the capability and efficiency of next-generation broadcasting to provide various services for multiple users. However, the strong self-interference (SI) coupled between the transmit and receive arrays is a significant challenge for simultaneous transmission and reception. This study considers multi-user oriented broadcast systems where the transmit array with subarray division allows simultaneous multi-beam transmission. Our objective is to mitigate SI from two aspects: 1) subarray assignment; 2) adaptive transmit beamforming (TxBF). We propose a min-SI TxBF design jointly with subarray assignment, which determines the best subarray assignment pattern and provides a closed-form solution of the optimal TxBF weight. Theoretical analysis and simulations show our design has a low algorithm complexity. Based on the measured SI channel data collected from a hardware prototype testbed, simulation results verify that the min-SI TxBF design with subarray assignment can effectively enhance SI cancelation.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 2","pages":"672-679"},"PeriodicalIF":3.2,"publicationDate":"2024-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144243909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distortion Propagation Model-Based V-PCC Rate Control for 3D Point Cloud Broadcasting 基于失真传播模型的三维点云广播V-PCC速率控制
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-12 DOI: 10.1109/TBC.2024.3511950
Zhanyuan Cai;Wenxu Gao;Ge Li;Wei Gao
For efficient point cloud broadcasting, point cloud compression technologies serve as the foundation, which plays a crucial role in immersive media communication and streaming. Video-based point cloud compression (V-PCC) is the recently developed standard by the Moving Picture Experts Group (MPEG) for dynamic point clouds. Its original fixed-ratio bit allocation (FR-BA) method in the unique all intra (AI) structure leads to a significant rate-distortion performance gap between the rate control manner and the fixed quantization parameters (FixedQP) scheme, as evidenced by significant increases in BD-Rate (Bjøntegaard Delta Rate) for both geometry and attribute. To address this issue, we propose a distortion propagation model-based frame-level bit allocation method that is specifically tailored for AI structure in V-PCC. First, the analysis is carried out for the distortion propagation model inside the group of pictures (GOP) for the AI configuration. Second, the skip ratio of 4x4 minimum coding units (CUs) is utilized to predict the distortion propagation factor. Third, the occupancy information is employed to refine the distortion propagation model and further enhance compression performance. Finally, experimental results demonstrate the effectiveness of the proposed distortion propagation model-based frame-level bit allocation method. Specifically, experimental results reveal that the proposed method achieves BD-Rate reductions of 0.92% and 4.85% in geometry and attribute, respectively, compared to the FR-BA method. Furthermore, with the introduction of distortion propagation factor prediction incorporating occupancy correction, the BD-Rate reductions are further extended to 2.16% and 6.13% in geometry and attribute, respectively.
为了实现高效的点云广播,点云压缩技术是其基础,在沉浸式媒体通信和流媒体中起着至关重要的作用。基于视频的点云压缩(V-PCC)是运动图像专家组(MPEG)最近为动态点云开发的标准。其原始的固定比比特分配(FR-BA)方法在独特的全内(AI)结构中导致速率控制方式与固定量化参数(FixedQP)方案之间存在显着的速率失真性能差距,这可以从几何和属性的BD-Rate (Bjøntegaard Delta rate)显着增加中得到证明。为了解决这个问题,我们提出了一种基于失真传播模型的帧级比特分配方法,该方法专门针对V-PCC中的AI结构量身定制。首先,对人工智能配置的图像组内失真传播模型(GOP)进行了分析。其次,利用4 × 4最小编码单元(CUs)的跳变比预测失真传播系数;第三,利用占用信息细化失真传播模型,进一步提高压缩性能。最后,实验结果验证了基于失真传播模型的帧级比特分配方法的有效性。实验结果表明,与FR-BA方法相比,该方法在几何和属性上分别降低了0.92%和4.85%的BD-Rate。此外,引入包含占用校正的失真传播因子预测后,BD-Rate在几何和属性上的降低幅度分别达到2.16%和6.13%。
{"title":"Distortion Propagation Model-Based V-PCC Rate Control for 3D Point Cloud Broadcasting","authors":"Zhanyuan Cai;Wenxu Gao;Ge Li;Wei Gao","doi":"10.1109/TBC.2024.3511950","DOIUrl":"https://doi.org/10.1109/TBC.2024.3511950","url":null,"abstract":"For efficient point cloud broadcasting, point cloud compression technologies serve as the foundation, which plays a crucial role in immersive media communication and streaming. Video-based point cloud compression (V-PCC) is the recently developed standard by the Moving Picture Experts Group (MPEG) for dynamic point clouds. Its original fixed-ratio bit allocation (FR-BA) method in the unique all intra (AI) structure leads to a significant rate-distortion performance gap between the rate control manner and the fixed quantization parameters (FixedQP) scheme, as evidenced by significant increases in BD-Rate (Bjøntegaard Delta Rate) for both geometry and attribute. To address this issue, we propose a distortion propagation model-based frame-level bit allocation method that is specifically tailored for AI structure in V-PCC. First, the analysis is carried out for the distortion propagation model inside the group of pictures (GOP) for the AI configuration. Second, the skip ratio of 4x4 minimum coding units (CUs) is utilized to predict the distortion propagation factor. Third, the occupancy information is employed to refine the distortion propagation model and further enhance compression performance. Finally, experimental results demonstrate the effectiveness of the proposed distortion propagation model-based frame-level bit allocation method. Specifically, experimental results reveal that the proposed method achieves BD-Rate reductions of 0.92% and 4.85% in geometry and attribute, respectively, compared to the FR-BA method. Furthermore, with the introduction of distortion propagation factor prediction incorporating occupancy correction, the BD-Rate reductions are further extended to 2.16% and 6.13% in geometry and attribute, respectively.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"180-192"},"PeriodicalIF":3.2,"publicationDate":"2024-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Broadcasting
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1