首页 > 最新文献

IEEE Transactions on Broadcasting最新文献

英文 中文
IEEE Transactions on Broadcasting Publication Information IEEE广播出版信息汇刊
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/TBC.2025.3640759
{"title":"IEEE Transactions on Broadcasting Publication Information","authors":"","doi":"10.1109/TBC.2025.3640759","DOIUrl":"https://doi.org/10.1109/TBC.2025.3640759","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"C2-C2"},"PeriodicalIF":4.8,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11302006","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
2025 Scott Helt Memorial Award for the Best Paper Published in IEEE Transactions on Broadcasting 2025年斯科特·海尔特纪念奖,在IEEE广播事务中发表的最佳论文
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/TBC.2025.3640887
{"title":"2025 Scott Helt Memorial Award for the Best Paper Published in IEEE Transactions on Broadcasting","authors":"","doi":"10.1109/TBC.2025.3640887","DOIUrl":"https://doi.org/10.1109/TBC.2025.3640887","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"1108-1110"},"PeriodicalIF":4.8,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11302029","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Broadcasting Information for Readers and Authors 面向读者和作者的广播信息IEEE汇刊
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/TBC.2025.3640761
{"title":"IEEE Transactions on Broadcasting Information for Readers and Authors","authors":"","doi":"10.1109/TBC.2025.3640761","DOIUrl":"https://doi.org/10.1109/TBC.2025.3640761","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"C3-C4"},"PeriodicalIF":4.8,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11302004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind Light Field Image Quality Assessment Using Multiplane Texture and Multilevel Wavelet Information 基于多平面纹理和多级小波信息的盲光场图像质量评价
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-06 DOI: 10.1109/TBC.2025.3627787
Zhengyu Zhang;Shishun Tian;Jianjun Xiang;Wenbin Zou;Luce Morin;Lu Zhang
Light Field Image (LFI) has garnered remarkable interest and fascination due to its burgeoning significance in immersive applications. Although the abundant information in LFIs enables a more immersive experience, it also poses a greater challenge for Light Field Image Quality Assessment (LFIQA), especially when reference information is inaccessible. In this paper, inspired by the holistic visual perception of high-dimensional LFIs and neuroscience studies on the Human Visual System (HVS), we propose a novel Blind Light Field image quality assessment metric by exploring MultiPlane Texture and Multilevel Wavelet Information, abbreviated as MPT-MWI-BLiF. Specifically, considering the texture sensitivity of the secondary visual cortex (V2), we first convert LFIs into multiple individual planes and capture textural variations from these planes. Then, the statistical histogram of textural variations for all planes is calculated as holistic textural variation features. In addition, motivated by the fact that neuronal responses in the visual cortex are frequency-dependent, we simulate this visual perception process by decomposing LFIs into multilevel wavelet subbands with Four-Dimensional Discrete Haar Wavelet Transform (4D-DHWT). After that, the subband geometric features of first-level 4D-DHWT subbands and the coefficient intensity features of second-level 4D-DHWT subbands are computed respectively. Finally, we combine all the extracted quality-aware features and employ the widely-used Support Vector Regression (SVR) to predict the perceptual quality of LFIs. To fully validate the effectiveness of the proposed metric, we perform extensive experiments on five representative LFIQA databases with two cross-validation methods. Experimental results demonstrate the superiority of the proposed metric in quality evaluation, as well as its low time complexity compared to other state-of-the-art metrics. The full code will be publicly available at https://github.com/ZhengyuZhang96/MPT-MWI-BLiF
光场图像(LFI)由于其在沉浸式应用中的重要性而引起了人们的极大兴趣和迷恋。虽然lfi中丰富的信息可以提供更身临其境的体验,但它也对光场图像质量评估(LFIQA)提出了更大的挑战,特别是在无法获得参考信息的情况下。本文受高维lfi整体视觉感知和人类视觉系统(HVS)神经科学研究的启发,提出了一种基于多平面纹理和多层次小波信息的盲光场图像质量评价方法,简称mpt - mwi - bif。具体而言,考虑到第二视觉皮层(V2)的纹理敏感性,我们首先将lfi转换为多个单独的平面,并从这些平面中捕获纹理变化。然后,计算各平面纹理变化的统计直方图作为整体纹理变化特征;此外,考虑到视觉皮层的神经元反应是频率依赖的,我们通过使用四维离散Haar小波变换(4D-DHWT)将lfi分解成多层次小波子带来模拟视觉感知过程。然后分别计算第一级4D-DHWT子带的子带几何特征和第二级4D-DHWT子带的系数强度特征。最后,我们结合所有提取的质量感知特征,并使用广泛使用的支持向量回归(SVR)来预测lfi的感知质量。为了充分验证所提出度量的有效性,我们使用两种交叉验证方法在五个代表性的LFIQA数据库上进行了广泛的实验。实验结果证明了该度量在质量评价方面的优越性,并且与其他最新度量相比具有较低的时间复杂度。完整的代码将在https://github.com/ZhengyuZhang96/MPT-MWI-BLiF上公开提供
{"title":"Blind Light Field Image Quality Assessment Using Multiplane Texture and Multilevel Wavelet Information","authors":"Zhengyu Zhang;Shishun Tian;Jianjun Xiang;Wenbin Zou;Luce Morin;Lu Zhang","doi":"10.1109/TBC.2025.3627787","DOIUrl":"https://doi.org/10.1109/TBC.2025.3627787","url":null,"abstract":"Light Field Image (LFI) has garnered remarkable interest and fascination due to its burgeoning significance in immersive applications. Although the abundant information in LFIs enables a more immersive experience, it also poses a greater challenge for Light Field Image Quality Assessment (LFIQA), especially when reference information is inaccessible. In this paper, inspired by the holistic visual perception of high-dimensional LFIs and neuroscience studies on the Human Visual System (HVS), we propose a novel Blind Light Field image quality assessment metric by exploring MultiPlane Texture and Multilevel Wavelet Information, abbreviated as MPT-MWI-BLiF. Specifically, considering the texture sensitivity of the secondary visual cortex (V2), we first convert LFIs into multiple individual planes and capture textural variations from these planes. Then, the statistical histogram of textural variations for all planes is calculated as holistic textural variation features. In addition, motivated by the fact that neuronal responses in the visual cortex are frequency-dependent, we simulate this visual perception process by decomposing LFIs into multilevel wavelet subbands with Four-Dimensional Discrete Haar Wavelet Transform (4D-DHWT). After that, the subband geometric features of first-level 4D-DHWT subbands and the coefficient intensity features of second-level 4D-DHWT subbands are computed respectively. Finally, we combine all the extracted quality-aware features and employ the widely-used Support Vector Regression (SVR) to predict the perceptual quality of LFIs. To fully validate the effectiveness of the proposed metric, we perform extensive experiments on five representative LFIQA databases with two cross-validation methods. Experimental results demonstrate the superiority of the proposed metric in quality evaluation, as well as its low time complexity compared to other state-of-the-art metrics. The full code will be publicly available at <uri>https://github.com/ZhengyuZhang96/MPT-MWI-BLiF</uri>","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"1092-1107"},"PeriodicalIF":4.8,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145765632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IRBFusion: Diffusion-Based Blind Image Super Resolution Using Unsupervised Learning and Bank of Restoration Networks IRBFusion:使用无监督学习和恢复网络库的基于扩散的盲图像超分辨率
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-30 DOI: 10.1109/TBC.2025.3622337
Morteza Poudineh;Alireza Esmaeilzehi;M. Omair Ahmad
Image super resolution focuses on increasing the spatial resolution of low-quality images and enhancing their visual quality. Since the image degradation process is unknown in real-life scenarios, it is crucial to perform image super resolution in a blind manner. Diffusion models have revolutionized the task of blind image super resolution in view of their powerful capability of producing realistic textures and structures. Design of the condition network is a key factor for diffusion models in providing high image super resolution performances. In this regard, we develop an effective image restoration bank by using a three-stage learning algorithm based on the idea of unsupervised learning, and feed its results, wherein visual artifacts are remarkably suppressed, to the condition network. The use of the unsupervised learning in the design of our image restoration bank guarantees that both diverse contextual information of visual signals, as well as, different degradation operations are considered for the task of blind image super resolution. Further, we guide the feature generation process of the condition network in such a way that the fidelity of the feature tensors produced for the task of image super resolution remains high. The results of extensive experiments show the superiority of our method over the state-of-the-art blind image super resolution schemes in the case of various benchmark datasets.
图像超分辨率的研究重点是提高低质量图像的空间分辨率,提高其视觉质量。由于在实际场景中图像的退化过程是未知的,因此以盲目的方式执行图像超分辨率是至关重要的。扩散模型具有生成逼真纹理和结构的强大能力,彻底改变了盲图像超分辨率的任务。条件网络的设计是扩散模型能否提供高图像超分辨性能的关键因素。在这方面,我们通过使用基于无监督学习思想的三阶段学习算法开发了一个有效的图像恢复库,并将其结果(其中视觉伪影被显著抑制)提供给条件网络。在我们的图像恢复库设计中使用无监督学习,保证了视觉信号的不同上下文信息以及盲图像超分辨率任务的不同退化操作。此外,我们以这样的方式指导条件网络的特征生成过程,即为图像超分辨率任务产生的特征张量的保真度仍然很高。大量的实验结果表明,在各种基准数据集的情况下,我们的方法优于最先进的盲图像超分辨率方案。
{"title":"IRBFusion: Diffusion-Based Blind Image Super Resolution Using Unsupervised Learning and Bank of Restoration Networks","authors":"Morteza Poudineh;Alireza Esmaeilzehi;M. Omair Ahmad","doi":"10.1109/TBC.2025.3622337","DOIUrl":"https://doi.org/10.1109/TBC.2025.3622337","url":null,"abstract":"Image super resolution focuses on increasing the spatial resolution of low-quality images and enhancing their visual quality. Since the image degradation process is unknown in real-life scenarios, it is crucial to perform image super resolution in a blind manner. Diffusion models have revolutionized the task of blind image super resolution in view of their powerful capability of producing realistic textures and structures. Design of the condition network is a key factor for diffusion models in providing high image super resolution performances. In this regard, we develop an effective image restoration bank by using a three-stage learning algorithm based on the idea of unsupervised learning, and feed its results, wherein visual artifacts are remarkably suppressed, to the condition network. The use of the unsupervised learning in the design of our image restoration bank guarantees that both diverse contextual information of visual signals, as well as, different degradation operations are considered for the task of blind image super resolution. Further, we guide the feature generation process of the condition network in such a way that the fidelity of the feature tensors produced for the task of image super resolution remains high. The results of extensive experiments show the superiority of our method over the state-of-the-art blind image super resolution schemes in the case of various benchmark datasets.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"1048-1064"},"PeriodicalIF":4.8,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Geographic Segmented Localcasting Co-Channel Interference Mitigation Using Iterative Joint Detection and Decoding 基于迭代联合检测和译码的地理分段局部广播同信道干扰抑制
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-16 DOI: 10.1109/TBC.2025.3579222
Hao Ju;Yin Xu;Dazhi He;Haoyang Li;Wenjun Zhang;Yiyan Wu
Geographic Segmented Localcasting (GSL) is an emerging Digital Terrestrial Television Broadcast (DTTB) physical layer operating mode. The system utilizes LDM-SFN to enable both wide-area Single Frequency Network (SFN) coverage and localized broadcast/multicast services within a single Radio Frequency (RF) broadcast channel. By combining SFN (core layer) and localcasting (enhanced layer) via Layered Division Multiplexing (LDM), GSL improves spectrum efficiency but faces challenges such as co-channel interference and SFN vs. Localcasting Channel Profile Mismatch (LCPM), which limit localcasting coverage. This paper presents two receiving methods that facilitate LDPC-coded LDM signal reception. Method 1 is the Multiple localcasting signals Iterative Joint Detection and Decoding (IJDD), which can mitigate severe co-channel interference with required Channel Status Information (CSI) of nearby localcasting transmitters. Method 2 is the Constellation Rotated IJDD (CR-IJDD), which can mitigate severe LCPM without CSI. The proposed methods enable decoding of both desired and interfering signals under high SNR conditions, enhancing spectrum reuse. Additionally, an Early Extrinsic Information Exchange for LDPC Iteration Reduction (EEIE-LIR) scheme is introduced to accelerate convergence and reduce receiver complexity. Evaluations based on ATSC 3.0 ModCods demonstrate that the proposed methods significantly improve spectrum efficiency and mitigate the co-channel interferences. The proposed technologies and can be extended to other DTTB systems and cell-based broadband wireless networks (e.g., the fifth generation (5G)/the sixth generation (6G), supporting seamless integration of broadcast, multicast, and unicast services.
地理分段本地广播(GSL)是一种新兴的数字地面电视广播(DTTB)物理层操作模式。该系统利用LDM-SFN实现广域单频网络(SFN)覆盖和在单个射频(RF)广播信道内的本地化广播/多播服务。通过分层分复用(LDM)将SFN(核心层)和本地广播(增强层)相结合,GSL提高了频谱效率,但面临着同信道干扰和SFN与本地广播信道配置不匹配(LCPM)等挑战,这些挑战限制了本地广播的覆盖范围。本文提出了两种方便ldpc编码LDM信号接收的接收方法。方法一是多本地广播信号迭代联合检测和解码(IJDD),该方法可以减轻附近本地广播发射机所需信道状态信息(CSI)的严重同信道干扰。方法二是星座旋转IJDD (CR-IJDD),它可以减轻严重的LCPM而不需要CSI。所提出的方法能够在高信噪比条件下对期望信号和干扰信号进行解码,从而提高频谱复用能力。此外,为了加快收敛速度和降低接收方复杂度,还引入了早期外部信息交换的LDPC迭代减少(EEIE-LIR)方案。基于ATSC 3.0 ModCods的评估表明,所提出的方法显著提高了频谱效率,减轻了同信道干扰。建议的技术可扩展至其他数字地面直播系统和基于蜂窝的宽带无线网络(例如,第五代(5G)/第六代(6G)),支持广播、多播和单播服务的无缝集成。
{"title":"Geographic Segmented Localcasting Co-Channel Interference Mitigation Using Iterative Joint Detection and Decoding","authors":"Hao Ju;Yin Xu;Dazhi He;Haoyang Li;Wenjun Zhang;Yiyan Wu","doi":"10.1109/TBC.2025.3579222","DOIUrl":"https://doi.org/10.1109/TBC.2025.3579222","url":null,"abstract":"Geographic Segmented Localcasting (GSL) is an emerging Digital Terrestrial Television Broadcast (DTTB) physical layer operating mode. The system utilizes LDM-SFN to enable both wide-area Single Frequency Network (SFN) coverage and localized broadcast/multicast services within a single Radio Frequency (RF) broadcast channel. By combining SFN (core layer) and localcasting (enhanced layer) via Layered Division Multiplexing (LDM), GSL improves spectrum efficiency but faces challenges such as co-channel interference and SFN vs. Localcasting Channel Profile Mismatch (LCPM), which limit localcasting coverage. This paper presents two receiving methods that facilitate LDPC-coded LDM signal reception. Method 1 is the Multiple localcasting signals Iterative Joint Detection and Decoding (IJDD), which can mitigate severe co-channel interference with required Channel Status Information (CSI) of nearby localcasting transmitters. Method 2 is the Constellation Rotated IJDD (CR-IJDD), which can mitigate severe LCPM without CSI. The proposed methods enable decoding of both desired and interfering signals under high SNR conditions, enhancing spectrum reuse. Additionally, an Early Extrinsic Information Exchange for LDPC Iteration Reduction (EEIE-LIR) scheme is introduced to accelerate convergence and reduce receiver complexity. Evaluations based on ATSC 3.0 ModCods demonstrate that the proposed methods significantly improve spectrum efficiency and mitigate the co-channel interferences. The proposed technologies and can be extended to other DTTB systems and cell-based broadband wireless networks (e.g., the fifth generation (5G)/the sixth generation (6G), supporting seamless integration of broadcast, multicast, and unicast services.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"941-953"},"PeriodicalIF":4.8,"publicationDate":"2025-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fusing Learning and Non-Learning: Hybrid CNN-Transformer Cooperative-Competitive Network for Underwater Image Enhancement 融合学习与非学习:用于水下图像增强的CNN-Transformer合作-竞争混合网络
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-09-23 DOI: 10.1109/TBC.2025.3611669
Xun Ji;Xu Wang;Li-Ying Hao;Chengtao Cai;Chengsong Dai;Ryan Wen Liu
Underwater image enhancement (UIE) aims to provide high-quality observations of challenging underwater scenarios, which is of great significance for various broadcast technologies. Extensive non-learning-based and learning-based UIE methods have been presented and applied. However, non-learning-based strategies typically struggle to demonstrate superior generalization capabilities, while learning-based strategies generally suffer from potential over- or under-enhancement due to the lack of sufficient prior knowledge. To address the challenges above, this paper presents a heuristic cooperative-competitive network, termed Co2Net. Specifically, our Co2Net integrates non-learning mechanisms into the deep learning framework to achieve information fusion from explainable prior knowledge and discernible hierarchical features, thereby facilitating promising and reasonable enhancement of degraded underwater images. Furthermore, our Co2Net adopts a hybrid convolutional neural network (CNN)-Transformer architecture, which comprises successive cooperative-competitive modules (Co2Ms) to achieve adequate extraction, representation, and transmission of both prior knowledge and discernible features. Comprehensive experiments are conducted to demonstrate the superiority and universality of our proposed Co2Net, and sufficient ablation studies are also performed to reveal the effectiveness of each component within our model. The source code is available at https://github.com/jixun-dmu/Co2Net
水下图像增强(UIE)旨在提供具有挑战性的水下场景的高质量观测结果,这对各种广播技术具有重要意义。广泛的非基于学习和基于学习的ie方法已经被提出和应用。然而,非基于学习的策略通常难以表现出卓越的泛化能力,而基于学习的策略通常由于缺乏足够的先验知识而遭受潜在的过度或不足增强。为了解决上述挑战,本文提出了一个启发式合作竞争网络,称为Co2Net。具体而言,我们的Co2Net将非学习机制集成到深度学习框架中,实现了可解释的先验知识和可识别的层次特征的信息融合,从而促进了降级水下图像的有希望和合理的增强。此外,我们的Co2Net采用混合卷积神经网络(CNN)-Transformer架构,该架构由连续的合作-竞争模块(Co2Ms)组成,以实现对先验知识和可识别特征的充分提取、表示和传输。我们进行了全面的实验,以证明我们提出的Co2Net的优越性和通用性,并进行了充分的消融研究,以揭示我们模型中每个组件的有效性。源代码可从https://github.com/jixun-dmu/Co2Net获得
{"title":"Fusing Learning and Non-Learning: Hybrid CNN-Transformer Cooperative-Competitive Network for Underwater Image Enhancement","authors":"Xun Ji;Xu Wang;Li-Ying Hao;Chengtao Cai;Chengsong Dai;Ryan Wen Liu","doi":"10.1109/TBC.2025.3611669","DOIUrl":"https://doi.org/10.1109/TBC.2025.3611669","url":null,"abstract":"Underwater image enhancement (UIE) aims to provide high-quality observations of challenging underwater scenarios, which is of great significance for various broadcast technologies. Extensive non-learning-based and learning-based UIE methods have been presented and applied. However, non-learning-based strategies typically struggle to demonstrate superior generalization capabilities, while learning-based strategies generally suffer from potential over- or under-enhancement due to the lack of sufficient prior knowledge. To address the challenges above, this paper presents a heuristic cooperative-competitive network, termed Co<sup>2</sup>Net. Specifically, our Co<sup>2</sup>Net integrates non-learning mechanisms into the deep learning framework to achieve information fusion from explainable prior knowledge and discernible hierarchical features, thereby facilitating promising and reasonable enhancement of degraded underwater images. Furthermore, our Co<sup>2</sup>Net adopts a hybrid convolutional neural network (CNN)-Transformer architecture, which comprises successive cooperative-competitive modules (Co<sup>2</sup>Ms) to achieve adequate extraction, representation, and transmission of both prior knowledge and discernible features. Comprehensive experiments are conducted to demonstrate the superiority and universality of our proposed Co<sup>2</sup>Net, and sufficient ablation studies are also performed to reveal the effectiveness of each component within our model. The source code is available at <uri>https://github.com/jixun-dmu/Co2Net</uri>","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"1065-1078"},"PeriodicalIF":4.8,"publicationDate":"2025-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FD-LSCIC: Frequency Decomposition-Based Learned Screen Content Image Compression 基于频率分解的学习屏幕内容图像压缩
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-09-19 DOI: 10.1109/TBC.2025.3609052
Shiqi Jiang;Hui Yuan;Shuai Li;Huanqiang Zeng;Sam Kwong
The learned image compression (LIC) methods have already surpassed traditional techniques in compressing natural scene (NS) images. However, directly applying these methods to screen content (SC) images, which possess distinct characteristics such as sharp edges, repetitive patterns, embedded text and graphics, yields suboptimal results. This paper addresses three key challenges in SC image compression: learning compact latent features, adapting quantization step sizes, and the lack of large SC datasets. To overcome these challenges, we propose a novel compression method that employs a multi-frequency two-stage octave residual block (MToRB) for feature extraction, a cascaded triple-scale feature fusion residual block (CTSFRB) for multi-scale feature integration and a multi-frequency context interaction module (MFCIM) to reduce inter-frequency correlations. Additionally, we introduce an adaptive quantization module that learns scaled uniform noise for each frequency component, enabling flexible control over quantization granularity. Furthermore, we construct a large SC image compression dataset (SDU-SCICD10K), which includes over 10,000 images spanning basic SC images, computer-rendered images, and mixed NS and SC images from both PC and mobile platforms. Experimental results demonstrate that our approach significantly improves SC image compression performance, outperforming traditional standards and state-of-the-art learning-based methods in terms of peak signal-to-noise ratio (PSNR) and multi-scale structural similarity (MS-SSIM). The source code will be opened later in https://github.com/SunshineSki/Screen-content-image-dataset/tree/main/SDU-SCICD10K
学习图像压缩(LIC)方法在自然场景图像压缩方面已经超越了传统的压缩技术。然而,直接将这些方法应用于屏幕内容(SC)图像,这些图像具有鲜明的特征,如尖锐的边缘、重复的图案、嵌入的文本和图形,会产生次优结果。本文解决了SC图像压缩中的三个关键挑战:学习紧凑的潜在特征,适应量化步长,以及缺乏大型SC数据集。为了克服这些挑战,我们提出了一种新的压缩方法,该方法使用多频两级倍频残差块(MToRB)进行特征提取,使用级联三尺度特征融合残差块(CTSFRB)进行多尺度特征集成,使用多频上下文交互模块(MFCIM)减少频间相关性。此外,我们还引入了一个自适应量化模块,该模块可以学习每个频率分量的缩放均匀噪声,从而可以灵活地控制量化粒度。此外,我们构建了一个大型SC图像压缩数据集(SDU-SCICD10K),其中包括超过10,000张图像,涵盖基本SC图像,计算机渲染图像以及来自PC和移动平台的混合NS和SC图像。实验结果表明,我们的方法显著提高了SC图像压缩性能,在峰值信噪比(PSNR)和多尺度结构相似性(MS-SSIM)方面优于传统标准和最先进的基于学习的方法。稍后将在https://github.com/SunshineSki/Screen-content-image-dataset/tree/main/SDU-SCICD10K中打开源代码
{"title":"FD-LSCIC: Frequency Decomposition-Based Learned Screen Content Image Compression","authors":"Shiqi Jiang;Hui Yuan;Shuai Li;Huanqiang Zeng;Sam Kwong","doi":"10.1109/TBC.2025.3609052","DOIUrl":"https://doi.org/10.1109/TBC.2025.3609052","url":null,"abstract":"The learned image compression (LIC) methods have already surpassed traditional techniques in compressing natural scene (NS) images. However, directly applying these methods to screen content (SC) images, which possess distinct characteristics such as sharp edges, repetitive patterns, embedded text and graphics, yields suboptimal results. This paper addresses three key challenges in SC image compression: learning compact latent features, adapting quantization step sizes, and the lack of large SC datasets. To overcome these challenges, we propose a novel compression method that employs a multi-frequency two-stage octave residual block (MToRB) for feature extraction, a cascaded triple-scale feature fusion residual block (CTSFRB) for multi-scale feature integration and a multi-frequency context interaction module (MFCIM) to reduce inter-frequency correlations. Additionally, we introduce an adaptive quantization module that learns scaled uniform noise for each frequency component, enabling flexible control over quantization granularity. Furthermore, we construct a large SC image compression dataset (SDU-SCICD10K), which includes over 10,000 images spanning basic SC images, computer-rendered images, and mixed NS and SC images from both PC and mobile platforms. Experimental results demonstrate that our approach significantly improves SC image compression performance, outperforming traditional standards and state-of-the-art learning-based methods in terms of peak signal-to-noise ratio (PSNR) and multi-scale structural similarity (MS-SSIM). The source code will be opened later in <uri>https://github.com/SunshineSki/Screen-content-image-dataset/tree/main/SDU-SCICD10K</uri>","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"1034-1047"},"PeriodicalIF":4.8,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rate-Distortion-Optimization-Driven Quantization Parameter Cascading for Screen Content Video Coding Using VVC 基于VVC的屏幕内容视频编码的率失真优化驱动量化参数级联
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-09-18 DOI: 10.1109/TBC.2025.3609039
Yanchao Gong;Yinghua Li;Baogui Li;Kaifang Yang;Nam Ling
Screen content videos (SCVs) have been widely used in television broadcasting, video conferencing, online education, and other fields. VVC is a new generation video coding standard for SCVs, where the quantization parameter (QP) is one of the key coding parameters that significantly affects the coding efficiency of SCVs. The method of selecting optimal QPs for pictures located at different temporal layers is called quantization parameter cascading (QPC). The QPC method recommended by VVC test model, i.e., VTM, does not take into account the impact of video content characteristics on QP selection, resulting in lower coding efficiency of SCVs. To address this issue, a QPC method driven by the rate-distortion (R-D) optimization for SCVs (QPC-SCV), was proposed. Combining experiments and the hybrid coding framework principles of VVC, a novel R-D cost function applicable to the SCV coding characteristics was first established and validated, where the spatiotemporal content characteristics of SCVs were evaluated to predict the model parameters. Then, a video motion form classification and particle swarm optimization were further proposed to effectively solve the R-D cost function and obtain optimized QPs. Compared with the QPC recommended by VTM, the QPC-SCV improves the coding efficiency of SCVs while reducing the coding time. For all test sequences, the average BD-rate corresponding to the QPC-SCV is −6.90%, and the average coding time is reduced by 4.56%.
屏幕内容视频(scv)已广泛应用于电视广播、视频会议、在线教育等领域。VVC是scv的新一代视频编码标准,其中量化参数(QP)是影响scv编码效率的关键编码参数之一。为不同时间层的图像选择最优qp的方法称为量化参数级联(QPC)。VVC测试模型推荐的QPC方法,即VTM,没有考虑视频内容特征对QP选择的影响,导致scv编码效率较低。为了解决这一问题,提出了一种基于速率失真(R-D)优化的QPC方法(QPC- scv)。结合实验和VVC的混合编码框架原理,首先建立并验证了一种适用于SCV编码特征的新型R-D代价函数,并对SCV的时空内容特征进行了评估,以预测模型参数。然后,进一步提出视频运动形式分类和粒子群优化方法,有效求解R-D代价函数,得到最优qp。与VTM推荐的QPC相比,QPC- scv在减少编码时间的同时提高了scv的编码效率。对于所有测试序列,QPC-SCV对应的平均bd率为- 6.90%,平均编码时间减少4.56%。
{"title":"Rate-Distortion-Optimization-Driven Quantization Parameter Cascading for Screen Content Video Coding Using VVC","authors":"Yanchao Gong;Yinghua Li;Baogui Li;Kaifang Yang;Nam Ling","doi":"10.1109/TBC.2025.3609039","DOIUrl":"https://doi.org/10.1109/TBC.2025.3609039","url":null,"abstract":"Screen content videos (SCVs) have been widely used in television broadcasting, video conferencing, online education, and other fields. VVC is a new generation video coding standard for SCVs, where the quantization parameter (QP) is one of the key coding parameters that significantly affects the coding efficiency of SCVs. The method of selecting optimal QPs for pictures located at different temporal layers is called quantization parameter cascading (QPC). The QPC method recommended by VVC test model, i.e., VTM, does not take into account the impact of video content characteristics on QP selection, resulting in lower coding efficiency of SCVs. To address this issue, a QPC method driven by the rate-distortion (R-D) optimization for SCVs (QPC-SCV), was proposed. Combining experiments and the hybrid coding framework principles of VVC, a novel R-D cost function applicable to the SCV coding characteristics was first established and validated, where the spatiotemporal content characteristics of SCVs were evaluated to predict the model parameters. Then, a video motion form classification and particle swarm optimization were further proposed to effectively solve the R-D cost function and obtain optimized QPs. Compared with the QPC recommended by VTM, the QPC-SCV improves the coding efficiency of SCVs while reducing the coding time. For all test sequences, the average BD-rate corresponding to the QPC-SCV is −6.90%, and the average coding time is reduced by 4.56%.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 4","pages":"993-1010"},"PeriodicalIF":4.8,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145766195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Broadcasting Information for Readers and Authors 面向读者和作者的广播信息IEEE汇刊
IF 4.8 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-09-05 DOI: 10.1109/TBC.2025.3603346
{"title":"IEEE Transactions on Broadcasting Information for Readers and Authors","authors":"","doi":"10.1109/TBC.2025.3603346","DOIUrl":"https://doi.org/10.1109/TBC.2025.3603346","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 3","pages":"C3-C4"},"PeriodicalIF":4.8,"publicationDate":"2025-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11152557","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Broadcasting
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1