首页 > 最新文献

IEEE Transactions on Broadcasting最新文献

英文 中文
Rate-Compatible Length-Scalable Quasi-Cyclic Spatially-Coupled LDPC Codes 速率兼容的长度可伸缩准循环空间耦合LDPC码
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-12 DOI: 10.1109/TBC.2024.3511916
Zhitong He;Kewu Peng;Jian Song
The capability of QC-SC-LDPC codes to be employed in broadcasting systems has been studied in previous research. However, the implementation-oriented features such as rate-compatibility and length-scalability for QC-SC-LDPC codes have not been well studied yet. In this paper, we first propose a new implementation-oriented structure of QC-SC-LDPC codes for broadcasting systems, with support for rate-compatibility and length-scalability. Then, the three-dimensional (3D-) grid-based (G-) progressive edge growth and lifting (PEGL) method is proposed to construct QC-SC-LDPC codes with that structure, which can achieve desirable performance across different code rates and code lengths within the given design complexity. Finally, a family of rate-compatible length-scalable QC-SC-LDPC codes are constructed via the 3D-G-PEGL method, and simulation results demonstrate the effectiveness of that method. Furthermore, the scaling behaviors of QC-SC-LDPC codes are observed from the provided simulation results.
在以往的研究中,对QC-SC-LDPC码在广播系统中的应用能力进行了研究。然而,QC-SC-LDPC码的码率兼容性和码长可扩展性等面向实现的特性还没有得到很好的研究。在本文中,我们首先提出了一种新的面向实现的广播系统QC-SC-LDPC码结构,支持速率兼容性和长度可扩展性。然后,提出了基于三维(3D-)网格(G-)渐进式边缘生长和提升(PEGL)方法,利用该结构构建QC-SC-LDPC代码,在给定的设计复杂度下,在不同码率和码长下都能获得理想的性能。最后,利用3D-G-PEGL方法构建了一组速率兼容的长度可扩展QC-SC-LDPC码,仿真结果验证了该方法的有效性。此外,通过仿真结果观察了QC-SC-LDPC码的标度行为。
{"title":"Rate-Compatible Length-Scalable Quasi-Cyclic Spatially-Coupled LDPC Codes","authors":"Zhitong He;Kewu Peng;Jian Song","doi":"10.1109/TBC.2024.3511916","DOIUrl":"https://doi.org/10.1109/TBC.2024.3511916","url":null,"abstract":"The capability of QC-SC-LDPC codes to be employed in broadcasting systems has been studied in previous research. However, the implementation-oriented features such as rate-compatibility and length-scalability for QC-SC-LDPC codes have not been well studied yet. In this paper, we first propose a new implementation-oriented structure of QC-SC-LDPC codes for broadcasting systems, with support for rate-compatibility and length-scalability. Then, the three-dimensional (3D-) grid-based (G-) progressive edge growth and lifting (PEGL) method is proposed to construct QC-SC-LDPC codes with that structure, which can achieve desirable performance across different code rates and code lengths within the given design complexity. Finally, a family of rate-compatible length-scalable QC-SC-LDPC codes are constructed via the 3D-G-PEGL method, and simulation results demonstrate the effectiveness of that method. Furthermore, the scaling behaviors of QC-SC-LDPC codes are observed from the provided simulation results.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"81-95"},"PeriodicalIF":3.2,"publicationDate":"2024-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Broadcasting Publication Information IEEE广播出版信息汇刊
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-11 DOI: 10.1109/TBC.2024.3495315
{"title":"IEEE Transactions on Broadcasting Publication Information","authors":"","doi":"10.1109/TBC.2024.3495315","DOIUrl":"https://doi.org/10.1109/TBC.2024.3495315","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 4","pages":"C2-C2"},"PeriodicalIF":3.2,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10791069","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
2024 Scott Helt Memorial Award for the Best Paper Published in the IEEE Transactions on Broadcasting 2024年斯科特·海尔特纪念奖,在IEEE广播事务中发表的最佳论文
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-11 DOI: 10.1109/TBC.2024.3492772
Presents the recipients of (Scott Helt Memorial Award) awards for (2024).
颁发(斯科特·海尔特纪念奖)(2024年)的获奖者。
{"title":"2024 Scott Helt Memorial Award for the Best Paper Published in the IEEE Transactions on Broadcasting","authors":"","doi":"10.1109/TBC.2024.3492772","DOIUrl":"https://doi.org/10.1109/TBC.2024.3492772","url":null,"abstract":"Presents the recipients of (Scott Helt Memorial Award) awards for (2024).","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 4","pages":"1316-1317"},"PeriodicalIF":3.2,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10790558","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Broadcasting Information for Authors IEEE作者广播信息汇刊
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-11 DOI: 10.1109/TBC.2024.3495317
{"title":"IEEE Transactions on Broadcasting Information for Authors","authors":"","doi":"10.1109/TBC.2024.3495317","DOIUrl":"https://doi.org/10.1109/TBC.2024.3495317","url":null,"abstract":"","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"70 4","pages":"C3-C4"},"PeriodicalIF":3.2,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10790559","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142810674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Omnidirectional Image Quality Assessment With Mutual Distillation 基于互蒸馏的全方位图像质量评价
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-03 DOI: 10.1109/TBC.2024.3503435
Pingchuan Ma;Lixiong Liu;Chengzhi Xiao;Dong Xu
There exists complementary relationship between different projection formats of omnidirectional images. However, most existing omnidirectional image quality assessment (OIQA) works only operate solely on single projection format, and rarely explore the solutions on different projection formats. To this end, we propose a mutual distillation-based omnidirectional image quality assessment method, abbreviated as MD-OIQA. The MD-OIQA explores the complementary relationship between different projection formats to improve the feature representation of omnidirectional images for quality prediction. Specifically, we separately feed equirectangular projection (ERP) and cubemap projection (CMP) images into two peer student networks to capture quality-aware features of specific projection contents. Meanwhile, we propose a self-adaptive mutual distillation module (SAMDM) that deploys mutual distillation at multiple network stages to achieve the mutual learning between the two networks. The proposed SAMDM is able to capture the useful knowledge from the dynamic optimized networks to improve the effect of mutual distillation by enhancing the feature interactions through a deep cross network and generating masks to efficiently capture the complementary information from different projection contents. Finally, the features extracted from single projection content are used for quality prediction. The experiment results on three public databases demonstrate that the proposed method can efficiently improve the model representation capability and achieves superior performance.
全向图像的不同投影格式之间存在互补关系。然而,现有的全向图像质量评估(OIQA)大多只适用于单一的投影格式,很少探索不同投影格式下的解决方案。为此,我们提出了一种基于互蒸馏的全方位图像质量评价方法,简称MD-OIQA。MD-OIQA探索不同投影格式之间的互补关系,以改善全向图像的特征表示,用于质量预测。具体来说,我们分别将等矩形投影(ERP)和立方体映射投影(CMP)图像馈送到两个对等学生网络中,以捕获特定投影内容的质量感知特征。同时,我们提出了一种自适应互蒸馏模块(SAMDM),该模块在多个网络阶段部署互蒸馏,以实现两个网络之间的相互学习。该方法通过深度交叉网络增强特征间的相互作用,并通过生成掩模来有效地捕获不同投影内容的互补信息,从而从动态优化的网络中捕获有用的知识,提高相互蒸馏的效果。最后,利用从单个投影内容中提取的特征进行质量预测。在三个公共数据库上的实验结果表明,该方法可以有效地提高模型表示能力,取得了较好的性能。
{"title":"Omnidirectional Image Quality Assessment With Mutual Distillation","authors":"Pingchuan Ma;Lixiong Liu;Chengzhi Xiao;Dong Xu","doi":"10.1109/TBC.2024.3503435","DOIUrl":"https://doi.org/10.1109/TBC.2024.3503435","url":null,"abstract":"There exists complementary relationship between different projection formats of omnidirectional images. However, most existing omnidirectional image quality assessment (OIQA) works only operate solely on single projection format, and rarely explore the solutions on different projection formats. To this end, we propose a mutual distillation-based omnidirectional image quality assessment method, abbreviated as MD-OIQA. The MD-OIQA explores the complementary relationship between different projection formats to improve the feature representation of omnidirectional images for quality prediction. Specifically, we separately feed equirectangular projection (ERP) and cubemap projection (CMP) images into two peer student networks to capture quality-aware features of specific projection contents. Meanwhile, we propose a self-adaptive mutual distillation module (SAMDM) that deploys mutual distillation at multiple network stages to achieve the mutual learning between the two networks. The proposed SAMDM is able to capture the useful knowledge from the dynamic optimized networks to improve the effect of mutual distillation by enhancing the feature interactions through a deep cross network and generating masks to efficiently capture the complementary information from different projection contents. Finally, the features extracted from single projection content are used for quality prediction. The experiment results on three public databases demonstrate that the proposed method can efficiently improve the model representation capability and achieves superior performance.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"264-276"},"PeriodicalIF":3.2,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VMG: Rethinking U-Net Architecture for Video Super-Resolution VMG:重新思考视频超分辨率的U-Net架构
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-21 DOI: 10.1109/TBC.2024.3486967
Jun Tang;Lele Niu;Linlin Liu;Hang Dai;Yong Ding
The U-Net architecture has exhibited significant efficacy across various vision tasks, yet its adaptation for Video Super-Resolution (VSR) remains underexplored. While the Video Restoration Transformer (VRT) introduced U-Net into the VSR domain, it poses challenges due to intricate design and substantial computational overhead. In this paper, we present VMG, a streamlined framework tailored for VSR. Through empirical analysis, we identify the crucial stages of the U-Net architecture contributing to performance enhancement in VSR tasks. Our optimized architecture substantially reduces model parameters and complexity while improving performance. Additionally, we introduce two key modules, namely the Gated MLP-like Mixer (GMM) and the Flow-Guided cross-attention Mixer (FGM), designed to enhance spatial and temporal feature aggregation. GMM dynamically encodes spatial correlations with linear complexity in space and time, and FGM leverages optical flow to capture motion variation and implement sparse attention to efficiently aggregate temporally related information. Extensive experiments demonstrate that VMG achieves nearly 70% reduction in GPU memory usage, 30% fewer parameters, and 10% lower computational complexity (FLOPs) compared to VRT, while yielding highly competitive or superior results across four benchmark datasets. Qualitative assessments reveal VMG’s ability to preserve remarkable details and sharp structures in the reconstructed videos. The code and pre-trained models are available at https://github.com/EasyVision-Ton/VMG.
U-Net架构在各种视觉任务中表现出显著的有效性,但其对视频超分辨率(VSR)的适应性仍有待探索。虽然视频恢复变压器(VRT)将U-Net引入了VSR领域,但由于复杂的设计和大量的计算开销,它带来了挑战。在本文中,我们提出了VMG,一个为VSR量身定制的流线型框架。通过实证分析,我们确定了有助于提高VSR任务性能的U-Net架构的关键阶段。我们优化的架构大大降低了模型参数和复杂性,同时提高了性能。此外,我们还介绍了两个关键模块,即门控MLP-like Mixer (GMM)和Flow-Guided cross-attention Mixer (FGM),旨在增强时空特征聚合。GMM动态编码具有空间和时间线性复杂性的空间相关性,FGM利用光流捕获运动变化并实现稀疏关注以有效聚合时间相关信息。大量实验表明,与VRT相比,VMG在GPU内存使用方面减少了近70%,参数减少了30%,计算复杂度(FLOPs)降低了10%,同时在四个基准数据集上产生了极具竞争力或更优的结果。定性评估显示VMG能够在重建的视频中保留显著的细节和清晰的结构。代码和预训练模型可在https://github.com/EasyVision-Ton/VMG上获得。
{"title":"VMG: Rethinking U-Net Architecture for Video Super-Resolution","authors":"Jun Tang;Lele Niu;Linlin Liu;Hang Dai;Yong Ding","doi":"10.1109/TBC.2024.3486967","DOIUrl":"https://doi.org/10.1109/TBC.2024.3486967","url":null,"abstract":"The U-Net architecture has exhibited significant efficacy across various vision tasks, yet its adaptation for Video Super-Resolution (VSR) remains underexplored. While the Video Restoration Transformer (VRT) introduced U-Net into the VSR domain, it poses challenges due to intricate design and substantial computational overhead. In this paper, we present VMG, a streamlined framework tailored for VSR. Through empirical analysis, we identify the crucial stages of the U-Net architecture contributing to performance enhancement in VSR tasks. Our optimized architecture substantially reduces model parameters and complexity while improving performance. Additionally, we introduce two key modules, namely the Gated MLP-like Mixer (GMM) and the Flow-Guided cross-attention Mixer (FGM), designed to enhance spatial and temporal feature aggregation. GMM dynamically encodes spatial correlations with linear complexity in space and time, and FGM leverages optical flow to capture motion variation and implement sparse attention to efficiently aggregate temporally related information. Extensive experiments demonstrate that VMG achieves nearly 70% reduction in GPU memory usage, 30% fewer parameters, and 10% lower computational complexity (FLOPs) compared to VRT, while yielding highly competitive or superior results across four benchmark datasets. Qualitative assessments reveal VMG’s ability to preserve remarkable details and sharp structures in the reconstructed videos. The code and pre-trained models are available at <uri>https://github.com/EasyVision-Ton/VMG</uri>.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"334-349"},"PeriodicalIF":3.2,"publicationDate":"2024-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparative Assessment of Physical Layer Performance: ATSC 3.0 vs. 5G Broadcast in Laboratory and Field Tests 物理层性能的比较评估:实验室和现场测试中的ATSC 3.0与5G广播
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-21 DOI: 10.1109/TBC.2024.3482183
Sunhyoung Kwon;Seok-Ki Ahn;Sungjun Ahn;Sungho Jeon;Sesh Simha;Mark Aitken;Anindya Saha;Prashant M. Maru;Parag Naik;Sung-Ik Park
This paper presents a comparative analysis of the physical layer performance of ATSC 3.0 and 3GPP 5G Broadcast through comprehensive laboratory and field tests. The study evaluates various reception scenarios, including fixed and mobile environments and various channel conditions, such as additive white Gaussian noise and mobile channels. Key performance metrics such as threshold of visibility (ToV) and erroneous second ratio (ESR) are measured to assess the reception quality of each standard. The results demonstrate that ATSC 3.0 generally outperforms 5G Broadcast due to its advanced bit-interleaved coded modulation and time interleaving techniques, effectively mitigating burst errors in mobile channels.
通过综合实验室和现场测试,对ATSC 3.0和3GPP 5G广播的物理层性能进行了对比分析。该研究评估了各种接收场景,包括固定和移动环境以及各种信道条件,如加性高斯白噪声和移动信道。测量关键性能指标,如可见阈值(ToV)和错误秒比(ESR),以评估每个标准的接收质量。结果表明,由于采用了先进的比特交错编码调制和时间交错技术,ATSC 3.0总体上优于5G广播,有效地减轻了移动信道中的突发错误。
{"title":"Comparative Assessment of Physical Layer Performance: ATSC 3.0 vs. 5G Broadcast in Laboratory and Field Tests","authors":"Sunhyoung Kwon;Seok-Ki Ahn;Sungjun Ahn;Sungho Jeon;Sesh Simha;Mark Aitken;Anindya Saha;Prashant M. Maru;Parag Naik;Sung-Ik Park","doi":"10.1109/TBC.2024.3482183","DOIUrl":"https://doi.org/10.1109/TBC.2024.3482183","url":null,"abstract":"This paper presents a comparative analysis of the physical layer performance of ATSC 3.0 and 3GPP 5G Broadcast through comprehensive laboratory and field tests. The study evaluates various reception scenarios, including fixed and mobile environments and various channel conditions, such as additive white Gaussian noise and mobile channels. Key performance metrics such as threshold of visibility (ToV) and erroneous second ratio (ESR) are measured to assess the reception quality of each standard. The results demonstrate that ATSC 3.0 generally outperforms 5G Broadcast due to its advanced bit-interleaved coded modulation and time interleaving techniques, effectively mitigating burst errors in mobile channels.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"2-10"},"PeriodicalIF":3.2,"publicationDate":"2024-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised 3D Point Cloud Reconstruction via Exploring Multi-View Consistency and Complementarity 基于多视图一致性和互补性的无监督三维点云重建
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-12 DOI: 10.1109/TBC.2024.3484269
Jiahui Song;Yonghong Hou;Bo Peng;Tianyi Qin;Qingming Huang;Jianjun Lei
Unsupervised 3D point cloud reconstruction has increasingly played an important role in 3D multimedia broadcasting, virtual reality, and augmented reality. Considering that multiple views collectively provide abundant object geometry and structure information, this paper proposes a novel Unsupervised Multi-View 3D Point Cloud Reconstruction Network (UMPR-Net) to reconstruct high-quality 3D point clouds by effectively exploring multi-view consistency and complementarity. In particular, by effectively perceiving the consistency of local object information contained in different views, a consistency-aware point cloud reconstruction module is designed to reconstruct 3D point clouds for each individual view. Additionally, a complementarity-oriented point cloud fusion module is presented to aggregate reliable complementary information explored from multiple point clouds corresponding to diverse views, thus ultimately obtaining a refined 3D point cloud. By projecting reconstructed 3D point clouds onto 2D planes and subsequently constraining the consistency between 2D projections and 2D supervision, the proposed UMPR-Net is encouraged to reconstruct high-quality 3D point clouds from multiple views. Experimental results on the synthetic and real-world datasets have validated the effectiveness of the proposed UMPR-Net.
无监督三维点云重建在三维多媒体广播、虚拟现实、增强现实等领域发挥着越来越重要的作用。考虑到多个视图共同提供了丰富的物体几何和结构信息,本文提出了一种新的无监督多视图三维点云重建网络(UMPR-Net),通过有效地探索多视图的一致性和互补性来重建高质量的三维点云。特别是,通过有效感知不同视图中包含的局部目标信息的一致性,设计了一致性感知点云重构模块,对每个单独视图进行三维点云重构。此外,提出了面向互补性的点云融合模块,对不同视点对应的多个点云探索出的可靠互补信息进行聚合,最终得到精细化的三维点云。通过将重建的三维点云投影到二维平面上,然后约束二维投影和二维监督之间的一致性,所提出的UMPR-Net被鼓励从多个视图重建高质量的三维点云。在合成数据集和实际数据集上的实验结果验证了所提出的UMPR-Net的有效性。
{"title":"Unsupervised 3D Point Cloud Reconstruction via Exploring Multi-View Consistency and Complementarity","authors":"Jiahui Song;Yonghong Hou;Bo Peng;Tianyi Qin;Qingming Huang;Jianjun Lei","doi":"10.1109/TBC.2024.3484269","DOIUrl":"https://doi.org/10.1109/TBC.2024.3484269","url":null,"abstract":"Unsupervised 3D point cloud reconstruction has increasingly played an important role in 3D multimedia broadcasting, virtual reality, and augmented reality. Considering that multiple views collectively provide abundant object geometry and structure information, this paper proposes a novel <underline>U</u>nsupervised <underline>M</u>ulti-View 3D <underline>P</u>oint Cloud <underline>R</u>econstruction <underline>Net</u>work (UMPR-Net) to reconstruct high-quality 3D point clouds by effectively exploring multi-view consistency and complementarity. In particular, by effectively perceiving the consistency of local object information contained in different views, a consistency-aware point cloud reconstruction module is designed to reconstruct 3D point clouds for each individual view. Additionally, a complementarity-oriented point cloud fusion module is presented to aggregate reliable complementary information explored from multiple point clouds corresponding to diverse views, thus ultimately obtaining a refined 3D point cloud. By projecting reconstructed 3D point clouds onto 2D planes and subsequently constraining the consistency between 2D projections and 2D supervision, the proposed UMPR-Net is encouraged to reconstruct high-quality 3D point clouds from multiple views. Experimental results on the synthetic and real-world datasets have validated the effectiveness of the proposed UMPR-Net.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"193-202"},"PeriodicalIF":3.2,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perception- and Fidelity-Aware Reduced-Reference Super-Resolution Image Quality Assessment 感知和保真度感知的低参考超分辨率图像质量评估
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-04 DOI: 10.1109/TBC.2024.3475820
Xinying Lin;Xuyang Liu;Hong Yang;Xiaohai He;Honggang Chen
With the advent of image super-resolution (SR) algorithms, how to evaluate the quality of generated SR images has become an urgent task. Although full-reference methods perform well in SR image quality assessment (SR-IQA), their reliance on high-resolution (HR) images limits their practical applicability. Leveraging available reconstruction information as much as possible for SR-IQA, such as low-resolution (LR) images and the scale factors, is a promising way to enhance assessment performance for SR-IQA without HR for reference. In this paper, we attempt to evaluate the perceptual quality and reconstruction fidelity of SR images considering LR images and scale factors. Specifically, we propose a novel dual-branch reduced-reference SR-IQA network, i.e., Perception- and Fidelity-aware SR-IQA (PFIQA). The perception-aware branch evaluates the perceptual quality of SR images by leveraging the merits of global modeling of Vision Transformer (ViT) and local relation of ResNet, and incorporating the scale factor to enable comprehensive visual perception. Meanwhile, the fidelity-aware branch assesses the reconstruction fidelity between LR and SR images through their visual perception. The combination of the two branches substantially aligns with the human visual system, enabling a comprehensive SR image evaluation. Experimental results indicate that our PFIQA outperforms current state-of-the-art models across three widely-used SR-IQA benchmarks. Notably, PFIQA excels in assessing the quality of real-world SR images. Our code is available at https://github.com/xinyouu/PFIQA.
随着图像超分辨率(SR)算法的出现,如何对生成的SR图像进行质量评价成为一个迫切需要解决的问题。尽管全参考方法在SR- iqa中表现良好,但其对高分辨率(HR)图像的依赖限制了其实际适用性。尽可能利用SR-IQA的现有重建信息,如低分辨率(LR)图像和比例因子,是在没有HR参考的情况下提高SR-IQA评估性能的一种有希望的方法。在本文中,我们尝试考虑LR图像和尺度因子来评估SR图像的感知质量和重建保真度。具体来说,我们提出了一种新的双分支简化参考SR-IQA网络,即感知和保真度感知SR-IQA (PFIQA)。感知分支利用视觉变换(Vision Transformer, ViT)的全局建模和ResNet的局部关系的优点,结合比例因子对SR图像的感知质量进行评价,实现全面的视觉感知。同时,保真度感知分支通过视觉感知来评估LR和SR图像之间的重建保真度。这两个分支的结合基本上与人类视觉系统保持一致,从而实现了全面的SR图像评估。实验结果表明,我们的PFIQA在三个广泛使用的SR-IQA基准测试中优于当前最先进的模型。值得注意的是,PFIQA在评估真实SR图像的质量方面表现出色。我们的代码可在https://github.com/xinyouu/PFIQA上获得。
{"title":"Perception- and Fidelity-Aware Reduced-Reference Super-Resolution Image Quality Assessment","authors":"Xinying Lin;Xuyang Liu;Hong Yang;Xiaohai He;Honggang Chen","doi":"10.1109/TBC.2024.3475820","DOIUrl":"https://doi.org/10.1109/TBC.2024.3475820","url":null,"abstract":"With the advent of image super-resolution (SR) algorithms, how to evaluate the quality of generated SR images has become an urgent task. Although full-reference methods perform well in SR image quality assessment (SR-IQA), their reliance on high-resolution (HR) images limits their practical applicability. Leveraging available reconstruction information as much as possible for SR-IQA, such as low-resolution (LR) images and the scale factors, is a promising way to enhance assessment performance for SR-IQA without HR for reference. In this paper, we attempt to evaluate the perceptual quality and reconstruction fidelity of SR images considering LR images and scale factors. Specifically, we propose a novel dual-branch reduced-reference SR-IQA network, <italic>i.e.</i>, Perception- and Fidelity-aware SR-IQA (PFIQA). The perception-aware branch evaluates the perceptual quality of SR images by leveraging the merits of global modeling of Vision Transformer (ViT) and local relation of ResNet, and incorporating the scale factor to enable comprehensive visual perception. Meanwhile, the fidelity-aware branch assesses the reconstruction fidelity between LR and SR images through their visual perception. The combination of the two branches substantially aligns with the human visual system, enabling a comprehensive SR image evaluation. Experimental results indicate that our PFIQA outperforms current state-of-the-art models across three widely-used SR-IQA benchmarks. Notably, PFIQA excels in assessing the quality of real-world SR images. Our code is available at <uri>https://github.com/xinyouu/PFIQA</uri>.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"323-333"},"PeriodicalIF":3.2,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
No-Reference Point Cloud Quality Assessment Through Structure Sampling and Clustering Based on Graph 基于图的结构采样聚类无参考点云质量评价
IF 3.2 1区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-10-29 DOI: 10.1109/TBC.2024.3482173
Xinqiang Wu;Zhouyan He;Gangyi Jiang;Mei Yu;Yang Song;Ting Luo
As a popular multimedia representation, 3D Point Clouds (PC) inevitably encounter distortion during their acquisition, processing, coding, and transmission, resulting in visual quality degradation. Therefore, it is critical to propose a Point Cloud Quality Assessment (PCQA) method to perceive the visual quality of PC. In this paper, we propose a no-reference PCQA method through structure sampling and clustering based on graph, which consists of two-stage pre-processing, quality feature extraction, attention-based feature fusion, and feature regression. For pre-processing, considering the Human Visual System (HVS) tendency to perceive distortions in both the global structure and local details of PCs, a two-stage sampling strategy is introduced. Specifically, to adapt to the irregular structure of PCs, it introduces structural key point sampling and local cluster to capture both global and local information, respectively, thereby facilitating more effective learning of distortion features. Then, in quality feature extraction, two modules are designed based on the two-stage pre-processing results (i.e., Global Feature Extraction (GFE) and Local Feature Extraction (LFE)) to respectively extract global and local quality features. Additionally, for attention-based feature fusion, a Unified Feature Integrator (UFI) module is proposed. This module enhances quality perception capability by integrating global features and individual local quality features and introduces the Transformer to interact with the integrated quality features. Finally, feature regression is conducted to map the final features into the quality score. The performance of the proposed method is tested on four publicly available databases, and the experimental results show that the proposed method is superior compared with existing state-of-the-art no-reference PCQA methods in most cases.
作为一种流行的多媒体表现形式,三维点云在其获取、处理、编码和传输过程中不可避免地会遇到失真,从而导致视觉质量下降。因此,提出一种点云质量评估(PCQA)方法来感知点云的视觉质量至关重要。本文提出了一种基于图的结构采样聚类的无参考PCQA方法,该方法由两阶段预处理、质量特征提取、基于注意力的特征融合和特征回归组成。在预处理方面,考虑到人类视觉系统(HVS)倾向于感知pc的全局结构和局部细节的扭曲,引入了一种两阶段采样策略。具体来说,为了适应pc机的不规则结构,引入了结构关键点采样和局部聚类,分别捕获全局和局部信息,从而更有效地学习畸变特征。然后,在质量特征提取方面,基于两阶段预处理结果设计了Global feature extraction (GFE)和Local feature extraction (LFE)两个模块,分别提取全局和局部质量特征。此外,针对基于注意力的特征融合,提出了统一特征集成(UFI)模块。该模块通过集成全局特征和单个局部质量特征来增强质量感知能力,并引入Transformer与集成质量特征进行交互。最后,进行特征回归,将最终的特征映射到质量分数中。在4个公开的数据库上测试了该方法的性能,实验结果表明,在大多数情况下,该方法优于现有的无参考PCQA方法。
{"title":"No-Reference Point Cloud Quality Assessment Through Structure Sampling and Clustering Based on Graph","authors":"Xinqiang Wu;Zhouyan He;Gangyi Jiang;Mei Yu;Yang Song;Ting Luo","doi":"10.1109/TBC.2024.3482173","DOIUrl":"https://doi.org/10.1109/TBC.2024.3482173","url":null,"abstract":"As a popular multimedia representation, 3D Point Clouds (PC) inevitably encounter distortion during their acquisition, processing, coding, and transmission, resulting in visual quality degradation. Therefore, it is critical to propose a Point Cloud Quality Assessment (PCQA) method to perceive the visual quality of PC. In this paper, we propose a no-reference PCQA method through structure sampling and clustering based on graph, which consists of two-stage pre-processing, quality feature extraction, attention-based feature fusion, and feature regression. For pre-processing, considering the Human Visual System (HVS) tendency to perceive distortions in both the global structure and local details of PCs, a two-stage sampling strategy is introduced. Specifically, to adapt to the irregular structure of PCs, it introduces structural key point sampling and local cluster to capture both global and local information, respectively, thereby facilitating more effective learning of distortion features. Then, in quality feature extraction, two modules are designed based on the two-stage pre-processing results (i.e., Global Feature Extraction (GFE) and Local Feature Extraction (LFE)) to respectively extract global and local quality features. Additionally, for attention-based feature fusion, a Unified Feature Integrator (UFI) module is proposed. This module enhances quality perception capability by integrating global features and individual local quality features and introduces the Transformer to interact with the integrated quality features. Finally, feature regression is conducted to map the final features into the quality score. The performance of the proposed method is tested on four publicly available databases, and the experimental results show that the proposed method is superior compared with existing state-of-the-art no-reference PCQA methods in most cases.","PeriodicalId":13159,"journal":{"name":"IEEE Transactions on Broadcasting","volume":"71 1","pages":"307-322"},"PeriodicalIF":3.2,"publicationDate":"2024-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143553458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Broadcasting
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1