首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
Camera-Based 3D Semantic Scene Completion With Sparse Guidance Network 利用稀疏制导网络实现基于摄像头的 3D 语义场景补全
Jianbiao Mei;Yu Yang;Mengmeng Wang;Junyu Zhu;Jongwon Ra;Yukai Ma;Laijian Li;Yong Liu
Semantic scene completion (SSC) aims to predict the semantic occupancy of each voxel in the entire 3D scene from limited observations, which is an emerging and critical task for autonomous driving. Recently, many studies have turned to camera-based SSC solutions due to the richer visual cues and cost-effectiveness of cameras. However, existing methods usually rely on sophisticated and heavy 3D models to process the lifted 3D features directly, which are not discriminative enough for clear segmentation boundaries. In this paper, we adopt the dense-sparse-dense design and propose a one-stage camera-based SSC framework, termed SGN, to propagate semantics from the semantic-aware seed voxels to the whole scene based on spatial geometry cues. Firstly, to exploit depth-aware context and dynamically select sparse seed voxels, we redesign the sparse voxel proposal network to process points generated by depth prediction directly with the coarse-to-fine paradigm. Furthermore, by designing hybrid guidance (sparse semantic and geometry guidance) and effective voxel aggregation for spatial geometry cues, we enhance the feature separation between different categories and expedite the convergence of semantic propagation. Finally, we devise the multi-scale semantic propagation module for flexible receptive fields while reducing the computation resources. Extensive experimental results on the SemanticKITTI and SSCBench-KITTI-360 datasets demonstrate the superiority of our SGN over existing state-of-the-art methods. And even our lightweight version SGN-L achieves notable scores of 14.80% mIoU and 45.45% IoU on SeamnticKITTI validation with only 12.5 M parameters and 7.16 G training memory. Code is available at https://github.com/Jieqianyu/SGN.
语义场景补全(SSC)旨在从有限的观测数据中预测整个三维场景中每个体素的语义占位情况,这是自动驾驶的一项新兴而关键的任务。最近,由于摄像头具有更丰富的视觉线索和成本效益,许多研究转向了基于摄像头的 SSC 解决方案。然而,现有的方法通常依赖于复杂而厚重的三维模型来直接处理提取的三维特征,而这些特征的判别能力不足以实现清晰的分割边界。在本文中,我们采用密集-稀疏-密集的设计,提出了一种基于摄像头的单级 SSC 框架(称为 SGN),根据空间几何线索将语义从语义感知种子体素传播到整个场景。首先,为了利用深度感知上下文并动态选择稀疏种子体素,我们重新设计了稀疏体素建议网络,以粗到细的范式直接处理深度预测生成的点。此外,通过设计混合引导(稀疏语义和几何引导)和有效的空间几何线索体素聚合,我们增强了不同类别之间的特征分离,并加快了语义传播的收敛。最后,我们为灵活的感受野设计了多尺度语义传播模块,同时减少了计算资源。在 SemanticKITTI 和 SSCBench-KITTI-360 数据集上的大量实验结果表明,我们的 SGN 优于现有的先进方法。即使是我们的轻量级版本SGN-L,在SeamnticKITTI验证中也取得了14.80% mIoU和45.45% IoU的显著成绩,而参数只有12.5 M,训练内存只有7.16 G。代码见 https://github.com/Jieqianyu/SGN。
{"title":"Camera-Based 3D Semantic Scene Completion With Sparse Guidance Network","authors":"Jianbiao Mei;Yu Yang;Mengmeng Wang;Junyu Zhu;Jongwon Ra;Yukai Ma;Laijian Li;Yong Liu","doi":"10.1109/TIP.2024.3461989","DOIUrl":"10.1109/TIP.2024.3461989","url":null,"abstract":"Semantic scene completion (SSC) aims to predict the semantic occupancy of each voxel in the entire 3D scene from limited observations, which is an emerging and critical task for autonomous driving. Recently, many studies have turned to camera-based SSC solutions due to the richer visual cues and cost-effectiveness of cameras. However, existing methods usually rely on sophisticated and heavy 3D models to process the lifted 3D features directly, which are not discriminative enough for clear segmentation boundaries. In this paper, we adopt the dense-sparse-dense design and propose a one-stage camera-based SSC framework, termed SGN, to propagate semantics from the semantic-aware seed voxels to the whole scene based on spatial geometry cues. Firstly, to exploit depth-aware context and dynamically select sparse seed voxels, we redesign the sparse voxel proposal network to process points generated by depth prediction directly with the coarse-to-fine paradigm. Furthermore, by designing hybrid guidance (sparse semantic and geometry guidance) and effective voxel aggregation for spatial geometry cues, we enhance the feature separation between different categories and expedite the convergence of semantic propagation. Finally, we devise the multi-scale semantic propagation module for flexible receptive fields while reducing the computation resources. Extensive experimental results on the SemanticKITTI and SSCBench-KITTI-360 datasets demonstrate the superiority of our SGN over existing state-of-the-art methods. And even our lightweight version SGN-L achieves notable scores of 14.80% mIoU and 45.45% IoU on SeamnticKITTI validation with only 12.5 M parameters and 7.16 G training memory. Code is available at \u0000<uri>https://github.com/Jieqianyu/SGN</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5468-5481"},"PeriodicalIF":0.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142325157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SWFormer: Stochastic Windows Convolutional Transformer for Hybrid Modality Hyperspectral Classification SWFormer:用于混合模态高光谱分类的随机窗口卷积变换器
Jiaojiao Li;Zhiyuan Zhang;Yuzhe Liu;Rui Song;Yunsong Li;Qian Du
Joint classification of hyperspectral images with hybrid modality can significantly enhance interpretation potentials, particularly when elevation information from the LiDAR sensor is integrated for outstanding performance. Recently, the transformer architecture was introduced to the HSI and LiDAR classification task, which has been verified as highly efficient. However, the existing naive transformer architectures suffer from two main drawbacks: 1) Inadequacy extraction for local spatial information and multi-scale information from HSI simultaneously. 2) The matrix calculation in the transformer consumes vast amounts of computing power. In this paper, we propose a novel Stochastic Window Transformer (SWFormer) framework to resolve these issues. First, the effective spatial and spectral feature projection networks are built independently based on hybrid-modal heterogeneous data composition using parallel feature extraction, which is conducive to excavating the perceptual features more representative along different dimensions. Furthermore, to construct local-global nonlinear feature maps more flexibly, we implement multi-scale strip convolution coupled with a transformer strategy. Moreover, in an innovative random window transformer structure, features are randomly masked to achieve sparse window pruning, alleviating the problem of information density redundancy, and reducing the parameters required for intensive attention. Finally, we designed a plug-and-play feature aggregation module that adapts domain offset between modal features adaptively to minimize semantic gaps between them and enhance the representational ability of the fusion feature. Three fiducial datasets demonstrate the effectiveness of the SWFormer in determining classification results.
采用混合模式对高光谱图像进行联合分类可显著提高解译潜力,尤其是在集成了激光雷达传感器的高程信息后,效果更为突出。最近,变压器架构被引入到高光谱和激光雷达分类任务中,其高效性已得到验证。然而,现有的天真变换器架构存在两个主要缺点:1) 无法同时从 HSI 中提取局部空间信息和多尺度信息。2) 变换器中的矩阵计算消耗大量计算能力。本文提出了一种新颖的随机窗口变换器(SWFormer)框架来解决这些问题。首先,利用并行特征提取技术,在混合模态异构数据组成的基础上,独立构建有效的空间和频谱特征投影网络,有利于从不同维度挖掘出更具代表性的感知特征。此外,为了更灵活地构建局部-全局非线性特征图,我们采用了多尺度带状卷积和变换器策略。此外,在创新的随机窗口变换器结构中,特征被随机屏蔽,实现了稀疏窗口剪枝,缓解了信息密度冗余问题,减少了密集关注所需的参数。最后,我们设计了一个即插即用的特征聚合模块,它能自适应地调整模态特征之间的域偏移,最大限度地减少它们之间的语义差距,增强融合特征的表征能力。三个固定数据集证明了 SWFormer 在确定分类结果方面的有效性。
{"title":"SWFormer: Stochastic Windows Convolutional Transformer for Hybrid Modality Hyperspectral Classification","authors":"Jiaojiao Li;Zhiyuan Zhang;Yuzhe Liu;Rui Song;Yunsong Li;Qian Du","doi":"10.1109/TIP.2024.3465038","DOIUrl":"10.1109/TIP.2024.3465038","url":null,"abstract":"Joint classification of hyperspectral images with hybrid modality can significantly enhance interpretation potentials, particularly when elevation information from the LiDAR sensor is integrated for outstanding performance. Recently, the transformer architecture was introduced to the HSI and LiDAR classification task, which has been verified as highly efficient. However, the existing naive transformer architectures suffer from two main drawbacks: 1) Inadequacy extraction for local spatial information and multi-scale information from HSI simultaneously. 2) The matrix calculation in the transformer consumes vast amounts of computing power. In this paper, we propose a novel Stochastic Window Transformer (SWFormer) framework to resolve these issues. First, the effective spatial and spectral feature projection networks are built independently based on hybrid-modal heterogeneous data composition using parallel feature extraction, which is conducive to excavating the perceptual features more representative along different dimensions. Furthermore, to construct local-global nonlinear feature maps more flexibly, we implement multi-scale strip convolution coupled with a transformer strategy. Moreover, in an innovative random window transformer structure, features are randomly masked to achieve sparse window pruning, alleviating the problem of information density redundancy, and reducing the parameters required for intensive attention. Finally, we designed a plug-and-play feature aggregation module that adapts domain offset between modal features adaptively to minimize semantic gaps between them and enhance the representational ability of the fusion feature. Three fiducial datasets demonstrate the effectiveness of the SWFormer in determining classification results.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5482-5495"},"PeriodicalIF":0.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142325161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SRNSD: Structure-Regularized Night-Time Self-Supervised Monocular Depth Estimation for Outdoor Scenes SRNSD:针对室外场景的结构规则化夜间自监督单目深度估计
Runmin Cong;Chunlei Wu;Xibin Song;Wei Zhang;Sam Kwong;Hongdong Li;Pan Ji
Deep CNNs have achieved impressive improvements for night-time self-supervised depth estimation form a monocular image. However, the performance degrades considerably compared to day-time depth estimation due to significant domain gaps, low visibility, and varying illuminations between day and night images. To address these challenges, we propose a novel night-time self-supervised monocular depth estimation framework with structure regularization, i.e., SRNSD, which incorporates three aspects of constraints for better performance, including feature and depth domain adaptation, image perspective constraint, and cropped multi-scale consistency loss. Specifically, we utilize adaptations of both feature and depth output spaces for better night-time feature extraction and depth map prediction, along with high- and low-frequency decoupling operations for better depth structure and texture recovery. Meanwhile, we employ an image perspective constraint to enhance the smoothness and obtain better depth maps in areas where the luminosity jumps change. Furthermore, we introduce a simple yet effective cropped multi-scale consistency loss that utilizes consistency among different scales of depth outputs for further optimization, refining the detailed textures and structures of predicted depth. Experimental results on different benchmarks with depth ranges of 40m and 60m, including Oxford RobotCar dataset, nuScenes dataset and CARLA-EPE dataset, demonstrate the superiority of our approach over state-of-the-art night-time self-supervised depth estimation approaches across multiple metrics, proving our effectiveness.
深度 CNN 在单目图像的夜间自监督深度估计方面取得了令人瞩目的进步。然而,与白天的深度估计相比,由于日夜图像之间存在明显的域差距、低能见度和不同的光照度,其性能大大降低。为了应对这些挑战,我们提出了一种具有结构正则化的新型夜间自监督单目深度估算框架,即 SRNSD,它结合了三个方面的约束条件以获得更好的性能,包括特征和深度域适应、图像透视约束和裁剪多尺度一致性损失。具体来说,我们利用特征和深度输出空间的适应性来实现更好的夜间特征提取和深度图预测,同时利用高频和低频解耦操作来实现更好的深度结构和纹理恢复。同时,我们采用图像透视约束来增强光滑度,并在亮度跃变区域获得更好的深度图。此外,我们还引入了一种简单而有效的裁剪多尺度一致性损失,利用不同尺度深度输出之间的一致性进行进一步优化,完善预测深度的细节纹理和结构。在深度范围为 40 米和 60 米的不同基准(包括牛津 RobotCar 数据集、nuScenes 数据集和 CARLA-EPE 数据集)上的实验结果表明,我们的方法在多个指标上都优于最先进的夜间自监督深度估计方法,证明了我们的有效性。
{"title":"SRNSD: Structure-Regularized Night-Time Self-Supervised Monocular Depth Estimation for Outdoor Scenes","authors":"Runmin Cong;Chunlei Wu;Xibin Song;Wei Zhang;Sam Kwong;Hongdong Li;Pan Ji","doi":"10.1109/TIP.2024.3465034","DOIUrl":"10.1109/TIP.2024.3465034","url":null,"abstract":"Deep CNNs have achieved impressive improvements for night-time self-supervised depth estimation form a monocular image. However, the performance degrades considerably compared to day-time depth estimation due to significant domain gaps, low visibility, and varying illuminations between day and night images. To address these challenges, we propose a novel night-time self-supervised monocular depth estimation framework with structure regularization, i.e., SRNSD, which incorporates three aspects of constraints for better performance, including feature and depth domain adaptation, image perspective constraint, and cropped multi-scale consistency loss. Specifically, we utilize adaptations of both feature and depth output spaces for better night-time feature extraction and depth map prediction, along with high- and low-frequency decoupling operations for better depth structure and texture recovery. Meanwhile, we employ an image perspective constraint to enhance the smoothness and obtain better depth maps in areas where the luminosity jumps change. Furthermore, we introduce a simple yet effective cropped multi-scale consistency loss that utilizes consistency among different scales of depth outputs for further optimization, refining the detailed textures and structures of predicted depth. Experimental results on different benchmarks with depth ranges of 40m and 60m, including Oxford RobotCar dataset, nuScenes dataset and CARLA-EPE dataset, demonstrate the superiority of our approach over state-of-the-art night-time self-supervised depth estimation approaches across multiple metrics, proving our effectiveness.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5538-5550"},"PeriodicalIF":0.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142325158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real-Time Predictive Condition Monitoring Using Multivariate Data 利用多变量数据进行实时预测性状态监测
Daniel Menges;Adil Rasheed;Harald Martens;Torbjørn Pedersen
This article presents an algorithmic framework for real-time condition monitoring and state forecasting using multivariate data demonstrated on thermal imagery data of a ship’s engine. The proposed method aims to improve the accuracy, efficiency, and robustness of condition monitoring and state predictions by identifying the most informative sampling locations of high-dimensional datasets and extracting the underlying dynamics of the system. The method is based on a combination of Proper Orthogonal Decomposition (POD), Optimal Sampling Location (OSL), and Dynamic Mode Decomposition (DMD), allowing the identification of key features in the system’s behavior and predicting future states. Based on thermal imagery data, it is shown how thermal areas of interest can be classified via POD. By extracting the POD modes of the data, dimensions can be drastically reduced and via OSL, optimal sampling locations are found. In addition, nonlinear kernel-based Support Vector Regression (SVR) is used to build models between the optimal locations, enabling the imputation of erroneous data to improve the overall robustness. To build predictive data-driven models, DMD is applied on the subspace obtained by OSL, which leads to an intensive lower demand of computational resources, making the proposed method real-time applicable. Furthermore, an unsupervised approach for anomaly detection is proposed using OSL. The anomaly detection framework is coupled with the state prediction framework, which extends the capabilities to real-time anomaly predictions. In summary, this study proposes a robust predictive condition monitoring framework for real-time risk assessment.
本文以船舶发动机的热图像数据为基础,介绍了一种利用多变量数据进行实时状态监测和状态预测的算法框架。所提出的方法旨在通过识别高维数据集中信息量最大的采样位置并提取系统的基本动态,提高状态监测和状态预测的准确性、效率和鲁棒性。该方法基于适当正交分解(POD)、最佳采样位置(OSL)和动态模式分解(DMD)的组合,可识别系统行为的关键特征并预测未来状态。基于热图像数据,我们展示了如何通过 POD 对感兴趣的热区域进行分类。通过提取数据的 POD 模式,可以大幅减少维数,并通过 OSL 找到最佳采样位置。此外,基于非线性核的支持向量回归(SVR)可用于在最佳位置之间建立模型,从而对错误数据进行估算,提高整体鲁棒性。为了建立预测性数据驱动模型,在 OSL 获得的子空间上应用了 DMD,从而降低了对计算资源的密集需求,使所提出的方法具有实时性。此外,还提出了一种利用 OSL 进行异常检测的无监督方法。异常检测框架与状态预测框架相结合,扩展了实时异常预测的能力。总之,本研究提出了一种用于实时风险评估的稳健预测性状态监测框架。
{"title":"Real-Time Predictive Condition Monitoring Using Multivariate Data","authors":"Daniel Menges;Adil Rasheed;Harald Martens;Torbjørn Pedersen","doi":"10.1109/TIP.2024.3468894","DOIUrl":"10.1109/TIP.2024.3468894","url":null,"abstract":"This article presents an algorithmic framework for real-time condition monitoring and state forecasting using multivariate data demonstrated on thermal imagery data of a ship’s engine. The proposed method aims to improve the accuracy, efficiency, and robustness of condition monitoring and state predictions by identifying the most informative sampling locations of high-dimensional datasets and extracting the underlying dynamics of the system. The method is based on a combination of Proper Orthogonal Decomposition (POD), Optimal Sampling Location (OSL), and Dynamic Mode Decomposition (DMD), allowing the identification of key features in the system’s behavior and predicting future states. Based on thermal imagery data, it is shown how thermal areas of interest can be classified via POD. By extracting the POD modes of the data, dimensions can be drastically reduced and via OSL, optimal sampling locations are found. In addition, nonlinear kernel-based Support Vector Regression (SVR) is used to build models between the optimal locations, enabling the imputation of erroneous data to improve the overall robustness. To build predictive data-driven models, DMD is applied on the subspace obtained by OSL, which leads to an intensive lower demand of computational resources, making the proposed method real-time applicable. Furthermore, an unsupervised approach for anomaly detection is proposed using OSL. The anomaly detection framework is coupled with the state prediction framework, which extends the capabilities to real-time anomaly predictions. In summary, this study proposes a robust predictive condition monitoring framework for real-time risk assessment.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5703-5714"},"PeriodicalIF":0.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142325156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LDPC Code-Based Distributed Source Coding With an Efficient Message Passing Mechanism for the Compression of Correlated Image Sources 基于 LDPC 码的分布式源编码与用于压缩相关图像源的高效信息传递机制
Hong Mo;Jianhua Chen
Different from traditional source coding techniques, distributed source coding (DSC) techniques rely on independent encoding at the encoding end but joint decoding at the decoding end to compress image sources which exhibit correlation. Channel code-based DSC techniques compress correlated image sources by fully utilizing such correlation. However, in addition to this correlation information, the current symbol of most real image sources is correlated to the preceding or subsequent symbols. Such correlation information should also play an important role for improving the compression performance of channel code-based DSC schemes. To this end, we present an efficient message passing mechanism for LDPC Code-based DSC schemes (ELCDSC) to compress correlated image sources. By utilizing this message passing mechanism, we enable LDPC code-based DSC techniques to not only make full use of the inter-source correlation to assist compression, but also integrate the intra-source correlation in each message passing iteration to improve the compression performance. It is the first that enables LDPC code-based DSC techniques to achieve the utilization of both intra- and inter-source correlations in the message passing mechanism. Experimental results reveal that ELCDSC significantly enhances the compression ratio of correlated image sources, surpassing other DSC schemes.
与传统的信源编码技术不同,分布式信源编码(DSC)技术依靠编码端独立编码和解码端联合解码来压缩呈现相关性的图像信源。基于信道编码的 DSC 技术通过充分利用这种相关性来压缩相关图像源。然而,除了这种相关性信息外,大多数真实图像源的当前符号还与前面或后面的符号相关。这种相关性信息也应在提高基于信道编码的 DSC 方案的压缩性能方面发挥重要作用。为此,我们为基于 LDPC 码的 DSC 方案(ELCDSC)提出了一种有效的信息传递机制,用于压缩相关图像源。通过利用这种信息传递机制,我们使基于 LDPC 码的 DSC 技术不仅能充分利用源间相关性来帮助压缩,还能在每次信息传递迭代中整合源内相关性,从而提高压缩性能。它首次使基于 LDPC 码的 DSC 技术在信息传递机制中实现了对源内和源间相关性的利用。实验结果表明,ELCDSC 显著提高了相关图像源的压缩率,超过了其他 DSC 方案。
{"title":"LDPC Code-Based Distributed Source Coding With an Efficient Message Passing Mechanism for the Compression of Correlated Image Sources","authors":"Hong Mo;Jianhua Chen","doi":"10.1109/TIP.2024.3463492","DOIUrl":"10.1109/TIP.2024.3463492","url":null,"abstract":"Different from traditional source coding techniques, distributed source coding (DSC) techniques rely on independent encoding at the encoding end but joint decoding at the decoding end to compress image sources which exhibit correlation. Channel code-based DSC techniques compress correlated image sources by fully utilizing such correlation. However, in addition to this correlation information, the current symbol of most real image sources is correlated to the preceding or subsequent symbols. Such correlation information should also play an important role for improving the compression performance of channel code-based DSC schemes. To this end, we present an efficient message passing mechanism for LDPC Code-based DSC schemes (ELCDSC) to compress correlated image sources. By utilizing this message passing mechanism, we enable LDPC code-based DSC techniques to not only make full use of the inter-source correlation to assist compression, but also integrate the intra-source correlation in each message passing iteration to improve the compression performance. It is the first that enables LDPC code-based DSC techniques to achieve the utilization of both intra- and inter-source correlations in the message passing mechanism. Experimental results reveal that ELCDSC significantly enhances the compression ratio of correlated image sources, surpassing other DSC schemes.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5354-5369"},"PeriodicalIF":0.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Subjective Quality Assessment of Compressed Tone-Mapped High Dynamic Range Videos 压缩色调映射高动态范围视频的主观质量评估
Abhinau K. Venkataramanan;Alan C. Bovik
High Dynamic Range (HDR) videos are able to represent wider ranges of contrasts and colors than Standard Dynamic Range (SDR) videos, giving more vivid experiences. Due to this, HDR videos are expected to grow into the dominant video modality of the future. However, HDR videos are incompatible with existing SDR displays, which form the majority of affordable consumer displays on the market. Because of this, HDR videos must be processed by tone-mapping them to reduced bit-depths to service a broad swath of SDR-limited video consumers. Here, we analyze the impact of tone-mapping operators on the visual quality of streaming HDR videos. To this end, we built the first large-scale subjectively annotated open-source database of compressed tone-mapped HDR videos, containing 15,000 tone-mapped sequences derived from 40 unique HDR source contents. The videos in the database were labeled with more than 750,000 subjective quality annotations, collected from more than 1,600 unique human observers. We demonstrate the usefulness of the new subjective database by benchmarking objective models of visual quality on it. We envision that the new LIVE Tone-Mapped HDR (LIVE-TMHDR) database will enable significant progress on HDR video tone mapping and quality assessment in the future. To this end, we make the database freely available to the community at https://live.ece.utexas.edu/research/LIVE_TMHDR/index.html.
与标准动态范围(SDR)视频相比,高动态范围(HDR)视频能够呈现更广泛的对比度和色彩,带来更生动的体验。因此,HDR 视频有望发展成为未来的主流视频模式。然而,HDR 视频与现有的 SDR 显示器不兼容,而市场上大多数经济型消费显示器都是 SDR 显示器。因此,在处理 HDR 视频时,必须将其色调映射到较低的比特深度,以便为广大受 SDR 限制的视频消费者提供服务。在此,我们分析了色调映射操作员对流式 HDR 视频视觉质量的影响。为此,我们建立了首个大规模主观注释的开源压缩音调映射 HDR 视频数据库,其中包含从 40 种独特 HDR 源内容中提取的 15,000 个音调映射序列。数据库中的视频标注了 750,000 多条主观质量注释,这些注释是从 1,600 多名独特的人类观察者那里收集来的。我们通过对视觉质量的客观模型进行基准测试,证明了新主观数据库的实用性。我们预计,新的 LIVE 色调映射 HDR(LIVE-TMHDR)数据库将在未来推动 HDR 视频色调映射和质量评估取得重大进展。为此,我们在 https://live.ece.utexas.edu/research/LIVE_TMHDR/index.html 上向社区免费提供该数据库。
{"title":"Subjective Quality Assessment of Compressed Tone-Mapped High Dynamic Range Videos","authors":"Abhinau K. Venkataramanan;Alan C. Bovik","doi":"10.1109/TIP.2024.3463418","DOIUrl":"10.1109/TIP.2024.3463418","url":null,"abstract":"High Dynamic Range (HDR) videos are able to represent wider ranges of contrasts and colors than Standard Dynamic Range (SDR) videos, giving more vivid experiences. Due to this, HDR videos are expected to grow into the dominant video modality of the future. However, HDR videos are incompatible with existing SDR displays, which form the majority of affordable consumer displays on the market. Because of this, HDR videos must be processed by tone-mapping them to reduced bit-depths to service a broad swath of SDR-limited video consumers. Here, we analyze the impact of tone-mapping operators on the visual quality of streaming HDR videos. To this end, we built the first large-scale subjectively annotated open-source database of compressed tone-mapped HDR videos, containing 15,000 tone-mapped sequences derived from 40 unique HDR source contents. The videos in the database were labeled with more than 750,000 subjective quality annotations, collected from more than 1,600 unique human observers. We demonstrate the usefulness of the new subjective database by benchmarking objective models of visual quality on it. We envision that the new LIVE Tone-Mapped HDR (LIVE-TMHDR) database will enable significant progress on HDR video tone mapping and quality assessment in the future. To this end, we make the database freely available to the community at \u0000<uri>https://live.ece.utexas.edu/research/LIVE_TMHDR/index.html</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5440-5455"},"PeriodicalIF":0.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Restructuring the Teacher and Student in Self-Distillation 在自我发散中重构师生关系
Yujie Zheng;Chong Wang;Chenchen Tao;Sunqi Lin;Jiangbo Qian;Jiafei Wu
Knowledge distillation aims to achieve model compression by transferring knowledge from complex teacher models to lightweight student models. To reduce reliance on pre-trained teacher models, self-distillation methods utilize knowledge from the model itself as additional supervision. However, their performance is limited by the same or similar network architecture between the teacher and student. In order to increase architecture variety, we propose a new self-distillation framework called restructured self-distillation (RSD), which involves restructuring both the teacher and student networks. The self-distilled model is expanded into a multi-branch topology to create a more powerful teacher. During training, diverse student sub-networks are generated by randomly discarding the teacher’s branches. Additionally, the teacher and student models are linked by a randomly inserted feature mixture block, introducing additional knowledge distillation in the mixed feature space. To avoid extra inference costs, the branches of the teacher model are then converted back to its original structure equivalently. Comprehensive experiments have demonstrated the effectiveness of our proposed framework for most architectures on CIFAR-10/100 and ImageNet datasets. Code is available at https://github.com/YujieZheng99/RSD.
知识蒸馏的目的是通过将复杂的教师模型中的知识转移到轻量级的学生模型中来实现模型压缩。为了减少对预训练教师模型的依赖,自蒸馏方法利用模型本身的知识作为额外的监督。然而,它们的性能受到教师和学生之间相同或相似网络架构的限制。为了增加架构的多样性,我们提出了一种新的自蒸馏框架,称为重组自蒸馏(RSD),其中涉及重组教师和学生网络。自蒸馏模型扩展为多分支拓扑结构,以创建更强大的教师网络。在训练过程中,通过随机丢弃教师的分支,生成多样化的学生子网络。此外,教师模型和学生模型通过随机插入的特征混合块连接起来,在混合特征空间中引入额外的知识提炼。为了避免额外的推理成本,教师模型的分支会被等效地转换回其原始结构。在 CIFAR-10/100 和 ImageNet 数据集上进行的综合实验证明了我们提出的框架对大多数架构的有效性。代码见 https://github.com/YujieZheng99/RSD。
{"title":"Restructuring the Teacher and Student in Self-Distillation","authors":"Yujie Zheng;Chong Wang;Chenchen Tao;Sunqi Lin;Jiangbo Qian;Jiafei Wu","doi":"10.1109/TIP.2024.3463421","DOIUrl":"10.1109/TIP.2024.3463421","url":null,"abstract":"Knowledge distillation aims to achieve model compression by transferring knowledge from complex teacher models to lightweight student models. To reduce reliance on pre-trained teacher models, self-distillation methods utilize knowledge from the model itself as additional supervision. However, their performance is limited by the same or similar network architecture between the teacher and student. In order to increase architecture variety, we propose a new self-distillation framework called restructured self-distillation (RSD), which involves restructuring both the teacher and student networks. The self-distilled model is expanded into a multi-branch topology to create a more powerful teacher. During training, diverse student sub-networks are generated by randomly discarding the teacher’s branches. Additionally, the teacher and student models are linked by a randomly inserted feature mixture block, introducing additional knowledge distillation in the mixed feature space. To avoid extra inference costs, the branches of the teacher model are then converted back to its original structure equivalently. Comprehensive experiments have demonstrated the effectiveness of our proposed framework for most architectures on CIFAR-10/100 and ImageNet datasets. Code is available at \u0000<uri>https://github.com/YujieZheng99/RSD</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5551-5563"},"PeriodicalIF":0.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis and Benchmarking of Extending Blind Face Image Restoration to Videos 将盲法人脸图像复原扩展到视频的分析和基准测试
Zhouxia Wang;Jiawei Zhang;Xintao Wang;Tianshui Chen;Ying Shan;Wenping Wang;Ping Luo
Recent progress in blind face restoration has resulted in producing high-quality restored results for static images. However, efforts to extend these advancements to video scenarios have been minimal, partly because of the absence of benchmarks that allow for a comprehensive and fair comparison. In this work, we first present a fair evaluation benchmark, in which we first introduce a Real-world Low-Quality Face Video benchmark (RFV-LQ), evaluate several leading image-based face restoration algorithms, and conduct a thorough systematical analysis of the benefits and challenges associated with extending blind face image restoration algorithms to degraded face videos. Our analysis identifies several key issues, primarily categorized into two aspects: significant jitters in facial components and noise-shape flickering between frames. To address these issues, we propose a Temporal Consistency Network (TCN) cooperated with alignment smoothing to reduce jitters and flickers in restored videos. TCN is a flexible component that can be seamlessly plugged into the most advanced face image restoration algorithms, ensuring the quality of image-based restoration is maintained as closely as possible. Extensive experiments have been conducted to evaluate the effectiveness and efficiency of our proposed TCN and alignment smoothing operation.
最近在盲目人脸修复方面取得的进展已经为静态图像提供了高质量的修复结果。然而,将这些进展扩展到视频场景的努力却微乎其微,部分原因是缺乏可进行全面公平比较的基准。在这项工作中,我们首先提出了一个公平的评估基准,在这个基准中,我们首先引入了真实世界低质量人脸视频基准(RFV-LQ),评估了几种领先的基于图像的人脸修复算法,并对将盲人脸图像修复算法扩展到降级人脸视频所带来的好处和挑战进行了全面系统的分析。我们的分析发现了几个关键问题,主要分为两个方面:面部组件的显著抖动和帧间的噪形闪烁。为了解决这些问题,我们提出了一种时间一致性网络(TCN),并将其与对齐平滑技术相结合,以减少修复视频中的抖动和闪烁。时间一致性网络是一个灵活的组件,可以无缝接入最先进的人脸图像修复算法,确保尽可能保持基于图像的修复质量。为了评估我们提出的 TCN 和对齐平滑操作的效果和效率,我们进行了广泛的实验。
{"title":"Analysis and Benchmarking of Extending Blind Face Image Restoration to Videos","authors":"Zhouxia Wang;Jiawei Zhang;Xintao Wang;Tianshui Chen;Ying Shan;Wenping Wang;Ping Luo","doi":"10.1109/TIP.2024.3463414","DOIUrl":"10.1109/TIP.2024.3463414","url":null,"abstract":"Recent progress in blind face restoration has resulted in producing high-quality restored results for static images. However, efforts to extend these advancements to video scenarios have been minimal, partly because of the absence of benchmarks that allow for a comprehensive and fair comparison. In this work, we first present a fair evaluation benchmark, in which we first introduce a Real-world Low-Quality Face Video benchmark (RFV-LQ), evaluate several leading image-based face restoration algorithms, and conduct a thorough systematical analysis of the benefits and challenges associated with extending blind face image restoration algorithms to degraded face videos. Our analysis identifies several key issues, primarily categorized into two aspects: significant jitters in facial components and noise-shape flickering between frames. To address these issues, we propose a Temporal Consistency Network (TCN) cooperated with alignment smoothing to reduce jitters and flickers in restored videos. TCN is a flexible component that can be seamlessly plugged into the most advanced face image restoration algorithms, ensuring the quality of image-based restoration is maintained as closely as possible. Extensive experiments have been conducted to evaluate the effectiveness and efficiency of our proposed TCN and alignment smoothing operation.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5676-5687"},"PeriodicalIF":0.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Boundary-Aware Prototype in Semi-Supervised Medical Image Segmentation 半监督医学图像分割中的边界感知原型
YongChao Wang;Bin Xiao;Xiuli Bi;Weisheng Li;Xinbo Gao
The true label plays an important role in semi-supervised medical image segmentation (SSMIS) because it can provide the most accurate supervision information when the label is limited. The popular SSMIS method trains labeled and unlabeled data separately, and the unlabeled data cannot be directly supervised by the true label. This limits the contribution of labels to model training. Is there an interactive mechanism that can break the separation between two types of data training to maximize the utilization of true labels? Inspired by this, we propose a novel consistency learning framework based on the non-parametric distance metric of boundary-aware prototypes to alleviate this problem. This method combines CNN-based linear classification and nearest neighbor-based non-parametric classification into one framework, encouraging the two segmentation paradigms to have similar predictions for the same input. More importantly, the prototype can be clustered from both labeled and unlabeled data features so that it can be seen as a bridge for interactive training between labeled and unlabeled data. When the prototype-based prediction is supervised by the true label, the supervisory signal can simultaneously affect the feature extraction process of both data. In addition, boundary-aware prototypes can explicitly model the differences in boundaries and centers of adjacent categories, so pixel-prototype contrastive learning is introduced to further improve the discriminability of features and make them more suitable for non-parametric distance measurement. Experiments show that although our method uses a modified lightweight UNet as the backbone, it outperforms the comparison method using a 3D VNet with more parameters.
真实标签在半监督医学图像分割(SSMIS)中发挥着重要作用,因为当标签有限时,它能提供最准确的监督信息。目前流行的 SSMIS 方法将有标签数据和无标签数据分开训练,无标签数据不能直接由真实标签监督。这就限制了标签对模型训练的贡献。有没有一种交互机制可以打破两类数据训练的分离,最大限度地利用真实标签呢?受此启发,我们提出了一种基于边界感知原型的非参数距离度量的新型一致性学习框架,以缓解这一问题。这种方法将基于 CNN 的线性分类和基于近邻的非参数分类结合到一个框架中,鼓励这两种分割范式对相同的输入做出相似的预测。更重要的是,原型可以从已标注和未标注的数据特征中进行聚类,因此它可以被视为标注数据和未标注数据之间交互式训练的桥梁。当基于原型的预测受到真实标签的监督时,监督信号可以同时影响两种数据的特征提取过程。此外,边界感知原型可以对相邻类别的边界和中心差异进行明确建模,因此引入了像素原型对比学习,以进一步提高特征的可辨别性,使其更适合非参数距离测量。实验表明,虽然我们的方法使用了改进的轻量级 UNet 作为骨干网,但其效果优于使用参数更多的 3D VNet 的对比方法。
{"title":"Boundary-Aware Prototype in Semi-Supervised Medical Image Segmentation","authors":"YongChao Wang;Bin Xiao;Xiuli Bi;Weisheng Li;Xinbo Gao","doi":"10.1109/TIP.2024.3463412","DOIUrl":"10.1109/TIP.2024.3463412","url":null,"abstract":"The true label plays an important role in semi-supervised medical image segmentation (SSMIS) because it can provide the most accurate supervision information when the label is limited. The popular SSMIS method trains labeled and unlabeled data separately, and the unlabeled data cannot be directly supervised by the true label. This limits the contribution of labels to model training. Is there an interactive mechanism that can break the separation between two types of data training to maximize the utilization of true labels? Inspired by this, we propose a novel consistency learning framework based on the non-parametric distance metric of boundary-aware prototypes to alleviate this problem. This method combines CNN-based linear classification and nearest neighbor-based non-parametric classification into one framework, encouraging the two segmentation paradigms to have similar predictions for the same input. More importantly, the prototype can be clustered from both labeled and unlabeled data features so that it can be seen as a bridge for interactive training between labeled and unlabeled data. When the prototype-based prediction is supervised by the true label, the supervisory signal can simultaneously affect the feature extraction process of both data. In addition, boundary-aware prototypes can explicitly model the differences in boundaries and centers of adjacent categories, so pixel-prototype contrastive learning is introduced to further improve the discriminability of features and make them more suitable for non-parametric distance measurement. Experiments show that although our method uses a modified lightweight UNet as the backbone, it outperforms the comparison method using a 3D VNet with more parameters.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5456-5467"},"PeriodicalIF":0.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CoBEV: Elevating Roadside 3D Object Detection With Depth and Height Complementarity CoBEV:利用深度和高度互补性提升路边 3D 物体检测能力
Hao Shi;Chengshan Pang;Jiaming Zhang;Kailun Yang;Yuhao Wu;Huajian Ni;Yining Lin;Rainer Stiefelhagen;Kaiwei Wang
Roadside camera-driven 3D object detection is a crucial task in intelligent transportation systems, which extends the perception range beyond the limitations of vision-centric vehicles and enhances road safety. While previous studies have limitations in using only depth or height information, we find both depth and height matter and they are in fact complementary. The depth feature encompasses precise geometric cues, whereas the height feature is primarily focused on distinguishing between various categories of height intervals, essentially providing semantic context. This insight motivates the development of Complementary-BEV (CoBEV), a novel end-to-end monocular 3D object detection framework that integrates depth and height to construct robust BEV representations. In essence, CoBEV estimates each pixel’s depth and height distribution and lifts the camera features into 3D space for lateral fusion using the newly proposed two-stage complementary feature selection (CFS) module. A BEV feature distillation framework is also seamlessly integrated to further enhance the detection accuracy from the prior knowledge of the fusion-modal CoBEV teacher. We conduct extensive experiments on the public 3D detection benchmarks of roadside camera-based DAIR-V2X-I and Rope3D, as well as the private Supremind-Road dataset, demonstrating that CoBEV not only achieves the accuracy of the new state-of-the-art, but also significantly advances the robustness of previous methods in challenging long-distance scenarios and noisy camera disturbance, and enhances generalization by a large margin in heterologous Settings with drastic changes in scene and camera parameters. For the first time, the vehicle AP score of a camera model reaches 80% on DAIR-V2X-I in terms of easy mode. The source code will be made publicly available at CoBEV.
路边摄像头驱动的三维物体检测是智能交通系统中的一项重要任务,它扩大了以视觉为中心的车辆的感知范围,提高了道路安全性。以往的研究存在只使用深度或高度信息的局限性,而我们发现深度和高度都很重要,而且它们实际上是互补的。深度特征包含精确的几何线索,而高度特征主要侧重于区分不同类别的高度区间,本质上是提供语义背景。这种洞察力促使我们开发了互补 BEV(CoBEV),这是一种新颖的端到端单目 3D 物体检测框架,它将深度和高度整合在一起,以构建稳健的 BEV 表示。从本质上讲,CoBEV 估算每个像素的深度和高度分布,并利用新提出的两阶段互补特征选择(CFS)模块将相机特征提升到三维空间进行横向融合。此外,还无缝集成了 BEV 特征提炼框架,利用融合模式 CoBEV 教师的先验知识进一步提高检测精度。我们在基于路边摄像头的 DAIR-V2X-I 和 Rope3D 公共三维检测基准以及私有 Supremind-Road 数据集上进行了广泛的实验,结果表明 CoBEV 不仅达到了新的一流水平的精度,而且在具有挑战性的长距离场景和高噪声摄像头干扰下显著提高了以前方法的鲁棒性,并在场景和摄像头参数急剧变化的异源设置中大幅增强了泛化能力。在 DAIR-V2X-I 简易模式下,摄像机模型的车辆 AP 得分首次达到 80%。源代码将在 CoBEV 公开发布。
{"title":"CoBEV: Elevating Roadside 3D Object Detection With Depth and Height Complementarity","authors":"Hao Shi;Chengshan Pang;Jiaming Zhang;Kailun Yang;Yuhao Wu;Huajian Ni;Yining Lin;Rainer Stiefelhagen;Kaiwei Wang","doi":"10.1109/TIP.2024.3463409","DOIUrl":"10.1109/TIP.2024.3463409","url":null,"abstract":"Roadside camera-driven 3D object detection is a crucial task in intelligent transportation systems, which extends the perception range beyond the limitations of vision-centric vehicles and enhances road safety. While previous studies have limitations in using only depth or height information, we find both depth and height matter and they are in fact complementary. The depth feature encompasses precise geometric cues, whereas the height feature is primarily focused on distinguishing between various categories of height intervals, essentially providing semantic context. This insight motivates the development of Complementary-BEV (CoBEV), a novel end-to-end monocular 3D object detection framework that integrates depth and height to construct robust BEV representations. In essence, CoBEV estimates each pixel’s depth and height distribution and lifts the camera features into 3D space for lateral fusion using the newly proposed two-stage complementary feature selection (CFS) module. A BEV feature distillation framework is also seamlessly integrated to further enhance the detection accuracy from the prior knowledge of the fusion-modal CoBEV teacher. We conduct extensive experiments on the public 3D detection benchmarks of roadside camera-based DAIR-V2X-I and Rope3D, as well as the private Supremind-Road dataset, demonstrating that CoBEV not only achieves the accuracy of the new state-of-the-art, but also significantly advances the robustness of previous methods in challenging long-distance scenarios and noisy camera disturbance, and enhances generalization by a large margin in \u0000<monospace>heterologous</monospace>\u0000 Settings with drastic changes in scene and camera parameters. For the first time, the vehicle AP score of a camera model reaches 80% on DAIR-V2X-I in terms of easy mode. The source code will be made publicly available at CoBEV.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"5424-5439"},"PeriodicalIF":0.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142317236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1