首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
EvRepSL: Event-Stream Representation via Self-Supervised Learning for Event-Based Vision EvRepSL:通过自监督学习进行事件流表示,实现基于事件的视觉效果
Qiang Qu;Xiaoming Chen;Yuk Ying Chung;Yiran Shen
Event-stream representation is the first step for many computer vision tasks using event cameras. It converts the asynchronous event-streams into a formatted structure so that conventional machine learning models can be applied easily. However, most of the state-of-the-art event-stream representations are manually designed and the quality of these representations cannot be guaranteed due to the noisy nature of event-streams. In this paper, we introduce a data-driven approach aiming at enhancing the quality of event-stream representations. Our approach commences with the introduction of a new event-stream representation based on spatial-temporal statistics, denoted as EvRep. Subsequently, we theoretically derive the intrinsic relationship between asynchronous event-streams and synchronous video frames. Building upon this theoretical relationship, we train a representation generator, RepGen, in a self-supervised learning manner accepting EvRep as input. Finally, the event-streams are converted to high-quality representations, termed as EvRepSL, by going through the learned RepGen (without the need of fine-tuning or retraining). Our methodology is rigorously validated through extensive evaluations on a variety of mainstream event-based classification and optical flow datasets (captured with various types of event cameras). The experimental results highlight not only our approach’s superior performance over existing event-stream representations but also its versatility, being agnostic to different event cameras and tasks.
事件流表示是许多使用事件摄像机的计算机视觉任务的第一步。它将异步事件流转换成格式化的结构,以便轻松应用传统的机器学习模型。然而,大多数最先进的事件流表示法都是人工设计的,而且由于事件流的噪声特性,这些表示法的质量无法得到保证。在本文中,我们介绍了一种数据驱动方法,旨在提高事件流表征的质量。随后,我们从理论上推导出异步事件流与同步视频帧之间的内在关系。在这一理论关系的基础上,我们以自我监督学习的方式训练表征生成器 RepGen,将 EvRep 作为输入。最后,通过学习到的 RepGen(无需微调或再训练)将事件流转换为高质量表示,称为 EvRepSL。通过对各种主流基于事件的分类和光流数据集(使用各种类型的事件相机捕获)进行广泛评估,我们的方法得到了严格验证。实验结果不仅凸显了我们的方法优于现有事件流表示法的性能,而且还凸显了它的通用性,即对不同的事件相机和任务都具有不可知性。
{"title":"EvRepSL: Event-Stream Representation via Self-Supervised Learning for Event-Based Vision","authors":"Qiang Qu;Xiaoming Chen;Yuk Ying Chung;Yiran Shen","doi":"10.1109/TIP.2024.3497795","DOIUrl":"10.1109/TIP.2024.3497795","url":null,"abstract":"Event-stream representation is the first step for many computer vision tasks using event cameras. It converts the asynchronous event-streams into a formatted structure so that conventional machine learning models can be applied easily. However, most of the state-of-the-art event-stream representations are manually designed and the quality of these representations cannot be guaranteed due to the noisy nature of event-streams. In this paper, we introduce a data-driven approach aiming at enhancing the quality of event-stream representations. Our approach commences with the introduction of a new event-stream representation based on spatial-temporal statistics, denoted as EvRep. Subsequently, we theoretically derive the intrinsic relationship between asynchronous event-streams and synchronous video frames. Building upon this theoretical relationship, we train a representation generator, RepGen, in a self-supervised learning manner accepting EvRep as input. Finally, the event-streams are converted to high-quality representations, termed as EvRepSL, by going through the learned RepGen (without the need of fine-tuning or retraining). Our methodology is rigorously validated through extensive evaluations on a variety of mainstream event-based classification and optical flow datasets (captured with various types of event cameras). The experimental results highlight not only our approach’s superior performance over existing event-stream representations but also its versatility, being agnostic to different event cameras and tasks.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6579-6591"},"PeriodicalIF":0.0,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142673314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pro2Diff: Proposal Propagation for Multi-Object Tracking via the Diffusion Model Pro2Diff:通过扩散模型进行多目标跟踪的提案传播
Hongmin Liu;Canbin Zhang;Bin Fan;Jinglin Xu
Multi-object tracking (MOT) aims to estimate the bounding boxes and ID labels of objects in videos. The challenging issue in this task is to alleviate competitive learning between the detection and tracking subtasks, for which, two-stage Tracking-By-Detection (TBD) optimizes the two subtasks individually, and the single-stage Joint Detection and Tracking (JDT) adjusts the complex network architectures finely in an end-to-end pipeline. In this paper, we propose a new MOT method, i.e., Proposal Propagation via Diffusion Models, called Pro2Diff, which integrates a diffusion model into the proposal propagation in multi-object tracking, focusing on the model training process rather than complex network design. Specifically, using a generative approach, Pro2Diff generates a considerable number of noisy proposals for the tracking image sequence in the forward process, and subsequently, Pro2Diff learns the discrepancies between these noisy proposals and the actual bounding boxes of the tracked objects, gradually optimizing these noisy proposals to obtain the initial sequence of real tracked objects. By introducing the denoising diffusion process into multi-object tracking, we have made three further important findings: 1) Generative methods can effectively handle multi-object tracking tasks; 2) Without the need to modify the model structure, we propose self-conditional proposal propagation to enhance model performance effectively during inference; 3) By adjusting the numbers of proposals and iterations appropriately for different tracking sequences, the optimal performance of the model can be achieved. Extensive experimental results on MOT17 and DanceTrack datasets demonstrate that Pro2Diff outperforms current end-to-end multi-object tracking methods. We achieve 61.9 HOTA on DanceTrack and 57.6 HOTA on MOT17, reaching the competitive result of the JDT approach.
多目标跟踪(MOT)旨在估计视频中物体的边界框和 ID 标签。在这项任务中,具有挑战性的问题是如何缓解检测和跟踪子任务之间的竞争性学习,为此,两阶段跟踪检测(Tracking-By-Detection,TBD)分别对这两个子任务进行优化,而单阶段联合检测和跟踪(Joint Detection and Tracking,JDT)则在端到端流水线中对复杂的网络架构进行精细调整。在本文中,我们提出了一种新的 MOT 方法,即通过扩散模型进行提议传播(Proposal Propagation via Diffusion Models),称为 Pro2Diff,它将扩散模型集成到多目标跟踪的提议传播中,重点关注模型训练过程而非复杂的网络设计。具体来说,Pro2Diff 采用生成式方法,在前向过程中为跟踪图像序列生成相当数量的噪声提议,随后,Pro2Diff 学习这些噪声提议与实际跟踪对象边界框之间的差异,逐步优化这些噪声提议,从而获得真实跟踪对象的初始序列。通过在多目标跟踪中引入去噪扩散过程,我们又有了三个重要发现:1)生成式方法可以有效地处理多目标跟踪任务;2)无需修改模型结构,我们提出了自条件提案传播法,可以在推理过程中有效地提高模型性能;3)通过针对不同的跟踪序列适当调整提案数和迭代数,可以实现模型的最佳性能。在 MOT17 和 DanceTrack 数据集上的大量实验结果表明,Pro2Diff 优于目前的端到端多目标跟踪方法。我们在 DanceTrack 上获得了 61.9 HOTA,在 MOT17 上获得了 57.6 HOTA,达到了 JDT 方法的竞争结果。
{"title":"Pro2Diff: Proposal Propagation for Multi-Object Tracking via the Diffusion Model","authors":"Hongmin Liu;Canbin Zhang;Bin Fan;Jinglin Xu","doi":"10.1109/TIP.2024.3494600","DOIUrl":"10.1109/TIP.2024.3494600","url":null,"abstract":"Multi-object tracking (MOT) aims to estimate the bounding boxes and ID labels of objects in videos. The challenging issue in this task is to alleviate competitive learning between the detection and tracking subtasks, for which, two-stage Tracking-By-Detection (TBD) optimizes the two subtasks individually, and the single-stage Joint Detection and Tracking (JDT) adjusts the complex network architectures finely in an end-to-end pipeline. In this paper, we propose a new MOT method, i.e., Proposal Propagation via Diffusion Models, called Pro2Diff, which integrates a diffusion model into the proposal propagation in multi-object tracking, focusing on the model training process rather than complex network design. Specifically, using a generative approach, Pro2Diff generates a considerable number of noisy proposals for the tracking image sequence in the forward process, and subsequently, Pro2Diff learns the discrepancies between these noisy proposals and the actual bounding boxes of the tracked objects, gradually optimizing these noisy proposals to obtain the initial sequence of real tracked objects. By introducing the denoising diffusion process into multi-object tracking, we have made three further important findings: 1) Generative methods can effectively handle multi-object tracking tasks; 2) Without the need to modify the model structure, we propose self-conditional proposal propagation to enhance model performance effectively during inference; 3) By adjusting the numbers of proposals and iterations appropriately for different tracking sequences, the optimal performance of the model can be achieved. Extensive experimental results on MOT17 and DanceTrack datasets demonstrate that Pro2Diff outperforms current end-to-end multi-object tracking methods. We achieve 61.9 HOTA on DanceTrack and 57.6 HOTA on MOT17, reaching the competitive result of the JDT approach.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6508-6520"},"PeriodicalIF":0.0,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142637275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced Multispectral Band-to-Band Registration Using Co-Occurrence Scale Space and Spatial Confined RANSAC Guided Segmented Affine Transformation 利用共现尺度空间和空间限制 RANSAC 引导的分段仿射变换增强多光谱波段到波段的配准
Indranil Misra;Mukesh Kumar Rohil;S. Manthira Moorthi;Debajyoti Dhar
Band-to-Band Registration (BBR) is a pre-requisite image processing operation essential for specific remote sensing multispectral sensors. BBR aims to align spectral wavelength channels at sub-pixel level accuracy over each other. The paper presents a novel BBR technique utilizing Co-occurrence Scale Space (CSS) for feature point detection and Spatial Confined RANSAC (SC-RANSAC) for removing outlier matched control points. Additionally, the Segmented Affine Transformation (SAT) model reduces distortion and ensures consistent BBR. The methodology developed is evaluated with Nano-MX multispectral images onboard the Indian Nano Satellite (INS-2B) covering diverse landscapes. BBR performance using the proposed method is also verified visually at a 4X zoom level on satellite scenes dominated by cloud pixels. The band misregistration effect on the Normalized Difference Vegetation Index (NDVI) from INS-2B is analyzed and cross-validated with the closest acquisition Landsat-9 OLI NDVI map before and after BBR correction. The experimental evaluation shows that the proposed BBR approach outperforms the state-of-the-art image registration techniques.
波段到波段配准(BBR)是特定遥感多光谱传感器必不可少的一项先决图像处理操作。波段对波段配准的目的是以亚像素级的精度将光谱波长通道相互配准。本文提出了一种新颖的 BBR 技术,利用共生尺度空间(CSS)进行特征点检测,并利用空间限制 RANSAC(SC-RANSAC)去除离群匹配控制点。此外,分段仿射变换 (SAT) 模型可减少失真并确保 BBR 的一致性。利用印度纳卫星(INS-2B)上的 Nano-MX 多光谱图像对所开发的方法进行了评估,这些图像覆盖了不同的地貌。在以云像素为主的卫星场景上,使用所提出方法的 BBR 性能也在 4 倍缩放级别上得到了直观验证。分析了波段错误注册对 INS-2B 归一化植被指数(NDVI)的影响,并与 BBR 校正前后最接近的 Landsat-9 OLI NDVI 地图进行了交叉验证。实验评估表明,所提出的 BBR 方法优于最先进的图像配准技术。
{"title":"Enhanced Multispectral Band-to-Band Registration Using Co-Occurrence Scale Space and Spatial Confined RANSAC Guided Segmented Affine Transformation","authors":"Indranil Misra;Mukesh Kumar Rohil;S. Manthira Moorthi;Debajyoti Dhar","doi":"10.1109/TIP.2024.3494555","DOIUrl":"10.1109/TIP.2024.3494555","url":null,"abstract":"Band-to-Band Registration (BBR) is a pre-requisite image processing operation essential for specific remote sensing multispectral sensors. BBR aims to align spectral wavelength channels at sub-pixel level accuracy over each other. The paper presents a novel BBR technique utilizing Co-occurrence Scale Space (CSS) for feature point detection and Spatial Confined RANSAC (SC-RANSAC) for removing outlier matched control points. Additionally, the Segmented Affine Transformation (SAT) model reduces distortion and ensures consistent BBR. The methodology developed is evaluated with Nano-MX multispectral images onboard the Indian Nano Satellite (INS-2B) covering diverse landscapes. BBR performance using the proposed method is also verified visually at a 4X zoom level on satellite scenes dominated by cloud pixels. The band misregistration effect on the Normalized Difference Vegetation Index (NDVI) from INS-2B is analyzed and cross-validated with the closest acquisition Landsat-9 OLI NDVI map before and after BBR correction. The experimental evaluation shows that the proposed BBR approach outperforms the state-of-the-art image registration techniques.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6521-6534"},"PeriodicalIF":0.0,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142637274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SegHSI: Semantic Segmentation of Hyperspectral Images With Limited Labeled Pixels SegHSI:利用有限的标记像素对高光谱图像进行语义分割
Huan Liu;Wei Li;Xiang-Gen Xia;Mengmeng Zhang;Zhengqi Guo;Lujie Song
Hyperspectral images (HSIs), with hundreds of narrow spectral bands, are increasingly used for ground object classification in remote sensing. However, many HSI classification models operate pixel-by-pixel, limiting the utilization of spatial information and resulting in increased inference time for the whole image. This paper proposes SegHSI, an effective and efficient end-to-end HSI segmentation model, alongside a novel training strategy. SegHSI adopts a head-free structure with cluster attention modules and spatial-aware feedforward networks (SA-FFN) for multiscale spatial encoding. Cluster attention encodes pixels through constructed clusters within the HSI, while SA-FFN integrates depth-wise convolution to enhance spatial context. Our training strategy utilizes a student-teacher model framework that combines labeled pixel class information with consistency learning on unlabeled pixels. Experiments on three public HSI datasets demonstrate that SegHSI not only surpasses other state-of-the-art models in segmentation accuracy but also achieves inference time at the scale of seconds, even reaching sub-second speeds for full-image classification. Code is available at https://github.com/huanliu233/SegHSI.
高光谱图像(HSI)具有数百个窄光谱带,越来越多地用于遥感中的地面物体分类。然而,许多 HSI 分类模型都是逐个像素进行操作,限制了空间信息的利用,导致整个图像的推理时间增加。本文提出了一种高效的端到端 HSI 分割模型 SegHSI 以及一种新颖的训练策略。SegHSI 采用无头结构,带有集群注意模块和空间感知前馈网络(SA-FFN),用于多尺度空间编码。集群注意通过在 HSI 中构建的集群对像素进行编码,而 SA-FFN 则整合了深度卷积以增强空间上下文。我们的训练策略采用学生-教师模型框架,将标记像素类别信息与未标记像素的一致性学习相结合。在三个公共 HSI 数据集上的实验表明,SegHSI 不仅在分割准确率上超越了其他最先进的模型,而且推理时间也达到了秒级,甚至在全图分类上达到了亚秒级的速度。代码见 https://github.com/huanliu233/SegHSI。
{"title":"SegHSI: Semantic Segmentation of Hyperspectral Images With Limited Labeled Pixels","authors":"Huan Liu;Wei Li;Xiang-Gen Xia;Mengmeng Zhang;Zhengqi Guo;Lujie Song","doi":"10.1109/TIP.2024.3492724","DOIUrl":"10.1109/TIP.2024.3492724","url":null,"abstract":"Hyperspectral images (HSIs), with hundreds of narrow spectral bands, are increasingly used for ground object classification in remote sensing. However, many HSI classification models operate pixel-by-pixel, limiting the utilization of spatial information and resulting in increased inference time for the whole image. This paper proposes SegHSI, an effective and efficient end-to-end HSI segmentation model, alongside a novel training strategy. SegHSI adopts a head-free structure with cluster attention modules and spatial-aware feedforward networks (SA-FFN) for multiscale spatial encoding. Cluster attention encodes pixels through constructed clusters within the HSI, while SA-FFN integrates depth-wise convolution to enhance spatial context. Our training strategy utilizes a student-teacher model framework that combines labeled pixel class information with consistency learning on unlabeled pixels. Experiments on three public HSI datasets demonstrate that SegHSI not only surpasses other state-of-the-art models in segmentation accuracy but also achieves inference time at the scale of seconds, even reaching sub-second speeds for full-image classification. Code is available at \u0000<uri>https://github.com/huanliu233/SegHSI</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6469-6482"},"PeriodicalIF":0.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142601116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Noisy-Aware Unsupervised Domain Adaptation for Scene Text Recognition 用于场景文本识别的噪声感知无监督域自适应技术
Xiao-Qian Liu;Peng-Fei Zhang;Xin Luo;Zi Huang;Xin-Shun Xu
Unsupervised Domain Adaptation (UDA) has shown promise in Scene Text Recognition (STR) by facilitating knowledge transfer from labeled synthetic text (source) to more challenging unlabeled real scene text (target). However, existing UDA-based STR methods fully rely on the pseudo-labels of target samples, which ignores the impact of domain gaps (inter-domain noise) and various natural environments (intra-domain noise), resulting in poor pseudo-label quality. In this paper, we propose a novel noisy-aware unsupervised domain adaptation framework tailored for STR, which aims to enhance model robustness against both inter- and intra-domain noise, thereby providing more precise pseudo-labels for target samples. Concretely, we propose a reweighting target pseudo-labels by estimating the entropy of refined probability distributions, which mitigates the impact of domain gaps on pseudo-labels. Additionally, a decoupled triple-P-N consistency matching module is proposed, which leverages data augmentation to increase data diversity, enhancing model robustness in diverse natural environments. Within this module, we design a low-confidence-based character negative learning, which is decoupled from high-confidence-based positive learning, thus improving sample utilization under scarce target samples. Furthermore, we extend our framework to the more challenging Source-Free UDA (SFUDA) setting, where only a pre-trained source model is available for adaptation, with no access to source data. Experimental results on benchmark datasets demonstrate the effectiveness of our framework. Under the SFUDA setting, our method exhibits faster convergence and superior performance with less training data than previous UDA-based STR methods. Our method surpasses representative STR methods, establishing new state-of-the-art results across multiple datasets.
无监督域自适应(UDA)可促进知识从有标签的合成文本(源文本)转移到更具挑战性的无标签真实场景文本(目标文本),因此在场景文本识别(STR)领域大有可为。然而,现有的基于 UDA 的 STR 方法完全依赖于目标样本的伪标签,忽略了领域间隙(领域间噪声)和各种自然环境(领域内噪声)的影响,导致伪标签质量低下。在本文中,我们提出了一种为 STR 量身定制的新型噪声感知无监督域适应框架,旨在增强模型对域间和域内噪声的鲁棒性,从而为目标样本提供更精确的伪标签。具体来说,我们提出了一种通过估计精炼概率分布的熵来重新加权目标伪标签的方法,从而减轻了领域间隙对伪标签的影响。此外,我们还提出了一个解耦的三重-P-N 一致性匹配模块,该模块利用数据扩增来增加数据多样性,从而增强模型在多样化自然环境中的鲁棒性。在该模块中,我们设计了基于低置信度的字符负学习,该学习与基于高置信度的正学习解耦,从而在目标样本稀缺的情况下提高了样本利用率。此外,我们还将框架扩展到更具挑战性的无源 UDA(SFUDA)环境中,在这种环境中,只有一个预先训练好的源模型可用于适配,而无法获取源数据。基准数据集上的实验结果证明了我们框架的有效性。在 SFUDA 环境下,与之前基于 UDA 的 STR 方法相比,我们的方法收敛速度更快,在使用较少训练数据的情况下性能更优。我们的方法超越了具有代表性的 STR 方法,在多个数据集上取得了新的一流成果。
{"title":"Noisy-Aware Unsupervised Domain Adaptation for Scene Text Recognition","authors":"Xiao-Qian Liu;Peng-Fei Zhang;Xin Luo;Zi Huang;Xin-Shun Xu","doi":"10.1109/TIP.2024.3492705","DOIUrl":"10.1109/TIP.2024.3492705","url":null,"abstract":"Unsupervised Domain Adaptation (UDA) has shown promise in Scene Text Recognition (STR) by facilitating knowledge transfer from labeled synthetic text (source) to more challenging unlabeled real scene text (target). However, existing UDA-based STR methods fully rely on the pseudo-labels of target samples, which ignores the impact of domain gaps (inter-domain noise) and various natural environments (intra-domain noise), resulting in poor pseudo-label quality. In this paper, we propose a novel noisy-aware unsupervised domain adaptation framework tailored for STR, which aims to enhance model robustness against both inter- and intra-domain noise, thereby providing more precise pseudo-labels for target samples. Concretely, we propose a reweighting target pseudo-labels by estimating the entropy of refined probability distributions, which mitigates the impact of domain gaps on pseudo-labels. Additionally, a decoupled triple-P-N consistency matching module is proposed, which leverages data augmentation to increase data diversity, enhancing model robustness in diverse natural environments. Within this module, we design a low-confidence-based character negative learning, which is decoupled from high-confidence-based positive learning, thus improving sample utilization under scarce target samples. Furthermore, we extend our framework to the more challenging Source-Free UDA (SFUDA) setting, where only a pre-trained source model is available for adaptation, with no access to source data. Experimental results on benchmark datasets demonstrate the effectiveness of our framework. Under the SFUDA setting, our method exhibits faster convergence and superior performance with less training data than previous UDA-based STR methods. Our method surpasses representative STR methods, establishing new state-of-the-art results across multiple datasets.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6550-6563"},"PeriodicalIF":0.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142601119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PVPUFormer: Probabilistic Visual Prompt Unified Transformer for Interactive Image Segmentation PVPUFormer:用于交互式图像分割的概率视觉提示统一变换器
Xu Zhang;Kailun Yang;Jiacheng Lin;Jin Yuan;Zhiyong Li;Shutao Li
Integration of diverse visual prompts like clicks, scribbles, and boxes in interactive image segmentation significantly facilitates users’ interaction as well as improves interaction efficiency. However, existing studies primarily encode the position or pixel regions of prompts without considering the contextual areas around them, resulting in insufficient prompt feedback, which is not conducive to performance acceleration. To tackle this problem, this paper proposes a simple yet effective Probabilistic Visual Prompt Unified Transformer (PVPUFormer) for interactive image segmentation, which allows users to flexibly input diverse visual prompts with the probabilistic prompt encoding and feature post-processing to excavate sufficient and robust prompt features for performance boosting. Specifically, we first propose a Probabilistic Prompt-unified Encoder (PPuE) to generate a unified one-dimensional vector by exploring both prompt and non-prompt contextual information, offering richer feedback cues to accelerate performance improvement. On this basis, we further present a Prompt-to-Pixel Contrastive (P2C) loss to accurately align both prompt and pixel features, bridging the representation gap between them to offer consistent feature representations for mask prediction. Moreover, our approach designs a Dual-cross Merging Attention (DMA) module to implement bidirectional feature interaction between image and prompt features, generating notable features for performance improvement. A comprehensive variety of experiments on several challenging datasets demonstrates that the proposed components achieve consistent improvements, yielding state-of-the-art interactive segmentation performance. Our code is available at https://github.com/XuZhang1211/PVPUFormer.
在交互式图像分割中整合点击、涂鸦和方框等多种视觉提示,可大大方便用户的交互,并提高交互效率。然而,现有研究主要对提示的位置或像素区域进行编码,而没有考虑提示周围的上下文区域,导致提示反馈不足,不利于提高性能。针对这一问题,本文提出了一种简单而有效的用于交互式图像分割的概率视觉提示统一变换器(PVPUFormer),允许用户灵活地输入多样化的视觉提示,并通过概率提示编码和特征后处理挖掘出充分而稳健的提示特征,以提高性能。具体来说,我们首先提出了概率提示统一编码器(PPuE),通过探索提示和非提示上下文信息生成统一的一维向量,提供更丰富的反馈线索,从而加速性能提升。在此基础上,我们进一步提出了 "提示到像素对比"(Prompt-to-Pixel Contrastive,P2C)损失,以精确调整提示和像素特征,弥合两者之间的表征差距,为掩码预测提供一致的特征表征。此外,我们的方法还设计了双交叉合并注意(DMA)模块,以实现图像和提示特征之间的双向特征交互,从而生成显著的特征来提高性能。在几个具有挑战性的数据集上进行的各种实验表明,所提出的组件实现了一致的改进,产生了最先进的交互式分割性能。我们的代码见 https://github.com/XuZhang1211/PVPUFormer。
{"title":"PVPUFormer: Probabilistic Visual Prompt Unified Transformer for Interactive Image Segmentation","authors":"Xu Zhang;Kailun Yang;Jiacheng Lin;Jin Yuan;Zhiyong Li;Shutao Li","doi":"10.1109/TIP.2024.3492713","DOIUrl":"10.1109/TIP.2024.3492713","url":null,"abstract":"Integration of diverse visual prompts like clicks, scribbles, and boxes in interactive image segmentation significantly facilitates users’ interaction as well as improves interaction efficiency. However, existing studies primarily encode the position or pixel regions of prompts without considering the contextual areas around them, resulting in insufficient prompt feedback, which is not conducive to performance acceleration. To tackle this problem, this paper proposes a simple yet effective Probabilistic Visual Prompt Unified Transformer (PVPUFormer) for interactive image segmentation, which allows users to flexibly input diverse visual prompts with the probabilistic prompt encoding and feature post-processing to excavate sufficient and robust prompt features for performance boosting. Specifically, we first propose a Probabilistic Prompt-unified Encoder (PPuE) to generate a unified one-dimensional vector by exploring both prompt and non-prompt contextual information, offering richer feedback cues to accelerate performance improvement. On this basis, we further present a Prompt-to-Pixel Contrastive (P2C) loss to accurately align both prompt and pixel features, bridging the representation gap between them to offer consistent feature representations for mask prediction. Moreover, our approach designs a Dual-cross Merging Attention (DMA) module to implement bidirectional feature interaction between image and prompt features, generating notable features for performance improvement. A comprehensive variety of experiments on several challenging datasets demonstrates that the proposed components achieve consistent improvements, yielding state-of-the-art interactive segmentation performance. Our code is available at \u0000<uri>https://github.com/XuZhang1211/PVPUFormer</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6455-6468"},"PeriodicalIF":0.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142601118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainability Enhanced Object Detection Transformer With Feature Disentanglement 利用特征分解增强物体检测变换器的可解释性
Wenlong Yu;Ruonan Liu;Dongyue Chen;Qinghua Hu
Explainability is a pivotal factor in determining whether a deep learning model can be authorized in critical applications. To enhance the explainability of models of end-to-end object DEtection with TRansformer (DETR), we introduce a disentanglement method that constrains the feature learning process, following a divide-and-conquer decoupling paradigm, similar to how people understand complex real-world problems. We first demonstrate the entangled property of the features between the extractor and detector and find that the regression function is a key factor contributing to the deterioration of disentangled feature activation. These highly entangled features always activate the local characteristics, making it difficult to cover the semantic information of an object, which also reduces the interpretability of single-backbone object detection models. Thus, an Explainability Enhanced object detection Transformer with feature Disentanglement (DETD) model is proposed, in which the Tensor Singular Value Decomposition (T-SVD) is used to produce feature bases and the Batch averaged Feature Spectral Penalization (BFSP) loss is introduced to constrain the disentanglement of the feature and balance the semantic activation. The proposed method is applied across three prominent backbones, two DETR variants, and a CNN based model. By combining two optimization techniques, extensive experiments on two datasets consistently demonstrate that the DETD model outperforms the counterpart in terms of object detection performance and feature disentanglement. The Grad-CAM visualizations demonstrate the enhancement of feature learning explainability in the disentanglement view.
可解释性是决定深度学习模型能否在关键应用中获得授权的关键因素。为了提高端到端对象检测与转换器(DETR)模型的可解释性,我们引入了一种解缠方法,该方法遵循分而治之的解耦范式,限制了特征学习过程,类似于人们理解复杂现实世界问题的方式。我们首先展示了提取器和检测器之间特征的纠缠特性,并发现回归函数是导致解纠缠特征激活恶化的关键因素。这些高度纠缠的特征总是激活局部特征,难以涵盖物体的语义信息,这也降低了单骨干物体检测模型的可解释性。因此,本文提出了一种带特征解缠的可解释性增强物体检测变换器(DETD)模型,其中使用张量奇异值分解(T-SVD)来生成特征基,并引入批量平均特征谱惩罚(BFSP)损失来约束特征的解缠并平衡语义激活。所提出的方法适用于三个突出的骨干网、两个 DETR 变体和一个基于 CNN 的模型。通过结合两种优化技术,在两个数据集上进行的大量实验一致表明,DETD 模型在物体检测性能和特征解缠方面优于对应模型。Grad-CAM 可视化展示了在解缠视图中特征学习可解释性的增强。
{"title":"Explainability Enhanced Object Detection Transformer With Feature Disentanglement","authors":"Wenlong Yu;Ruonan Liu;Dongyue Chen;Qinghua Hu","doi":"10.1109/TIP.2024.3492733","DOIUrl":"10.1109/TIP.2024.3492733","url":null,"abstract":"Explainability is a pivotal factor in determining whether a deep learning model can be authorized in critical applications. To enhance the explainability of models of end-to-end object DEtection with TRansformer (DETR), we introduce a disentanglement method that constrains the feature learning process, following a divide-and-conquer decoupling paradigm, similar to how people understand complex real-world problems. We first demonstrate the entangled property of the features between the extractor and detector and find that the regression function is a key factor contributing to the deterioration of disentangled feature activation. These highly entangled features always activate the local characteristics, making it difficult to cover the semantic information of an object, which also reduces the interpretability of single-backbone object detection models. Thus, an Explainability Enhanced object detection Transformer with feature Disentanglement (DETD) model is proposed, in which the Tensor Singular Value Decomposition (T-SVD) is used to produce feature bases and the Batch averaged Feature Spectral Penalization (BFSP) loss is introduced to constrain the disentanglement of the feature and balance the semantic activation. The proposed method is applied across three prominent backbones, two DETR variants, and a CNN based model. By combining two optimization techniques, extensive experiments on two datasets consistently demonstrate that the DETD model outperforms the counterpart in terms of object detection performance and feature disentanglement. The Grad-CAM visualizations demonstrate the enhancement of feature learning explainability in the disentanglement view.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6439-6454"},"PeriodicalIF":0.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142601130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Constructing Diverse Inlier Consistency for Partial Point Cloud Registration 为部分点云注册构建多样的离层一致性
Yu-Xin Zhang;Jie Gui;James Tin-Yau Kwok
Partial point cloud registration aims to align partial scans into a shared coordinate system. While learning-based partial point cloud registration methods have achieved remarkable progress, they often fail to take full advantage of the relative positional relationships both within (intra-) and between (inter-) point clouds. This oversight hampers their ability to accurately identify overlapping regions and search for reliable correspondences. To address these limitations, a diverse inlier consistency (DIC) method has been proposed that adaptively embeds the positional information of a reliable correspondence in the intra- and inter-point cloud. Firstly, a diverse inlier consistency-driven region perception (DICdRP) module is devised, which encodes the positional information of the selected correspondence within the intra-point cloud. This module enhances the sensitivity of all points to overlapping regions by recognizing the position of the selected correspondence. Secondly, a diverse inlier consistency-aware correspondence search (DICaCS) module is developed, which leverages relative positions in the inter-point cloud. This module studies an inter-point cloud DIC weight to supervise correspondence compatibility, allowing for precise identification of correspondences and effective outlier filtration. Thirdly, diverse information is integrated throughout our framework to achieve a more holistic and detailed registration process. Extensive experiments on object-level and scene-level datasets demonstrate the superior performance of the proposed algorithm. The code is available at https://github.com/yxzhang15/DIC.
局部点云注册的目的是将局部扫描对齐到一个共享坐标系中。虽然基于学习的部分点云注册方法取得了显著进展,但它们往往无法充分利用点云内部和点云之间的相对位置关系。这种疏忽妨碍了它们准确识别重叠区域和搜索可靠对应关系的能力。为了解决这些局限性,我们提出了一种多样化离群值一致性(DIC)方法,该方法能自适应地在点云内部和点云之间嵌入可靠对应关系的位置信息。首先,设计了一个多样化离群值一致性驱动的区域感知(DICdRP)模块,该模块将所选对应点的位置信息嵌入点内云。该模块通过识别所选对应点的位置,提高所有点对重叠区域的敏感度。其次,我们还开发了一个利用点间云中相对位置的多样化离群一致性感知对应搜索(DICaCS)模块。该模块研究点间云 DIC 权重,以监督对应关系的兼容性,从而精确识别对应关系并有效过滤异常值。第三,我们在整个框架中整合了各种信息,以实现更全面、更详细的注册过程。在对象级和场景级数据集上的广泛实验证明了所提算法的卓越性能。代码见 https://github.com/yxzhang15/DIC。
{"title":"Constructing Diverse Inlier Consistency for Partial Point Cloud Registration","authors":"Yu-Xin Zhang;Jie Gui;James Tin-Yau Kwok","doi":"10.1109/TIP.2024.3492700","DOIUrl":"10.1109/TIP.2024.3492700","url":null,"abstract":"Partial point cloud registration aims to align partial scans into a shared coordinate system. While learning-based partial point cloud registration methods have achieved remarkable progress, they often fail to take full advantage of the relative positional relationships both within (intra-) and between (inter-) point clouds. This oversight hampers their ability to accurately identify overlapping regions and search for reliable correspondences. To address these limitations, a diverse inlier consistency (DIC) method has been proposed that adaptively embeds the positional information of a reliable correspondence in the intra- and inter-point cloud. Firstly, a diverse inlier consistency-driven region perception (DICdRP) module is devised, which encodes the positional information of the selected correspondence within the intra-point cloud. This module enhances the sensitivity of all points to overlapping regions by recognizing the position of the selected correspondence. Secondly, a diverse inlier consistency-aware correspondence search (DICaCS) module is developed, which leverages relative positions in the inter-point cloud. This module studies an inter-point cloud DIC weight to supervise correspondence compatibility, allowing for precise identification of correspondences and effective outlier filtration. Thirdly, diverse information is integrated throughout our framework to achieve a more holistic and detailed registration process. Extensive experiments on object-level and scene-level datasets demonstrate the superior performance of the proposed algorithm. The code is available at \u0000<uri>https://github.com/yxzhang15/DIC</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6535-6549"},"PeriodicalIF":0.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142601132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cost Volume Aggregation in Stereo Matching Revisited: A Disparity Classification Perspective 再论立体匹配中的成本量聚合:差异分类视角
Yun Wang;Longguang Wang;Kunhong Li;Yongjian Zhang;Dapeng Oliver Wu;Yulan Guo
Cost aggregation plays a critical role in existing stereo matching methods. In this paper, we revisit cost aggregation in stereo matching from disparity classification and propose a generic yet efficient Disparity Context Aggregation (DCA) module to improve the performance of CNN-based methods. Our approach is based on an insight that a coarse disparity class prior is beneficial to disparity regression. To obtain such a prior, we first classify pixels in an image into several disparity classes and treat pixels within the same class as homogeneous regions. We then generate homogeneous region representations and incorporate these representations into the cost volume to suppress irrelevant information while enhancing the matching ability for cost aggregation. With the help of homogeneous region representations, efficient and informative cost aggregation can be achieved with only a shallow 3D CNN. Our DCA module is fully-differentiable and well-compatible with different network architectures, which can be seamlessly plugged into existing networks to improve performance with small additional overheads. It is demonstrated that our DCA module can effectively exploit disparity class priors to improve the performance of cost aggregation. Based on our DCA, we design a highly accurate network named DCANet, which achieves state-of-the-art performance on several benchmarks.
成本聚合在现有的立体匹配方法中起着至关重要的作用。在本文中,我们从差异分类的角度重新审视了立体匹配中的成本聚合,并提出了一种通用而高效的差异上下文聚合(DCA)模块,以提高基于 CNN 的方法的性能。我们的方法基于一种见解,即粗略的差异类别先验有利于差异回归。为了获得这样的先验,我们首先将图像中的像素划分为多个差异类别,并将同一类别中的像素视为同质区域。然后,我们生成同质区域表征,并将这些表征纳入成本量,以抑制无关信息,同时增强成本聚合的匹配能力。在同质区域表示法的帮助下,只需一个浅层三维 CNN 就能实现高效且信息丰富的成本汇总。我们的 DCA 模块是完全可变的,与不同的网络架构兼容,可无缝插入现有网络,以较小的额外开销提高性能。实验证明,我们的 DCA 模块可以有效利用差异类先验来提高成本聚合的性能。基于我们的 DCA,我们设计了一个名为 DCANet 的高精度网络,该网络在多个基准测试中取得了最先进的性能。
{"title":"Cost Volume Aggregation in Stereo Matching Revisited: A Disparity Classification Perspective","authors":"Yun Wang;Longguang Wang;Kunhong Li;Yongjian Zhang;Dapeng Oliver Wu;Yulan Guo","doi":"10.1109/TIP.2024.3484251","DOIUrl":"10.1109/TIP.2024.3484251","url":null,"abstract":"Cost aggregation plays a critical role in existing stereo matching methods. In this paper, we revisit cost aggregation in stereo matching from disparity classification and propose a generic yet efficient Disparity Context Aggregation (DCA) module to improve the performance of CNN-based methods. Our approach is based on an insight that a coarse disparity class prior is beneficial to disparity regression. To obtain such a prior, we first classify pixels in an image into several disparity classes and treat pixels within the same class as homogeneous regions. We then generate homogeneous region representations and incorporate these representations into the cost volume to suppress irrelevant information while enhancing the matching ability for cost aggregation. With the help of homogeneous region representations, efficient and informative cost aggregation can be achieved with only a shallow 3D CNN. Our DCA module is fully-differentiable and well-compatible with different network architectures, which can be seamlessly plugged into existing networks to improve performance with small additional overheads. It is demonstrated that our DCA module can effectively exploit disparity class priors to improve the performance of cost aggregation. Based on our DCA, we design a highly accurate network named DCANet, which achieves state-of-the-art performance on several benchmarks.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6425-6438"},"PeriodicalIF":0.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142599222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Smooth Tensor Product for Tensor Completion 用于张量补全的平滑张量积
Tongle Wu;Jicong Fan
Low-rank tensor completion (LRTC) has shown promise in processing incomplete visual data, yet it often overlooks the inherent local smooth structures in images and videos. Recent advances in LRTC, integrating total variation regularization to capitalize on the local smoothness, have yielded notable improvements. Nonetheless, these methods are limited to exploiting local smoothness within the original data space, neglecting the latent factor space of tensors. More seriously, there is a lack of theoretical backing for the role of local smoothness in enhancing recovery performance. In response, this paper introduces an innovative tensor completion model that concurrently leverages the global low-rank structure of the original tensor and the local smooth structure of its factor tensors. Our objective is to learn a low-rank tensor that decomposes into two factor tensors, each exhibiting sufficient local smoothness. We propose an efficient alternating direction method of multipliers to optimize our model. Further, we establish generalization error bounds for smooth factor-based tensor completion methods across various decomposition frameworks. These bounds are significantly tighter than existing baselines. We conduct extensive inpainting experiments on color images, multispectral images, and videos, which demonstrate the efficacy and superiority of our method. Additionally, our approach shows a low sensitivity to hyper-parameter settings, enhancing its convenience and reliability for practical applications.
低秩张量补全(LRTC)在处理不完整的视觉数据方面大有可为,但它往往忽略了图像和视频中固有的局部平滑结构。低秩张量补全(LRTC)的最新进展是整合总变异正则化,以利用局部平滑性,取得了显著的改进。然而,这些方法仅限于利用原始数据空间内的局部平滑性,忽略了张量的潜在因子空间。更严重的是,局部平滑性在提高恢复性能方面的作用缺乏理论支持。为此,本文介绍了一种创新的张量补全模型,它能同时利用原始张量的全局低秩结构及其因子张量的局部平滑结构。我们的目标是学习一种能分解成两个因子张量的低秩张量,每个因子张量都表现出足够的局部平滑性。我们提出了一种高效的交替方向乘法来优化我们的模型。此外,我们还为各种分解框架中基于平滑因子的张量补全方法建立了广义误差边界。这些界限比现有的基线严格得多。我们在彩色图像、多光谱图像和视频上进行了广泛的内绘实验,证明了我们方法的有效性和优越性。此外,我们的方法对超参数设置的敏感度很低,从而提高了实际应用的便利性和可靠性。
{"title":"Smooth Tensor Product for Tensor Completion","authors":"Tongle Wu;Jicong Fan","doi":"10.1109/TIP.2024.3489272","DOIUrl":"10.1109/TIP.2024.3489272","url":null,"abstract":"Low-rank tensor completion (LRTC) has shown promise in processing incomplete visual data, yet it often overlooks the inherent local smooth structures in images and videos. Recent advances in LRTC, integrating total variation regularization to capitalize on the local smoothness, have yielded notable improvements. Nonetheless, these methods are limited to exploiting local smoothness within the original data space, neglecting the latent factor space of tensors. More seriously, there is a lack of theoretical backing for the role of local smoothness in enhancing recovery performance. In response, this paper introduces an innovative tensor completion model that concurrently leverages the global low-rank structure of the original tensor and the local smooth structure of its factor tensors. Our objective is to learn a low-rank tensor that decomposes into two factor tensors, each exhibiting sufficient local smoothness. We propose an efficient alternating direction method of multipliers to optimize our model. Further, we establish generalization error bounds for smooth factor-based tensor completion methods across various decomposition frameworks. These bounds are significantly tighter than existing baselines. We conduct extensive inpainting experiments on color images, multispectral images, and videos, which demonstrate the efficacy and superiority of our method. Additionally, our approach shows a low sensitivity to hyper-parameter settings, enhancing its convenience and reliability for practical applications.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":"33 ","pages":"6483-6496"},"PeriodicalIF":0.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142599221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1