首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
CalibNet: Dual-Branch Cross-Modal Calibration for RGB-D Salient Instance Segmentation CalibNet:用于 RGB-D 突出实例分割的双分支跨模态校准。
Jialun Pei;Tao Jiang;He Tang;Nian Liu;Yueming Jin;Deng-Ping Fan;Pheng-Ann Heng
In this study, we propose a novel approach for RGB-D salient instance segmentation using a dual-branch cross-modal feature calibration architecture called CalibNet. Our method simultaneously calibrates depth and RGB features in the kernel and mask branches to generate instance-aware kernels and mask features. CalibNet consists of three simple modules, a dynamic interactive kernel (DIK) and a weight-sharing fusion (WSF), which work together to generate effective instance-aware kernels and integrate cross-modal features. To improve the quality of depth features, we incorporate a depth similarity assessment (DSA) module prior to DIK and WSF. In addition, we further contribute a new DSIS dataset, which contains 1,940 images with elaborate instance-level annotations. Extensive experiments on three challenging benchmarks show that CalibNet yields a promising result, i.e., 58.0% AP with $320times 480$ input size on the COME15K-E test set, which significantly surpasses the alternative frameworks. Our code and dataset will be publicly available at: https://github.com/PJLallen/CalibNet.
在本研究中,我们提出了一种使用名为 CalibNet 的双分支跨模态特征校准架构进行 RGB-D 突出实例分割的新方法。我们的方法在内核和掩码分支中同时校准深度和 RGB 特征,以生成实例感知内核和掩码特征。CalibNet 由三个简单的模块组成:动态交互内核(DIK)和权重共享融合(WSF),它们共同作用生成有效的实例感知内核并整合跨模态特征。为了提高深度特征的质量,我们在 DIK 和 WSF 之前加入了深度相似性评估(DSA)模块。此外,我们还进一步贡献了一个新的 DSIS 数据集,该数据集包含 1,940 张带有详细实例级注释的图像。在三个具有挑战性的基准上进行的广泛实验表明,CalibNet 取得了可喜的成果,即在 COME15K-E 测试集上,输入大小为 320×480 的 AP 为 58.0%,大大超过了其他框架。我们的代码和数据集将在以下网站公开:https://github.com/PJLallen/CalibNet。
{"title":"CalibNet: Dual-Branch Cross-Modal Calibration for RGB-D Salient Instance Segmentation","authors":"Jialun Pei;Tao Jiang;He Tang;Nian Liu;Yueming Jin;Deng-Ping Fan;Pheng-Ann Heng","doi":"10.1109/TIP.2024.3432328","DOIUrl":"10.1109/TIP.2024.3432328","url":null,"abstract":"In this study, we propose a novel approach for RGB-D salient instance segmentation using a dual-branch cross-modal feature calibration architecture called CalibNet. Our method simultaneously calibrates depth and RGB features in the kernel and mask branches to generate instance-aware kernels and mask features. CalibNet consists of three simple modules, a dynamic interactive kernel (DIK) and a weight-sharing fusion (WSF), which work together to generate effective instance-aware kernels and integrate cross-modal features. To improve the quality of depth features, we incorporate a depth similarity assessment (DSA) module prior to DIK and WSF. In addition, we further contribute a new DSIS dataset, which contains 1,940 images with elaborate instance-level annotations. Extensive experiments on three challenging benchmarks show that CalibNet yields a promising result, i.e., 58.0% AP with \u0000<inline-formula> <tex-math>$320times 480$ </tex-math></inline-formula>\u0000 input size on the COME15K-E test set, which significantly surpasses the alternative frameworks. Our code and dataset will be publicly available at: \u0000<uri>https://github.com/PJLallen/CalibNet</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141794319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Student Network Under Universal Label Noise 通用标签噪声下的学习学生网络。
Jialiang Tang;Ning Jiang;Hongyuan Zhu;Joey Tianyi Zhou;Chen Gong
Data-free knowledge distillation aims to learn a small student network from a large pre-trained teacher network without the aid of original training data. Recent works propose to gather alternative data from the Internet for training student network. In a more realistic scenario, the data on the Internet contains two types of label noise, namely: 1) closed-set label noise, where some examples belong to the known categories but are mislabeled; and 2) open-set label noise, where the true labels of some mislabeled examples are outside the known categories. However, the latter is largely ignored by existing works, leading to limited student network performance. Therefore, this paper proposes a novel data-free knowledge distillation paradigm by utilizing a webly-collected dataset under universal label noise, which means both closed-set and open-set label noise should be tackled. Specifically, we first split the collected noisy dataset into clean set, closed noisy set, and open noisy set based on the prediction uncertainty of various data types. For the closed-set noisy examples, their labels are refined by teacher network. Meanwhile, a noise-robust hybrid contrastive learning is performed on the clean set and refined closed noisy set to encourage student network to learn the categorical and instance knowledge inherited by teacher network. For the open-set noisy examples unexplored by previous work, we regard them as unlabeled and conduct self-supervised learning on them to enrich the supervision signal for student network. Intensive experimental results on image classification tasks demonstrate that our approach can achieve superior performance to state-of-the-art data-free knowledge distillation methods.
无数据知识提炼旨在不借助原始训练数据,从预先训练好的大型教师网络中学习小型学生网络。最近的研究提出从互联网上收集替代数据来训练学生网络。在更现实的情况下,互联网上的数据包含两种类型的标签噪声,即1) 封闭集标签噪声,即某些示例属于已知类别,但被错误标记;以及 2) 开放集标签噪声,即某些被错误标记示例的真实标签不属于已知类别。然而,现有研究在很大程度上忽略了后者,导致学生网络性能有限。因此,本文提出了一种新颖的无数据知识提炼范式,即利用网络收集的数据集来处理普遍标签噪声,也就是同时处理封闭集和开放集标签噪声。具体来说,我们首先根据各种数据类型的预测不确定性,将收集到的噪声数据集分成干净集、封闭噪声集和开放噪声集。对于封闭集的噪声示例,其标签由教师网络完善。同时,在干净集和精炼的封闭噪声集上执行噪声稳健混合对比学习,以鼓励学生网络学习教师网络继承的分类和实例知识。对于前人未探索过的开放噪声集示例,我们将其视为未标记示例,并对其进行自我监督学习,以丰富学生网络的监督信号。在图像分类任务上的大量实验结果表明,我们的方法可以取得优于最先进的无数据知识提炼方法的性能。
{"title":"Learning Student Network Under Universal Label Noise","authors":"Jialiang Tang;Ning Jiang;Hongyuan Zhu;Joey Tianyi Zhou;Chen Gong","doi":"10.1109/TIP.2024.3430539","DOIUrl":"10.1109/TIP.2024.3430539","url":null,"abstract":"Data-free knowledge distillation aims to learn a small student network from a large pre-trained teacher network without the aid of original training data. Recent works propose to gather alternative data from the Internet for training student network. In a more realistic scenario, the data on the Internet contains two types of label noise, namely: 1) closed-set label noise, where some examples belong to the known categories but are mislabeled; and 2) open-set label noise, where the true labels of some mislabeled examples are outside the known categories. However, the latter is largely ignored by existing works, leading to limited student network performance. Therefore, this paper proposes a novel data-free knowledge distillation paradigm by utilizing a webly-collected dataset under universal label noise, which means both closed-set and open-set label noise should be tackled. Specifically, we first split the collected noisy dataset into clean set, closed noisy set, and open noisy set based on the prediction uncertainty of various data types. For the closed-set noisy examples, their labels are refined by teacher network. Meanwhile, a noise-robust hybrid contrastive learning is performed on the clean set and refined closed noisy set to encourage student network to learn the categorical and instance knowledge inherited by teacher network. For the open-set noisy examples unexplored by previous work, we regard them as unlabeled and conduct self-supervised learning on them to enrich the supervision signal for student network. Intensive experimental results on image classification tasks demonstrate that our approach can achieve superior performance to state-of-the-art data-free knowledge distillation methods.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141794320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Event-based Optical Flow via Transforming into Motion-dependent View. 通过转换为运动视图实现基于事件的光流。
Zengyu Wan, Yang Wang, Zhai Wei, Ganchao Tan, Yang Cao, Zheng-Jun Zha

Event cameras respond to temporal dynamics, helping to resolve ambiguities in spatio-temporal changes for optical flow estimation. However, the unique spatio-temporal event distribution challenges the feature extraction, and the direct construction of motion representation through the orthogonal view is less than ideal due to the entanglement of appearance and motion. This paper proposes to transform the orthogonal view into a motion-dependent one for enhancing event-based motion representation and presents a Motion View-based Network (MV-Net) for practical optical flow estimation. Specifically, this motion-dependent view transformation is achieved through the Event View Transformation Module, which captures the relationship between the steepest temporal changes and motion direction, incorporating these temporal cues into the view transformation process for feature gathering. This module includes two phases: extracting the temporal evolution clues by central difference operation in the extraction phase and capturing the motion pattern by evolution-guided deformable convolution in the perception phase. Besides, the MV-Net constructs an eccentric downsampling process to avoid response weakening from the sparsity of events in the downsampling stage. The whole network is trained end-to-end in a self-supervised manner, and the evaluations conducted on four challenging datasets reveal the superior performance of the proposed model compared to state-of-the-art (SOTA) methods.

事件摄像机能对时间动态做出反应,有助于解决光流估计中时空变化的模糊性。然而,独特的时空事件分布给特征提取带来了挑战,而且由于外观和运动的纠缠,通过正交视图直接构建运动表示法并不理想。本文提出将正交视图转换为与运动相关的视图,以增强基于事件的运动表示,并提出了一种基于运动视图的网络(MV-Net),用于实际的光流估计。具体来说,这种与运动相关的视图转换是通过事件视图转换模块实现的,该模块捕捉最陡峭的时间变化与运动方向之间的关系,并将这些时间线索纳入视图转换过程以收集特征。该模块包括两个阶段:在提取阶段,通过中心差分运算提取时间演化线索;在感知阶段,通过演化引导的可变形卷积捕捉运动模式。此外,MV-网络还构建了一个偏心下采样过程,以避免在下采样阶段因事件稀疏而导致响应减弱。整个网络是以自我监督的方式进行端到端训练的,在四个具有挑战性的数据集上进行的评估表明,与最先进的(SOTA)方法相比,所提出的模型具有更优越的性能。
{"title":"Event-based Optical Flow via Transforming into Motion-dependent View.","authors":"Zengyu Wan, Yang Wang, Zhai Wei, Ganchao Tan, Yang Cao, Zheng-Jun Zha","doi":"10.1109/TIP.2024.3426469","DOIUrl":"https://doi.org/10.1109/TIP.2024.3426469","url":null,"abstract":"<p><p>Event cameras respond to temporal dynamics, helping to resolve ambiguities in spatio-temporal changes for optical flow estimation. However, the unique spatio-temporal event distribution challenges the feature extraction, and the direct construction of motion representation through the orthogonal view is less than ideal due to the entanglement of appearance and motion. This paper proposes to transform the orthogonal view into a motion-dependent one for enhancing event-based motion representation and presents a Motion View-based Network (MV-Net) for practical optical flow estimation. Specifically, this motion-dependent view transformation is achieved through the Event View Transformation Module, which captures the relationship between the steepest temporal changes and motion direction, incorporating these temporal cues into the view transformation process for feature gathering. This module includes two phases: extracting the temporal evolution clues by central difference operation in the extraction phase and capturing the motion pattern by evolution-guided deformable convolution in the perception phase. Besides, the MV-Net constructs an eccentric downsampling process to avoid response weakening from the sparsity of events in the downsampling stage. The whole network is trained end-to-end in a self-supervised manner, and the evaluations conducted on four challenging datasets reveal the superior performance of the proposed model compared to state-of-the-art (SOTA) methods.</p>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141768324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Embeddable Implicit IUVD Representation for Part-Based 3D Human Surface Reconstruction 用于基于部件的三维人体表面重建的可嵌入式隐式 IUVD 表示法
Baoxing Li;Yong Deng;Yehui Yang;Xu Zhao
To reconstruct a 3D human surface from a single image, it is crucial to simultaneously consider human pose, shape, and clothing details. Recent approaches have combined parametric body models (such as SMPL), which capture body pose and shape priors, with neural implicit functions that flexibly learn clothing details. However, this combined representation introduces additional computation, e.g. signed distance calculation in 3D body feature extraction, leading to redundancy in the implicit query-and-infer process and failing to preserve the underlying body shape prior. To address these issues, we propose a novel IUVD-Feedback representation, consisting of an IUVD occupancy function and a feedback query algorithm. This representation replaces the time-consuming signed distance calculation with a simple linear transformation in the IUVD space, leveraging the SMPL UV maps. Additionally, it reduces redundant query points through a feedback mechanism, leading to more reasonable 3D body features and more effective query points, thereby preserving the parametric body prior. Moreover, the IUVD-Feedback representation can be embedded into any existing implicit human reconstruction pipeline without requiring modifications to the trained neural networks. Experiments on the THuman2.0 dataset demonstrate that the proposed IUVD-Feedback representation improves the robustness of results and achieves three times faster acceleration in the query-and-infer process. Furthermore, this representation holds potential for generative applications by leveraging its inherent semantic information from the parametric body model.
要从单张图像重建三维人体表面,必须同时考虑人体姿势、形状和服装细节。最近的方法结合了参数化人体模型(如 SMPL)和神经隐函数,前者可捕捉人体姿势和形状先验,后者可灵活学习服装细节。然而,这种组合表示法引入了额外的计算,例如三维身体特征提取中的签名距离计算,导致隐式查询和推理过程中出现冗余,无法保留底层的身体形状先验。为了解决这些问题,我们提出了一种新颖的 IUVD 反馈表示法,它由 IUVD 占有函数和反馈查询算法组成。这种表示方法利用 SMPL UV 地图,用 IUVD 空间中的简单线性变换取代了耗时的符号距离计算。此外,它还通过反馈机制减少了冗余查询点,从而获得更合理的三维人体特征和更有效的查询点,从而保留了参数化人体先验。此外,IUVD-反馈表示法可以嵌入到任何现有的隐式人体重建管道中,而无需修改经过训练的神经网络。在 THuman2.0 数据集上的实验表明,所提出的 IUVD-Feedback 表示法提高了结果的鲁棒性,并在查询和推理过程中实现了三倍的加速。此外,通过利用参数化人体模型的固有语义信息,该表示法还具有生成应用的潜力。
{"title":"An Embeddable Implicit IUVD Representation for Part-Based 3D Human Surface Reconstruction","authors":"Baoxing Li;Yong Deng;Yehui Yang;Xu Zhao","doi":"10.1109/TIP.2024.3430073","DOIUrl":"10.1109/TIP.2024.3430073","url":null,"abstract":"To reconstruct a 3D human surface from a single image, it is crucial to simultaneously consider human pose, shape, and clothing details. Recent approaches have combined parametric body models (such as SMPL), which capture body pose and shape priors, with neural implicit functions that flexibly learn clothing details. However, this combined representation introduces additional computation, e.g. signed distance calculation in 3D body feature extraction, leading to redundancy in the implicit query-and-infer process and failing to preserve the underlying body shape prior. To address these issues, we propose a novel IUVD-Feedback representation, consisting of an IUVD occupancy function and a feedback query algorithm. This representation replaces the time-consuming signed distance calculation with a simple linear transformation in the IUVD space, leveraging the SMPL UV maps. Additionally, it reduces redundant query points through a feedback mechanism, leading to more reasonable 3D body features and more effective query points, thereby preserving the parametric body prior. Moreover, the IUVD-Feedback representation can be embedded into any existing implicit human reconstruction pipeline without requiring modifications to the trained neural networks. Experiments on the THuman2.0 dataset demonstrate that the proposed IUVD-Feedback representation improves the robustness of results and achieves three times faster acceleration in the query-and-infer process. Furthermore, this representation holds potential for generative applications by leveraging its inherent semantic information from the parametric body model.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141768323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MAS-CL: An End-to-End Multi-Atlas Supervised Contrastive Learning Framework for Brain ROI Segmentation MAS-CL:用于大脑 ROI 分割的端到端多图谱监督对比学习框架。
Liang Sun;Yanling Fu;Junyong Zhao;Wei Shao;Qi Zhu;Daoqiang Zhang
Brain region-of-interest (ROI) segmentation with magnetic resonance (MR) images is a basic prerequisite step for brain analysis. The main problem with using deep learning for brain ROI segmentation is the lack of sufficient annotated data. To address this issue, in this paper, we propose a simple multi-atlas supervised contrastive learning framework (MAS-CL) for brain ROI segmentation with MR images in an end-to-end manner. Specifically, our MAS-CL framework mainly consists of two steps, including 1) a multi-atlas supervised contrastive learning method to learn the latent representation using a limited amount of voxel-level labeling brain MR images, and 2) brain ROI segmentation based on the pre-trained backbone using our MSA-CL method. Specifically, different from traditional contrastive learning, in our proposed method, we use multi-atlas supervised information to pre-train the backbone for learning the latent representation of input MR image, i.e., the correlation of each sample pair is defined by using the label maps of input MR image and atlas images. Then, we extend the pre-trained backbone to segment brain ROI with MR images. We perform our proposed MAS-CL framework with five segmentation methods on LONI-LPBA40, IXI, OASIS, ADNI, and CC359 datasets for brain ROI segmentation with MR images. Various experimental results suggested that our proposed MAS-CL framework can significantly improve the segmentation performance on these five datasets.
磁共振(MR)图像的脑兴趣区(ROI)分割是脑分析的基本前提步骤。使用深度学习进行脑兴趣区分割的主要问题是缺乏足够的注释数据。为了解决这个问题,我们在本文中提出了一个简单的多图谱监督对比学习框架(MAS-CL),用于以端到端的方式对磁共振图像进行大脑 ROI 分割。具体来说,我们的 MAS-CL 框架主要包括两个步骤:1)利用有限的体素级标注脑部 MR 图像,采用多图谱监督对比学习方法学习潜表征;2)利用我们的 MSA-CL 方法,基于预训练的骨干进行脑部 ROI 分割。具体来说,与传统的对比学习不同,在我们提出的方法中,我们使用多图集监督信息来预训练骨干,以学习输入 MR 图像的潜表征,即使用输入 MR 图像和图集图像的标签图来定义每个样本对的相关性。然后,我们将预先训练好的反骨干扩展到用磁共振图像分割大脑 ROI。我们在 LONI-LPBA40、IXI、OASIS、ADNI 和 CC359 数据集上使用我们提出的 MAS-CL 框架和五种分割方法对磁共振图像进行脑 ROI 分割。各种实验结果表明,我们提出的 MAS-CL 框架能显著提高这五个数据集的分割性能。
{"title":"MAS-CL: An End-to-End Multi-Atlas Supervised Contrastive Learning Framework for Brain ROI Segmentation","authors":"Liang Sun;Yanling Fu;Junyong Zhao;Wei Shao;Qi Zhu;Daoqiang Zhang","doi":"10.1109/TIP.2024.3431451","DOIUrl":"10.1109/TIP.2024.3431451","url":null,"abstract":"Brain region-of-interest (ROI) segmentation with magnetic resonance (MR) images is a basic prerequisite step for brain analysis. The main problem with using deep learning for brain ROI segmentation is the lack of sufficient annotated data. To address this issue, in this paper, we propose a simple multi-atlas supervised contrastive learning framework (MAS-CL) for brain ROI segmentation with MR images in an end-to-end manner. Specifically, our MAS-CL framework mainly consists of two steps, including 1) a multi-atlas supervised contrastive learning method to learn the latent representation using a limited amount of voxel-level labeling brain MR images, and 2) brain ROI segmentation based on the pre-trained backbone using our MSA-CL method. Specifically, different from traditional contrastive learning, in our proposed method, we use multi-atlas supervised information to pre-train the backbone for learning the latent representation of input MR image, i.e., the correlation of each sample pair is defined by using the label maps of input MR image and atlas images. Then, we extend the pre-trained backbone to segment brain ROI with MR images. We perform our proposed MAS-CL framework with five segmentation methods on LONI-LPBA40, IXI, OASIS, ADNI, and CC359 datasets for brain ROI segmentation with MR images. Various experimental results suggested that our proposed MAS-CL framework can significantly improve the segmentation performance on these five datasets.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141763553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multispectral Image Stitching via Global-Aware Quadrature Pyramid Regression 通过全局感知正交金字塔回归实现多光谱图像拼接
Zhiying Jiang;Zengxi Zhang;Jinyuan Liu;Xin Fan;Risheng Liu
Image stitching is a critical task in panorama perception that involves combining images captured from different viewing positions to reconstruct a wider field-of-view (FOV) image. Existing visible image stitching methods suffer from performance drops under severe conditions since environmental factors can easily impair visible images. In contrast, infrared images possess greater penetrating ability and are less affected by environmental factors. Therefore, we propose an infrared and visible image-based multispectral image stitching method to achieve all-weather, broad FOV scene perception. Specifically, based on two pairs of infrared and visible images, we employ the salient structural information from the infrared images and the textual details from the visible images to infer the correspondences within different modality-specific features. For this purpose, a multiscale progressive mechanism coupled with quadrature correlation is exploited to improve regression in different modalities. Exploiting the complementary properties, accurate and credible homography can be obtained by integrating the deformation parameters of the two modalities to compensate for the missing modality-specific information. A global-aware guided reconstruction module is established to generate an informative and broad scene, wherein the attentive features of different viewpoints are introduced to fuse the source images with a more seamless and comprehensive appearance. We construct a high-quality infrared and visible stitching dataset for evaluation, including real-world and synthetic sets. The qualitative and quantitative results demonstrate that the proposed method outperforms the intuitive cascaded fusion-stitching procedure, achieving more robust and credible panorama generation. Code and dataset are available at https://github.com/Jzy2017/MSGA.
图像拼接是全景感知中的一项关键任务,它涉及将从不同观察位置捕捉到的图像进行组合,以重建更宽视场(FOV)的图像。现有的可见光图像拼接方法在恶劣条件下性能下降,因为环境因素很容易损害可见光图像。相比之下,红外图像具有更强的穿透能力,受环境因素的影响较小。因此,我们提出了一种基于红外和可见光图像的多光谱图像拼接方法,以实现全天候、宽视场的场景感知。具体来说,基于两对红外图像和可见光图像,我们利用红外图像中的突出结构信息和可见光图像中的文字细节来推断不同模态特征之间的对应关系。为此,我们利用多尺度渐进机制和正交相关性来改进不同模态的回归。利用互补特性,通过整合两种模态的变形参数来补偿缺失的特定模态信息,从而获得准确可信的同源性。我们建立了一个全局感知的引导重建模块,以生成信息丰富的广阔场景,其中引入了不同视角的注意特征,从而以更加无缝和全面的外观融合源图像。我们构建了一个高质量的红外和可见光拼接数据集进行评估,其中包括真实世界和合成集。定性和定量结果表明,所提出的方法优于直观的级联融合-拼接程序,能生成更强大、更可信的全景图。代码和数据集见 https://github.com/Jzy2017/MSGA。
{"title":"Multispectral Image Stitching via Global-Aware Quadrature Pyramid Regression","authors":"Zhiying Jiang;Zengxi Zhang;Jinyuan Liu;Xin Fan;Risheng Liu","doi":"10.1109/TIP.2024.3430532","DOIUrl":"10.1109/TIP.2024.3430532","url":null,"abstract":"Image stitching is a critical task in panorama perception that involves combining images captured from different viewing positions to reconstruct a wider field-of-view (FOV) image. Existing visible image stitching methods suffer from performance drops under severe conditions since environmental factors can easily impair visible images. In contrast, infrared images possess greater penetrating ability and are less affected by environmental factors. Therefore, we propose an infrared and visible image-based multispectral image stitching method to achieve all-weather, broad FOV scene perception. Specifically, based on two pairs of infrared and visible images, we employ the salient structural information from the infrared images and the textual details from the visible images to infer the correspondences within different modality-specific features. For this purpose, a multiscale progressive mechanism coupled with quadrature correlation is exploited to improve regression in different modalities. Exploiting the complementary properties, accurate and credible homography can be obtained by integrating the deformation parameters of the two modalities to compensate for the missing modality-specific information. A global-aware guided reconstruction module is established to generate an informative and broad scene, wherein the attentive features of different viewpoints are introduced to fuse the source images with a more seamless and comprehensive appearance. We construct a high-quality infrared and visible stitching dataset for evaluation, including real-world and synthetic sets. The qualitative and quantitative results demonstrate that the proposed method outperforms the intuitive cascaded fusion-stitching procedure, achieving more robust and credible panorama generation. Code and dataset are available at \u0000<uri>https://github.com/Jzy2017/MSGA</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141754775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spiking Transfer Learning From RGB Image to Neuromorphic Event Stream 从 RGB 图像到神经形态事件流的尖峰转移学习
Qiugang Zhan;Guisong Liu;Xiurui Xie;Ran Tao;Malu Zhang;Huajin Tang
Recent advances in bio-inspired vision with event cameras and associated spiking neural networks (SNNs) have provided promising solutions for low-power consumption neuromorphic tasks. However, as the research of event cameras is still in its infancy, the amount of labeled event stream data is much less than that of the RGB database. The traditional method of converting static images into event streams by simulation to increase the sample size cannot simulate the characteristics of event cameras such as high temporal resolution. To take advantage of both the rich knowledge in labeled RGB images and the features of the event camera, we propose a transfer learning method from the RGB to the event domain in this paper. Specifically, we first introduce a transfer learning framework named R2ETL (RGB to Event Transfer Learning), including a novel encoding alignment module and a feature alignment module. Then, we introduce the temporal centered kernel alignment (TCKA) loss function to improve the efficiency of transfer learning. It aligns the distribution of temporal neuron states by adding a temporal learning constraint. Finally, we theoretically analyze the amount of data required by the deep neuromorphic model to prove the necessity of our method. Numerous experiments demonstrate that our proposed framework outperforms the state-of-the-art SNN and artificial neural network (ANN) models trained on event streams, including N-MNIST, CIFAR10-DVS and N-Caltech101. This indicates that the R2ETL framework is able to leverage the knowledge of labeled RGB images to help the training of SNN on event streams.
最近,生物启发视觉领域的进展是利用事件相机和相关的尖峰神经网络(SNN),为低功耗神经形态任务提供了前景广阔的解决方案。然而,由于事件相机的研究仍处于起步阶段,标注的事件流数据量远低于 RGB 数据库。通过模拟将静态图像转换为事件流以增加样本量的传统方法无法模拟事件相机的高时间分辨率等特性。为了充分利用 RGB 图像中的丰富知识和事件摄像机的特点,我们在本文中提出了一种从 RGB 到事件域的迁移学习方法。具体来说,我们首先引入了一个名为 R2ETL(RGB 到事件迁移学习)的迁移学习框架,其中包括一个新颖的编码对齐模块和一个特征对齐模块。然后,我们引入了时间中心核对齐(TCKA)损失函数,以提高迁移学习的效率。它通过添加时空学习约束来对齐时空神经元状态的分布。最后,我们从理论上分析了深度神经形态模型所需的数据量,以证明我们的方法的必要性。大量实验证明,我们提出的框架优于在事件流上训练的最先进的 SNN 和人工神经网络 (ANN) 模型,包括 N-MNIST、CIFAR10-DVS 和 N-Caltech101。这表明 R2ETL 框架能够利用标注 RGB 图像的知识来帮助在事件流上训练 SNN。
{"title":"Spiking Transfer Learning From RGB Image to Neuromorphic Event Stream","authors":"Qiugang Zhan;Guisong Liu;Xiurui Xie;Ran Tao;Malu Zhang;Huajin Tang","doi":"10.1109/TIP.2024.3430043","DOIUrl":"10.1109/TIP.2024.3430043","url":null,"abstract":"Recent advances in bio-inspired vision with event cameras and associated spiking neural networks (SNNs) have provided promising solutions for low-power consumption neuromorphic tasks. However, as the research of event cameras is still in its infancy, the amount of labeled event stream data is much less than that of the RGB database. The traditional method of converting static images into event streams by simulation to increase the sample size cannot simulate the characteristics of event cameras such as high temporal resolution. To take advantage of both the rich knowledge in labeled RGB images and the features of the event camera, we propose a transfer learning method from the RGB to the event domain in this paper. Specifically, we first introduce a transfer learning framework named R2ETL (RGB to Event Transfer Learning), including a novel encoding alignment module and a feature alignment module. Then, we introduce the temporal centered kernel alignment (TCKA) loss function to improve the efficiency of transfer learning. It aligns the distribution of temporal neuron states by adding a temporal learning constraint. Finally, we theoretically analyze the amount of data required by the deep neuromorphic model to prove the necessity of our method. Numerous experiments demonstrate that our proposed framework outperforms the state-of-the-art SNN and artificial neural network (ANN) models trained on event streams, including N-MNIST, CIFAR10-DVS and N-Caltech101. This indicates that the R2ETL framework is able to leverage the knowledge of labeled RGB images to help the training of SNN on event streams.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141753572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spiking Tucker Fusion Transformer for Audio-Visual Zero-Shot Learning 用于视听零点学习的尖峰塔克融合变压器
Wenrui Li;Penghong Wang;Ruiqin Xiong;Xiaopeng Fan
The spiking neural networks (SNNs) that efficiently encode temporal sequences have shown great potential in extracting audio-visual joint feature representations. However, coupling SNNs (binary spike sequences) with transformers (float-point sequences) to jointly explore the temporal-semantic information still facing challenges. In this paper, we introduce a novel Spiking Tucker Fusion Transformer (STFT) for audio-visual zero-shot learning (ZSL). The STFT leverage the temporal and semantic information from different time steps to generate robust representations. The time-step factor (TSF) is introduced to dynamically synthesis the subsequent inference information. To guide the formation of input membrane potentials and reduce the spike noise, we propose a global-local pooling (GLP) which combines the max and average pooling operations. Furthermore, the thresholds of the spiking neurons are dynamically adjusted based on semantic and temporal cues. Integrating the temporal and semantic information extracted by SNNs and Transformers are difficult due to the increased number of parameters in a straightforward bilinear model. To address this, we introduce a temporal-semantic Tucker fusion module, which achieves multi-scale fusion of SNN and Transformer outputs while maintaining full second-order interactions. Our experimental results demonstrate the effectiveness of the proposed approach in achieving state-of-the-art performance in three benchmark datasets. The harmonic mean (HM) improvement of VGGSound, UCF101 and ActivityNet are around 15.4%, 3.9%, and 14.9%, respectively.
能有效编码时间序列的尖峰神经网络(SNN)在提取视听联合特征表征方面显示出巨大潜力。然而,将尖峰神经网络(二进制尖峰序列)与变换器(浮点序列)耦合以共同探索时间语义信息仍面临挑战。在本文中,我们介绍了一种用于视听零点学习(ZSL)的新型尖峰塔克融合变换器(STFT)。STFT 利用不同时间步长的时间和语义信息生成稳健的表征。引入时间步长因子(TSF)可动态合成后续推理信息。为了引导输入膜电位的形成并降低尖峰噪声,我们提出了全局-局部池化(GLP),它结合了最大池化和平均池化操作。此外,尖峰神经元的阈值会根据语义和时间线索进行动态调整。由于直接双线性模型中的参数数量增加,因此很难整合 SNN 和 Transformers 提取的时间和语义信息。为了解决这个问题,我们引入了时空-语义塔克融合模块,该模块可实现 SNN 和 Transformer 输出的多尺度融合,同时保持完整的二阶交互。我们的实验结果表明,所提出的方法在三个基准数据集上取得了最先进的性能。VGGSound、UCF101 和 ActivityNet 的谐波平均值(HM)分别提高了约 15.4%、3.9% 和 14.9%。
{"title":"Spiking Tucker Fusion Transformer for Audio-Visual Zero-Shot Learning","authors":"Wenrui Li;Penghong Wang;Ruiqin Xiong;Xiaopeng Fan","doi":"10.1109/TIP.2024.3430080","DOIUrl":"10.1109/TIP.2024.3430080","url":null,"abstract":"The spiking neural networks (SNNs) that efficiently encode temporal sequences have shown great potential in extracting audio-visual joint feature representations. However, coupling SNNs (binary spike sequences) with transformers (float-point sequences) to jointly explore the temporal-semantic information still facing challenges. In this paper, we introduce a novel Spiking Tucker Fusion Transformer (STFT) for audio-visual zero-shot learning (ZSL). The STFT leverage the temporal and semantic information from different time steps to generate robust representations. The time-step factor (TSF) is introduced to dynamically synthesis the subsequent inference information. To guide the formation of input membrane potentials and reduce the spike noise, we propose a global-local pooling (GLP) which combines the max and average pooling operations. Furthermore, the thresholds of the spiking neurons are dynamically adjusted based on semantic and temporal cues. Integrating the temporal and semantic information extracted by SNNs and Transformers are difficult due to the increased number of parameters in a straightforward bilinear model. To address this, we introduce a temporal-semantic Tucker fusion module, which achieves multi-scale fusion of SNN and Transformer outputs while maintaining full second-order interactions. Our experimental results demonstrate the effectiveness of the proposed approach in achieving state-of-the-art performance in three benchmark datasets. The harmonic mean (HM) improvement of VGGSound, UCF101 and ActivityNet are around 15.4%, 3.9%, and 14.9%, respectively.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141753573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Multi-Modal Spatial–Temporal Contexts for High-Performance RGB-T Tracking 为高性能 RGB-T 跟踪探索多模态时空语境
Tianlu Zhang;Qiang Jiao;Qiang Zhang;Jungong Han
In RGB-T tracking, there exist rich spatial relationships between the target and backgrounds within multi-modal data as well as sound consistencies of spatial relationships among successive frames, which are crucial for boosting the tracking performance. However, most existing RGB-T trackers overlook such multi-modal spatial relationships and temporal consistencies within RGB-T videos, hindering them from robust tracking and practical applications in complex scenarios. In this paper, we propose a novel Multi-modal Spatial-Temporal Context (MMSTC) network for RGB-T tracking, which employs a Transformer architecture for the construction of reliable multi-modal spatial context information and the effective propagation of temporal context information. Specifically, a Multi-modal Transformer Encoder (MMTE) is designed to achieve the encoding of reliable multi-modal spatial contexts as well as the fusion of multi-modal features. Furthermore, a Quality-aware Transformer Decoder (QATD) is proposed to effectively propagate the tracking cues from historical frames to the current frame, which facilitates the object searching process. Moreover, the proposed MMSTC network can be easily extended to various tracking frameworks. New state-of-the-art results on five prevalent RGB-T tracking benchmarks demonstrate the superiorities of our proposed trackers over existing ones.
在 RGB-T 跟踪中,多模态数据中的目标与背景之间存在丰富的空间关系,连续帧之间的空间关系也具有良好的一致性,这对提高跟踪性能至关重要。然而,大多数现有的 RGB-T 追踪器都忽略了 RGB-T 视频中的这种多模态空间关系和时间一致性,阻碍了它们在复杂场景中的稳健追踪和实际应用。在本文中,我们提出了一种用于 RGB-T 跟踪的新型多模态空间-时间上下文(MMSTC)网络,该网络采用变换器架构来构建可靠的多模态空间上下文信息,并有效传播时间上下文信息。具体来说,设计了一个多模态变换器编码器(MMTE),以实现可靠的多模态空间上下文编码以及多模态特征融合。此外,还提出了质量感知变换器解码器(QATD),以有效地将历史帧的跟踪线索传播到当前帧,从而促进物体搜索过程。此外,所提出的 MMSTC 网络可轻松扩展到各种跟踪框架。在五个流行的 RGB-T 跟踪基准上取得的新的先进结果表明,我们提出的跟踪器优于现有的跟踪器。
{"title":"Exploring Multi-Modal Spatial–Temporal Contexts for High-Performance RGB-T Tracking","authors":"Tianlu Zhang;Qiang Jiao;Qiang Zhang;Jungong Han","doi":"10.1109/TIP.2024.3428316","DOIUrl":"10.1109/TIP.2024.3428316","url":null,"abstract":"In RGB-T tracking, there exist rich spatial relationships between the target and backgrounds within multi-modal data as well as sound consistencies of spatial relationships among successive frames, which are crucial for boosting the tracking performance. However, most existing RGB-T trackers overlook such multi-modal spatial relationships and temporal consistencies within RGB-T videos, hindering them from robust tracking and practical applications in complex scenarios. In this paper, we propose a novel Multi-modal Spatial-Temporal Context (MMSTC) network for RGB-T tracking, which employs a Transformer architecture for the construction of reliable multi-modal spatial context information and the effective propagation of temporal context information. Specifically, a Multi-modal Transformer Encoder (MMTE) is designed to achieve the encoding of reliable multi-modal spatial contexts as well as the fusion of multi-modal features. Furthermore, a Quality-aware Transformer Decoder (QATD) is proposed to effectively propagate the tracking cues from historical frames to the current frame, which facilitates the object searching process. Moreover, the proposed MMSTC network can be easily extended to various tracking frameworks. New state-of-the-art results on five prevalent RGB-T tracking benchmarks demonstrate the superiorities of our proposed trackers over existing ones.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141728471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Remote Sensing Change Detection With Bitemporal and Differential Feature Interactive Perception 利用位时和差异特征交互感知进行遥感变化检测
Hao Chang;Peijin Wang;Wenhui Diao;Guangluan Xu;Xian Sun
Recently, the transformer has achieved notable success in remote sensing (RS) change detection (CD). Its outstanding long-distance modeling ability can effectively recognize the change of interest (CoI). However, in order to obtain the precise pixel-level change regions, many methods directly integrate the stacked transformer blocks into the UNet-style structure, which causes the high computation costs. Besides, the existing methods generally consider bitemporal or differential features separately, which makes the utilization of ground semantic information still insufficient. In this paper, we propose the multiscale dual-space interactive perception network (MDIPNet) to fill these two gaps. On the one hand, we simplify the stacked multi-head transformer blocks into the single-layer single-head attention module and further introduce the lightweight parallel fusion module (LPFM) to perform the efficient information integration. On the other hand, based on the simplified attention mechanism, we propose the cross-space perception module (CSPM) to connect the bitemporal and differential feature spaces, which can help our model suppress the pseudo changes and mine the more abundant semantic consistency of CoI. Extensive experiment results on three challenging datasets and one urban expansion scene indicate that compared with the mainstream CD methods, our MDIPNet obtains the state-of-the-art (SOTA) performance while further controlling the computation costs.
最近,变压器在遥感(RS)变化探测(CD)方面取得了显著的成功。其出色的远距离建模能力可以有效识别感兴趣的变化(CoI)。然而,为了获得精确的像素级变化区域,许多方法直接将堆叠的变换器块集成到 UNet 样式的结构中,导致计算成本较高。此外,现有方法一般都是单独考虑位时或差分特征,对地面语义信息的利用还不够充分。本文提出了多尺度双空间交互感知网络(MDIPNet)来填补这两个空白。一方面,我们将堆叠的多头变换模块简化为单层单头注意模块,并进一步引入轻量级并行融合模块(LPFM),以实现高效的信息整合。另一方面,在简化注意力机制的基础上,我们提出了跨空间感知模块(CSPM)来连接位时特征空间和差分特征空间,这可以帮助我们的模型抑制伪变化,挖掘出更丰富的CoI语义一致性。在三个具有挑战性的数据集和一个城市扩张场景上的大量实验结果表明,与主流的 CD 方法相比,我们的 MDIPNet 在进一步控制计算成本的同时,获得了最先进的性能(SOTA)。
{"title":"Remote Sensing Change Detection With Bitemporal and Differential Feature Interactive Perception","authors":"Hao Chang;Peijin Wang;Wenhui Diao;Guangluan Xu;Xian Sun","doi":"10.1109/TIP.2024.3424335","DOIUrl":"10.1109/TIP.2024.3424335","url":null,"abstract":"Recently, the transformer has achieved notable success in remote sensing (RS) change detection (CD). Its outstanding long-distance modeling ability can effectively recognize the change of interest (CoI). However, in order to obtain the precise pixel-level change regions, many methods directly integrate the stacked transformer blocks into the UNet-style structure, which causes the high computation costs. Besides, the existing methods generally consider bitemporal or differential features separately, which makes the utilization of ground semantic information still insufficient. In this paper, we propose the multiscale dual-space interactive perception network (MDIPNet) to fill these two gaps. On the one hand, we simplify the stacked multi-head transformer blocks into the single-layer single-head attention module and further introduce the lightweight parallel fusion module (LPFM) to perform the efficient information integration. On the other hand, based on the simplified attention mechanism, we propose the cross-space perception module (CSPM) to connect the bitemporal and differential feature spaces, which can help our model suppress the pseudo changes and mine the more abundant semantic consistency of CoI. Extensive experiment results on three challenging datasets and one urban expansion scene indicate that compared with the mainstream CD methods, our MDIPNet obtains the state-of-the-art (SOTA) performance while further controlling the computation costs.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141621991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1