首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
BinsFormer: Revisiting Adaptive Bins for Monocular Depth Estimation BinsFormer:重新审视用于单目深度估算的自适应分区
Zhenyu Li;Xuyang Wang;Xianming Liu;Junjun Jiang
Monocular depth estimation (MDE) is a fundamental task in computer vision and has drawn increasing attention. Recently, some methods reformulate it as a classification-regression task to boost the model performance, where continuous depth is estimated via a linear combination of predicted probability distributions and discrete bins. In this paper, we present a novel framework called BinsFormer, tailored for the classification-regression-based depth estimation. It mainly focuses on two crucial components in the specific task: 1) proper generation of adaptive bins; and 2) sufficient interaction between probability distribution and bins predictions. To specify, we employ a Transformer decoder to generate bins, novelly viewing it as a direct set-to-set prediction problem. We further integrate a multi-scale decoder structure to achieve a comprehensive understanding of spatial geometry information and estimate depth maps in a coarse-to-fine manner. Moreover, an extra scene understanding query is proposed to improve the estimation accuracy, which turns out that models can implicitly learn useful information from the auxiliary environment classification task. Extensive experiments on the KITTI, NYU, and SUN RGB-D datasets demonstrate that BinsFormer surpasses state-of-the-art MDE methods with prominent margins. Code and pretrained models are made publicly available at https://github.com/zhyever/ Monocular-Depth-Estimation-Toolbox/tree/main/configs/ binsformer.
单目深度估计(MDE)是计算机视觉中的一项基本任务,受到越来越多的关注。最近,一些方法将其重新表述为分类-回归任务,以提高模型性能,其中连续深度是通过预测概率分布和离散分层的线性组合来估计的。在本文中,我们提出了一种名为 BinsFormer 的新框架,专门用于基于分类回归的深度估算。它主要关注特定任务中的两个关键部分:1)自适应分层的正确生成;2)概率分布和分层预测之间的充分互动。为了具体说明这一点,我们采用了变换器解码器(Transformer decoder)来生成分集,并将其视为一个直接的集对集预测问题。我们进一步整合了多尺度解码器结构,以实现对空间几何信息的全面理解,并以从粗到细的方式估算深度图。此外,我们还提出了一个额外的场景理解查询来提高估计精度,结果发现模型可以从辅助环境分类任务中隐含地学习有用的信息。在 KITTI、NYU 和 SUN RGB-D 数据集上进行的大量实验表明,BinsFormer 超越了最先进的 MDE 方法,而且优势明显。代码和预训练模型可在 https://github.com/zhyever/Monocular-Depth-Estimation-Toolbox/tree/ main/configs/binsformer 上公开获取。
{"title":"BinsFormer: Revisiting Adaptive Bins for Monocular Depth Estimation","authors":"Zhenyu Li;Xuyang Wang;Xianming Liu;Junjun Jiang","doi":"10.1109/TIP.2024.3416065","DOIUrl":"10.1109/TIP.2024.3416065","url":null,"abstract":"Monocular depth estimation (MDE) is a fundamental task in computer vision and has drawn increasing attention. Recently, some methods reformulate it as a classification-regression task to boost the model performance, where continuous depth is estimated via a linear combination of predicted probability distributions and discrete bins. In this paper, we present a novel framework called BinsFormer, tailored for the classification-regression-based depth estimation. It mainly focuses on two crucial components in the specific task: 1) proper generation of adaptive bins; and 2) sufficient interaction between probability distribution and bins predictions. To specify, we employ a Transformer decoder to generate bins, novelly viewing it as a direct set-to-set prediction problem. We further integrate a multi-scale decoder structure to achieve a comprehensive understanding of spatial geometry information and estimate depth maps in a coarse-to-fine manner. Moreover, an extra scene understanding query is proposed to improve the estimation accuracy, which turns out that models can implicitly learn useful information from the auxiliary environment classification task. Extensive experiments on the KITTI, NYU, and SUN RGB-D datasets demonstrate that BinsFormer surpasses state-of-the-art MDE methods with prominent margins. Code and pretrained models are made publicly available at \u0000<uri>https://github.com/zhyever/</uri>\u0000 Monocular-Depth-Estimation-Toolbox/tree/main/configs/ binsformer.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141447878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Causality-Enhanced Multiple Instance Learning With Graph Convolutional Networks for Parkinsonian Freezing-of-Gait Assessment 利用图卷积网络进行因果关系增强型多实例学习,用于帕金森病步态冻结评估。
Rui Guo;Zheng Xie;Chencheng Zhang;Xiaohua Qian
Freezing of gait (FoG) is a common disabling symptom of Parkinson’s disease (PD). It is clinically characterized by sudden and transient walking interruptions for specific human body parts, and it presents the localization in time and space. Due to the difficulty in extracting global fine-grained features from lengthy videos, developing an automated five-point FoG scoring system is quite challenging. Therefore, we propose a novel video-based automated five-classification FoG assessment method with a causality-enhanced multiple-instance-learning graph convolutional network (GCN). This method involves developing a temporal segmentation GCN to segment each video into three motion stages for stage-level feature modeling, followed by a multiple-instance-learning framework to divide each stage into short clips for instance-level feature extraction. Subsequently, an uncertainty-driven multiple-instance-learning GCN is developed to capture spatial and temporal fine-grained features through GCN scheme and uncertainty learning, respectively, for acquiring global representations. Finally, a causality-enhanced graph generation strategy is proposed to exploit causal inference for mining and enhancing human structures causally related to clinical assessment, thereby extracting spatial causal features. Extensive experimental results demonstrate the excellent performance of the proposed method on five-classification FoG assessment with an accuracy of 62.72% and an acceptable accuracy of 91.32%, which is confirmed by independent testing. Additionally, it enables temporal and spatial localization of FoG events to a certain extent, facilitating reasonable clinical interpretations. In conclusion, our method provides a valuable tool for automated FoG assessment in PD, and the proposed causality-related component exhibits promising potential for extension to other general and medical fine-grained action recognition tasks.
步态冻结(FoG)是帕金森病(PD)常见的致残症状。它的临床特征是人体特定部位突然出现短暂的行走中断,并呈现出时间和空间上的定位。由于难以从冗长的视频中提取全局精细特征,开发五点 FoG 自动评分系统颇具挑战性。因此,我们利用因果增强多实例学习图卷积网络(GCN)提出了一种新颖的基于视频的 FoG 自动五分类评估方法。该方法包括开发一个时间分割 GCN,将每个视频分割成三个运动阶段,进行阶段级特征建模,然后利用多实例学习框架将每个阶段划分为短片段,进行实例级特征提取。随后,开发了不确定性驱动的多实例学习 GCN,通过 GCN 方案和不确定性学习分别捕捉空间和时间细粒度特征,以获取全局表征。最后,提出了一种因果增强图生成策略,利用因果推理挖掘和增强与临床评估有因果关系的人体结构,从而提取空间因果特征。广泛的实验结果表明,所提出的方法在五分类 FoG 评估中表现出色,准确率达到 62.72%,可接受的准确率为 91.32%,这一点得到了独立测试的证实。此外,它还在一定程度上实现了 FoG 事件的时间和空间定位,为合理的临床解释提供了便利。总之,我们的方法为帕金森病患者的 FoG 自动评估提供了一种有价值的工具,而且所提出的因果关系相关组件在扩展到其他一般和医疗细粒度动作识别任务方面具有广阔的前景。
{"title":"Causality-Enhanced Multiple Instance Learning With Graph Convolutional Networks for Parkinsonian Freezing-of-Gait Assessment","authors":"Rui Guo;Zheng Xie;Chencheng Zhang;Xiaohua Qian","doi":"10.1109/TIP.2024.3416052","DOIUrl":"10.1109/TIP.2024.3416052","url":null,"abstract":"Freezing of gait (FoG) is a common disabling symptom of Parkinson’s disease (PD). It is clinically characterized by sudden and transient walking interruptions for specific human body parts, and it presents the localization in time and space. Due to the difficulty in extracting global fine-grained features from lengthy videos, developing an automated five-point FoG scoring system is quite challenging. Therefore, we propose a novel video-based automated five-classification FoG assessment method with a causality-enhanced multiple-instance-learning graph convolutional network (GCN). This method involves developing a temporal segmentation GCN to segment each video into three motion stages for stage-level feature modeling, followed by a multiple-instance-learning framework to divide each stage into short clips for instance-level feature extraction. Subsequently, an uncertainty-driven multiple-instance-learning GCN is developed to capture spatial and temporal fine-grained features through GCN scheme and uncertainty learning, respectively, for acquiring global representations. Finally, a causality-enhanced graph generation strategy is proposed to exploit causal inference for mining and enhancing human structures causally related to clinical assessment, thereby extracting spatial causal features. Extensive experimental results demonstrate the excellent performance of the proposed method on five-classification FoG assessment with an accuracy of 62.72% and an acceptable accuracy of 91.32%, which is confirmed by independent testing. Additionally, it enables temporal and spatial localization of FoG events to a certain extent, facilitating reasonable clinical interpretations. In conclusion, our method provides a valuable tool for automated FoG assessment in PD, and the proposed causality-related component exhibits promising potential for extension to other general and medical fine-grained action recognition tasks.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141447879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PIG: Prompt Images Guidance for Night-Time Scene Parsing PIG:夜间场景解析的提示图像指导。
Zhifeng Xie;Rui Qiu;Sen Wang;Xin Tan;Yuan Xie;Lizhuang Ma
Night-time scene parsing aims to extract pixel-level semantic information in night images, aiding downstream tasks in understanding scene object distribution. Due to limited labeled night image datasets, unsupervised domain adaptation (UDA) has become the predominant method for studying night scenes. UDA typically relies on paired day-night image pairs to guide adaptation, but this approach hampers dataset construction and restricts generalization across night scenes in different datasets. Moreover, UDA, focusing on network architecture and training strategies, faces difficulties in handling classes with few domain similarities. In this paper, we leverage Prompt Images Guidance (PIG) to enhance UDA with supplementary night knowledge. We propose a Night-Focused Network (NFNet) to learn night-specific features from both target domain images and prompt images. To generate high-quality pseudo-labels, we propose Pseudo-label Fusion via Domain Similarity Guidance (FDSG). Classes with fewer domain similarities are predicted by NFNet, which excels in parsing night features, while classes with more domain similarities are predicted by UDA, which has rich labeled semantics. Additionally, we propose two data augmentation strategies: the Prompt Mixture Strategy (PMS) and the Alternate Mask Strategy (AMS), aimed at mitigating the overfitting of the NFNet to a few prompt images. We conduct extensive experiments on four night-time datasets: NightCity, NightCity+, Dark Zurich, and ACDC. The results indicate that utilizing PIG can enhance the parsing accuracy of UDA. The code is available at https://github.com/qiurui4shu/PIG.
夜景解析旨在提取夜景图像中像素级的语义信息,帮助下游任务理解场景物体的分布。由于标注的夜景图像数据集有限,无监督领域适应(UDA)已成为研究夜景的主要方法。UDA 通常依赖成对的昼夜图像来指导自适应,但这种方法阻碍了数据集的构建,并限制了不同数据集中夜景的通用性。此外,UDA 以网络架构和训练策略为重点,在处理领域相似性较低的类别时面临困难。在本文中,我们利用 "提示图像引导"(PIG)来增强 UDA 的夜景知识。我们提出了夜间聚焦网络(NFNet),从目标域图像和提示图像中学习夜间特定特征。为了生成高质量的伪标签,我们提出了通过领域相似性引导进行伪标签融合(FDSG)。领域相似性较低的类别由擅长解析夜间特征的 NFNet 预测,而领域相似性较高的类别则由具有丰富标签语义的 UDA 预测。此外,我们还提出了两种数据增强策略:提示混合策略(PMS)和替代掩码策略(AMS),旨在减轻 NFNet 对少数提示图像的过度拟合。我们在四个夜间数据集上进行了广泛的实验:NightCity、NightCity+、Dark Zurich 和 ACDC。结果表明,利用 PIG 可以提高 UDA 的解析准确性。代码见 https://github.com/qiurui4shu/PIG。
{"title":"PIG: Prompt Images Guidance for Night-Time Scene Parsing","authors":"Zhifeng Xie;Rui Qiu;Sen Wang;Xin Tan;Yuan Xie;Lizhuang Ma","doi":"10.1109/TIP.2024.3415963","DOIUrl":"10.1109/TIP.2024.3415963","url":null,"abstract":"Night-time scene parsing aims to extract pixel-level semantic information in night images, aiding downstream tasks in understanding scene object distribution. Due to limited labeled night image datasets, unsupervised domain adaptation (UDA) has become the predominant method for studying night scenes. UDA typically relies on paired day-night image pairs to guide adaptation, but this approach hampers dataset construction and restricts generalization across night scenes in different datasets. Moreover, UDA, focusing on network architecture and training strategies, faces difficulties in handling classes with few domain similarities. In this paper, we leverage Prompt Images Guidance (PIG) to enhance UDA with supplementary night knowledge. We propose a Night-Focused Network (NFNet) to learn night-specific features from both target domain images and prompt images. To generate high-quality pseudo-labels, we propose Pseudo-label Fusion via Domain Similarity Guidance (FDSG). Classes with fewer domain similarities are predicted by NFNet, which excels in parsing night features, while classes with more domain similarities are predicted by UDA, which has rich labeled semantics. Additionally, we propose two data augmentation strategies: the Prompt Mixture Strategy (PMS) and the Alternate Mask Strategy (AMS), aimed at mitigating the overfitting of the NFNet to a few prompt images. We conduct extensive experiments on four night-time datasets: NightCity, NightCity+, Dark Zurich, and ACDC. The results indicate that utilizing PIG can enhance the parsing accuracy of UDA. The code is available at \u0000<uri>https://github.com/qiurui4shu/PIG</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141447881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalization Beyond Feature Alignment: Concept Activation-Guided Contrastive Learning 超越特征对齐的泛化:概念激活引导的对比学习
Yibing Liu;Chris Xing Tian;Haoliang Li;Shiqi Wang
Learning invariant representations via contrastive learning has seen state-of-the-art performance in domain generalization (DG). Despite such success, in this paper, we find that its core learning strategy – feature alignment – could heavily hinder model generalization. Drawing insights in neuron interpretability, we characterize this problem from a neuron activation view. Specifically, by treating feature elements as neuron activation states, we show that conventional alignment methods tend to deteriorate the diversity of learned invariant features, as they indiscriminately minimize all neuron activation differences. This instead ignores rich relations among neurons – many of them often identify the same visual concepts despite differing activation patterns. With this finding, we present a simple yet effective approach, Concept Contrast (CoCo), which relaxes element-wise feature alignments by contrasting high-level concepts encoded in neurons. Our CoCo performs in a plug-and-play fashion, thus it can be integrated into any contrastive method in DG. We evaluate CoCo over four canonical contrastive methods, showing that CoCo promotes the diversity of feature representations and consistently improves model generalization capability. By decoupling this success through neuron coverage analysis, we further find that CoCo potentially invokes more meaningful neurons during training, thereby improving model learning.
通过对比学习来学习不变表征,在领域泛化(DG)方面取得了最先进的性能。尽管取得了这样的成功,但在本文中,我们发现其核心学习策略--特征对齐--可能会严重阻碍模型泛化。借鉴神经元可解释性的观点,我们从神经元激活的角度来描述这一问题。具体来说,通过将特征元素视为神经元激活状态,我们发现传统的配准方法往往会恶化所学不变特征的多样性,因为它们会不加区分地最小化所有神经元激活差异。这反而忽略了神经元之间的丰富关系--尽管激活模式不同,但许多神经元往往能识别相同的视觉概念。有鉴于此,我们提出了一种简单而有效的方法--概念对比法(Concept Contrast,CoCo),它通过对比神经元编码的高级概念来放松元素特征排列。我们的 CoCo 以即插即用的方式运行,因此可以集成到 DG 中的任何对比方法中。我们对四种典型对比方法的 CoCo 进行了评估,结果表明,CoCo 促进了特征表征的多样性,并持续提高了模型的泛化能力。通过神经元覆盖率分析对这一成功进行解耦,我们进一步发现 CoCo 有可能在训练过程中调用更多有意义的神经元,从而改善模型学习。
{"title":"Generalization Beyond Feature Alignment: Concept Activation-Guided Contrastive Learning","authors":"Yibing Liu;Chris Xing Tian;Haoliang Li;Shiqi Wang","doi":"10.1109/TIP.2024.3416873","DOIUrl":"10.1109/TIP.2024.3416873","url":null,"abstract":"Learning invariant representations via contrastive learning has seen state-of-the-art performance in domain generalization (DG). Despite such success, in this paper, we find that its core learning strategy – feature alignment – could heavily hinder model generalization. Drawing insights in neuron interpretability, we characterize this problem from a neuron activation view. Specifically, by treating feature elements as neuron activation states, we show that conventional alignment methods tend to deteriorate the diversity of learned invariant features, as they indiscriminately minimize all neuron activation differences. This instead ignores rich relations among neurons – many of them often identify the same visual concepts despite differing activation patterns. With this finding, we present a simple yet effective approach, Concept Contrast (CoCo), which relaxes element-wise feature alignments by contrasting high-level concepts encoded in neurons. Our CoCo performs in a plug-and-play fashion, thus it can be integrated into any contrastive method in DG. We evaluate CoCo over four canonical contrastive methods, showing that CoCo promotes the diversity of feature representations and consistently improves model generalization capability. By decoupling this success through neuron coverage analysis, we further find that CoCo potentially invokes more meaningful neurons during training, thereby improving model learning.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141447880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incrementally Adapting Pretrained Model Using Network Prior for Multi-Focus Image Fusion 利用网络先验增量调整预训练模型,实现多焦点图像融合
Xingyu Hu;Junjun Jiang;Chenyang Wang;Xianming Liu;Jiayi Ma
Multi-focus image fusion can fuse the clear parts of two or more source images captured at the same scene with different focal lengths into an all-in-focus image. On the one hand, previous supervised learning-based multi-focus image fusion methods relying on synthetic datasets have a clear distribution shift with real scenarios. On the other hand, unsupervised learning-based multi-focus image fusion methods can well adapt to the observed images but lack the general knowledge of defocus blur that can be learned from paired data. To avoid the problems of existing methods, this paper presents a novel multi-focus image fusion model by considering both the general knowledge brought by the supervised pretrained backbone and the extrinsic priors optimized on specific testing sample to improve the performance of image fusion. To be specific, the Incremental Network Prior Adaptation (INPA) framework is proposed to incrementally integrate features extracted from the pretrained strong baselines into a tiny prior network (6.9% parameters of the backbone network) to boost the performance for test samples. We evaluate our method on both synthetic and real-world public datasets (Lytro, MFI-WHU, and Real-MFF) and show that our method outperforms existing supervised learning-based methods and unsupervised learning based methods.
多焦图像融合可以将同一场景中两幅或多幅不同焦距的源图像的清晰部分融合成一幅全焦图像。一方面,以往基于监督学习的多焦点图像融合方法依赖于合成数据集,与真实场景的分布存在明显偏移。另一方面,基于无监督学习的多焦点图像融合方法能很好地适应观察到的图像,但缺乏从配对数据中学习到的关于离焦模糊的一般知识。为了避免现有方法存在的问题,本文提出了一种新的多焦点图像融合模型,既考虑了监督预训练骨干带来的一般知识,也考虑了在特定测试样本上优化的外在前验,以提高图像融合的性能。具体来说,我们提出了增量网络先验适应(INPA)框架,将从预训练的强基线中提取的特征逐步整合到一个微小的先验网络(占骨干网络参数的 6.9%)中,以提高测试样本的性能。我们在合成数据集和真实世界公共数据集(Lytro、MFI-WHU 和 Real-MFF)上对我们的方法进行了评估,结果表明我们的方法优于现有的基于监督学习的方法和基于无监督学习的方法。
{"title":"Incrementally Adapting Pretrained Model Using Network Prior for Multi-Focus Image Fusion","authors":"Xingyu Hu;Junjun Jiang;Chenyang Wang;Xianming Liu;Jiayi Ma","doi":"10.1109/TIP.2024.3409940","DOIUrl":"10.1109/TIP.2024.3409940","url":null,"abstract":"Multi-focus image fusion can fuse the clear parts of two or more source images captured at the same scene with different focal lengths into an all-in-focus image. On the one hand, previous supervised learning-based multi-focus image fusion methods relying on synthetic datasets have a clear distribution shift with real scenarios. On the other hand, unsupervised learning-based multi-focus image fusion methods can well adapt to the observed images but lack the general knowledge of defocus blur that can be learned from paired data. To avoid the problems of existing methods, this paper presents a novel multi-focus image fusion model by considering both the general knowledge brought by the supervised pretrained backbone and the extrinsic priors optimized on specific testing sample to improve the performance of image fusion. To be specific, the Incremental Network Prior Adaptation (INPA) framework is proposed to incrementally integrate features extracted from the pretrained strong baselines into a tiny prior network (6.9% parameters of the backbone network) to boost the performance for test samples. We evaluate our method on both synthetic and real-world public datasets (Lytro, MFI-WHU, and Real-MFF) and show that our method outperforms existing supervised learning-based methods and unsupervised learning based methods.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141437949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Condition Latent Diffusion Network for Scene-Aware Neural Human Motion Prediction 用于场景感知神经人体运动预测的多条件潜在扩散网络
Xuehao Gao;Yang Yang;Yang Wu;Shaoyi Du;Guo-Jun Qi
Inferring 3D human motion is fundamental in many applications, including understanding human activity and analyzing one’s intention. While many fruitful efforts have been made to human motion prediction, most approaches focus on pose-driven prediction and inferring human motion in isolation from the contextual environment, thus leaving the body location movement in the scene behind. However, real-world human movements are goal-directed and highly influenced by the spatial layout of their surrounding scenes. In this paper, instead of planning future human motion in a “dark” room, we propose a Multi-Condition Latent Diffusion network (MCLD) that reformulates the human motion prediction task as a multi-condition joint inference problem based on the given historical 3D body motion and the current 3D scene contexts. Specifically, instead of directly modeling joint distribution over the raw motion sequences, MCLD performs a conditional diffusion process within the latent embedding space, characterizing the cross-modal mapping from the past body movement and current scene context condition embeddings to the future human motion embedding. Extensive experiments on large-scale human motion prediction datasets demonstrate that our MCLD achieves significant improvements over the state-of-the-art methods on both realistic and diverse predictions.
推断三维人体运动是许多应用的基础,包括理解人体活动和分析人的意图。虽然人们在人类运动预测方面做出了许多卓有成效的努力,但大多数方法都侧重于姿势驱动预测和脱离上下文环境推断人类运动,从而将场景中的身体位置运动抛在脑后。然而,现实世界中的人体运动是以目标为导向的,受周围场景空间布局的影响很大。在本文中,我们提出了一种多条件潜在扩散网络(MCLD),而不是在一个 "黑暗 "的房间中规划未来的人体运动,它将人体运动预测任务重新表述为一个基于给定的历史三维人体运动和当前三维场景上下文的多条件联合推理问题。具体来说,MCLD 并不直接对原始运动序列的联合分布建模,而是在潜在嵌入空间内执行条件扩散过程,描述从过去的身体运动和当前场景条件嵌入到未来人体运动嵌入的跨模态映射。在大规模人体运动预测数据集上进行的广泛实验表明,我们的 MCLD 在现实和多样化预测方面都比最先进的方法有显著改进。
{"title":"Multi-Condition Latent Diffusion Network for Scene-Aware Neural Human Motion Prediction","authors":"Xuehao Gao;Yang Yang;Yang Wu;Shaoyi Du;Guo-Jun Qi","doi":"10.1109/TIP.2024.3414935","DOIUrl":"10.1109/TIP.2024.3414935","url":null,"abstract":"Inferring 3D human motion is fundamental in many applications, including understanding human activity and analyzing one’s intention. While many fruitful efforts have been made to human motion prediction, most approaches focus on pose-driven prediction and inferring human motion in isolation from the contextual environment, thus leaving the body location movement in the scene behind. However, real-world human movements are goal-directed and highly influenced by the spatial layout of their surrounding scenes. In this paper, instead of planning future human motion in a “dark” room, we propose a Multi-Condition Latent Diffusion network (MCLD) that reformulates the human motion prediction task as a multi-condition joint inference problem based on the given historical 3D body motion and the current 3D scene contexts. Specifically, instead of directly modeling joint distribution over the raw motion sequences, MCLD performs a conditional diffusion process within the latent embedding space, characterizing the cross-modal mapping from the past body movement and current scene context condition embeddings to the future human motion embedding. Extensive experiments on large-scale human motion prediction datasets demonstrate that our MCLD achieves significant improvements over the state-of-the-art methods on both realistic and diverse predictions.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141433611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-Supervised Learning With Heterogeneous Distribution Consistency for Visible Infrared Person Re-Identification 利用异质分布一致性进行可见红外人员再识别的半监督学习。
Ziyu Wei;Xi Yang;Nannan Wang;Xinbo Gao
Visible infrared person re-identification (VI-ReID) exposes considerable challenges because of the modality gaps between the person images captured by daytime visible cameras and nighttime infrared cameras. Several fully-supervised VI-ReID methods have improved the performance with extensive labeled heterogeneous images. However, the identity of the person is difficult to obtain in real-world situations, especially at night. Limited known identities and large modality discrepancies impede the effectiveness of the model to a great extent. In this paper, we propose a novel Semi-Supervised Learning framework with Heterogeneous Distribution Consistency (HDC-SSL) for VI-ReID. Specifically, through investigating the confidence distribution of heterogeneous images, we introduce a Gaussian Mixture Model-based Pseudo Labeling (GMM-PL) method, which adaptively adjusts different thresholds for each modality to label the identity. Moreover, to facilitate the representation learning of unutilized data whose prediction is lower than the threshold, Modality Consistency Regularization (MCR) is proposed to ensure the prediction consistency of the cross-modality pedestrian images and handle the modality variance. Extensive experiments with different label settings on two VI-ReID datasets demonstrate the effectiveness of our method. Particularly, HDC-SSL achieves competitive performance with state-of-the-art fully-supervised VI-ReID methods on RegDB dataset with only 1 visible label and 1 infrared label per class.
由于白天可见光摄像机和夜间红外摄像机拍摄的人物图像之间存在模态差距,可见光红外人物再识别(VI-ReID)面临着巨大挑战。有几种完全监督的 VI-ReID 方法通过大量标注的异构图像提高了性能。然而,在现实世界中,尤其是在夜间,很难获得人物的身份。有限的已知身份和巨大的模态差异在很大程度上阻碍了模型的有效性。在本文中,我们为 VI-ReID 提出了一种新颖的具有异构分布一致性(HDC-SSL)的半监督学习框架。具体来说,通过研究异构图像的置信度分布,我们引入了一种基于高斯混杂模型的伪标签(GMM-PL)方法,该方法可自适应地调整每种模态的不同阈值来标记身份。此外,为了便于对预测值低于阈值的未利用数据进行表示学习,我们提出了模态一致性正则化(MCR),以确保跨模态行人图像的预测一致性,并处理模态方差。在两个 VI-ReID 数据集上使用不同标签设置进行的大量实验证明了我们方法的有效性。特别是在 RegDB 数据集上,HDC-SSL 在每类只有 1 个可见光标签和 1 个红外标签的情况下,取得了与最先进的全监督 VI-ReID 方法相当的性能。
{"title":"Semi-Supervised Learning With Heterogeneous Distribution Consistency for Visible Infrared Person Re-Identification","authors":"Ziyu Wei;Xi Yang;Nannan Wang;Xinbo Gao","doi":"10.1109/TIP.2024.3414938","DOIUrl":"10.1109/TIP.2024.3414938","url":null,"abstract":"Visible infrared person re-identification (VI-ReID) exposes considerable challenges because of the modality gaps between the person images captured by daytime visible cameras and nighttime infrared cameras. Several fully-supervised VI-ReID methods have improved the performance with extensive labeled heterogeneous images. However, the identity of the person is difficult to obtain in real-world situations, especially at night. Limited known identities and large modality discrepancies impede the effectiveness of the model to a great extent. In this paper, we propose a novel Semi-Supervised Learning framework with Heterogeneous Distribution Consistency (HDC-SSL) for VI-ReID. Specifically, through investigating the confidence distribution of heterogeneous images, we introduce a Gaussian Mixture Model-based Pseudo Labeling (GMM-PL) method, which adaptively adjusts different thresholds for each modality to label the identity. Moreover, to facilitate the representation learning of unutilized data whose prediction is lower than the threshold, Modality Consistency Regularization (MCR) is proposed to ensure the prediction consistency of the cross-modality pedestrian images and handle the modality variance. Extensive experiments with different label settings on two VI-ReID datasets demonstrate the effectiveness of our method. Particularly, HDC-SSL achieves competitive performance with state-of-the-art fully-supervised VI-ReID methods on RegDB dataset with only 1 visible label and 1 infrared label per class.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141433612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
L₀ Gradient-Regularization and Scale Space Representation Model for Cartoon and Texture Decomposition 用于卡通和纹理分解的 L0 梯度规整化和尺度空间表示模型
Huan Pan;You-Wei Wen;Ya Huang
In this paper, we consider decomposing an image into its cartoon and texture components. Traditional methods, which mainly rely on the gradient amplitude of images to distinguish between these components, often show limitations in decomposing small-scale, high-contrast texture patterns and large-scale, low-contrast structural components. Specifically, these methods tend to decompose the former to the cartoon image and the latter to the texture image, neglecting the scale features inherent in both components. To overcome these challenges, we introduce a new variational model which incorporates an $L_{0}$ -based total variation norm for the cartoon component and an $L_{2}$ norm for the scale space representation of the texture component. We show that the texture component has a small $L_{2}$ norm in the scale space representation. We apply a quadratic penalty function to handle the non-separable $L_{0}$ norm minimization problem. Numerical experiments are given to illustrate the efficiency and effectiveness of our approach.
在本文中,我们考虑将图像分解为卡通成分和纹理成分。传统方法主要依靠图像的梯度振幅来区分卡通和纹理,但在分解小尺度、高对比度的纹理图案和大尺度、低对比度的结构成分时往往会表现出局限性。具体来说,这些方法倾向于将前者分解为卡通图像,而将后者分解为纹理图像,从而忽略了这两种成分固有的尺度特征。为了克服这些挑战,我们引入了一种新的变分模型,该模型将基于 L0 的总变分规范用于卡通分量,将 L2 规范用于纹理分量的尺度空间表示。我们证明,纹理分量在比例空间表示中的 L2 规范较小。我们采用二次惩罚函数来处理不可分割的 L0 准则最小化问题。我们给出了数值实验来说明我们方法的效率和有效性。
{"title":"L₀ Gradient-Regularization and Scale Space Representation Model for Cartoon and Texture Decomposition","authors":"Huan Pan;You-Wei Wen;Ya Huang","doi":"10.1109/TIP.2024.3403505","DOIUrl":"10.1109/TIP.2024.3403505","url":null,"abstract":"In this paper, we consider decomposing an image into its cartoon and texture components. Traditional methods, which mainly rely on the gradient amplitude of images to distinguish between these components, often show limitations in decomposing small-scale, high-contrast texture patterns and large-scale, low-contrast structural components. Specifically, these methods tend to decompose the former to the cartoon image and the latter to the texture image, neglecting the scale features inherent in both components. To overcome these challenges, we introduce a new variational model which incorporates an \u0000<inline-formula> <tex-math>$L_{0}$ </tex-math></inline-formula>\u0000-based total variation norm for the cartoon component and an \u0000<inline-formula> <tex-math>$L_{2}$ </tex-math></inline-formula>\u0000 norm for the scale space representation of the texture component. We show that the texture component has a small \u0000<inline-formula> <tex-math>$L_{2}$ </tex-math></inline-formula>\u0000 norm in the scale space representation. We apply a quadratic penalty function to handle the non-separable \u0000<inline-formula> <tex-math>$L_{0}$ </tex-math></inline-formula>\u0000 norm minimization problem. Numerical experiments are given to illustrate the efficiency and effectiveness of our approach.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141433610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FF-LPD: A Real-Time Frame-by-Frame License Plate Detector With Knowledge Distillation and Feature Propagation FF-LPD:利用知识提炼和特征传播的实时逐帧车牌检测器。
Haoxuan Ding;Junyu Gao;Yuan Yuan;Qi Wang
With the increasing availability of cameras in vehicles, obtaining license plate (LP) information via on-board cameras has become feasible in traffic scenarios. LPs play a pivotal role in vehicle identification, making automatic LP detection (ALPD) a crucial area within traffic analysis. Recent advancements in deep learning have spurred a surge of studies in ALPD. However, the computational limitations of on-board devices hinder the performance of real-time ALPD systems for moving vehicles. Therefore, we propose a real-time frame-by-frame LP detector focusing on real-time accurate LP detection. Specifically, video frames are categorized into keyframes and non-keyframes. Keyframes are processed by a deeper network (high-level stream), while non-keyframes are handled by a lightweight network (low-level stream), significantly enhancing efficiency. To achieve accurate detection, we design a knowledge distillation strategy to boost the performance of low-level stream and a feature propagation method to introduce the temporal clues in video LP detection. Our contributions are: (1) A real-time frame-by-frame LP detector for video LP detection is proposed, achieving a competitive performance with popular one-stage LP detectors. (2) A simple feature-based knowledge distillation strategy is introduced to improve the low-level stream performance. (3) A spatial-temporal attention feature propagation method is designed to refine the features from non-keyframes guided by the memory features from keyframes, leveraging the inherent temporal correlation in videos. The ablation studies show the effectiveness of knowledge distillation strategy and feature propagation method.
随着车载摄像头的普及,在交通场景中通过车载摄像头获取车牌(LP)信息变得可行。LP 在车辆识别中起着举足轻重的作用,因此自动 LP 检测(ALPD)成为交通分析中的一个重要领域。深度学习的最新进展推动了 ALPD 的研究热潮。然而,车载设备的计算局限性阻碍了移动车辆实时 ALPD 系统的性能。因此,我们提出了一种实时逐帧 LP 检测器,专注于实时准确的 LP 检测。具体来说,视频帧分为关键帧和非关键帧。关键帧由较深的网络(高层流)处理,而非关键帧则由轻量级网络(低层流)处理,从而大大提高了效率。为了实现精确检测,我们设计了一种知识提炼策略来提高低级流的性能,并设计了一种特征传播方法来在视频 LP 检测中引入时间线索。我们的贡献在于(1) 提出了一种用于视频 LP 检测的实时逐帧 LP 检测器,其性能可与流行的单级 LP 检测器相媲美。(2) 引入了一种简单的基于特征的知识提炼策略,以提高底层流的性能。(3) 设计了一种空间-时间注意力特征传播方法,利用视频固有的时间相关性,在关键帧记忆特征的引导下提炼非关键帧的特征。消融研究表明了知识提炼策略和特征传播方法的有效性。
{"title":"FF-LPD: A Real-Time Frame-by-Frame License Plate Detector With Knowledge Distillation and Feature Propagation","authors":"Haoxuan Ding;Junyu Gao;Yuan Yuan;Qi Wang","doi":"10.1109/TIP.2024.3414269","DOIUrl":"10.1109/TIP.2024.3414269","url":null,"abstract":"With the increasing availability of cameras in vehicles, obtaining license plate (LP) information via on-board cameras has become feasible in traffic scenarios. LPs play a pivotal role in vehicle identification, making automatic LP detection (ALPD) a crucial area within traffic analysis. Recent advancements in deep learning have spurred a surge of studies in ALPD. However, the computational limitations of on-board devices hinder the performance of real-time ALPD systems for moving vehicles. Therefore, we propose a real-time frame-by-frame LP detector focusing on real-time accurate LP detection. Specifically, video frames are categorized into keyframes and non-keyframes. Keyframes are processed by a deeper network (high-level stream), while non-keyframes are handled by a lightweight network (low-level stream), significantly enhancing efficiency. To achieve accurate detection, we design a knowledge distillation strategy to boost the performance of low-level stream and a feature propagation method to introduce the temporal clues in video LP detection. Our contributions are: (1) A real-time frame-by-frame LP detector for video LP detection is proposed, achieving a competitive performance with popular one-stage LP detectors. (2) A simple feature-based knowledge distillation strategy is introduced to improve the low-level stream performance. (3) A spatial-temporal attention feature propagation method is designed to refine the features from non-keyframes guided by the memory features from keyframes, leveraging the inherent temporal correlation in videos. The ablation studies show the effectiveness of knowledge distillation strategy and feature propagation method.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141428585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SemiRS-COC: Semi-Supervised Classification for Complex Remote Sensing Scenes With Cross-Object Consistency SemiRS-COC:具有跨对象一致性的复杂遥感场景半监督分类。
Qiang Liu;Jun Yue;Yang Kuang;Weiying Xie;Leyuan Fang
Semi-supervised learning (SSL), which aims to learn with limited labeled data and massive amounts of unlabeled data, offers a promising approach to exploit the massive amounts of satellite Earth observation images. The fundamental concept underlying most state-of-the-art SSL methods involves generating pseudo-labels for unlabeled data based on image-level predictions. However, complex remote sensing (RS) scene images frequently encounter challenges, such as interference from multiple background objects and significant intra-class differences, resulting in unreliable pseudo-labels. In this paper, we propose the SemiRS-COC, a novel semi-supervised classification method for complex RS scenes. Inspired by the idea that neighboring objects in feature space should share consistent semantic labels, SemiRS-COC utilizes the similarity between foreground objects in RS images to generate reliable object-level pseudo-labels, effectively addressing the issues of multiple background objects and significant intra-class differences in complex RS images. Specifically, we first design a Local Self-Learning Object Perception (LSLOP) mechanism, which transforms multiple background objects interference of RS images into usable annotation information, enhancing the model’s object perception capability. Furthermore, we present a Cross-Object Consistency Pseudo-Labeling (COCPL) strategy, which generates reliable object-level pseudo-labels by comparing the similarity of foreground objects across different RS images, effectively handling significant intra-class differences. Extensive experiments demonstrate that our proposed method achieves excellent performance compared to state-of-the-art methods on three widely-adopted RS datasets.
半监督学习(SSL)旨在利用有限的标记数据和海量的非标记数据进行学习,为利用海量卫星地球观测图像提供了一种前景广阔的方法。大多数最先进的半监督学习方法的基本概念是根据图像级预测为未标记数据生成伪标签。然而,复杂的遥感(RS)场景图像经常会遇到各种挑战,例如来自多个背景物体的干扰和显著的类内差异,从而导致伪标签不可靠。在本文中,我们提出了针对复杂遥感场景的新型半监督分类方法 SemiRS-COC。SemiRS-COC 受特征空间中相邻对象应共享一致语义标签的思想启发,利用 RS 图像中前景对象之间的相似性生成可靠的对象级伪标签,有效解决了复杂 RS 图像中的多背景对象和显著类内差异问题。具体来说,我们首先设计了一种局部自学习物体感知(LSLOP)机制,将 RS 图像中的多重背景物体干扰转化为可用的标注信息,增强了模型的物体感知能力。此外,我们还提出了一种跨对象一致性伪标注(COCPL)策略,通过比较不同 RS 图像中前景对象的相似性来生成可靠的对象级伪标注,从而有效处理显著的类内差异。广泛的实验证明,在三个广泛采用的 RS 数据集上,与最先进的方法相比,我们提出的方法取得了卓越的性能。
{"title":"SemiRS-COC: Semi-Supervised Classification for Complex Remote Sensing Scenes With Cross-Object Consistency","authors":"Qiang Liu;Jun Yue;Yang Kuang;Weiying Xie;Leyuan Fang","doi":"10.1109/TIP.2024.3414122","DOIUrl":"10.1109/TIP.2024.3414122","url":null,"abstract":"Semi-supervised learning (SSL), which aims to learn with limited labeled data and massive amounts of unlabeled data, offers a promising approach to exploit the massive amounts of satellite Earth observation images. The fundamental concept underlying most state-of-the-art SSL methods involves generating pseudo-labels for unlabeled data based on image-level predictions. However, complex remote sensing (RS) scene images frequently encounter challenges, such as interference from multiple background objects and significant intra-class differences, resulting in unreliable pseudo-labels. In this paper, we propose the SemiRS-COC, a novel semi-supervised classification method for complex RS scenes. Inspired by the idea that neighboring objects in feature space should share consistent semantic labels, SemiRS-COC utilizes the similarity between foreground objects in RS images to generate reliable object-level pseudo-labels, effectively addressing the issues of multiple background objects and significant intra-class differences in complex RS images. Specifically, we first design a Local Self-Learning Object Perception (LSLOP) mechanism, which transforms multiple background objects interference of RS images into usable annotation information, enhancing the model’s object perception capability. Furthermore, we present a Cross-Object Consistency Pseudo-Labeling (COCPL) strategy, which generates reliable object-level pseudo-labels by comparing the similarity of foreground objects across different RS images, effectively handling significant intra-class differences. Extensive experiments demonstrate that our proposed method achieves excellent performance compared to state-of-the-art methods on three widely-adopted RS datasets.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141428586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1