首页 > 最新文献

IEEE transactions on image processing : a publication of the IEEE Signal Processing Society最新文献

英文 中文
Semantic-aware Message Broadcasting for Efficient Unsupervised Domain Adaptation. 面向高效无监督领域适应的语义感知信息广播。
Xin Li, Cuiling Lan, Guoqiang Wei, Zhibo Chen

Vision transformer has demonstrated great potential in abundant vision tasks. However, it also inevitably suffers from poor generalization capability when the distribution shift occurs in testing (i.e., out-of-distribution data). To mitigate this issue, we propose a novel method, Semantic-aware Message Broadcasting (SAMB), which enables more informative and flexible feature alignment for unsupervised domain adaptation (UDA). Particularly, we study the attention module in the vision transformer and notice that the alignment space using one global class token lacks enough flexibility, where it interacts information with all image tokens in the same manner but ignores the rich semantics of different regions. In this paper, we aim to improve the richness of the alignment features by enabling semantic-aware adaptive message broadcasting. Particularly, we introduce a group of learned group tokens as nodes to aggregate the global information from all image tokens, but encourage different group tokens to adaptively focus on the message broadcasting to different semantic regions. In this way, our message broadcasting encourages the group tokens to learn more informative and diverse information for effective domain alignment. Moreover, we systematically study the effects of adversarial-based feature alignment (ADA) and pseudo-label based self-training (PST) on UDA. We find that one simple two-stage training strategy with the cooperation of ADA and PST can further improve the adaptation capability of the vision transformer. Extensive experiments on DomainNet, OfficeHome, and VisDA-2017 demonstrate the effectiveness of our methods for UDA.

视觉变换器在丰富的视觉任务中展现了巨大的潜力。然而,当测试中出现分布偏移(即分布外数据)时,它也不可避免地存在泛化能力差的问题。为了缓解这一问题,我们提出了一种新方法--语义感知信息广播(SAMB),它能为无监督领域适应(UDA)提供更多信息和更灵活的特征配准。我们特别研究了视觉转换器中的注意力模块,发现使用一个全局类标记的配准空间缺乏足够的灵活性,它以相同的方式与所有图像标记进行信息交互,却忽略了不同区域的丰富语义。在本文中,我们旨在通过实现语义感知的自适应信息广播来提高配准特征的丰富性。特别是,我们引入了一组学习到的组标记作为节点,汇总来自所有图像标记的全局信息,但鼓励不同的组标记自适应地将信息广播重点放在不同的语义区域。通过这种方式,我们的信息广播鼓励组标记学习更多不同的信息,从而实现有效的领域对齐。此外,我们还系统地研究了基于对抗的特征对齐(ADA)和基于伪标签的自我训练(PST)对 UDA 的影响。我们发现,一个简单的两阶段训练策略(ADA 和 PST)可以进一步提高视觉转换器的适应能力。在 DomainNet、OfficeHome 和 VisDA-2017 上进行的大量实验证明了我们的方法对 UDA 的有效性。
{"title":"Semantic-aware Message Broadcasting for Efficient Unsupervised Domain Adaptation.","authors":"Xin Li, Cuiling Lan, Guoqiang Wei, Zhibo Chen","doi":"10.1109/TIP.2024.3437212","DOIUrl":"https://doi.org/10.1109/TIP.2024.3437212","url":null,"abstract":"<p><p>Vision transformer has demonstrated great potential in abundant vision tasks. However, it also inevitably suffers from poor generalization capability when the distribution shift occurs in testing (i.e., out-of-distribution data). To mitigate this issue, we propose a novel method, Semantic-aware Message Broadcasting (SAMB), which enables more informative and flexible feature alignment for unsupervised domain adaptation (UDA). Particularly, we study the attention module in the vision transformer and notice that the alignment space using one global class token lacks enough flexibility, where it interacts information with all image tokens in the same manner but ignores the rich semantics of different regions. In this paper, we aim to improve the richness of the alignment features by enabling semantic-aware adaptive message broadcasting. Particularly, we introduce a group of learned group tokens as nodes to aggregate the global information from all image tokens, but encourage different group tokens to adaptively focus on the message broadcasting to different semantic regions. In this way, our message broadcasting encourages the group tokens to learn more informative and diverse information for effective domain alignment. Moreover, we systematically study the effects of adversarial-based feature alignment (ADA) and pseudo-label based self-training (PST) on UDA. We find that one simple two-stage training strategy with the cooperation of ADA and PST can further improve the adaptation capability of the vision transformer. Extensive experiments on DomainNet, OfficeHome, and VisDA-2017 demonstrate the effectiveness of our methods for UDA.</p>","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141908632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Common Semantics via Optimal Transport for Contrastive Multi-View Clustering 通过最佳传输学习共同语义,实现多视角对比聚类
Qian Zhang;Lin Zhang;Ran Song;Runmin Cong;Yonghuai Liu;Wei Zhang
Multi-view clustering aims to learn discriminative representations from multi-view data. Although existing methods show impressive performance by leveraging contrastive learning to tackle the representation gap between every two views, they share the common limitation of not performing semantic alignment from a global perspective, resulting in the undermining of semantic patterns in multi-view data. This paper presents CSOT, namely Common Semantics via Optimal Transport, to boost contrastive multi-view clustering via semantic learning in a common space that integrates all views. Through optimal transport, the samples in multiple views are mapped to the joint clusters which represent the multi-view semantic patterns in the common space. With the semantic assignment derived from the optimal transport plan, we design a semantic learning module where the soft assignment vector works as a global supervision to enforce the model to learn consistent semantics among all views. Moreover, we propose a semantic-aware re-weighting strategy to treat samples differently according to their semantic significance, which improves the effectiveness of cross-view contrastive representation learning. Extensive experimental results demonstrate that CSOT achieves the state-of-the-art clustering performance.
多视图聚类旨在从多视图数据中学习判别表征。虽然现有的方法通过利用对比学习来解决每两个视图之间的表征差距,从而显示出令人印象深刻的性能,但它们都有一个共同的局限性,即没有从全局角度进行语义对齐,导致多视图数据中的语义模式受到破坏。本文提出的 CSOT(即通过最优传输的通用语义)通过在整合所有视图的通用空间中进行语义学习来促进对比性多视图聚类。通过最优传输,多个视图中的样本被映射到共同空间中代表多视图语义模式的联合聚类中。利用从最优传输计划中得出的语义分配,我们设计了一个语义学习模块,其中软分配向量作为全局监督,强制模型在所有视图中学习一致的语义。此外,我们还提出了一种语义感知再加权策略,根据样本的语义意义对其进行不同处理,从而提高了跨视图对比表征学习的效果。广泛的实验结果表明,CSOT 实现了最先进的聚类性能。
{"title":"Learning Common Semantics via Optimal Transport for Contrastive Multi-View Clustering","authors":"Qian Zhang;Lin Zhang;Ran Song;Runmin Cong;Yonghuai Liu;Wei Zhang","doi":"10.1109/TIP.2024.3436615","DOIUrl":"10.1109/TIP.2024.3436615","url":null,"abstract":"Multi-view clustering aims to learn discriminative representations from multi-view data. Although existing methods show impressive performance by leveraging contrastive learning to tackle the representation gap between every two views, they share the common limitation of not performing semantic alignment from a global perspective, resulting in the undermining of semantic patterns in multi-view data. This paper presents CSOT, namely Common Semantics via Optimal Transport, to boost contrastive multi-view clustering via semantic learning in a common space that integrates all views. Through optimal transport, the samples in multiple views are mapped to the joint clusters which represent the multi-view semantic patterns in the common space. With the semantic assignment derived from the optimal transport plan, we design a semantic learning module where the soft assignment vector works as a global supervision to enforce the model to learn consistent semantics among all views. Moreover, we propose a semantic-aware re-weighting strategy to treat samples differently according to their semantic significance, which improves the effectiveness of cross-view contrastive representation learning. Extensive experimental results demonstrate that CSOT achieves the state-of-the-art clustering performance.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141908631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unfolded Proximal Neural Networks for Robust Image Gaussian Denoising 用于鲁棒图像高斯去噪的折叠近端神经网络
Hoang Trieu Vy Le;Audrey Repetti;Nelly Pustelnik
A common approach to solve inverse imaging problems relies on finding a maximum a posteriori (MAP) estimate of the original unknown image, by solving a minimization problem. In this context, iterative proximal algorithms are widely used, enabling to handle non-smooth functions and linear operators. Recently, these algorithms have been paired with deep learning strategies, to further improve the estimate quality. In particular, proximal neural networks (PNNs) have been introduced, obtained by unrolling a proximal algorithm as for finding a MAP estimate, but over a fixed number of iterations, with learned linear operators and parameters. As PNNs are based on optimization theory, they are very flexible, and can be adapted to any image restoration task, as soon as a proximal algorithm can solve it. They further have much lighter architectures than traditional networks. In this article we propose a unified framework to build PNNs for the Gaussian denoising task, based on both the dual-FB and the primal-dual Chambolle-Pock algorithms. We further show that accelerated inertial versions of these algorithms enable skip connections in the associated NN layers. We propose different learning strategies for our PNN framework, and investigate their robustness (Lipschitz property) and denoising efficiency. Finally, we assess the robustness of our PNNs when plugged in a forward-backward algorithm for an image deblurring problem.
解决逆成像问题的常用方法是通过解决最小化问题,找到原始未知图像的最大后验(MAP)估计值。在这种情况下,迭代近似算法被广泛使用,能够处理非光滑函数和线性算子。最近,这些算法与深度学习策略搭配使用,进一步提高了估计质量。特别是,近端神经网络(PNN)被引入,它是通过展开近端算法来获得的,就像寻找 MAP 估计值一样,但要经过固定次数的迭代,并使用学习到的线性算子和参数。由于 PNN 以优化理论为基础,因此非常灵活,只要近似算法能够解决,就能适用于任何图像修复任务。与传统网络相比,它们的架构更加轻巧。在这篇文章中,我们提出了一个统一的框架,以双 FB 算法和原始双 Chambolle-Pock 算法为基础,为高斯去噪任务构建 PNN。我们进一步证明,这些算法的加速惯性版本可以在相关的 NN 层中实现跳过连接。我们为 PNN 框架提出了不同的学习策略,并研究了它们的鲁棒性(Lipschitz 属性)和去噪效率。最后,我们评估了我们的 PNN 在插入图像去模糊问题的前向后向算法时的鲁棒性。
{"title":"Unfolded Proximal Neural Networks for Robust Image Gaussian Denoising","authors":"Hoang Trieu Vy Le;Audrey Repetti;Nelly Pustelnik","doi":"10.1109/TIP.2024.3437219","DOIUrl":"10.1109/TIP.2024.3437219","url":null,"abstract":"A common approach to solve inverse imaging problems relies on finding a maximum a posteriori (MAP) estimate of the original unknown image, by solving a minimization problem. In this context, iterative proximal algorithms are widely used, enabling to handle non-smooth functions and linear operators. Recently, these algorithms have been paired with deep learning strategies, to further improve the estimate quality. In particular, proximal neural networks (PNNs) have been introduced, obtained by unrolling a proximal algorithm as for finding a MAP estimate, but over a fixed number of iterations, with learned linear operators and parameters. As PNNs are based on optimization theory, they are very flexible, and can be adapted to any image restoration task, as soon as a proximal algorithm can solve it. They further have much lighter architectures than traditional networks. In this article we propose a unified framework to build PNNs for the Gaussian denoising task, based on both the dual-FB and the primal-dual Chambolle-Pock algorithms. We further show that accelerated inertial versions of these algorithms enable skip connections in the associated NN layers. We propose different learning strategies for our PNN framework, and investigate their robustness (Lipschitz property) and denoising efficiency. Finally, we assess the robustness of our PNNs when plugged in a forward-backward algorithm for an image deblurring problem.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141903949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph Embedding Interclass Relation-Aware Adaptive Network for Cross-Scene Classification of Multisource Remote Sensing Data 用于多源遥感数据跨场景分类的图嵌入类间关系感知自适应网络
Teng Yang;Song Xiao;Jiahui Qu;Wenqian Dong;Qian Du;Yunsong Li
The unsupervised domain adaptation (UDA) based cross-scene remote sensing image classification has recently become an appealing research topic, since it is a valid solution to unsupervised scene classification by exploiting well-labeled data from another scene. Despite its good performance in reducing domain shifts, UDA in multisource data scenarios is hindered by several critical challenges. The first one is the heterogeneity inherent in multisource data complicates domain alignment. The second challenge is the incomplete representation of feature distribution caused by the neglect of the contribution from global information. The third challenge is the inaccuracies in alignment due to errors in establishing target domain conditional distributions. Since UDA does not guarantee the complete consistency of the distribution of the two domains, networks using simple classifiers are still affected by domain shifts, resulting in poor performance. In this paper, we propose a graph embedding interclass relation-aware adaptive network (GeIraA-Net) for unsupervised classification of multi-source remote sensing data, which facilitates knowledge transfer at the class level for two domains by leveraging aligned features to perceive inter-class relation. More specifically, a graph-based progressive hierarchical feature extraction network is constructed, capable of capturing both local and global features of multisource data, thereby consolidating comprehensive domain information within a unified feature space. To deal with the imprecise alignment of data distribution, a joint de-scrambling alignment strategy is designed to utilize the features obtained by a three-step pseudo-label generation module for more delicate domain calibration. Moreover, an adaptive inter-class topology based classifier is constructed to further improve the classification accuracy by making the classifier domain adaptive at the category level. The experimental results show that GeIraA-Net has significant advantages over the current state-of-the-art cross-scene classification methods.
基于无监督域自适应(UDA)的跨场景遥感图像分类最近成为一个颇具吸引力的研究课题,因为它是利用来自另一场景的标记良好的数据进行无监督场景分类的有效解决方案。尽管 UDA 在减少领域偏移方面表现出色,但在多源数据场景中却受到几个关键挑战的阻碍。首先,多源数据固有的异质性使域对齐变得复杂。第二个挑战是由于忽略了全局信息的贡献而导致特征分布的不完整呈现。第三个挑战是在建立目标域条件分布时出现错误,导致配准不准确。由于 UDA 不能保证两个域的分布完全一致,因此使用简单分类器的网络仍然会受到域偏移的影响,导致性能不佳。在本文中,我们提出了一种用于多源遥感数据无监督分类的图嵌入类间关系感知自适应网络(GeIraA-Net),它通过利用对齐特征感知类间关系,促进了两个域的类级知识转移。更具体地说,它构建了一个基于图的渐进式分层特征提取网络,能够捕捉多源数据的局部和全局特征,从而在统一的特征空间内整合综合领域信息。针对数据分布不精确的配准问题,设计了一种联合去扰配准策略,利用三步伪标签生成模块获得的特征进行更精细的领域配准。此外,还构建了基于类间拓扑结构的自适应分类器,通过使分类器域在类别级别上自适应,进一步提高分类精度。实验结果表明,与目前最先进的跨场景分类方法相比,GeIraA-Net 具有显著的优势。
{"title":"Graph Embedding Interclass Relation-Aware Adaptive Network for Cross-Scene Classification of Multisource Remote Sensing Data","authors":"Teng Yang;Song Xiao;Jiahui Qu;Wenqian Dong;Qian Du;Yunsong Li","doi":"10.1109/TIP.2024.3422881","DOIUrl":"10.1109/TIP.2024.3422881","url":null,"abstract":"The unsupervised domain adaptation (UDA) based cross-scene remote sensing image classification has recently become an appealing research topic, since it is a valid solution to unsupervised scene classification by exploiting well-labeled data from another scene. Despite its good performance in reducing domain shifts, UDA in multisource data scenarios is hindered by several critical challenges. The first one is the heterogeneity inherent in multisource data complicates domain alignment. The second challenge is the incomplete representation of feature distribution caused by the neglect of the contribution from global information. The third challenge is the inaccuracies in alignment due to errors in establishing target domain conditional distributions. Since UDA does not guarantee the complete consistency of the distribution of the two domains, networks using simple classifiers are still affected by domain shifts, resulting in poor performance. In this paper, we propose a graph embedding interclass relation-aware adaptive network (GeIraA-Net) for unsupervised classification of multi-source remote sensing data, which facilitates knowledge transfer at the class level for two domains by leveraging aligned features to perceive inter-class relation. More specifically, a graph-based progressive hierarchical feature extraction network is constructed, capable of capturing both local and global features of multisource data, thereby consolidating comprehensive domain information within a unified feature space. To deal with the imprecise alignment of data distribution, a joint de-scrambling alignment strategy is designed to utilize the features obtained by a three-step pseudo-label generation module for more delicate domain calibration. Moreover, an adaptive inter-class topology based classifier is constructed to further improve the classification accuracy by making the classifier domain adaptive at the category level. The experimental results show that GeIraA-Net has significant advantages over the current state-of-the-art cross-scene classification methods.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141899233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph-DETR4D: Spatio-Temporal Graph Modeling for Multi-View 3D Object Detection Graph-DETR4D:用于多视角 3D 物体检测的时空图建模
Zehui Chen;Zheng Chen;Zhenyu Li;Shiquan Zhang;Liangji Fang;Qinhong Jiang;Feng Wu;Feng Zhao
Multi-View 3D object detection (MV3D) has made tremendous progress by leveraging multiple perspective features through surrounding cameras. Despite demonstrating promising prospects in various applications, accurately detecting objects through camera view in the 3D space is extremely difficult due to the ill-posed issue in monocular depth estimation. Recently, Graph-DETR3D presents a novel graph-based 3D-2D query paradigm in aggregating multi-view images for 3D object detection and achieves competitive performance. Although it enriches the query representations with 2D image features through a learnable 3D graph, it still suffers from limited depth and velocity estimation abilities due to the adoption of a single-frame input setting. To solve this problem, we introduce a unified spatial-temporal graph modeling framework to fully leverage the multi-view imagery cues under the multi-frame inputs setting. Thanks to the flexibility and sparsity of the dynamic graph architecture, we lift the original 3D graph into the 4D space with an effective attention mechanism to automatically perceive imagery information at both spatial and temporal levels. Moreover, considering the main latency bottleneck lies in the image backbone, we propose a novel dense-sparse distillation framework for multi-view 3D object detection, to reduce the computational budget while sacrificing no detection accuracy, making it more suitable for real-world deployment. To this end, we propose Graph-DETR4D, a faster and stronger multi-view 3D object detection framework, built on top of Graph-DETR3D. Extensive experiments on nuScenes and Waymo benchmarks demonstrate the effectiveness and efficiency of Graph-DETR4D. Notably, our best model achieves 62.0% NDS on nuScenes test leaderboard. Code is available at https://github.com/zehuichen123/Graph-DETR4D.
多视角三维物体检测(MV3D)通过利用周围摄像机的多视角特征取得了巨大进步。尽管在各种应用中展示了广阔的前景,但由于单目深度估计中存在的问题,在三维空间中通过摄像头视图精确检测物体极为困难。最近,Graph-DETR3D [12] 提出了一种新颖的基于图的 3D-2D 查询范例,用于聚合多视角图像进行 3D 物体检测,并取得了极具竞争力的性能。虽然它通过可学习的三维图用二维图像特征丰富了查询表示,但由于采用了单帧输入设置,其深度和速度估计能力仍然有限。为了解决这个问题,我们引入了一个统一的时空图建模框架,以便在多帧输入设置下充分利用多视角图像线索。得益于动态图架构的灵活性和稀疏性,我们将原始的三维图提升到四维空间,并通过有效的注意力机制自动感知空间和时间层面的图像信息。此外,考虑到主要的延迟瓶颈在于图像骨干网,我们提出了一种新颖的密集-稀疏蒸馏框架,用于多视角三维物体检测,在不牺牲检测精度的前提下减少计算预算,使其更适合现实世界的部署。为此,我们在 Graph-DETR3D 的基础上提出了更快、更强的多视角三维物体检测框架 Graph-DETR4D。在 nuScenes 和 Waymo 基准上进行的大量实验证明了 Graph-DETR4D 的有效性和效率。值得注意的是,我们的最佳模型在 nuScenes 测试排行榜上取得了 62.0% 的 NDS。代码见 https://github.com/zehuichen123/Graph-DETR4D。
{"title":"Graph-DETR4D: Spatio-Temporal Graph Modeling for Multi-View 3D Object Detection","authors":"Zehui Chen;Zheng Chen;Zhenyu Li;Shiquan Zhang;Liangji Fang;Qinhong Jiang;Feng Wu;Feng Zhao","doi":"10.1109/TIP.2024.3430473","DOIUrl":"10.1109/TIP.2024.3430473","url":null,"abstract":"Multi-View 3D object detection (MV3D) has made tremendous progress by leveraging multiple perspective features through surrounding cameras. Despite demonstrating promising prospects in various applications, accurately detecting objects through camera view in the 3D space is extremely difficult due to the ill-posed issue in monocular depth estimation. Recently, Graph-DETR3D presents a novel graph-based 3D-2D query paradigm in aggregating multi-view images for 3D object detection and achieves competitive performance. Although it enriches the query representations with 2D image features through a learnable 3D graph, it still suffers from limited depth and velocity estimation abilities due to the adoption of a single-frame input setting. To solve this problem, we introduce a unified spatial-temporal graph modeling framework to fully leverage the multi-view imagery cues under the multi-frame inputs setting. Thanks to the flexibility and sparsity of the dynamic graph architecture, we lift the original 3D graph into the 4D space with an effective attention mechanism to automatically perceive imagery information at both spatial and temporal levels. Moreover, considering the main latency bottleneck lies in the image backbone, we propose a novel dense-sparse distillation framework for multi-view 3D object detection, to reduce the computational budget while sacrificing no detection accuracy, making it more suitable for real-world deployment. To this end, we propose Graph-DETR4D, a faster and stronger multi-view 3D object detection framework, built on top of Graph-DETR3D. Extensive experiments on nuScenes and Waymo benchmarks demonstrate the effectiveness and efficiency of Graph-DETR4D. Notably, our best model achieves 62.0% NDS on nuScenes test leaderboard. Code is available at \u0000<uri>https://github.com/zehuichen123/Graph-DETR4D</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141879978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Model Attention Expansion for Few-Shot Class-Incremental Learning 少镜头类增量学习的注意力扩展模型
Xuan Wang;Zhong Ji;Yunlong Yu;Yanwei Pang;Jungong Han
Few-Shot Class-Incremental Learning (FSCIL) aims at incrementally learning new knowledge from limited training examples without forgetting previous knowledge. However, we observe that existing methods face a challenge known as supervision collapse, where the model disproportionately emphasizes class-specific features of base classes at the detriment of novel class representations, leading to restricted cognitive capabilities. To alleviate this issue, we propose a new framework, Model aTtention Expansion for Few-Shot Class-Incremental Learning (MTE-FSCIL), aimed at expanding the model attention fields to improve transferability without compromising the discriminative capability for base classes. Specifically, the framework adopts a dual-stage training strategy, comprising pre-training and meta-training stages. In the pre-training stage, we present a new regularization technique, named the Reserver (RS) loss, to expand the global perception and reduce over-reliance on class-specific features by amplifying feature map activations. During the meta-training stage, we introduce the Repeller (RP) loss, a novel pair-based loss that promotes variation in representations and improves the model’s recognition of sample uniqueness by scattering intra-class samples within the embedding space. Furthermore, we propose a Transformational Adaptation (TA) strategy to enable continuous incorporation of new knowledge from downstream tasks, thus facilitating cross-task knowledge transfer. Extensive experimental results on mini-ImageNet, CIFAR100, and CUB200 datasets demonstrate that our proposed framework consistently outperforms the state-of-the-art methods.
少量类增量学习(FSCIL)旨在从有限的训练实例中增量学习新知识,同时不遗忘以前的知识。然而,我们发现现有的方法面临着一个被称为 "监督崩溃"(supervision collapse)的挑战,即模型不成比例地强调基础类的特定类特征,而忽略了新的类表征,从而导致认知能力受限。为了缓解这一问题,我们提出了一个新的框架,即 "少量类增量学习的注意力扩展模型(MTE-FSCIL)",旨在扩展模型的注意力领域,以提高可迁移性,同时不影响对基类的判别能力。具体来说,该框架采用双阶段训练策略,包括预训练和元训练阶段。在预训练阶段,我们提出了一种新的正则化技术,名为 "Reserver(RS)损失",以扩大全局感知,并通过放大特征图激活来减少对特定类别特征的过度依赖。在元训练阶段,我们引入了 Repeller(RP)损失,这是一种基于配对的新型损失,通过在嵌入空间内分散类内样本,促进表征的变化并提高模型对样本唯一性的识别能力。此外,我们还提出了一种转换适应(TA)策略,以持续吸收来自下游任务的新知识,从而促进跨任务知识转移。在 mini-ImageNet、CIFAR100 和 CUB200 数据集上的大量实验结果表明,我们提出的框架始终优于最先进的方法。
{"title":"Model Attention Expansion for Few-Shot Class-Incremental Learning","authors":"Xuan Wang;Zhong Ji;Yunlong Yu;Yanwei Pang;Jungong Han","doi":"10.1109/TIP.2024.3434475","DOIUrl":"10.1109/TIP.2024.3434475","url":null,"abstract":"Few-Shot Class-Incremental Learning (FSCIL) aims at incrementally learning new knowledge from limited training examples without forgetting previous knowledge. However, we observe that existing methods face a challenge known as supervision collapse, where the model disproportionately emphasizes class-specific features of base classes at the detriment of novel class representations, leading to restricted cognitive capabilities. To alleviate this issue, we propose a new framework, Model aTtention Expansion for Few-Shot Class-Incremental Learning (MTE-FSCIL), aimed at expanding the model attention fields to improve transferability without compromising the discriminative capability for base classes. Specifically, the framework adopts a dual-stage training strategy, comprising pre-training and meta-training stages. In the pre-training stage, we present a new regularization technique, named the Reserver (RS) loss, to expand the global perception and reduce over-reliance on class-specific features by amplifying feature map activations. During the meta-training stage, we introduce the Repeller (RP) loss, a novel pair-based loss that promotes variation in representations and improves the model’s recognition of sample uniqueness by scattering intra-class samples within the embedding space. Furthermore, we propose a Transformational Adaptation (TA) strategy to enable continuous incorporation of new knowledge from downstream tasks, thus facilitating cross-task knowledge transfer. Extensive experimental results on mini-ImageNet, CIFAR100, and CUB200 datasets demonstrate that our proposed framework consistently outperforms the state-of-the-art methods.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141876990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perception-Distortion Balanced Super-Resolution: A Multi-Objective Optimization Perspective 感知-失真平衡超分辨率:多目标优化视角。
Lingchen Sun;Jie Liang;Shuaizheng Liu;Hongwei Yong;Lei Zhang
High perceptual quality and low distortion degree are two important goals in image restoration tasks such as super-resolution (SR). Most of the existing SR methods aim to achieve these goals by minimizing the corresponding yet conflicting losses, such as the $ell _{1}$ loss and the adversarial loss. Unfortunately, the commonly used gradient-based optimizers, such as Adam, are hard to balance these objectives due to the opposite gradient decent directions of the contradictory losses. In this paper, we formulate the perception-distortion trade-off in SR as a multi-objective optimization problem and develop a new optimizer by integrating the gradient-free evolutionary algorithm (EA) with gradient-based Adam, where EA and Adam focus on the divergence and convergence of the optimization directions respectively. As a result, a population of optimal models with different perception-distortion preferences is obtained. We then design a fusion network to merge these models into a single stronger one for an effective perception-distortion trade-off. Experiments demonstrate that with the same backbone network, the perception-distortion balanced SR model trained by our method can achieve better perceptual quality than its competitors while attaining better reconstruction fidelity. Codes and models can be found at https://github.com/csslc/EA-Adam.
高感知质量和低失真度是超分辨率(SR)等图像复原任务的两个重要目标。现有的大多数 SR 方法都是通过最小化相应但相互冲突的损失(如 ℓ1 损失和对抗损失)来实现这些目标的。遗憾的是,常用的基于梯度的优化器(如 Adam)很难兼顾这些目标,因为相互矛盾的损失具有相反的梯度分布方向。本文将 SR 中的感知-失真权衡问题表述为一个多目标优化问题,并通过将无梯度进化算法(EA)与基于梯度的 Adam 相结合,开发了一种新的优化器,其中 EA 和 Adam 分别关注优化方向的发散性和收敛性。因此,我们得到了具有不同感知失真偏好的最优模型群。然后,我们设计了一个融合网络,将这些模型合并成一个更强的模型,以实现有效的感知-失真权衡。实验证明,在相同的骨干网络下,通过我们的方法训练出的感知-失真平衡的 SR 模型可以获得比竞争对手更好的感知质量,同时达到更好的重建保真度。代码和模型见 https://github.com/csslc/EA-Adam。
{"title":"Perception-Distortion Balanced Super-Resolution: A Multi-Objective Optimization Perspective","authors":"Lingchen Sun;Jie Liang;Shuaizheng Liu;Hongwei Yong;Lei Zhang","doi":"10.1109/TIP.2024.3434426","DOIUrl":"10.1109/TIP.2024.3434426","url":null,"abstract":"High perceptual quality and low distortion degree are two important goals in image restoration tasks such as super-resolution (SR). Most of the existing SR methods aim to achieve these goals by minimizing the corresponding yet conflicting losses, such as the \u0000<inline-formula> <tex-math>$ell _{1}$ </tex-math></inline-formula>\u0000 loss and the adversarial loss. Unfortunately, the commonly used gradient-based optimizers, such as Adam, are hard to balance these objectives due to the opposite gradient decent directions of the contradictory losses. In this paper, we formulate the perception-distortion trade-off in SR as a multi-objective optimization problem and develop a new optimizer by integrating the gradient-free evolutionary algorithm (EA) with gradient-based Adam, where EA and Adam focus on the divergence and convergence of the optimization directions respectively. As a result, a population of optimal models with different perception-distortion preferences is obtained. We then design a fusion network to merge these models into a single stronger one for an effective perception-distortion trade-off. Experiments demonstrate that with the same backbone network, the perception-distortion balanced SR model trained by our method can achieve better perceptual quality than its competitors while attaining better reconstruction fidelity. Codes and models can be found at \u0000<uri>https://github.com/csslc/EA-Adam</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141876991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Inspector for Face Forgery Detection: Defending Against Adversarial Attacks From Coarse to Fine 人脸伪造检测检查员:从粗到细抵御对抗性攻击
Ruiyang Xia;Dawei Zhou;Decheng Liu;Jie Li;Lin Yuan;Nannan Wang;Xinbo Gao
The emergence of face forgery has raised global concerns on social security, thereby facilitating the research on automatic forgery detection. Although current forgery detectors have demonstrated promising performance in determining authenticity, their susceptibility to adversarial perturbations remains insufficiently addressed. Given the nuanced discrepancies between real and fake instances are essential in forgery detection, previous defensive paradigms based on input processing and adversarial training tend to disrupt these discrepancies. For the detectors, the learning difficulty is thus increased, and the natural accuracy is dramatically decreased. To achieve adversarial defense without changing the instances as well as the detectors, a novel defensive paradigm called Inspector is designed specifically for face forgery detectors. Specifically, Inspector defends against adversarial attacks in a coarse-to-fine manner. In the coarse defense stage, adversarial instances with evident perturbations are directly identified and filtered out. Subsequently, in the fine defense stage, the threats from adversarial instances with imperceptible perturbations are further detected and eliminated. Experimental results across different types of face forgery datasets and detectors demonstrate that our method achieves state-of-the-art performances against various types of adversarial perturbations while better preserving natural accuracy. Code is available on https://github.com/xarryon/Inspector.
人脸伪造现象的出现引发了全球对社会安全的关注,从而促进了对自动伪造检测的研究。尽管目前的伪造检测器在判断真伪方面表现出了良好的性能,但它们易受对抗性扰动影响的问题仍未得到充分解决。鉴于真假实例之间的细微差别对伪造检测至关重要,以往基于输入处理和对抗训练的防御范式往往会破坏这些差异。对于检测器来说,学习难度会因此增加,自然准确率也会大大降低。为了在不改变实例和检测器的情况下实现对抗性防御,我们专门为人脸伪造检测器设计了一种名为 Inspector 的新型防御范式。具体来说,Inspector 以从粗到细的方式防御对抗性攻击。在粗略防御阶段,具有明显扰动的对抗实例会被直接识别并过滤掉。然后,在精细防御阶段,进一步检测和消除来自不易察觉扰动的对抗实例的威胁。在不同类型的人脸伪造数据集和检测器上的实验结果表明,我们的方法在对抗各种类型的对抗性扰动的同时,还能更好地保持自然准确性,达到了最先进的性能。代码见 https://github.com/xarryon/Inspector。
{"title":"Inspector for Face Forgery Detection: Defending Against Adversarial Attacks From Coarse to Fine","authors":"Ruiyang Xia;Dawei Zhou;Decheng Liu;Jie Li;Lin Yuan;Nannan Wang;Xinbo Gao","doi":"10.1109/TIP.2024.3434388","DOIUrl":"10.1109/TIP.2024.3434388","url":null,"abstract":"The emergence of face forgery has raised global concerns on social security, thereby facilitating the research on automatic forgery detection. Although current forgery detectors have demonstrated promising performance in determining authenticity, their susceptibility to adversarial perturbations remains insufficiently addressed. Given the nuanced discrepancies between real and fake instances are essential in forgery detection, previous defensive paradigms based on input processing and adversarial training tend to disrupt these discrepancies. For the detectors, the learning difficulty is thus increased, and the natural accuracy is dramatically decreased. To achieve adversarial defense without changing the instances as well as the detectors, a novel defensive paradigm called Inspector is designed specifically for face forgery detectors. Specifically, Inspector defends against adversarial attacks in a coarse-to-fine manner. In the coarse defense stage, adversarial instances with evident perturbations are directly identified and filtered out. Subsequently, in the fine defense stage, the threats from adversarial instances with imperceptible perturbations are further detected and eliminated. Experimental results across different types of face forgery datasets and detectors demonstrate that our method achieves state-of-the-art performances against various types of adversarial perturbations while better preserving natural accuracy. Code is available on \u0000<uri>https://github.com/xarryon/Inspector</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141876989","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SelfGCN: Graph Convolution Network With Self-Attention for Skeleton-Based Action Recognition SelfGCN:基于骨架的动作识别图卷积网络(Graph Convolution Network with Self-Attention for Skeleton-based Action Recognition)。
Zhize Wu;Pengpeng Sun;Xin Chen;Keke Tang;Tong Xu;Le Zou;Xiaofeng Wang;Ming Tan;Fan Cheng;Thomas Weise
Graph Convolutional Networks (GCNs) are widely used for skeleton-based action recognition and achieved remarkable performance. Due to the locality of graph convolution, GCNs can only utilize short-range node dependencies but fail to model long-range node relationships. In addition, existing graph convolution based methods normally use a uniform skeleton topology for all frames, which limits the ability of feature learning. To address these issues, we present the Graph Convolution Network with Self-Attention (SelfGCN), which consists of a mixing features across self-attention and graph convolution (MFSG) module and a temporal-specific spatial self-attention (TSSA) module. The MFSG module models local and global relationships between joints by executing graph convolution and self-attention branches in parallel. Its bi-directional interactive learning strategy utilizes complementary clues in the channel dimensions and the spatial dimensions across both of these branches. The TSSA module uses self-attention to learn the spatial relationships between joints of each frame in a skeleton sequence. It also models the unique spatial features of the single frames. We conduct extensive experiments on three popular benchmark datasets, NTU RGB+D, NTU RGB+D120, and Northwestern-UCLA. The results of the experiment demonstrate that our method achieves or exceeds the record accuracies on all three benchmarks. Our project website is available at https://github.com/SunPengP/SelfGCN.
图卷积网络(Graph Convolutional Networks,GCNs)被广泛用于基于骨骼的动作识别,并取得了显著的性能。由于图卷积的局部性,GCN 只能利用短程节点依赖关系,而无法模拟长程节点关系。此外,现有的基于图卷积的方法通常对所有帧使用统一的骨架拓扑结构,这限制了特征学习的能力。为了解决这些问题,我们提出了具有自注意力的图卷积网络(SelfGCN),它由一个跨自注意力和图卷积的混合特征(MFSG)模块和一个特定于时间的空间自注意力(TSSA)模块组成。MFSG 模块通过并行执行图卷积和自我注意分支来模拟关节之间的局部和全局关系。它的双向互动学习策略利用通道维度和空间维度的互补线索贯穿这两个分支。TSSA 模块利用自我注意来学习骨架序列中每个帧的关节之间的空间关系。它还对单帧的独特空间特征进行建模。我们在 NTU RGB+D、NTU RGB+D120 和 Northwestern-UCLA 这三个流行的基准数据集上进行了广泛的实验。实验结果表明,我们的方法在所有三个基准数据集上都达到或超过了创纪录的精确度。我们的项目网站是 https://github.com/SunPengP/SelfGCN。
{"title":"SelfGCN: Graph Convolution Network With Self-Attention for Skeleton-Based Action Recognition","authors":"Zhize Wu;Pengpeng Sun;Xin Chen;Keke Tang;Tong Xu;Le Zou;Xiaofeng Wang;Ming Tan;Fan Cheng;Thomas Weise","doi":"10.1109/TIP.2024.3433581","DOIUrl":"10.1109/TIP.2024.3433581","url":null,"abstract":"Graph Convolutional Networks (GCNs) are widely used for skeleton-based action recognition and achieved remarkable performance. Due to the locality of graph convolution, GCNs can only utilize short-range node dependencies but fail to model long-range node relationships. In addition, existing graph convolution based methods normally use a uniform skeleton topology for all frames, which limits the ability of feature learning. To address these issues, we present the Graph Convolution Network with Self-Attention (SelfGCN), which consists of a mixing features across self-attention and graph convolution (MFSG) module and a temporal-specific spatial self-attention (TSSA) module. The MFSG module models local and global relationships between joints by executing graph convolution and self-attention branches in parallel. Its bi-directional interactive learning strategy utilizes complementary clues in the channel dimensions and the spatial dimensions across both of these branches. The TSSA module uses self-attention to learn the spatial relationships between joints of each frame in a skeleton sequence. It also models the unique spatial features of the single frames. We conduct extensive experiments on three popular benchmark datasets, NTU RGB+D, NTU RGB+D120, and Northwestern-UCLA. The results of the experiment demonstrate that our method achieves or exceeds the record accuracies on all three benchmarks. Our project website is available at \u0000<uri>https://github.com/SunPengP/SelfGCN</uri>\u0000.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141861926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Blind Super-Resolution Network for Spatial-Specific and Spatial-Agnostic Degradations 针对空间特定和空间诊断衰减的自适应盲超分辨率网络
Weilei Wen;Chunle Guo;Wenqi Ren;Hongpeng Wang;Xiuli Shao
Prior methodologies have disregarded the diversities among distinct degradation types during image reconstruction, employing a uniform network model to handle multiple deteriorations. Nevertheless, we discover that prevalent degradation modalities, including sampling, blurring, and noise, can be roughly categorized into two classes. We classify the first class as spatial-agnostic dominant degradations, less affected by regional changes in image space, such as downsampling and noise degradation. The second class degradation type is intimately associated with the spatial position of the image, such as blurring, and we identify them as spatial-specific dominant degradations. We introduce a dynamic filter network integrating global and local branches to address these two degradation types. This network can greatly alleviate the practical degradation problem. Specifically, the global dynamic filtering layer can perceive the spatial-agnostic dominant degradation in different images by applying weights generated by the attention mechanism to multiple parallel standard convolution kernels, enhancing the network’s representation ability. Meanwhile, the local dynamic filtering layer converts feature maps of the image into a spatially specific dynamic filtering operator, which performs spatially specific convolution operations on the image features to handle spatial-specific dominant degradations. By effectively integrating both global and local dynamic filtering operators, our proposed method outperforms state-of-the-art blind super-resolution algorithms in both synthetic and real image datasets.
之前的方法在图像重建过程中忽略了不同退化类型之间的多样性,采用统一的网络模型来处理多种退化。然而,我们发现,包括采样、模糊和噪声在内的普遍退化模式可大致分为两类。我们将第一类归为空间无关的主要劣化,受图像空间区域变化的影响较小,例如下采样和噪声劣化。第二类退化类型与图像的空间位置密切相关,如模糊,我们将其确定为特定空间的主导退化。我们引入了一个动态滤波网络,整合了全局和局部分支,以解决这两种劣化类型。该网络能极大地缓解实际退化问题。具体来说,全局动态滤波层可以通过将注意力机制产生的权重应用于多个并行标准卷积核来感知不同图像中的空间无关主导退化,从而增强网络的表示能力。同时,局部动态滤波层将图像的特征图转换成空间特定的动态滤波算子,对图像特征进行空间特定的卷积运算,以处理空间特定的主导退化。通过有效整合全局和局部动态滤波算子,我们提出的方法在合成和真实图像数据集上都优于最先进的盲超分辨率算法。
{"title":"Adaptive Blind Super-Resolution Network for Spatial-Specific and Spatial-Agnostic Degradations","authors":"Weilei Wen;Chunle Guo;Wenqi Ren;Hongpeng Wang;Xiuli Shao","doi":"10.1109/TIP.2024.3425169","DOIUrl":"10.1109/TIP.2024.3425169","url":null,"abstract":"Prior methodologies have disregarded the diversities among distinct degradation types during image reconstruction, employing a uniform network model to handle multiple deteriorations. Nevertheless, we discover that prevalent degradation modalities, including sampling, blurring, and noise, can be roughly categorized into two classes. We classify the first class as spatial-agnostic dominant degradations, less affected by regional changes in image space, such as downsampling and noise degradation. The second class degradation type is intimately associated with the spatial position of the image, such as blurring, and we identify them as spatial-specific dominant degradations. We introduce a dynamic filter network integrating global and local branches to address these two degradation types. This network can greatly alleviate the practical degradation problem. Specifically, the global dynamic filtering layer can perceive the spatial-agnostic dominant degradation in different images by applying weights generated by the attention mechanism to multiple parallel standard convolution kernels, enhancing the network’s representation ability. Meanwhile, the local dynamic filtering layer converts feature maps of the image into a spatially specific dynamic filtering operator, which performs spatially specific convolution operations on the image features to handle spatial-specific dominant degradations. By effectively integrating both global and local dynamic filtering operators, our proposed method outperforms state-of-the-art blind super-resolution algorithms in both synthetic and real image datasets.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141857447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1