首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
Transformer Module Networks for Systematic Generalization in Visual Question Answering. 在视觉问题解答中实现系统泛化的变压器模块网络
Pub Date : 2024-08-06 DOI: 10.1109/TPAMI.2024.3438887
Moyuru Yamada, Vanessa D'Amario, Kentaro Takemoto, Xavier Boix, Tomotake Sasaki

Transformers achieve great performance on Visual Question Answering (VQA). However, their systematic generalization capabilities, i.e., handling novel combinations of known concepts, is unclear. We reveal that Neural Module Networks (NMNs), i.e., question-specific compositions of modules that tackle a sub-task, achieve better or similar systematic generalization performance than the conventional Transformers, even though NMNs' modules are CNN-based. In order to address this shortcoming of Transformers with respect to NMNs, in this paper we investigate whether and how modularity can bring benefits to Transformers. Namely, we introduce Transformer Module Network (TMN), a novel NMN based on compositions of Transformer modules. TMNs achieve state-of-the-art systematic generalization performance in three VQA datasets, improving more than 30% over standard Transformers for novel compositions of sub-tasks. We show that not only the module composition but also the module specialization for each sub-task are the key of such performance gain.

变换器在视觉问题解答(VQA)中表现出色。然而,它们的系统泛化能力,即处理已知概念的新组合的能力尚不清楚。我们发现,神经模块网络(NMN),即针对特定问题的模块组合,能够处理子任务,与传统的变形器相比,具有更好或相似的系统泛化性能,尽管 NMN 的模块是基于 CNN 的。为了解决变形器相对于 NMNs 的这一不足,我们在本文中研究了模块化能否以及如何为变形器带来好处。也就是说,我们引入了变形模块网络(TMN),这是一种基于变形模块组合的新型 NMN。TMN 在三个 VQA 数据集中实现了最先进的系统泛化性能,在子任务的新颖组合方面比标准 Transformers 提高了 30% 以上。我们的研究表明,这种性能提升的关键不仅在于模块组成,还在于每个子任务的模块专业化。
{"title":"Transformer Module Networks for Systematic Generalization in Visual Question Answering.","authors":"Moyuru Yamada, Vanessa D'Amario, Kentaro Takemoto, Xavier Boix, Tomotake Sasaki","doi":"10.1109/TPAMI.2024.3438887","DOIUrl":"10.1109/TPAMI.2024.3438887","url":null,"abstract":"<p><p>Transformers achieve great performance on Visual Question Answering (VQA). However, their systematic generalization capabilities, i.e., handling novel combinations of known concepts, is unclear. We reveal that Neural Module Networks (NMNs), i.e., question-specific compositions of modules that tackle a sub-task, achieve better or similar systematic generalization performance than the conventional Transformers, even though NMNs' modules are CNN-based. In order to address this shortcoming of Transformers with respect to NMNs, in this paper we investigate whether and how modularity can bring benefits to Transformers. Namely, we introduce Transformer Module Network (TMN), a novel NMN based on compositions of Transformer modules. TMNs achieve state-of-the-art systematic generalization performance in three VQA datasets, improving more than 30% over standard Transformers for novel compositions of sub-tasks. We show that not only the module composition but also the module specialization for each sub-task are the key of such performance gain.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141899236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CamoFormer: Masked Separable Attention for Camouflaged Object Detection. CamoFormer:用于伪装物体检测的屏蔽可分离注意力
Pub Date : 2024-08-05 DOI: 10.1109/TPAMI.2024.3438565
Bowen Yin, Xuying Zhang, Deng-Ping Fan, Shaohui Jiao, Ming-Ming Cheng, Luc Van Gool, Qibin Hou

How to identify and segment camouflaged objects from the background is challenging. Inspired by the multi-head self-attention in Transformers, we present a simple masked separable attention (MSA) for camouflaged object detection. We first separate the multi-head self-attention into three parts, which are responsible for distinguishing the camouflaged objects from the background using different mask strategies. Furthermore, we propose to capture high-resolution semantic representations progressively based on a simple top-down decoder with the proposed MSA to attain precise segmentation results. These structures plus a backbone encoder form a new model, dubbed CamoFormer. Extensive experiments show that CamoFormer achieves new state-of-the-art performance on three widely-used camouflaged object detection benchmarks. To better evaluate the performance of the proposed CamoFormer around the border regions, we propose to use two new metrics, i.e. BR-M and BR-F. There are on average  ∼ 5% relative improvements over previous methods in terms of S-measure and weighted F-measure. Our code is available at https://github.com/HVision-NKU/CamoFormer.

如何从背景中识别和分割伪装物体是一项挑战。受《变形金刚》中多头自我注意的启发,我们提出了一种用于伪装物体检测的简单遮罩可分离注意(MSA)。我们首先将多头自我注意分为三个部分,这三个部分负责使用不同的掩码策略将伪装物体从背景中区分出来。此外,我们还建议在简单的自上而下解码器基础上,利用所提出的 MSA 逐步捕捉高分辨率语义表征,以获得精确的分割结果。这些结构加上主干编码器构成了一个新模型,被称为 CamoFormer。广泛的实验表明,CamoFormer 在三个广泛使用的伪装物体检测基准上取得了新的一流性能。为了更好地评估 CamoFormer 在边界区域的性能,我们建议使用两个新指标,即 BR-M 和 BR-F。与之前的方法相比,在 S-度量和加权 F-度量方面平均有 5%的相对改进。我们的代码见 https://github.com/HVision-NKU/CamoFormer。
{"title":"CamoFormer: Masked Separable Attention for Camouflaged Object Detection.","authors":"Bowen Yin, Xuying Zhang, Deng-Ping Fan, Shaohui Jiao, Ming-Ming Cheng, Luc Van Gool, Qibin Hou","doi":"10.1109/TPAMI.2024.3438565","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3438565","url":null,"abstract":"<p><p>How to identify and segment camouflaged objects from the background is challenging. Inspired by the multi-head self-attention in Transformers, we present a simple masked separable attention (MSA) for camouflaged object detection. We first separate the multi-head self-attention into three parts, which are responsible for distinguishing the camouflaged objects from the background using different mask strategies. Furthermore, we propose to capture high-resolution semantic representations progressively based on a simple top-down decoder with the proposed MSA to attain precise segmentation results. These structures plus a backbone encoder form a new model, dubbed CamoFormer. Extensive experiments show that CamoFormer achieves new state-of-the-art performance on three widely-used camouflaged object detection benchmarks. To better evaluate the performance of the proposed CamoFormer around the border regions, we propose to use two new metrics, i.e. BR-M and BR-F. There are on average  ∼ 5% relative improvements over previous methods in terms of S-measure and weighted F-measure. Our code is available at https://github.com/HVision-NKU/CamoFormer.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141895073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Random Permutation Set Reasoning. 随机排列组合推理
Pub Date : 2024-08-05 DOI: 10.1109/TPAMI.2024.3438349
Jixiang Deng, Yong Deng, Jian- Bo Yang

In artificial intelligence, it is crucial for pattern recognition systems to process data with uncertain information, necessitating uncertainty reasoning approaches such as evidence theory. As an orderable extension of evidence theory, random permutation set (RPS) theory has received increasing attention. However, RPS theory lacks a suitable generation method for the element order of permutation mass function (PMF) and an efficient determination method for the fusion order of permutation orthogonal sum (POS). To solve these two issues, this paper proposes a reasoning model for RPS theory, called random permutation set reasoning (RPSR). RPSR consists of three techniques, including RPS generation method (RPSGM), RPSR rule of combination, and ordered probability transformation (OPT). Specifically, RPSGM can construct RPS based on Gaussian discriminant model and weight analysis; RPSR rule incorporates POS with reliability vector, which can combine RPS sources with reliability in fusion order; OPT is used to convert RPS into a probability distribution for the final decision. Besides, numerical examples are provided to illustrate the proposed RPSR. Moreover, the proposed RPSR is applied to classification problems. An RPSR-based classification algorithm (RPSRCA) and its hyperparameter tuning method are presented. The results demonstrate the efficiency and stability of RPSRCA compared to existing classifiers.

在人工智能领域,模式识别系统处理具有不确定信息的数据至关重要,这就需要证据理论等不确定性推理方法。作为证据理论的有序扩展,随机置换集(RPS)理论受到越来越多的关注。然而,RPS 理论缺乏一种合适的置换质量函数(PMF)元素阶的生成方法,也缺乏一种有效的置换正交和(POS)融合阶的确定方法。为了解决这两个问题,本文提出了一种 RPS 理论的推理模型,称为随机置换集推理(RPSR)。RPSR 由三种技术组成,包括 RPS 生成法(RPSGM)、RPSR 组合规则和有序概率变换(OPT)。具体来说,RPSGM 可基于高斯判别模型和权重分析构建 RPS;RPSR 规则将 POS 与可靠性向量相结合,可将具有可靠性的 RPS 来源按融合顺序组合起来;OPT 用于将 RPS 转化为概率分布,供最终决策使用。此外,还提供了数值示例来说明所提出的 RPSR。此外,还将所提出的 RPSR 应用于分类问题。介绍了基于 RPSR 的分类算法(RPSRCA)及其超参数调整方法。结果表明,与现有分类器相比,RPSRCA 具有高效性和稳定性。
{"title":"Random Permutation Set Reasoning.","authors":"Jixiang Deng, Yong Deng, Jian- Bo Yang","doi":"10.1109/TPAMI.2024.3438349","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3438349","url":null,"abstract":"<p><p>In artificial intelligence, it is crucial for pattern recognition systems to process data with uncertain information, necessitating uncertainty reasoning approaches such as evidence theory. As an orderable extension of evidence theory, random permutation set (RPS) theory has received increasing attention. However, RPS theory lacks a suitable generation method for the element order of permutation mass function (PMF) and an efficient determination method for the fusion order of permutation orthogonal sum (POS). To solve these two issues, this paper proposes a reasoning model for RPS theory, called random permutation set reasoning (RPSR). RPSR consists of three techniques, including RPS generation method (RPSGM), RPSR rule of combination, and ordered probability transformation (OPT). Specifically, RPSGM can construct RPS based on Gaussian discriminant model and weight analysis; RPSR rule incorporates POS with reliability vector, which can combine RPS sources with reliability in fusion order; OPT is used to convert RPS into a probability distribution for the final decision. Besides, numerical examples are provided to illustrate the proposed RPSR. Moreover, the proposed RPSR is applied to classification problems. An RPSR-based classification algorithm (RPSRCA) and its hyperparameter tuning method are presented. The results demonstrate the efficiency and stability of RPSRCA compared to existing classifiers.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141895075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gradient Harmonization in Unsupervised Domain Adaptation. 无监督领域适应中的梯度协调。
Pub Date : 2024-08-05 DOI: 10.1109/TPAMI.2024.3438154
Fuxiang Huang, Suqi Song, Lei Zhang

Unsupervised domain adaptation (UDA) intends to transfer knowledge from a labeled source domain to an unlabeled target domain. Many current methods focus on learning feature representations that are both discriminative for classification and invariant across domains by simultaneously optimizing domain alignment and classification tasks. However, these methods often overlook a crucial challenge: the inherent conflict between these two tasks during gradient-based optimization. In this paper, we delve into this issue and introduce two effective solutions known as Gradient Harmonization, including GH and GH++, to mitigate the conflict between domain alignment and classification tasks. GH operates by altering the gradient angle between different tasks from an obtuse angle to an acute angle, thus resolving the conflict and trade-offing the two tasks in a coordinated manner. Yet, this would cause both tasks to deviate from their original optimization directions. We thus further propose an improved version, GH++, which adjusts the gradient angle between tasks from an obtuse angle to a vertical angle. This not only eliminates the conflict but also minimizes deviation from the original gradient directions. Finally, for optimization convenience and efficiency, we evolve the gradient harmonization strategies into a dynamically weighted loss function using an integral operator on the harmonized gradient. Notably, GH/GH++ are orthogonal to UDA and can be seamlessly integrated into most existing UDA models. Theoretical insights and experimental analyses demonstrate that the proposed approaches not only enhance popular UDA baselines but also improve recent state-of-the-art models.

无监督领域适应(UDA)旨在将知识从有标签的源领域转移到无标签的目标领域。目前的许多方法都侧重于通过同时优化域对齐和分类任务来学习既能区分分类又能跨域不变的特征表征。然而,这些方法往往忽略了一个关键挑战:在基于梯度的优化过程中,这两个任务之间存在内在冲突。在本文中,我们深入探讨了这一问题,并介绍了两种有效的解决方案,即梯度协调(Gradient Harmonization),包括 GH 和 GH++,以缓解领域对齐和分类任务之间的冲突。GH 通过改变不同任务之间的梯度角,从钝角变为锐角,从而解决冲突,并以协调的方式权衡两个任务。然而,这将导致两个任务偏离原来的优化方向。因此,我们进一步提出了改进版 GH++,将任务间的梯度角从钝角调整为垂直角。这不仅消除了冲突,还最大限度地减少了对原始梯度方向的偏离。最后,为了优化的方便性和效率,我们利用协调梯度上的积分算子,将梯度协调策略演化为动态加权损失函数。值得注意的是,GH/GH++ 与 UDA 是正交的,可以无缝集成到大多数现有的 UDA 模型中。理论见解和实验分析表明,所提出的方法不仅增强了流行的 UDA 基线,还改进了最新的先进模型。
{"title":"Gradient Harmonization in Unsupervised Domain Adaptation.","authors":"Fuxiang Huang, Suqi Song, Lei Zhang","doi":"10.1109/TPAMI.2024.3438154","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3438154","url":null,"abstract":"<p><p>Unsupervised domain adaptation (UDA) intends to transfer knowledge from a labeled source domain to an unlabeled target domain. Many current methods focus on learning feature representations that are both discriminative for classification and invariant across domains by simultaneously optimizing domain alignment and classification tasks. However, these methods often overlook a crucial challenge: the inherent conflict between these two tasks during gradient-based optimization. In this paper, we delve into this issue and introduce two effective solutions known as Gradient Harmonization, including GH and GH++, to mitigate the conflict between domain alignment and classification tasks. GH operates by altering the gradient angle between different tasks from an obtuse angle to an acute angle, thus resolving the conflict and trade-offing the two tasks in a coordinated manner. Yet, this would cause both tasks to deviate from their original optimization directions. We thus further propose an improved version, GH++, which adjusts the gradient angle between tasks from an obtuse angle to a vertical angle. This not only eliminates the conflict but also minimizes deviation from the original gradient directions. Finally, for optimization convenience and efficiency, we evolve the gradient harmonization strategies into a dynamically weighted loss function using an integral operator on the harmonized gradient. Notably, GH/GH++ are orthogonal to UDA and can be seamlessly integrated into most existing UDA models. Theoretical insights and experimental analyses demonstrate that the proposed approaches not only enhance popular UDA baselines but also improve recent state-of-the-art models.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141895074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Essential Number of Principal Components and Nearly Training-Free Model for Spectral Analysis. 用于频谱分析的基本主成分数和近乎无训练模型
Pub Date : 2024-08-02 DOI: 10.1109/TPAMI.2024.3436860
Yifeng Bie, Shuai You, Xinrui Li, Xuekui Zhang, Tao Lu

Learning-enabled spectroscopic analysis, promising for automated real-time analysis of chemicals, is facing several challenges. Firstly, a typical machine learning model requires a large number of training samples that physical systems can not provide. Secondly, it requires the testing samples to be in range with the training samples, which often is not the case in the real world. Further, a spectroscopy device is limited by its memory size, computing power, and battery capacity. That requires highly efficient learning models for on-site analysis. In this paper, by analyzing multi-gas mixtures and multi-molecule suspensions, we first show that orders of magnitude reduction of data dimension can be achieved as the number of principal components that need to be retained is the same as the independent constituents in the mixture. From this principle, we designed highly compact models in which the essential principal components can be directly extracted from the interrelations between the individual chemical properties and principal components; and only a few training samples are required. Our model can predict the constituent concentrations that have not been seen in the training dataset and provide estimations of measurement noises. This approach can be extended as an effectively standardized method for principle component extraction.

具有学习功能的光谱分析有望实现化学物质的自动实时分析,但目前面临着一些挑战。首先,典型的机器学习模型需要大量的训练样本,而物理系统无法提供。其次,它要求测试样本与训练样本在一定范围内,而现实世界中往往不存在这种情况。此外,光谱设备还受到内存大小、计算能力和电池容量的限制。这就需要高效的学习模型来进行现场分析。在本文中,通过分析多气体混合物和多分子悬浮液,我们首先表明,由于需要保留的主成分数量与混合物中的独立成分数量相同,因此可以实现数据维度的数量级缩减。根据这一原理,我们设计了高度紧凑的模型,可以直接从单个化学特性和主成分之间的相互关系中提取基本主成分,而且只需要少量训练样本。我们的模型可以预测训练数据集中未出现的成分浓度,并提供测量噪声估计。这种方法可以扩展为一种有效的标准化原理成分提取方法。
{"title":"Essential Number of Principal Components and Nearly Training-Free Model for Spectral Analysis.","authors":"Yifeng Bie, Shuai You, Xinrui Li, Xuekui Zhang, Tao Lu","doi":"10.1109/TPAMI.2024.3436860","DOIUrl":"10.1109/TPAMI.2024.3436860","url":null,"abstract":"<p><p>Learning-enabled spectroscopic analysis, promising for automated real-time analysis of chemicals, is facing several challenges. Firstly, a typical machine learning model requires a large number of training samples that physical systems can not provide. Secondly, it requires the testing samples to be in range with the training samples, which often is not the case in the real world. Further, a spectroscopy device is limited by its memory size, computing power, and battery capacity. That requires highly efficient learning models for on-site analysis. In this paper, by analyzing multi-gas mixtures and multi-molecule suspensions, we first show that orders of magnitude reduction of data dimension can be achieved as the number of principal components that need to be retained is the same as the independent constituents in the mixture. From this principle, we designed highly compact models in which the essential principal components can be directly extracted from the interrelations between the individual chemical properties and principal components; and only a few training samples are required. Our model can predict the constituent concentrations that have not been seen in the training dataset and provide estimations of measurement noises. This approach can be extended as an effectively standardized method for principle component extraction.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141879980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UNK-VQA: A Dataset and a Probe into the Abstention Ability of Multi-modal Large Models. UNK-VQA:多模式大型模型的数据集和弃权能力探究
Pub Date : 2024-08-02 DOI: 10.1109/TPAMI.2024.3437288
Yangyang Guo, Fangkai Jiao, Zhiqi Shen, Liqiang Nie, Mohan Kankanhalli

Teaching Visual Question Answering (VQA) models to refrain from answering unanswerable questions is necessary for building a trustworthy AI system. Existing studies, though have explored various aspects of VQA but somewhat ignored this particular attribute. This paper aims to bridge the research gap by contributing a comprehensive dataset, called UNK-VQA. The dataset is specifically designed to address the challenge of questions that models do not know. To this end, we first augment the existing data via deliberate perturbations on either the image or question. In specific, we carefully ensure that the question-image semantics remain close to the original unperturbed distribution. By this means, the identification of unanswerable questions becomes challenging, setting our dataset apart from others that involve mere image replacement. We then extensively evaluate the zero- and few-shot performance of several emerging multi-modal large models and discover their significant limitations when applied to our dataset. Additionally, we also propose a straightforward method to tackle these unanswerable questions. This dataset, we believe, will serve as a valuable benchmark for enhancing the abstention capability of VQA models, thereby leading to increased trustworthiness of AI systems. We have made the dataset available to facilitate further exploration in this area.

要建立一个值得信赖的人工智能系统,就必须教会视觉问题解答(VQA)模型避免回答无法回答的问题。现有的研究虽然探索了 VQA 的各个方面,但在一定程度上忽略了这一特殊属性。本文旨在通过提供一个名为 UNK-VQA 的综合数据集来弥补这一研究空白。该数据集专门用于解决模型不知道的问题。为此,我们首先通过故意扰动图像或问题来增强现有数据。具体来说,我们会仔细确保问题-图像语义仍然接近原始的未扰动分布。通过这种方法,识别无法回答的问题就变得具有挑战性,从而使我们的数据集有别于其他仅涉及图像替换的数据集。然后,我们广泛评估了几种新兴多模态大型模型的零次和少次性能,并发现它们在应用于我们的数据集时存在明显的局限性。此外,我们还提出了一种直接的方法来解决这些无法回答的问题。我们相信,该数据集将成为增强 VQA 模型弃权能力的宝贵基准,从而提高人工智能系统的可信度。我们提供了该数据集,以促进在这一领域的进一步探索。
{"title":"UNK-VQA: A Dataset and a Probe into the Abstention Ability of Multi-modal Large Models.","authors":"Yangyang Guo, Fangkai Jiao, Zhiqi Shen, Liqiang Nie, Mohan Kankanhalli","doi":"10.1109/TPAMI.2024.3437288","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3437288","url":null,"abstract":"<p><p>Teaching Visual Question Answering (VQA) models to refrain from answering unanswerable questions is necessary for building a trustworthy AI system. Existing studies, though have explored various aspects of VQA but somewhat ignored this particular attribute. This paper aims to bridge the research gap by contributing a comprehensive dataset, called UNK-VQA. The dataset is specifically designed to address the challenge of questions that models do not know. To this end, we first augment the existing data via deliberate perturbations on either the image or question. In specific, we carefully ensure that the question-image semantics remain close to the original unperturbed distribution. By this means, the identification of unanswerable questions becomes challenging, setting our dataset apart from others that involve mere image replacement. We then extensively evaluate the zero- and few-shot performance of several emerging multi-modal large models and discover their significant limitations when applied to our dataset. Additionally, we also propose a straightforward method to tackle these unanswerable questions. This dataset, we believe, will serve as a valuable benchmark for enhancing the abstention capability of VQA models, thereby leading to increased trustworthiness of AI systems. We have made the dataset available to facilitate further exploration in this area.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141879981","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UniMiSS+: Universal Medical Self-Supervised Learning From Cross-Dimensional Unpaired Data. UniMiSS+:从跨维非配对数据进行通用医学自我监督学习
Pub Date : 2024-07-31 DOI: 10.1109/TPAMI.2024.3436105
Yutong Xie, Jianpeng Zhang, Yong Xia, Qi Wu

Self-supervised learning (SSL) opens up huge opportunities for medical image analysis that is well known for its lack of annotations. However, aggregating massive (unlabeled) 3D medical images like computerized tomography (CT) remains challenging due to its high imaging cost and privacy restrictions. In our pilot study, we advocated bringing a wealth of 2D images like chest X-rays as compensation for the lack of 3D data, aiming to build a universal medical self-supervised representation learning framework, called UniMiSS. Especially, we designed a pyramid U- like medical Transformer (MiT) as the backbone to make UniMiSS possible to perform SSL with both 2D and 3D images. Consequently, the predecessor UniMiSS has two obvious merits compared to current 3D-specific SSL: (1) more effective - superior to learning strong representations, benefiting from more and diverse data; and (2) more versatile - suitable for various downstream tasks without the restriction on the dimensionality barrier. Unfortunately, UniMiSS did not dig deeply into the intrinsic anatomy correlation between 2D medical images and 3D volumes due to the lack of paired multi-modal/dimension patient data. In this extension paper, we propose the UniMiSS+, in which we introduce the digitally reconstructed radiographs (DRR) technology to simulate X-ray images from a CT volume to access paired CT and X-ray data. Benefiting from the paired group, we introduce an extra pair- wise constraint to boost the cross-modality correlation learning, which also can be adopted as a cross-dimension regularization to further improve the representations. We conduct expensive experiments on multiple 3D/2D medical image analysis tasks, including segmentation and classification. The results show that the proposed UniMiSS+ achieves promising performance on various downstream tasks, not only outperforming the ImageNet pre-training and other advanced SSL counterparts substantially but also improving the predecessor UniMiSS pre-training. Code is available at: https://github.com/YtongXie/UniMiSS-code.

自监督学习(SSL)为医学图像分析带来了巨大的机遇,众所周知,医学图像缺乏注释。然而,由于成像成本高和隐私限制,聚合计算机断层扫描(CT)等海量(无标注)三维医学图像仍具有挑战性。在我们的试点研究中,我们主张利用胸部 X 光片等丰富的二维图像来弥补三维数据的不足,旨在建立一个通用的医学自监督表示学习框架,称为 UniMiSS。特别是,我们设计了一个金字塔 U 型医疗转换器(MiT)作为骨干,使 UniMiSS 可以同时使用二维和三维图像执行 SSL。因此,UniMiSS 的前身与目前的三维专用 SSL 相比有两个明显的优点:(1) 更有效--优于学习强表征,受益于更多、更多样化的数据;(2) 更通用--适用于各种下游任务,不受维度障碍的限制。遗憾的是,由于缺乏配对的多模态/多维度患者数据,UniMiSS 并未深入挖掘二维医学图像与三维体积之间的内在解剖关联性。在这篇扩展论文中,我们提出了 UniMiSS+,其中引入了数字重建射线照片(DRR)技术,从 CT 卷中模拟 X 射线图像,以获取成对的 CT 和 X 射线数据。得益于配对组,我们引入了额外的配对约束来增强跨模态相关性学习,这也可以作为一种跨维度正则化来进一步改进表征。我们在多个三维/二维医学图像分析任务上进行了昂贵的实验,包括分割和分类。结果表明,所提出的 UniMiSS+ 在各种下游任务上都取得了可喜的性能,不仅大大优于 ImageNet 预训练和其他先进的 SSL 对应算法,而且还改进了前身 UniMiSS 预训练。代码见:https://github.com/YtongXie/UniMiSS-code。
{"title":"UniMiSS+: Universal Medical Self-Supervised Learning From Cross-Dimensional Unpaired Data.","authors":"Yutong Xie, Jianpeng Zhang, Yong Xia, Qi Wu","doi":"10.1109/TPAMI.2024.3436105","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3436105","url":null,"abstract":"<p><p>Self-supervised learning (SSL) opens up huge opportunities for medical image analysis that is well known for its lack of annotations. However, aggregating massive (unlabeled) 3D medical images like computerized tomography (CT) remains challenging due to its high imaging cost and privacy restrictions. In our pilot study, we advocated bringing a wealth of 2D images like chest X-rays as compensation for the lack of 3D data, aiming to build a universal medical self-supervised representation learning framework, called UniMiSS. Especially, we designed a pyramid U- like medical Transformer (MiT) as the backbone to make UniMiSS possible to perform SSL with both 2D and 3D images. Consequently, the predecessor UniMiSS has two obvious merits compared to current 3D-specific SSL: (1) more effective - superior to learning strong representations, benefiting from more and diverse data; and (2) more versatile - suitable for various downstream tasks without the restriction on the dimensionality barrier. Unfortunately, UniMiSS did not dig deeply into the intrinsic anatomy correlation between 2D medical images and 3D volumes due to the lack of paired multi-modal/dimension patient data. In this extension paper, we propose the UniMiSS+, in which we introduce the digitally reconstructed radiographs (DRR) technology to simulate X-ray images from a CT volume to access paired CT and X-ray data. Benefiting from the paired group, we introduce an extra pair- wise constraint to boost the cross-modality correlation learning, which also can be adopted as a cross-dimension regularization to further improve the representations. We conduct expensive experiments on multiple 3D/2D medical image analysis tasks, including segmentation and classification. The results show that the proposed UniMiSS+ achieves promising performance on various downstream tasks, not only outperforming the ImageNet pre-training and other advanced SSL counterparts substantially but also improving the predecessor UniMiSS pre-training. Code is available at: https://github.com/YtongXie/UniMiSS-code.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141861980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Prompt Search. 神经提示搜索
Pub Date : 2024-07-30 DOI: 10.1109/TPAMI.2024.3435939
Yuanhan Zhang, Kaiyang Zhou, Ziwei Liu

The size of vision models has grown exponentially over the last few years, especially after the emergence of Vision Transformer. This has motivated the development of parameter-efficient tuning methods, such as learning adapter layers or visual prompt tokens, which allow a tiny portion of model parameters to be trained whereas the vast majority obtained from pre-training are frozen. However, designing a proper tuning method is non-trivial: one might need to try out a lengthy list of design choices, not to mention that each downstream dataset often requires custom designs. In this paper, we view the existing parameter-efficient tuning methods as "prompt modules" and propose Neural prOmpt seArcH (NOAH), a novel approach that learns, for large vision models, the optimal design of prompt modules through a neural architecture search algorithm, specifically for each downstream dataset. By conducting extensive experiments on over 20 vision datasets, we demonstrate that NOAH (i) is superior to individual prompt modules, (ii) has good few-shot learning ability, and (iii) is domain-generalizable. The code and models are available at https://github.com/ZhangYuanhan-AI/NOAH.

在过去几年中,视觉模型的规模呈指数级增长,尤其是在视觉转换器(Vision Transformer)出现之后。这推动了参数效率调整方法的发展,例如学习适配器层或视觉提示标记,这些方法允许对极小部分模型参数进行训练,而通过预训练获得的绝大部分参数则被冻结。然而,设计一种合适的调整方法并非易事:我们可能需要尝试一长串的设计选择,更不用说每个下游数据集通常都需要定制设计。在本文中,我们将现有的参数高效调整方法视为 "提示模块",并提出了神经提示模块(NOAH),这是一种新颖的方法,通过神经架构搜索算法学习大型视觉模型的最佳提示模块设计,特别适用于每个下游数据集。通过在 20 多个视觉数据集上进行广泛实验,我们证明了 NOAH (i) 优于单个提示模块,(ii) 具有良好的少量学习能力,(iii) 具有领域通用性。代码和模型可在 https://github.com/ZhangYuanhan-AI/NOAH 上获取。
{"title":"Neural Prompt Search.","authors":"Yuanhan Zhang, Kaiyang Zhou, Ziwei Liu","doi":"10.1109/TPAMI.2024.3435939","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3435939","url":null,"abstract":"<p><p>The size of vision models has grown exponentially over the last few years, especially after the emergence of Vision Transformer. This has motivated the development of parameter-efficient tuning methods, such as learning adapter layers or visual prompt tokens, which allow a tiny portion of model parameters to be trained whereas the vast majority obtained from pre-training are frozen. However, designing a proper tuning method is non-trivial: one might need to try out a lengthy list of design choices, not to mention that each downstream dataset often requires custom designs. In this paper, we view the existing parameter-efficient tuning methods as \"prompt modules\" and propose Neural prOmpt seArcH (NOAH), a novel approach that learns, for large vision models, the optimal design of prompt modules through a neural architecture search algorithm, specifically for each downstream dataset. By conducting extensive experiments on over 20 vision datasets, we demonstrate that NOAH (i) is superior to individual prompt modules, (ii) has good few-shot learning ability, and (iii) is domain-generalizable. The code and models are available at https://github.com/ZhangYuanhan-AI/NOAH.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141857453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Diffusion Model Translator for Efficient Image-to-Image Translation. 用于高效图像到图像翻译的扩散模型翻译器
Pub Date : 2024-07-30 DOI: 10.1109/TPAMI.2024.3435448
Mengfei Xia, Yu Zhou, Ran Yi, Yong-Jin Liu, Wenping Wang

Applying diffusion models to image-to-image translation (I2I) has recently received increasing attention due to its practical applications. Previous attempts inject information from the source image into each denoising step for an iterative refinement, thus resulting in a time-consuming implementation. We propose an efficient method that equips a diffusion model with a lightweight translator, dubbed a Diffusion Model Translator (DMT), to accomplish I2I. Specifically, we first offer theoretical justification that in employing the pioneering DDPM work for the I2I task, it is both feasible and sufficient to transfer the distribution from one domain to another only at some intermediate step. We further observe that the translation performance highly depends on the chosen timestep for domain transfer, and therefore propose a practical strategy to automatically select an appropriate timestep for a given task. We evaluate our approach on a range of I2I applications, including image stylization, image colorization, segmentation to image, and sketch to image, to validate its efficacy and general utility. The comparisons show that our DMT surpasses existing methods in both quality and efficiency. Code will be made publicly available.

将扩散模型应用于图像到图像的转换(I2I),因其实际应用而受到越来越多的关注。以往的尝试是在每个去噪步骤中注入源图像信息,进行迭代改进,因此实施起来非常耗时。我们提出了一种高效的方法,为扩散模型配备一个轻量级翻译器,称为扩散模型翻译器(DMT),以实现 I2I。具体来说,我们首先从理论上证明,在将开创性的 DDPM 工作用于 I2I 任务时,仅在某个中间步骤将分布从一个域转移到另一个域既可行又充分。我们进一步观察到,翻译性能在很大程度上取决于所选择的域转移时间步,因此我们提出了一种实用策略,可为给定任务自动选择合适的时间步。我们在一系列 I2I 应用中评估了我们的方法,包括图像风格化、图像着色、图像分割和图像素描,以验证其有效性和通用性。比较结果表明,我们的 DMT 在质量和效率上都超过了现有方法。代码将公开发布。
{"title":"A Diffusion Model Translator for Efficient Image-to-Image Translation.","authors":"Mengfei Xia, Yu Zhou, Ran Yi, Yong-Jin Liu, Wenping Wang","doi":"10.1109/TPAMI.2024.3435448","DOIUrl":"10.1109/TPAMI.2024.3435448","url":null,"abstract":"<p><p>Applying diffusion models to image-to-image translation (I2I) has recently received increasing attention due to its practical applications. Previous attempts inject information from the source image into each denoising step for an iterative refinement, thus resulting in a time-consuming implementation. We propose an efficient method that equips a diffusion model with a lightweight translator, dubbed a Diffusion Model Translator (DMT), to accomplish I2I. Specifically, we first offer theoretical justification that in employing the pioneering DDPM work for the I2I task, it is both feasible and sufficient to transfer the distribution from one domain to another only at some intermediate step. We further observe that the translation performance highly depends on the chosen timestep for domain transfer, and therefore propose a practical strategy to automatically select an appropriate timestep for a given task. We evaluate our approach on a range of I2I applications, including image stylization, image colorization, segmentation to image, and sketch to image, to validate its efficacy and general utility. The comparisons show that our DMT surpasses existing methods in both quality and efficiency. Code will be made publicly available.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141857450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MoIL: Momentum Imitation Learning for Efficient Vision-Language Adaptation. MoIL:高效视觉语言适应的动量模仿学习。
Pub Date : 2024-07-30 DOI: 10.1109/TPAMI.2024.3435790
Gen Luo, Yiyi Zhou, Minglang Huang, Tianhe Ren, Xiaoshuai Sun, Rongrong Ji

Pre-training and fine-tuning have been the de-facto paradigm in vision-language domains. Along with the rapid growth of model sizes, fully fine-tuning these large-scale vision-language pre-training (VLP) models requires prohibitively expensive storage costs. To address this issue, recent advances in NLP offer a promising and efficient adaptation approach called LoRA, which aims to approximate the fine-tuning of large pre-trained model by updating low-rank parameters. Despite its effectiveness, we identify that LoRA suffers a large approximation error on VLP models and its optimization is also inefficient, which greatly limits its performance upper bound. In this paper, we mathematically prove that the approximation error of low-rank adaptation can be optimized by a new optimization objective, i.e., the weight distance between LoRA and fine-tuning. Based on this finding, we propose a novel PETL method for VLP models, namely momentum imitation learning (MoIL). Specifically, MoIL formulates PETL as a weight imitation learning process and directly optimize the approximation error bound of the low-rank adaptation. Based on this training scheme, we also explore a new hybrid approximation function to reduce the learning difficulty of low-rank adaptations. With these two novel designs, MoIL can greatly improve the optimization efficiency of the low-rank parameters on VLP models. We validate MoIL on three VLP models ranging from end-to-end network to two-stage network, and conduct extensive experiments on four VL tasks. Experimental results demonstrate superior performance and optimization efficiency of MoIL than existing PETL methods. For instance, by updating only 6.23% parameters, MoIL can even outperform full tuning by +2.3% on image-text matching task. Meanwhile, its inference efficiency and generalization ability is also validated by multiple VLP models, e.g., VLMO and VinVL.

预训练和微调一直是视觉语言领域的事实范式。随着模型规模的快速增长,对这些大规模视觉语言预训练(VLP)模型进行完全微调需要高昂的存储成本,令人望而却步。为了解决这个问题,NLP 领域的最新进展提供了一种前景广阔的高效适应方法--LoRA,其目的是通过更新低等级参数来近似微调大型预训练模型。尽管这种方法很有效,但我们发现 LoRA 在 VLP 模型上存在很大的近似误差,而且其优化效率也很低,这大大限制了其性能上限。在本文中,我们用数学方法证明了低阶适应的近似误差可以通过一个新的优化目标来优化,即 LoRA 与微调之间的权重距离。基于这一发现,我们为 VLP 模型提出了一种新的 PETL 方法,即动量模仿学习(MoIL)。具体来说,MoIL 将 PETL 表述为权重模仿学习过程,并直接优化低阶适应的近似误差约束。在此训练方案的基础上,我们还探索了一种新的混合近似函数,以降低低阶适应的学习难度。通过这两种新颖的设计,MoIL 可以大大提高 VLP 模型低阶参数的优化效率。我们在从端到端网络到两阶段网络的三个 VLP 模型上验证了 MoIL,并在四个 VL 任务上进行了广泛的实验。实验结果表明,MoIL 的性能和优化效率优于现有的 PETL 方法。例如,只需更新 6.23% 的参数,MoIL 在图像-文本匹配任务中的性能甚至比完全调整方法高出 2.3%。同时,MoIL 的推理效率和泛化能力也得到了多个 VLP 模型(如 VLMO 和 VinVL)的验证。
{"title":"MoIL: Momentum Imitation Learning for Efficient Vision-Language Adaptation.","authors":"Gen Luo, Yiyi Zhou, Minglang Huang, Tianhe Ren, Xiaoshuai Sun, Rongrong Ji","doi":"10.1109/TPAMI.2024.3435790","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3435790","url":null,"abstract":"<p><p>Pre-training and fine-tuning have been the de-facto paradigm in vision-language domains. Along with the rapid growth of model sizes, fully fine-tuning these large-scale vision-language pre-training (VLP) models requires prohibitively expensive storage costs. To address this issue, recent advances in NLP offer a promising and efficient adaptation approach called LoRA, which aims to approximate the fine-tuning of large pre-trained model by updating low-rank parameters. Despite its effectiveness, we identify that LoRA suffers a large approximation error on VLP models and its optimization is also inefficient, which greatly limits its performance upper bound. In this paper, we mathematically prove that the approximation error of low-rank adaptation can be optimized by a new optimization objective, i.e., the weight distance between LoRA and fine-tuning. Based on this finding, we propose a novel PETL method for VLP models, namely momentum imitation learning (MoIL). Specifically, MoIL formulates PETL as a weight imitation learning process and directly optimize the approximation error bound of the low-rank adaptation. Based on this training scheme, we also explore a new hybrid approximation function to reduce the learning difficulty of low-rank adaptations. With these two novel designs, MoIL can greatly improve the optimization efficiency of the low-rank parameters on VLP models. We validate MoIL on three VLP models ranging from end-to-end network to two-stage network, and conduct extensive experiments on four VL tasks. Experimental results demonstrate superior performance and optimization efficiency of MoIL than existing PETL methods. For instance, by updating only 6.23% parameters, MoIL can even outperform full tuning by +2.3% on image-text matching task. Meanwhile, its inference efficiency and generalization ability is also validated by multiple VLP models, e.g., VLMO and VinVL.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141857452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1