首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
TCFormer: Visual Recognition via Token Clustering Transformer. TCFormer:通过令牌聚类进行视觉识别变换器
Pub Date : 2024-07-11 DOI: 10.1109/TPAMI.2024.3425768
Wang Zeng, Sheng Jin, Lumin Xu, Wentao Liu, Chen Qian, Wanli Ouyang, Ping Luo, Xiaogang Wang

Transformers are widely used in computer vision areas and have achieved remarkable success. Most state-of-the-art approaches split images into regular grids and represent each grid region with a vision token. However, fixed token distribution disregards the semantic meaning of different image regions, resulting in sub-optimal performance. To address this issue, we propose the Token Clustering Transformer (TCFormer), which generates dynamic vision tokens based on semantic meaning. Our dynamic tokens possess two crucial characteristics: (1) Representing image regions with similar semantic meanings using the same vision token, even if those regions are not adjacent, and (2) concentrating on regions with valuable details and represent them using fine tokens. Through extensive experimentation across various applications, including image classification, human pose estimation, semantic segmentation, and object detection, we demonstrate the effectiveness of our TCFormer. The code and models for this work are available at https://github.com/zengwang430521/TCFormer.

变换器被广泛应用于计算机视觉领域,并取得了显著的成就。大多数最先进的方法都是将图像分割成规则的网格,并用视觉标记来表示每个网格区域。然而,固定的标记分布忽略了不同图像区域的语义,导致性能未达到最佳。为了解决这个问题,我们提出了标记聚类转换器(TCFormer),它能根据语义生成动态视觉标记。我们的动态标记具有两个关键特征:(1) 使用相同的视觉标记来表示具有相似语义的图像区域,即使这些区域并不相邻;(2) 专注于具有有价值细节的区域,并使用精细标记来表示它们。通过对图像分类、人体姿态估计、语义分割和物体检测等各种应用的广泛实验,我们证明了 TCFormer 的有效性。这项工作的代码和模型可在 https://github.com/zengwang430521/TCFormer 上获取。
{"title":"TCFormer: Visual Recognition via Token Clustering Transformer.","authors":"Wang Zeng, Sheng Jin, Lumin Xu, Wentao Liu, Chen Qian, Wanli Ouyang, Ping Luo, Xiaogang Wang","doi":"10.1109/TPAMI.2024.3425768","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3425768","url":null,"abstract":"<p><p>Transformers are widely used in computer vision areas and have achieved remarkable success. Most state-of-the-art approaches split images into regular grids and represent each grid region with a vision token. However, fixed token distribution disregards the semantic meaning of different image regions, resulting in sub-optimal performance. To address this issue, we propose the Token Clustering Transformer (TCFormer), which generates dynamic vision tokens based on semantic meaning. Our dynamic tokens possess two crucial characteristics: (1) Representing image regions with similar semantic meanings using the same vision token, even if those regions are not adjacent, and (2) concentrating on regions with valuable details and represent them using fine tokens. Through extensive experimentation across various applications, including image classification, human pose estimation, semantic segmentation, and object detection, we demonstrate the effectiveness of our TCFormer. The code and models for this work are available at https://github.com/zengwang430521/TCFormer.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141592376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The effects of experiment duration and supertrial analysis on EEG classification methods. 实验持续时间和超时空分析对脑电图分类方法的影响。
Pub Date : 2024-07-10 DOI: 10.1109/TPAMI.2024.3426296
Simone Palazzo, Concetto Spampinato, Isaak Kavasidis, Daniela Giordano, Joseph Schmidt, Mubarak Shah

Bharadwaj et al. [1] present a comments paper evaluating the classification accuracy of several state-of-the-art methods using EEG data averaged over random class samples. According to the results, some of the methods achieve above-chance accuracy, while the method proposed in [2], that is the target of their analysis, does not. In this rebuttal, we address these claims and explain why they are not grounded in the cognitive neuroscience literature, and why the evaluation procedure is ineffective and unfair.

Bharadwaj 等人[1]发表了一篇评论文章,利用随机类样本的平均脑电图数据评估了几种最先进方法的分类准确性。结果显示,其中一些方法达到了高于概率的准确度,而他们分析的目标--[2] 中提出的方法却没有达到。在这篇反驳文章中,我们将针对这些说法,解释为什么它们在认知神经科学文献中没有依据,以及为什么评估程序是无效和不公平的。
{"title":"The effects of experiment duration and supertrial analysis on EEG classification methods.","authors":"Simone Palazzo, Concetto Spampinato, Isaak Kavasidis, Daniela Giordano, Joseph Schmidt, Mubarak Shah","doi":"10.1109/TPAMI.2024.3426296","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3426296","url":null,"abstract":"<p><p>Bharadwaj et al. [1] present a comments paper evaluating the classification accuracy of several state-of-the-art methods using EEG data averaged over random class samples. According to the results, some of the methods achieve above-chance accuracy, while the method proposed in [2], that is the target of their analysis, does not. In this rebuttal, we address these claims and explain why they are not grounded in the cognitive neuroscience literature, and why the evaluation procedure is ineffective and unfair.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141581864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ask Questions with Double Hints: Visual Question Generation with Answer-awareness and Region-reference. 用双重提示提问:具有答案感知和区域参考功能的可视化问题生成。
Pub Date : 2024-07-09 DOI: 10.1109/TPAMI.2024.3425222
Kai Shen, Lingfei Wu, Siliang Tang, Fangli Xu, Bo Long, Yueting Zhuang, Jian Pei

The visual question generation (VQG) task aims to generate human-like questions from an image and potentially other side information (e.g. answer type). Previous works on VQG fall in two aspects: i) They suffer from one image to many questions mapping problem, which leads to the failure of generating referential and meaningful questions from an image. ii) They fail to model complex implicit relations among the visual objects in an image and also overlook potential interactions between the side information and image. To address these limitations, we first propose a novel learning paradigm to generate visual questions with answer-awareness and region-reference. Concretely, we aim to ask the right visual questions with Double Hints - textual answers and visual regions of interests, which could effectively mitigate the existing one-to-many mapping issue. Particularly, we develop a simple methodology to self-learn the visual hints without introducing any additional human annotations. Furthermore, to capture these sophisticated relationships, we propose a new double-hints guided Graph-to-Sequence learning framework, which first models them as a dynamic graph and learns the implicit topology end-to-end, and then utilizes a graph-to-sequence model to generate the questions with double hints. Experimental results demonstrate the priority of our proposed method.

视觉问题生成(VQG)任务旨在从图像和潜在的其他侧面信息(如答案类型)中生成类似人类的问题。以往的视觉问题生成工作存在两个方面的问题:i) 它们存在一图多问的映射问题,导致无法从图像中生成有参考价值和意义的问题;ii) 它们未能模拟图像中视觉对象之间复杂的隐含关系,也忽略了侧面信息与图像之间潜在的交互作用。为了解决这些局限性,我们首先提出了一种新颖的学习范式,以生成具有答案感知和区域参照功能的视觉问题。具体来说,我们的目标是通过双重提示(文本答案和感兴趣的视觉区域)提出正确的视觉问题,从而有效缓解现有的一对多映射问题。特别是,我们开发了一种简单的方法来自我学习视觉提示,而无需引入任何额外的人工注释。此外,为了捕捉这些复杂的关系,我们提出了一种新的双重提示引导的图到序列学习框架,该框架首先将它们建模为动态图,并端到端学习隐含拓扑结构,然后利用图到序列模型生成带有双重提示的问题。实验结果证明了我们提出的方法的优先性。
{"title":"Ask Questions with Double Hints: Visual Question Generation with Answer-awareness and Region-reference.","authors":"Kai Shen, Lingfei Wu, Siliang Tang, Fangli Xu, Bo Long, Yueting Zhuang, Jian Pei","doi":"10.1109/TPAMI.2024.3425222","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3425222","url":null,"abstract":"<p><p>The visual question generation (VQG) task aims to generate human-like questions from an image and potentially other side information (e.g. answer type). Previous works on VQG fall in two aspects: i) They suffer from one image to many questions mapping problem, which leads to the failure of generating referential and meaningful questions from an image. ii) They fail to model complex implicit relations among the visual objects in an image and also overlook potential interactions between the side information and image. To address these limitations, we first propose a novel learning paradigm to generate visual questions with answer-awareness and region-reference. Concretely, we aim to ask the right visual questions with Double Hints - textual answers and visual regions of interests, which could effectively mitigate the existing one-to-many mapping issue. Particularly, we develop a simple methodology to self-learn the visual hints without introducing any additional human annotations. Furthermore, to capture these sophisticated relationships, we propose a new double-hints guided Graph-to-Sequence learning framework, which first models them as a dynamic graph and learns the implicit topology end-to-end, and then utilizes a graph-to-sequence model to generate the questions with double hints. Experimental results demonstrate the priority of our proposed method.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141565414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-Fidelity and Efficient Pluralistic Image Completion with Transformers. 利用变换器实现高保真、高效的多元图像补全。
Pub Date : 2024-07-09 DOI: 10.1109/TPAMI.2024.3424835
Ziyu Wan, Jingbo Zhang, Dongdong Chen, Jing Liao

Image completion has made tremendous progress with convolutional neural networks (CNNs), because of their powerful texture modeling capacity. However, due to some inherent properties (e.g., local inductive prior, spatial-invariant kernels), CNNs do not perform well in understanding global structures or naturally support pluralistic completion. Recently, transformers demonstrate their power in modeling the long-term relationship and generating diverse results, but their computation complexity is quadratic to input length, thus hampering the application in processing high-resolution images. This paper brings the best of both worlds to pluralistic image completion: appearance prior reconstruction with transformer and texture replenishment with CNN. The former transformer recovers pluralistic coherent structures together with some coarse textures, while the latter CNN enhances the local texture details of coarse priors guided by the high-resolution masked images. To decode diversified outputs from transformers, auto-regressive sampling is the most common method, but with extremely low efficiency. We further overcome this issue by proposing a new decoding strategy, temperature annealing probabilistic sampling (TAPS), which firstly achieves more than 70× speedup of inference at most, meanwhile maintaining the high quality and diversity of the sampled global structures. Moreover, we find the full CNN architecture will lead to suboptimal solutions for guided upsampling. To render more realistic and coherent contents, we design a novel module, named texture-aware guided attention, to concurrently consider the procedures of texture copy and generation, meanwhile raising several important modifications to solve the boundary artifacts. Through dense experiments, we found the proposed method vastly outperforms state-of-the-art methods in terms of four aspects: 1) large performance boost on image fidelity even compared to deterministic completion methods; 2) better diversity and higher fidelity for pluralistic completion; 3) exceptional generalization ability on large masks and generic dataset, like ImageNet. 4) Much higher decoding efficiency over previous auto-regressive based methods.

卷积神经网络(CNN)具有强大的纹理建模能力,在图像补全方面取得了巨大进步。然而,由于一些固有特性(如局部归纳先验、空间不变核),卷积神经网络在理解全局结构方面表现不佳,也不能自然地支持多元补全。最近,变换器展示了其在长期关系建模和生成多样化结果方面的能力,但其计算复杂度是输入长度的二次方,因此阻碍了其在处理高分辨率图像方面的应用。本文为多元图像补全带来了两全其美的方法:利用变换器进行外观先验重建,利用 CNN 进行纹理补全。前者利用变换器恢复多元相干结构和一些粗纹理,后者利用 CNN 在高分辨率遮蔽图像的引导下增强粗先验的局部纹理细节。要解码变换器的多样化输出,自动回归采样是最常用的方法,但效率极低。我们进一步克服了这一问题,提出了一种新的解码策略--温度退火概率采样(TAPS),它首先使推理速度最多提高了 70 倍以上,同时保持了采样全局结构的高质量和多样性。此外,我们还发现全 CNN 架构会导致引导上采样的次优解。为了呈现更真实、更连贯的内容,我们设计了一个名为 "纹理感知引导注意力 "的新模块,同时考虑纹理复制和生成的过程,并提出了几个重要的修改来解决边界伪影问题。通过密集的实验,我们发现所提出的方法在四个方面大大优于最先进的方法:1) 与确定性补全方法相比,在图像保真度方面有很大的性能提升;2) 多元补全方法有更好的多样性和更高的保真度;3) 在大型掩码和通用数据集(如 ImageNet)上有卓越的泛化能力。4) 解码效率远高于之前基于自动回归的方法。
{"title":"High-Fidelity and Efficient Pluralistic Image Completion with Transformers.","authors":"Ziyu Wan, Jingbo Zhang, Dongdong Chen, Jing Liao","doi":"10.1109/TPAMI.2024.3424835","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3424835","url":null,"abstract":"<p><p>Image completion has made tremendous progress with convolutional neural networks (CNNs), because of their powerful texture modeling capacity. However, due to some inherent properties (e.g., local inductive prior, spatial-invariant kernels), CNNs do not perform well in understanding global structures or naturally support pluralistic completion. Recently, transformers demonstrate their power in modeling the long-term relationship and generating diverse results, but their computation complexity is quadratic to input length, thus hampering the application in processing high-resolution images. This paper brings the best of both worlds to pluralistic image completion: appearance prior reconstruction with transformer and texture replenishment with CNN. The former transformer recovers pluralistic coherent structures together with some coarse textures, while the latter CNN enhances the local texture details of coarse priors guided by the high-resolution masked images. To decode diversified outputs from transformers, auto-regressive sampling is the most common method, but with extremely low efficiency. We further overcome this issue by proposing a new decoding strategy, temperature annealing probabilistic sampling (TAPS), which firstly achieves more than 70× speedup of inference at most, meanwhile maintaining the high quality and diversity of the sampled global structures. Moreover, we find the full CNN architecture will lead to suboptimal solutions for guided upsampling. To render more realistic and coherent contents, we design a novel module, named texture-aware guided attention, to concurrently consider the procedures of texture copy and generation, meanwhile raising several important modifications to solve the boundary artifacts. Through dense experiments, we found the proposed method vastly outperforms state-of-the-art methods in terms of four aspects: 1) large performance boost on image fidelity even compared to deterministic completion methods; 2) better diversity and higher fidelity for pluralistic completion; 3) exceptional generalization ability on large masks and generic dataset, like ImageNet. 4) Much higher decoding efficiency over previous auto-regressive based methods.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141565415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-serial Quantization-aware Deep Optics for Snapshot Hyperspectral Imaging. 用于快照高光谱成像的非序列量化感知深度光学。
Pub Date : 2024-07-09 DOI: 10.1109/TPAMI.2024.3425512
Lizhi Wang, Lingen Li, Weitao Song, Lei Zhang, Zhiwei Xiong, Hua Huang

Deep optics has been endeavoring to capture hyperspectral images of dynamic scenes, where the optical encoder plays an essential role in deciding the imaging performance. Our key insight is that the optical encoder of a deep optics system is expected to keep fabrication-friendliness and decoder-friendliness, to be faithfully realized in the implementation phase and fully interacted with the decoder in the design phase, respectively. In this paper, we propose the non-serial quantization-aware deep optics (NSQDO), which consists of the fabrication-friendly quantization-aware model (QAM) and the decoder-friendly non-serial manner (NSM). The QAM integrates the quantization process into the optimization and adaptively adjusts the physical height of each quantization level, reducing the deviation of the physical encoder from the numerical simulation through the awareness of and adaptation to the quantization operation of the DOE physical structure. The NSM bridges the encoder and the decoder with full interaction through bidirectional hint connections and flexibilize the connections with a gating mechanism, boosting the power of joint optimization in deep optics. The proposed NSQDO improves the fabrication-friendliness and decoder-friendliness of the encoder and develops the deep optics framework to be more practical and powerful. Extensive synthetic simulation and real hardware experiments demonstrate the superior performance of the proposed method.

深度光学一直致力于捕捉动态场景的高光谱图像,其中光学编码器在决定成像性能方面起着至关重要的作用。我们的主要观点是,深度光学系统的光编码器应保持制造友好性和解码友好性,分别在实现阶段和设计阶段与解码器充分互动。本文提出了非串行量化感知深度光学(NSQDO),它由便于制造的量化感知模型(QAM)和便于解码器的非串行方式(NSM)组成。QAM 将量化过程集成到优化中,并自适应地调整每个量化级的物理高度,通过感知和适应 DOE 物理结构的量化操作,减少物理编码器与数值模拟的偏差。NSM 通过双向提示连接将编码器和解码器连接起来,实现充分互动,并通过门控机制灵活连接,增强了深度光学中的联合优化能力。所提出的 NSQDO 改善了编码器的制造友好性和解码器友好性,使深度光学框架更加实用和强大。广泛的合成仿真和实际硬件实验证明了所提方法的优越性能。
{"title":"Non-serial Quantization-aware Deep Optics for Snapshot Hyperspectral Imaging.","authors":"Lizhi Wang, Lingen Li, Weitao Song, Lei Zhang, Zhiwei Xiong, Hua Huang","doi":"10.1109/TPAMI.2024.3425512","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3425512","url":null,"abstract":"<p><p>Deep optics has been endeavoring to capture hyperspectral images of dynamic scenes, where the optical encoder plays an essential role in deciding the imaging performance. Our key insight is that the optical encoder of a deep optics system is expected to keep fabrication-friendliness and decoder-friendliness, to be faithfully realized in the implementation phase and fully interacted with the decoder in the design phase, respectively. In this paper, we propose the non-serial quantization-aware deep optics (NSQDO), which consists of the fabrication-friendly quantization-aware model (QAM) and the decoder-friendly non-serial manner (NSM). The QAM integrates the quantization process into the optimization and adaptively adjusts the physical height of each quantization level, reducing the deviation of the physical encoder from the numerical simulation through the awareness of and adaptation to the quantization operation of the DOE physical structure. The NSM bridges the encoder and the decoder with full interaction through bidirectional hint connections and flexibilize the connections with a gating mechanism, boosting the power of joint optimization in deep optics. The proposed NSQDO improves the fabrication-friendliness and decoder-friendliness of the encoder and develops the deep optics framework to be more practical and powerful. Extensive synthetic simulation and real hardware experiments demonstrate the superior performance of the proposed method.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141565416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D Reconstruction from a Single Sketch via View-dependent Depth Sampling. 通过视图深度采样从单张草图进行三维重建
Pub Date : 2024-07-08 DOI: 10.1109/TPAMI.2024.3424404
Chenjian Gao, Xilin Wang, Qian Yu, Lu Sheng, Jing Zhang, Xiaoguang Han, Yi-Zhe Song, Dong Xu

Reconstructing a 3D shape based on a single sketch image is challenging due to the inherent sparsity and ambiguity present in sketches. Existing methods lose fine details when extracting features to predict 3D objects from sketches. Upon analyzing the 3D-to-2D projection process, we observe that the density map, characterizing the distribution of 2D point clouds, can serve as a proxy to facilitate the reconstruction process. In this work, we propose a novel sketch-based 3D reconstruction model named SketchSampler. It initiates the process by translating a sketch through an image translation network into a more informative 2D representation, which is then used to generate a density map. Subsequently, a two-stage probabilistic sampling process is employed to reconstruct a 3D point cloud: firstly, recovering the 2D points (i.e., the x and y coordinates) by sampling the density map; and secondly, predicting the depth (i.e., the z coordinate) by sampling the depth values along the ray determined by each 2D point. Additionally, we convert the reconstructed point cloud into a 3D mesh for wider applications. To reduce ambiguity, we incorporate hidden lines in sketches. Experimental results demonstrate that our proposed approach significantly outperforms other baseline methods.

由于草图固有的稀疏性和模糊性,根据单张草图图像重建三维形状具有挑战性。现有方法在从草图中提取特征来预测三维物体时会丢失一些细节。在分析三维到二维的投影过程时,我们观察到,表征二维点云分布的密度图可以作为一个代理来促进重建过程。在这项工作中,我们提出了一种名为 SketchSampler 的基于草图的新型三维重建模型。该模型通过图像转换网络将草图转换为信息量更大的二维表示,然后生成密度图。随后,采用两阶段概率采样过程来重建三维点云:首先,通过对密度图进行采样来恢复二维点(即 x 和 y 坐标);其次,通过对每个二维点确定的射线沿线的深度值进行采样来预测深度(即 z 坐标)。此外,我们还将重建的点云转换为三维网格,以便进行更广泛的应用。为了减少模糊性,我们在草图中加入了隐藏线。实验结果表明,我们提出的方法明显优于其他基准方法。
{"title":"3D Reconstruction from a Single Sketch via View-dependent Depth Sampling.","authors":"Chenjian Gao, Xilin Wang, Qian Yu, Lu Sheng, Jing Zhang, Xiaoguang Han, Yi-Zhe Song, Dong Xu","doi":"10.1109/TPAMI.2024.3424404","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3424404","url":null,"abstract":"<p><p>Reconstructing a 3D shape based on a single sketch image is challenging due to the inherent sparsity and ambiguity present in sketches. Existing methods lose fine details when extracting features to predict 3D objects from sketches. Upon analyzing the 3D-to-2D projection process, we observe that the density map, characterizing the distribution of 2D point clouds, can serve as a proxy to facilitate the reconstruction process. In this work, we propose a novel sketch-based 3D reconstruction model named SketchSampler. It initiates the process by translating a sketch through an image translation network into a more informative 2D representation, which is then used to generate a density map. Subsequently, a two-stage probabilistic sampling process is employed to reconstruct a 3D point cloud: firstly, recovering the 2D points (i.e., the x and y coordinates) by sampling the density map; and secondly, predicting the depth (i.e., the z coordinate) by sampling the depth values along the ray determined by each 2D point. Additionally, we convert the reconstructed point cloud into a 3D mesh for wider applications. To reduce ambiguity, we incorporate hidden lines in sketches. Experimental results demonstrate that our proposed approach significantly outperforms other baseline methods.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141560602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Instance Consistency Regularization for Semi-Supervised 3D Instance Segmentation. 半监督三维实例分割的实例一致性正则化
Pub Date : 2024-07-05 DOI: 10.1109/TPAMI.2024.3424243
Yizheng Wu, Zhiyu Pan, Kewei Wang, Xingyi Li, Jiahao Cui, Liwen Xiao, Guosheng Lin, Zhiguo Cao

Large-scale datasets with point-wise semantic and instance labels are crucial to 3D instance segmentation but also expensive. To leverage unlabeled data, previous semi-supervised 3D instance segmentation approaches have explored self-training frameworks, which rely on high-quality pseudo labels for consistency regularization. They intuitively utilize both instance and semantic pseudo labels in a joint learning manner. However, semantic pseudo labels contain numerous noise derived from the imbalanced category distribution and natural confusion of similar but distinct categories, which leads to severe collapses in self-training. Motivated by the observation that 3D instances are non-overlapping and spatially separable, we ask whether we can solely rely on instance consistency regularization for improved semi-supervised segmentation. To this end, we propose a novel self-training network InsTeacher3D to explore and exploit pure instance knowledge from unlabeled data. We first build a parallel base 3D instance segmentation model DKNet, which distinguishes each instance from the others via discriminative instance kernels without reliance on semantic segmentation. Based on DKNet, we further design a novel instance consistency regularization framework to generate and leverage high-quality instance pseudo labels. Experimental results on multiple large-scale datasets show that the InsTeacher3D significantly outperforms prior state-of-the-art semi-supervised approaches.

带有点式语义和实例标签的大规模数据集对三维实例分割至关重要,但成本也很高。为了利用无标注数据,以前的半监督三维实例分割方法探索了自我训练框架,该框架依赖高质量伪标签进行一致性正则化。它们以联合学习的方式直观地利用实例和语义伪标签。然而,语义伪标签包含大量噪声,这些噪声来自不平衡的类别分布和相似但不同类别的自然混淆,从而导致自我训练的严重崩溃。我们观察到三维实例是不重叠的,并且在空间上是可分离的,受此启发,我们提出了一个问题:我们是否可以仅仅依靠实例一致性正则化来改进半监督分割。为此,我们提出了一种新颖的自我训练网络 InsTeacher3D,以探索和利用来自无标记数据的纯实例知识。我们首先建立了一个并行基础三维实例分割模型 DKNet,该模型通过判别实例核将每个实例与其他实例区分开来,而无需依赖语义分割。在 DKNet 的基础上,我们进一步设计了一个新颖的实例一致性正则化框架,以生成和利用高质量的实例伪标签。在多个大规模数据集上的实验结果表明,InsTeacher3D 的性能明显优于之前最先进的半监督方法。
{"title":"Instance Consistency Regularization for Semi-Supervised 3D Instance Segmentation.","authors":"Yizheng Wu, Zhiyu Pan, Kewei Wang, Xingyi Li, Jiahao Cui, Liwen Xiao, Guosheng Lin, Zhiguo Cao","doi":"10.1109/TPAMI.2024.3424243","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3424243","url":null,"abstract":"<p><p>Large-scale datasets with point-wise semantic and instance labels are crucial to 3D instance segmentation but also expensive. To leverage unlabeled data, previous semi-supervised 3D instance segmentation approaches have explored self-training frameworks, which rely on high-quality pseudo labels for consistency regularization. They intuitively utilize both instance and semantic pseudo labels in a joint learning manner. However, semantic pseudo labels contain numerous noise derived from the imbalanced category distribution and natural confusion of similar but distinct categories, which leads to severe collapses in self-training. Motivated by the observation that 3D instances are non-overlapping and spatially separable, we ask whether we can solely rely on instance consistency regularization for improved semi-supervised segmentation. To this end, we propose a novel self-training network InsTeacher3D to explore and exploit pure instance knowledge from unlabeled data. We first build a parallel base 3D instance segmentation model DKNet, which distinguishes each instance from the others via discriminative instance kernels without reliance on semantic segmentation. Based on DKNet, we further design a novel instance consistency regularization framework to generate and leverage high-quality instance pseudo labels. Experimental results on multiple large-scale datasets show that the InsTeacher3D significantly outperforms prior state-of-the-art semi-supervised approaches.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141539128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models. Adan:更快优化深度模型的自适应内斯特罗夫动量算法
Pub Date : 2024-07-04 DOI: 10.1109/TPAMI.2024.3423382
Xingyu Xie, Pan Zhou, Huan Li, Zhouchen Lin, Shuicheng Yan

In deep learning, different kinds of deep networks typically need different optimizers, which have to be chosen after multiple trials, making the training process inefficient. To relieve this issue and consistently improve the model training speed across deep networks, we propose the ADAptive Nesterov momentum algorithm, Adan for short. Adan first reformulates the vanilla Nesterov acceleration to develop a new Nesterov momentum estimation (NME) method, which avoids the extra overhead of computing gradient at the extrapolation point. Then Adan adopts NME to estimate the gradient's first- and second-order moments in adaptive gradient algorithms for convergence acceleration. Besides, we prove that Adan finds an ϵ-approximate first-order stationary point within O(ϵ-3.5) stochastic gradient complexity on the non-convex stochastic problems (e.g.deep learning problems), matching the best-known lower bound. Extensive experimental results show that Adan consistently surpasses the corresponding SoTA optimizers on vision, language, and RL tasks and sets new SoTAs for many popular networks and frameworks, eg ResNet, ConvNext, ViT, Swin, MAE, DETR, GPT-2, Transformer-XL, and BERT. More surprisingly, Adan can use half of the training cost (epochs) of SoTA optimizers to achieve higher or comparable performance on ViT, GPT-2, MAE, etc, and also shows great tolerance to a large range of minibatch size, e.g.from 1k to 32k. Code is released at https://github.com/sail-sg/Adan, and has been used in multiple popular deep learning frameworks or projects.

在深度学习中,不同类型的深度网络通常需要不同的优化器,而这些优化器必须在多次试验后才能选择,这使得训练过程效率低下。为了缓解这一问题,并持续提高各种深度网络的模型训练速度,我们提出了 ADAptive Nesterov 动量算法,简称 Adan。Adan 首先对 vanilla 内斯特罗夫加速算法进行了重构,开发出一种新的内斯特罗夫动量估计(NME)方法,避免了在外推法点计算梯度的额外开销。然后,Adan 在自适应梯度算法中采用 NME 估算梯度的一阶和二阶矩,以加速收敛。此外,我们还证明了 Adan 在非凸随机问题(如深度学习问题)上能在 O(ϵ-3.5) 随机梯度复杂度内找到一个近似的一阶静止点,与最著名的下限相匹配。广泛的实验结果表明,在视觉、语言和 RL 任务上,Adan 始终超越了相应的 SoTA 优化器,并为 ResNet、ConvNext、ViT、Swin、MAE、DETR、GPT-2、Transformer-XL 和 BERT 等许多流行网络和框架设定了新的 SoTA。更令人惊讶的是,Adan 可以使用 SoTA 优化器一半的训练成本(epochs),在 ViT、GPT-2、MAE 等系统上实现更高或相当的性能,而且还显示出对大量迷你批大小(如从 1k 到 32k)的极大耐受性。代码发布于 https://github.com/sail-sg/Adan,已在多个流行的深度学习框架或项目中使用。
{"title":"Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models.","authors":"Xingyu Xie, Pan Zhou, Huan Li, Zhouchen Lin, Shuicheng Yan","doi":"10.1109/TPAMI.2024.3423382","DOIUrl":"10.1109/TPAMI.2024.3423382","url":null,"abstract":"<p><p>In deep learning, different kinds of deep networks typically need different optimizers, which have to be chosen after multiple trials, making the training process inefficient. To relieve this issue and consistently improve the model training speed across deep networks, we propose the ADAptive Nesterov momentum algorithm, Adan for short. Adan first reformulates the vanilla Nesterov acceleration to develop a new Nesterov momentum estimation (NME) method, which avoids the extra overhead of computing gradient at the extrapolation point. Then Adan adopts NME to estimate the gradient's first- and second-order moments in adaptive gradient algorithms for convergence acceleration. Besides, we prove that Adan finds an ϵ-approximate first-order stationary point within O(ϵ<sup>-3.5</sup>) stochastic gradient complexity on the non-convex stochastic problems (e.g.deep learning problems), matching the best-known lower bound. Extensive experimental results show that Adan consistently surpasses the corresponding SoTA optimizers on vision, language, and RL tasks and sets new SoTAs for many popular networks and frameworks, eg ResNet, ConvNext, ViT, Swin, MAE, DETR, GPT-2, Transformer-XL, and BERT. More surprisingly, Adan can use half of the training cost (epochs) of SoTA optimizers to achieve higher or comparable performance on ViT, GPT-2, MAE, etc, and also shows great tolerance to a large range of minibatch size, e.g.from 1k to 32k. Code is released at https://github.com/sail-sg/Adan, and has been used in multiple popular deep learning frameworks or projects.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141536217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D-PSSIM: Projective Structural Similarity for 3D Mesh Quality Assessment Robust to Topological Irregularities. 3D-PSSIM:针对拓扑不规则性的三维网格质量评估投影结构相似性。
Pub Date : 2024-07-03 DOI: 10.1109/TPAMI.2024.3422490
Seongmin Lee, Jiwoo Kang, Sanghoon Lee, Weisi Lin, Alan Conrad Bovik

Despite acceleration in the use of 3D meshes, it is difficult to find effective mesh quality assessment algorithms that can produce predictions highly correlated with human subjective opinions. Defining mesh quality features is challenging due to the irregular topology of meshes, which are defined on vertices and triangles. To address this, we propose a novel 3D projective structural similarity index ( 3D- PSSIM) for meshes that is robust to differences in mesh topology. We address topological differences between meshes by introducing multi-view and multi-layer projections that can densely represent the mesh textures and geometrical shapes irrespective of mesh topology. It also addresses occlusion problems that occur during projection. We propose visual sensitivity weights that capture the perceptual sensitivity to the degree of mesh surface curvature. 3D- PSSIM computes perceptual quality predictions by aggregating quality-aware features that are computed in multiple projective spaces onto the mesh domain, rather than on 2D spaces. This allows 3D- PSSIM to determine which parts of a mesh surface are distorted by geometric or color impairments. Experimental results show that 3D- PSSIM can predict mesh quality with high correlation against human subjective judgments, across the presence of noise, even when there are large topological differences, outperforming existing mesh quality assessment models.

尽管三维网格的使用在加速,但很难找到有效的网格质量评估算法,以产生与人类主观意见高度相关的预测结果。由于网格的拓扑结构不规则,且网格是根据顶点和三角形定义的,因此定义网格质量特征具有挑战性。为此,我们提出了一种新颖的网格三维投影结构相似性指数(3D- PSSIM),它对网格拓扑结构的差异具有鲁棒性。我们通过引入多视角和多层投影来解决网格之间的拓扑差异,这种投影可以密集地表示网格纹理和几何形状,而与网格拓扑无关。它还能解决投影过程中出现的遮挡问题。我们提出了视觉灵敏度权重,以捕捉对网格表面曲率程度的感知灵敏度。3D- PSSIM 通过将在多个投影空间中计算出的质量感知特征聚合到网格域而不是二维空间中,来计算感知质量预测。这样,3D- PSSIM 就能确定网格表面的哪些部分因几何或色彩缺陷而失真。实验结果表明,即使存在较大的拓扑差异,3D- PSSIM 也能预测出与人类主观判断高度相关的网格质量,超越了现有的网格质量评估模型。
{"title":"3D-PSSIM: Projective Structural Similarity for 3D Mesh Quality Assessment Robust to Topological Irregularities.","authors":"Seongmin Lee, Jiwoo Kang, Sanghoon Lee, Weisi Lin, Alan Conrad Bovik","doi":"10.1109/TPAMI.2024.3422490","DOIUrl":"https://doi.org/10.1109/TPAMI.2024.3422490","url":null,"abstract":"<p><p>Despite acceleration in the use of 3D meshes, it is difficult to find effective mesh quality assessment algorithms that can produce predictions highly correlated with human subjective opinions. Defining mesh quality features is challenging due to the irregular topology of meshes, which are defined on vertices and triangles. To address this, we propose a novel 3D projective structural similarity index ( 3D- PSSIM) for meshes that is robust to differences in mesh topology. We address topological differences between meshes by introducing multi-view and multi-layer projections that can densely represent the mesh textures and geometrical shapes irrespective of mesh topology. It also addresses occlusion problems that occur during projection. We propose visual sensitivity weights that capture the perceptual sensitivity to the degree of mesh surface curvature. 3D- PSSIM computes perceptual quality predictions by aggregating quality-aware features that are computed in multiple projective spaces onto the mesh domain, rather than on 2D spaces. This allows 3D- PSSIM to determine which parts of a mesh surface are distorted by geometric or color impairments. Experimental results show that 3D- PSSIM can predict mesh quality with high correlation against human subjective judgments, across the presence of noise, even when there are large topological differences, outperforming existing mesh quality assessment models.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141499993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial Steerability of GANs via Self-Supervision from Discriminator. 通过鉴别器的自我监督实现 GAN 的空间可控性
Pub Date : 2024-07-03 DOI: 10.1109/TPAMI.2024.3422820
Jianyuan Wang, Lalit Bhagat, Ceyuan Yang, Yinghao Xu, Yujun Shen, Hongdong Li, Bolei Zhou

Generative models make huge progress to the photorealistic image synthesis in recent years. To enable human to steer the image generation process and customize the output, many works explore the interpretable dimensions of the latent space in GANs. Existing methods edit the attributes of the output image such as orientation or color scheme by varying the latent code along certain directions. However, these methods usually require additional human annotations for each pretrained model, and they mostly focus on editing global attributes. In this work, we propose a self-supervised approach to improve the spatial steerability of GANs without searching for steerable directions in the latent space or requiring extra annotations. Specifically, we design randomly sampled Gaussian heatmaps to be encoded into the intermediate layers of generative models as spatial inductive bias. Along with training the GAN model from scratch, these heatmaps are being aligned with the emerging attention of the GAN's discriminator in a self-supervised learning manner. During inference, users can interact with the spatial heatmaps in an intuitive manner, enabling them to edit the output image by adjusting the scene layout, moving, or removing objects. Moreover, we incorporate DragGAN into our framework, which facilitates fine-grained manipulation within a reasonable time and supports a coarse-to-fine editing process. Extensive experiments show that the proposed method not only enables spatial editing over human faces, animal faces, outdoor scenes, and complicated multi-object indoor scenes but also brings improvement in synthesis quality. Code, models, and demo video are available at https://genforce.github.io/SpatialGAN/.

近年来,生成模型在逼真图像合成方面取得了巨大进步。为了让人类能够引导图像生成过程并定制输出结果,许多研究都在探索 GAN 中潜在空间的可解释维度。现有方法通过沿特定方向改变潜码来编辑输出图像的属性,如方向或配色方案。然而,这些方法通常需要对每个预训练模型进行额外的人工注释,而且它们大多侧重于编辑全局属性。在这项工作中,我们提出了一种自监督方法来提高 GAN 的空间可转向性,而无需在潜空间中搜索可转向的方向,也不需要额外的注释。具体来说,我们设计了随机采样的高斯热图,作为空间归纳偏置编码到生成模型的中间层。在从头开始训练 GAN 模型的同时,这些热图将以自我监督学习的方式与 GAN 识别器的新兴注意力保持一致。在推理过程中,用户可以以直观的方式与空间热图进行交互,通过调整场景布局、移动或删除对象来编辑输出图像。此外,我们还将 DragGAN 纳入了我们的框架,这有助于在合理的时间内进行细粒度操作,并支持从粗到细的编辑过程。大量实验表明,所提出的方法不仅能对人脸、动物脸、室外场景和复杂的多物体室内场景进行空间编辑,还能提高合成质量。代码、模型和演示视频请访问 https://genforce.github.io/SpatialGAN/。
{"title":"Spatial Steerability of GANs via Self-Supervision from Discriminator.","authors":"Jianyuan Wang, Lalit Bhagat, Ceyuan Yang, Yinghao Xu, Yujun Shen, Hongdong Li, Bolei Zhou","doi":"10.1109/TPAMI.2024.3422820","DOIUrl":"10.1109/TPAMI.2024.3422820","url":null,"abstract":"<p><p>Generative models make huge progress to the photorealistic image synthesis in recent years. To enable human to steer the image generation process and customize the output, many works explore the interpretable dimensions of the latent space in GANs. Existing methods edit the attributes of the output image such as orientation or color scheme by varying the latent code along certain directions. However, these methods usually require additional human annotations for each pretrained model, and they mostly focus on editing global attributes. In this work, we propose a self-supervised approach to improve the spatial steerability of GANs without searching for steerable directions in the latent space or requiring extra annotations. Specifically, we design randomly sampled Gaussian heatmaps to be encoded into the intermediate layers of generative models as spatial inductive bias. Along with training the GAN model from scratch, these heatmaps are being aligned with the emerging attention of the GAN's discriminator in a self-supervised learning manner. During inference, users can interact with the spatial heatmaps in an intuitive manner, enabling them to edit the output image by adjusting the scene layout, moving, or removing objects. Moreover, we incorporate DragGAN into our framework, which facilitates fine-grained manipulation within a reasonable time and supports a coarse-to-fine editing process. Extensive experiments show that the proposed method not only enables spatial editing over human faces, animal faces, outdoor scenes, and complicated multi-object indoor scenes but also brings improvement in synthesis quality. Code, models, and demo video are available at https://genforce.github.io/SpatialGAN/.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141499994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1