首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
Outlier-Aware Contrastive Learning. 异常值感知对比学习。
IF 18.6 Pub Date : 2026-03-03 DOI: 10.1109/TPAMI.2026.3669598
Jen-Tzung Chien, Kuan Chen

Contrastive learning aims to learn an embedding space with sample discrimination where similar samples attract together while dissimilar samples repulse apart. However, the issue of sampling bias likely happens and degrades the classification performance when a contrast model is trained with the leakage caused by similar samples but from different classes or dissimilar samples from the same class. Out-of-distribution (OOD) detection provides a meaningful scheme to detect and mask those false negative samples for debiasing in an outlier-aware contrastive loss for high-fidelity contrastive learning. Sample debiasing is feasible to reduce the upper bound of contrastive loss. Also, the previous OOD detector was trained from auxiliary collection of OOD samples. In real world, the prior knowledge of OOD samples is commonly unavailable. This study presents new outlier-aware detection and contrast models through generation and augmentation of those samples near the boundary between in-distribution (ID) and OOD. These synthesized samples are located right outside ID, and their Gaussian embeddings sufficiently reflect OOD behaviors. An OOD detector is learned by using ID samples and synthesized OOD samples with the learning objective towards contrastive OOD detection and debiased contrast model. The experiments are conducted to illustrate the merit of the proposed outlier-aware contrastive learning.

对比学习的目的是学习一个具有样本判别的嵌入空间,相似的样本相互吸引,不同的样本相互排斥。然而,当使用来自不同类别的相似样本或来自同一类别的不同样本引起的泄漏训练对比模型时,可能会出现抽样偏差问题并降低分类性能。分布外检测(out -distribution, OOD)为高保真对比学习提供了一种有意义的方案来检测和掩盖那些假阴性样本,以便在离群值感知的对比损失中去偏。样品去偏对于降低对比损耗上界是可行的。此外,以前的OOD检测器是通过辅助收集OOD样本来训练的。在现实世界中,OOD样本的先验知识通常是不可用的。本研究通过生成和增强分布中(ID)和OOD之间边界附近的样本,提出了新的异常值感知检测和对比模型。这些合成的样品位于ID的正外侧,它们的高斯嵌入充分反映了OOD行为。利用ID样本和合成OOD样本学习OOD检测器,学习目标为对比OOD检测和去偏对比模型。通过实验验证了这种异常值感知对比学习的优点。
{"title":"Outlier-Aware Contrastive Learning.","authors":"Jen-Tzung Chien, Kuan Chen","doi":"10.1109/TPAMI.2026.3669598","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3669598","url":null,"abstract":"<p><p>Contrastive learning aims to learn an embedding space with sample discrimination where similar samples attract together while dissimilar samples repulse apart. However, the issue of sampling bias likely happens and degrades the classification performance when a contrast model is trained with the leakage caused by similar samples but from different classes or dissimilar samples from the same class. Out-of-distribution (OOD) detection provides a meaningful scheme to detect and mask those false negative samples for debiasing in an outlier-aware contrastive loss for high-fidelity contrastive learning. Sample debiasing is feasible to reduce the upper bound of contrastive loss. Also, the previous OOD detector was trained from auxiliary collection of OOD samples. In real world, the prior knowledge of OOD samples is commonly unavailable. This study presents new outlier-aware detection and contrast models through generation and augmentation of those samples near the boundary between in-distribution (ID) and OOD. These synthesized samples are located right outside ID, and their Gaussian embeddings sufficiently reflect OOD behaviors. An OOD detector is learned by using ID samples and synthesized OOD samples with the learning objective towards contrastive OOD detection and debiased contrast model. The experiments are conducted to illustrate the merit of the proposed outlier-aware contrastive learning.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147349807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Near-Perfect Clustering Based on Recursive Binary Splitting Using Max-MMD. 基于Max-MMD递归二值分割的近完美聚类。
IF 18.6 Pub Date : 2026-03-03 DOI: 10.1109/TPAMI.2026.3669975
Sourav Chakrabarty, Anirvan Chakraborty, Shyamal K De

We develop novel clustering algorithms for functional data when the number of clusters $K$ is unspecified and also when it is specified. These algorithms are developed based on the Maximum Mean Discrepancy (MMD) measure between the empirical distributions associated with two sets of observations. The algorithms recursively use a binary splitting strategy to partition the dataset into two subgroups such that they are maximally separated in terms of an appropriate weighted MMD measure. When $K$ is unspecified, the proposed clustering algorithm has an additional step to check whether a group of observations obtained by the binary splitting technique consists of observations from a single population. We also learn $K$ directly from the data using this algorithm. When $K$ is specified, a modification of the previous algorithm is proposed which consists of an additional step of merging subgroups which are similar in terms of the weighted MMD distance. The theoretical properties of the proposed algorithms are investigated in an oracle scenario that requires the knowledge of the empirical distributions of the observations from different populations involved. In this setting, we prove that the algorithm proposed when $K$ is unspecified achieves perfect clustering while the algorithm proposed when $K$ is specified has the perfect order preserving (POP) property. Extensive real and simulated data analyses using a variety of models having location difference as well as scale difference show near-perfect clustering performance of both the algorithms which improve upon the state-of-the-art clustering methods for functional data.

我们开发了新的聚类算法,用于功能数据在簇数K$未指定和指定时的聚类。这些算法是基于与两组观测值相关的经验分布之间的最大平均差异(MMD)度量而开发的。这些算法递归地使用二进制分割策略将数据集划分为两个子组,以便根据适当的加权MMD度量最大限度地分离它们。当$K$未指定时,本文提出的聚类算法有一个额外的步骤来检查由二元分割技术获得的一组观测值是否由单个总体的观测值组成。我们还使用该算法直接从数据中学习K。当$K$指定时,提出了对先前算法的改进,其中包括合并加权MMD距离相似的子组的额外步骤。所提出的算法的理论性质在oracle场景中进行了研究,该场景需要了解来自不同种群的观测值的经验分布。在这种情况下,我们证明了当$K$不确定时提出的算法实现了完美聚类,而当$K$指定时提出的算法具有完美保序(POP)性质。使用具有位置差异和尺度差异的各种模型进行了广泛的真实和模拟数据分析,结果表明这两种算法的聚类性能接近完美,这两种算法改进了最先进的功能数据聚类方法。
{"title":"Near-Perfect Clustering Based on Recursive Binary Splitting Using Max-MMD.","authors":"Sourav Chakrabarty, Anirvan Chakraborty, Shyamal K De","doi":"10.1109/TPAMI.2026.3669975","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3669975","url":null,"abstract":"<p><p>We develop novel clustering algorithms for functional data when the number of clusters $K$ is unspecified and also when it is specified. These algorithms are developed based on the Maximum Mean Discrepancy (MMD) measure between the empirical distributions associated with two sets of observations. The algorithms recursively use a binary splitting strategy to partition the dataset into two subgroups such that they are maximally separated in terms of an appropriate weighted MMD measure. When $K$ is unspecified, the proposed clustering algorithm has an additional step to check whether a group of observations obtained by the binary splitting technique consists of observations from a single population. We also learn $K$ directly from the data using this algorithm. When $K$ is specified, a modification of the previous algorithm is proposed which consists of an additional step of merging subgroups which are similar in terms of the weighted MMD distance. The theoretical properties of the proposed algorithms are investigated in an oracle scenario that requires the knowledge of the empirical distributions of the observations from different populations involved. In this setting, we prove that the algorithm proposed when $K$ is unspecified achieves perfect clustering while the algorithm proposed when $K$ is specified has the perfect order preserving (POP) property. Extensive real and simulated data analyses using a variety of models having location difference as well as scale difference show near-perfect clustering performance of both the algorithms which improve upon the state-of-the-art clustering methods for functional data.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147349792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature Compression for Cloud-Edge Multimodal 3D Object Detection. 云边缘多模态3D物体检测的特征压缩。
IF 18.6 Pub Date : 2026-03-03 DOI: 10.1109/TPAMI.2026.3669471
Chongzhen Tian, Zhengxin Li, Hui Yuan, Raouf Hamzaoui, Liquan Shen, Sam Kwong

Machine vision systems, which can efficiently manage extensive visual perception tasks, are becoming increasingly popular in industrial production and daily life. Due to the challenge of simultaneously obtaining accurate depth and texture information with a single sensor, multimodal data captured by cameras and LiDAR is commonly used to enhance performance. Additionally, cloud-edge cooperation has emerged as a novel computing approach to improve user experience and ensure data security in machine vision systems. This paper proposes a pioneering solution to address the feature compression problem in multimodal 3D object detection. Given a sparse tensor-based object detection network at the edge device, we introduce two modes to accommodate different application requirements: Transmission-Friendly Feature Compression (T-FFC) and Accuracy-Friendly Feature Compression (A-FFC). In T-FFC mode, only the output of the last layer of the network's backbone is transmitted from the edge device. The received feature is processed at the cloud device through a channel expansion module and two spatial upsampling modules to generate multi-scale features. In A-FFC mode, we expand upon the T-FFC mode by transmitting two additional types of features. These added features enable the cloud device to generate more accurate multi-scale features. Experimental results on the KITTI dataset using the VirConv-L detection network showed that T-FFC was able to compress the features by a factor of 4933 with less than a 3% reduction in detection performance. On the other hand, A-FFC compressed the features by a factor of about 733 with almost no degradation in detection performance. We also designed optional residual extraction and 3D object reconstruction modules to facilitate the reconstruction of detected objects. The reconstructed objects effectively reflected the shape, occlusion, and details of the original objects.

机器视觉系统可以有效地管理广泛的视觉感知任务,在工业生产和日常生活中越来越受欢迎。由于使用单个传感器同时获得准确的深度和纹理信息的挑战,通常使用相机和激光雷达捕获的多模态数据来提高性能。此外,云边缘合作已经成为一种新的计算方法,可以改善用户体验并确保机器视觉系统中的数据安全。针对多模态三维目标检测中的特征压缩问题,提出了一种开创性的解决方案。给定边缘设备上基于稀疏张量的目标检测网络,我们引入了两种模式来适应不同的应用需求:传输友好特征压缩(T-FFC)和精度友好特征压缩(a - ffc)。在T-FFC模式下,只有网络骨干网最后一层的输出从边缘设备传输出去。所接收的特征在云设备上通过通道扩展模块和两个空间上采样模块进行处理,生成多尺度特征。在A-FFC模式中,我们通过传输两种附加类型的特征来扩展T-FFC模式。这些添加的功能使云设备能够生成更精确的多尺度特征。使用virconvl检测网络在KITTI数据集上的实验结果表明,T-FFC能够将特征压缩4933倍,而检测性能降低不到3%。另一方面,a - ffc将特征压缩了约733倍,而检测性能几乎没有下降。我们还设计了可选的残差提取和三维物体重建模块,便于对检测到的物体进行重建。重建的物体有效地反映了原始物体的形状、遮挡和细节。
{"title":"Feature Compression for Cloud-Edge Multimodal 3D Object Detection.","authors":"Chongzhen Tian, Zhengxin Li, Hui Yuan, Raouf Hamzaoui, Liquan Shen, Sam Kwong","doi":"10.1109/TPAMI.2026.3669471","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3669471","url":null,"abstract":"<p><p>Machine vision systems, which can efficiently manage extensive visual perception tasks, are becoming increasingly popular in industrial production and daily life. Due to the challenge of simultaneously obtaining accurate depth and texture information with a single sensor, multimodal data captured by cameras and LiDAR is commonly used to enhance performance. Additionally, cloud-edge cooperation has emerged as a novel computing approach to improve user experience and ensure data security in machine vision systems. This paper proposes a pioneering solution to address the feature compression problem in multimodal 3D object detection. Given a sparse tensor-based object detection network at the edge device, we introduce two modes to accommodate different application requirements: Transmission-Friendly Feature Compression (T-FFC) and Accuracy-Friendly Feature Compression (A-FFC). In T-FFC mode, only the output of the last layer of the network's backbone is transmitted from the edge device. The received feature is processed at the cloud device through a channel expansion module and two spatial upsampling modules to generate multi-scale features. In A-FFC mode, we expand upon the T-FFC mode by transmitting two additional types of features. These added features enable the cloud device to generate more accurate multi-scale features. Experimental results on the KITTI dataset using the VirConv-L detection network showed that T-FFC was able to compress the features by a factor of 4933 with less than a 3% reduction in detection performance. On the other hand, A-FFC compressed the features by a factor of about 733 with almost no degradation in detection performance. We also designed optional residual extraction and 3D object reconstruction modules to facilitate the reconstruction of detected objects. The reconstructed objects effectively reflected the shape, occlusion, and details of the original objects.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147349819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatio-temporal Decoupled Knowledge Compensator for Few-Shot Action Recognition. 基于时空解耦的小动作识别知识补偿。
IF 18.6 Pub Date : 2026-03-02 DOI: 10.1109/TPAMI.2026.3669254
Hongyu Qu, Xiangbo Shu, Rui Yan, Hailiang Gao, Wenguan Wang, Jinhui Tang

Few-Shot Action Recognition (FSAR) is a challenging task that requires recognizing novel action categories with a few labeled videos. Recent works typically apply semantically coarse category names as auxiliary contexts to guide the learning of discriminative visual features. However, such context provided by the action names is too limited to provide sufficient background knowledge for capturing novel spatial and temporal concepts in actions. In this paper, we propose DiST, an innovative Decomposition-incorporation framework for FSAR that makes use of decoupled Spatial and Temporal knowledge provided by large language models to learn expressive multi-granularity prototypes. In the decomposition stage, we decouple vanilla action names into diverse spatio-temporal attribute descriptions (action-related knowledge). Such commonsense knowledge complements semantic contexts from spatial and temporal perspectives. In the incorporation stage, we propose Spatial/Temporal Knowledge Compensators (SKC/TKC) to discover discriminative object-level and frame-level prototypes, respectively. In SKC, object-level prototypes adaptively aggregate important patch tokens under the guidance of spatial knowledge. Moreover, in TKC, frame-level prototypes utilize temporal attributes to assist in inter-frame temporal relation modeling. These learned prototypes thus provide transparency in capturing fine-grained spatial details and diverse temporal patterns. Experimental results show DiST achieves state-of-the-art results on five standard FSAR datasets.

少镜头动作识别(FSAR)是一项具有挑战性的任务,需要识别新的动作类别与少数标记视频。近年来的研究通常采用语义粗糙的类别名称作为辅助语境来指导识别性视觉特征的学习。然而,由动作名称提供的上下文太过有限,无法提供足够的背景知识来捕捉动作中的新空间和时间概念。在本文中,我们提出了一种创新的FSAR分解合并框架DiST,它利用大型语言模型提供的解耦时空知识来学习具有表现力的多粒度原型。在分解阶段,我们将普通的动作名称解耦为不同的时空属性描述(与动作相关的知识)。这些常识性知识从空间和时间的角度补充了语义上下文。在整合阶段,我们提出了空间/时间知识补偿器(SKC/TKC)来分别发现区分性的对象级和框架级原型。在SKC中,对象级原型在空间知识的引导下自适应地聚合重要的patch token。此外,在TKC中,帧级原型利用时间属性来辅助帧间时间关系建模。因此,这些学习的原型为捕获细粒度的空间细节和不同的时间模式提供了透明度。实验结果表明,DiST在5个标准FSAR数据集上取得了最先进的结果。
{"title":"Spatio-temporal Decoupled Knowledge Compensator for Few-Shot Action Recognition.","authors":"Hongyu Qu, Xiangbo Shu, Rui Yan, Hailiang Gao, Wenguan Wang, Jinhui Tang","doi":"10.1109/TPAMI.2026.3669254","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3669254","url":null,"abstract":"<p><p>Few-Shot Action Recognition (FSAR) is a challenging task that requires recognizing novel action categories with a few labeled videos. Recent works typically apply semantically coarse category names as auxiliary contexts to guide the learning of discriminative visual features. However, such context provided by the action names is too limited to provide sufficient background knowledge for capturing novel spatial and temporal concepts in actions. In this paper, we propose DiST, an innovative Decomposition-incorporation framework for FSAR that makes use of decoupled Spatial and Temporal knowledge provided by large language models to learn expressive multi-granularity prototypes. In the decomposition stage, we decouple vanilla action names into diverse spatio-temporal attribute descriptions (action-related knowledge). Such commonsense knowledge complements semantic contexts from spatial and temporal perspectives. In the incorporation stage, we propose Spatial/Temporal Knowledge Compensators (SKC/TKC) to discover discriminative object-level and frame-level prototypes, respectively. In SKC, object-level prototypes adaptively aggregate important patch tokens under the guidance of spatial knowledge. Moreover, in TKC, frame-level prototypes utilize temporal attributes to assist in inter-frame temporal relation modeling. These learned prototypes thus provide transparency in capturing fine-grained spatial details and diverse temporal patterns. Experimental results show DiST achieves state-of-the-art results on five standard FSAR datasets.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147345981","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Generating Realistic 3D Semantic Training Data for Autonomous Driving. 面向自动驾驶生成逼真的三维语义训练数据。
IF 18.6 Pub Date : 2026-03-02 DOI: 10.1109/TPAMI.2026.3669002
Lucas Nunes, Rodrigo Marcuzzi, Jens Behley, Cyrill Stachniss

Semantic scene understanding is crucial for robotics and computer vision applications. In autonomous driving, 3D semantic segmentation plays an important role for enabling safe navigation. Despite significant advances in the field, the complexity of collecting and annotating 3D data is a bottleneck in this developments. To overcome that data annotation limitation, synthetic simulated data has been used to generate annotated data on demand. There is still, however, a domain gap between real and simulated data. More recently, diffusion models have been in the spotlight, enabling close-to-real data synthesis. Those generative models have been recently applied to the 3D data domain for generating scene-scale data with semantic annotations. Still, those methods either rely on image projection or decoupled models trained with different resolutions in a coarse-to-fine manner. Such intermediary representations impact the generated data quality due to errors added in those transformations. In this work, we propose a novel approach able to generate 3D semantic scene-scale data without relying on any projection or decoupled trained multi-resolution models, achieving more realistic semantic scene data generation compared to previous state-of-the-art methods. Besides improving 3D semantic scene-scale data synthesis, we thoroughly evaluate the use of the synthetic scene samples as labeled data to train a semantic segmentation network. In our experiments, we show that using the synthetic annotated data generated by our method as training data together with the real semantic segmentation labels, leads to an improvement in the semantic segmentation model performance. Our results show the potential of generated scene-scale point clouds to generate more training data to extend existing datasets, reducing the data annotation effort. Our code is available at https://github.com/PRBonn/3DiSS.

语义场景理解对于机器人和计算机视觉应用至关重要。在自动驾驶中,三维语义分割对于实现安全导航具有重要作用。尽管该领域取得了重大进展,但收集和注释3D数据的复杂性是这一发展的瓶颈。为了克服数据注释的限制,合成模拟数据被用于按需生成注释数据。然而,真实数据和模拟数据之间仍然存在领域差距。最近,扩散模型已经成为焦点,使接近真实的数据合成成为可能。这些生成模型最近被应用于三维数据领域,用于生成带有语义注释的场景尺度数据。尽管如此,这些方法要么依赖于图像投影,要么依赖于用不同分辨率以粗到精的方式训练的解耦模型。由于这些转换中添加的错误,这种中间表示会影响生成的数据质量。在这项工作中,我们提出了一种新的方法,能够在不依赖任何投影或解耦训练的多分辨率模型的情况下生成3D语义场景数据,与以前最先进的方法相比,实现更真实的语义场景数据生成。除了改进三维语义场景尺度的数据合成外,我们还全面评估了将合成的场景样本作为标记数据来训练语义分割网络的方法。实验表明,将本文方法生成的合成标注数据与真实的语义切分标签一起作为训练数据,可以提高语义切分模型的性能。我们的研究结果表明,生成的场景尺度点云可以生成更多的训练数据来扩展现有的数据集,从而减少数据注释的工作量。我们的代码可在https://github.com/PRBonn/3DiSS上获得。
{"title":"Towards Generating Realistic 3D Semantic Training Data for Autonomous Driving.","authors":"Lucas Nunes, Rodrigo Marcuzzi, Jens Behley, Cyrill Stachniss","doi":"10.1109/TPAMI.2026.3669002","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3669002","url":null,"abstract":"<p><p>Semantic scene understanding is crucial for robotics and computer vision applications. In autonomous driving, 3D semantic segmentation plays an important role for enabling safe navigation. Despite significant advances in the field, the complexity of collecting and annotating 3D data is a bottleneck in this developments. To overcome that data annotation limitation, synthetic simulated data has been used to generate annotated data on demand. There is still, however, a domain gap between real and simulated data. More recently, diffusion models have been in the spotlight, enabling close-to-real data synthesis. Those generative models have been recently applied to the 3D data domain for generating scene-scale data with semantic annotations. Still, those methods either rely on image projection or decoupled models trained with different resolutions in a coarse-to-fine manner. Such intermediary representations impact the generated data quality due to errors added in those transformations. In this work, we propose a novel approach able to generate 3D semantic scene-scale data without relying on any projection or decoupled trained multi-resolution models, achieving more realistic semantic scene data generation compared to previous state-of-the-art methods. Besides improving 3D semantic scene-scale data synthesis, we thoroughly evaluate the use of the synthetic scene samples as labeled data to train a semantic segmentation network. In our experiments, we show that using the synthetic annotated data generated by our method as training data together with the real semantic segmentation labels, leads to an improvement in the semantic segmentation model performance. Our results show the potential of generated scene-scale point clouds to generate more training data to extend existing datasets, reducing the data annotation effort. Our code is available at https://github.com/PRBonn/3DiSS.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147346050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Continuous Wasserstein Barycenter Space for Generalized All-in-One Image Restoration. 广义一体化图像恢复的连续Wasserstein重心空间学习。
IF 18.6 Pub Date : 2026-03-02 DOI: 10.1109/TPAMI.2026.3669121
Xiaole Tang, Xiaoyi He, Jiayi Xu, Xiang Gu, Jian Sun

Despite substantial advances in all-in-one image restoration for addressing diverse degradations within a unified model, existing methods remain vulnerable to out-of-distribution degradations, thereby limiting their generalization in real-world scenarios. To tackle the challenge, this work is motivated by the intuition that multisource degraded feature distributions are induced by different degradation-specific shifts from an underlying degradation-agnostic distribution, and recovering such a shared distribution is thus crucial for achieving generalization across degradations. With this insight, we propose BaryIR, a representation learning framework that aligns multisource degraded features in the Wasserstein barycenter (WB) space, which models a degradation-agnostic distribution by minimizing the average of Wasserstein distances to multisource degraded distributions. We further introduce residual subspaces, whose embeddings are mutually contrasted while remaining orthogonal to the WB embeddings. Consequently, BaryIR explicitly decouples two orthogonal spaces: a WB space that encodes the degradation-agnostic invariant contents shared across degradations, and residual subspaces that adaptively preserve the degradation-specific knowledge. This disentanglement mitigates overfitting to in-distribution degradations and enables adaptive restoration grounded on the degradation-agnostic shared invariance. Extensive experiments demonstrate that BaryIR performs competitively against state-of-the-art all-in-one methods. Notably, BaryIR generalizes well to unseen degradations (e.g., types and levels) and shows remarkable robustness in learning generalized features, even when trained on limited degradation types and evaluated on real-world data with mixed degradations.

尽管在统一模型中解决多种退化的一体化图像恢复方面取得了实质性进展,但现有方法仍然容易受到分布外退化的影响,从而限制了它们在现实场景中的泛化。为了应对这一挑战,这项工作的动机是基于这样一种直觉,即多源退化特征分布是由潜在的退化不可知分布的不同退化特定转移引起的,因此恢复这种共享分布对于实现跨退化的泛化至关重要。基于这一见解,我们提出了BaryIR,这是一个表征学习框架,它将Wasserstein质心(WB)空间中的多源退化特征对齐,通过最小化与多源退化分布的Wasserstein距离的平均值来建模退化不可知分布。我们进一步引入残差子空间,残差子空间的嵌入相互对比,同时与WB嵌入保持正交。因此,BaryIR显式地解耦了两个正交空间:编码跨退化共享的退化不可知不变内容的WB空间,以及自适应地保留退化特定知识的残差子空间。这种解纠缠减轻了分布内退化的过拟合,并使基于退化不可知的共享不变性的自适应恢复成为可能。广泛的实验表明,BaryIR与最先进的一体化方法相比具有竞争力。值得注意的是,BaryIR可以很好地泛化到看不见的退化(例如,类型和水平),并且在学习广义特征方面表现出显著的鲁棒性,即使在有限的退化类型上进行训练并在混合退化的真实数据上进行评估时也是如此。
{"title":"Learning Continuous Wasserstein Barycenter Space for Generalized All-in-One Image Restoration.","authors":"Xiaole Tang, Xiaoyi He, Jiayi Xu, Xiang Gu, Jian Sun","doi":"10.1109/TPAMI.2026.3669121","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3669121","url":null,"abstract":"<p><p>Despite substantial advances in all-in-one image restoration for addressing diverse degradations within a unified model, existing methods remain vulnerable to out-of-distribution degradations, thereby limiting their generalization in real-world scenarios. To tackle the challenge, this work is motivated by the intuition that multisource degraded feature distributions are induced by different degradation-specific shifts from an underlying degradation-agnostic distribution, and recovering such a shared distribution is thus crucial for achieving generalization across degradations. With this insight, we propose BaryIR, a representation learning framework that aligns multisource degraded features in the Wasserstein barycenter (WB) space, which models a degradation-agnostic distribution by minimizing the average of Wasserstein distances to multisource degraded distributions. We further introduce residual subspaces, whose embeddings are mutually contrasted while remaining orthogonal to the WB embeddings. Consequently, BaryIR explicitly decouples two orthogonal spaces: a WB space that encodes the degradation-agnostic invariant contents shared across degradations, and residual subspaces that adaptively preserve the degradation-specific knowledge. This disentanglement mitigates overfitting to in-distribution degradations and enables adaptive restoration grounded on the degradation-agnostic shared invariance. Extensive experiments demonstrate that BaryIR performs competitively against state-of-the-art all-in-one methods. Notably, BaryIR generalizes well to unseen degradations (e.g., types and levels) and shows remarkable robustness in learning generalized features, even when trained on limited degradation types and evaluated on real-world data with mixed degradations.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147345990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaborative Feedback Discriminative Propagation for Video Super-Resolution. 视频超分辨率协同反馈判别传播。
IF 18.6 Pub Date : 2026-02-27 DOI: 10.1109/TPAMI.2026.3668757
Hao Li, Xiang Chen, Jiangxin Dong, Jinhui Tang, Jinshan Pan

The key success of existing video super-resolution (VSR) methods stems mainly from exploring spatial and temporal information that is usually achieved by a temporal propagation with alignment strategies. However, inaccurate alignment usually leads to significant artifacts that will be accumulated during propagation and thus affect video restoration. Moreover, only propagating the same timestep features forward or backward does not handle the videos with complex motion or occlusion. To address these issues, we propose a collaborative feedback discriminative (CFD) method to correct inaccurate aligned features and better model spatial and temporal information for VSR. Specifically, we first develop a discriminative alignment correction (DAC) method to reduce the influences of the artifacts caused by inaccurate alignment. Then, we propose a collaborative feedback propagation (CFP) module based on feedback and gating mechanisms to explore spatial and temporal information of different timestep features from forward and backward propagation simultaneously. Finally, we embed the proposed DAC and CFP into commonly used VSR networks to verify the effectiveness of our method. Experimental results demonstrate that our method improves the performance of existing VSR models while maintaining a lower model complexity.

现有视频超分辨率(VSR)方法的成功关键在于对空间和时间信息的探索,而这些信息通常是通过带有对齐策略的时间传播来实现的。然而,不准确的对准通常会导致显著的伪影,这些伪影将在传播过程中积累,从而影响视频恢复。此外,仅向前或向后传播相同的时间步长特征并不能处理具有复杂运动或遮挡的视频。为了解决这些问题,我们提出了一种协同反馈判别(CFD)方法来纠正不准确的对齐特征并更好地建模VSR的时空信息。具体而言,我们首先开发了一种判别对准校正(DAC)方法,以减少由于不准确对准引起的伪影的影响。在此基础上,提出了一种基于反馈和门控机制的协同反馈传播(CFP)模块,对不同时间步长特征的时空信息进行前向和后向同步传播。最后,我们将提出的DAC和CFP嵌入到常用的VSR网络中,以验证我们方法的有效性。实验结果表明,该方法在保持较低模型复杂度的同时,提高了现有VSR模型的性能。
{"title":"Collaborative Feedback Discriminative Propagation for Video Super-Resolution.","authors":"Hao Li, Xiang Chen, Jiangxin Dong, Jinhui Tang, Jinshan Pan","doi":"10.1109/TPAMI.2026.3668757","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3668757","url":null,"abstract":"<p><p>The key success of existing video super-resolution (VSR) methods stems mainly from exploring spatial and temporal information that is usually achieved by a temporal propagation with alignment strategies. However, inaccurate alignment usually leads to significant artifacts that will be accumulated during propagation and thus affect video restoration. Moreover, only propagating the same timestep features forward or backward does not handle the videos with complex motion or occlusion. To address these issues, we propose a collaborative feedback discriminative (CFD) method to correct inaccurate aligned features and better model spatial and temporal information for VSR. Specifically, we first develop a discriminative alignment correction (DAC) method to reduce the influences of the artifacts caused by inaccurate alignment. Then, we propose a collaborative feedback propagation (CFP) module based on feedback and gating mechanisms to explore spatial and temporal information of different timestep features from forward and backward propagation simultaneously. Finally, we embed the proposed DAC and CFP into commonly used VSR networks to verify the effectiveness of our method. Experimental results demonstrate that our method improves the performance of existing VSR models while maintaining a lower model complexity.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147319254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UDFStudio: A Unified Framework of Datasets, Benchmarks and Generative Models for Unsigned Distance Functions. UDFStudio:无符号距离函数的数据集、基准和生成模型的统一框架。
IF 18.6 Pub Date : 2026-02-27 DOI: 10.1109/TPAMI.2026.3668763
Junsheng Zhou, Weiqi Zhang, Baorui Ma, Kanle Shi, Yu-Shen Liu, Zhizhong Han

Unsigned distance functions (UDFs) have emerged as powerful representation for modeling and reconstructing geometries with open surfaces. However, the development of 3D generative models for UDFs remains largely unexplored, limiting current methods from generating diverse open-surface 3D content. Moreover, mainstream 3D datasets predominantly consist of watertight meshes, revealing a critical challenge: the absence of standardized datasets and benchmarks specifically tailored for open-surface generation and reconstruction. In this paper, we begin by introducing UDiFF, a novel diffusion-based 3D generative model specifically designed for UDFs. UDiFF supports both conditional and unconditional generation of textured 3D shapes with open surfaces. At its core, UDiFF generates UDFs in the spatial-frequency domain using a learnable wavelet transform. Instead of relying on manually selected wavelet transforms, which are labor-intensive and prone to information loss, we introduce a data-driven approach that learns the optimal wavelet transformation from UDFs datasets. Beyond UDiFF, we present the UWings dataset, comprising 1,509 high-quality 3D open surface models of winged creatures. Using UWings, we establish comprehensive benchmarks for evaluating both generative and reconstruction methods based on UDFs.

无符号距离函数(Unsigned distance functions, udf)作为一种强大的表示形式出现,用于建模和重建具有开放表面的几何形状。然而,udf的3D生成模型的开发在很大程度上仍未被探索,这限制了当前生成各种开放表面3D内容的方法。此外,主流的3D数据集主要由水密网格组成,这揭示了一个关键的挑战:缺乏专门为开放表面生成和重建定制的标准化数据集和基准。在本文中,我们首先介绍了UDiFF,这是一种专门为udf设计的基于扩散的新型3D生成模型。UDiFF支持条件和无条件生成具有开放表面的纹理3D形状。UDiFF的核心是使用可学习的小波变换在空频域中生成udf。而不是依赖于人工选择的小波变换,这是劳动密集型的,容易导致信息丢失,我们引入了一种数据驱动的方法,从udf数据集学习最优小波变换。除了UDiFF,我们还提供了UWings数据集,包括1,509个高质量的有翼生物3D开放表面模型。使用UWings,我们建立了综合的基准来评估基于udf的生成和重建方法。
{"title":"UDFStudio: A Unified Framework of Datasets, Benchmarks and Generative Models for Unsigned Distance Functions.","authors":"Junsheng Zhou, Weiqi Zhang, Baorui Ma, Kanle Shi, Yu-Shen Liu, Zhizhong Han","doi":"10.1109/TPAMI.2026.3668763","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3668763","url":null,"abstract":"<p><p>Unsigned distance functions (UDFs) have emerged as powerful representation for modeling and reconstructing geometries with open surfaces. However, the development of 3D generative models for UDFs remains largely unexplored, limiting current methods from generating diverse open-surface 3D content. Moreover, mainstream 3D datasets predominantly consist of watertight meshes, revealing a critical challenge: the absence of standardized datasets and benchmarks specifically tailored for open-surface generation and reconstruction. In this paper, we begin by introducing UDiFF, a novel diffusion-based 3D generative model specifically designed for UDFs. UDiFF supports both conditional and unconditional generation of textured 3D shapes with open surfaces. At its core, UDiFF generates UDFs in the spatial-frequency domain using a learnable wavelet transform. Instead of relying on manually selected wavelet transforms, which are labor-intensive and prone to information loss, we introduce a data-driven approach that learns the optimal wavelet transformation from UDFs datasets. Beyond UDiFF, we present the UWings dataset, comprising 1,509 high-quality 3D open surface models of winged creatures. Using UWings, we establish comprehensive benchmarks for evaluating both generative and reconstruction methods based on UDFs.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147319279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CLIP-Actor-X: Text-driven 4D Human Avatar Generation via Cross-modal Synthesis-through-Optimization. CLIP-Actor-X:文本驱动的4D人类化身生成通过跨模态合成-通过优化。
IF 18.6 Pub Date : 2026-02-23 DOI: 10.1109/TPAMI.2026.3665111
Kim Youwang, Taehyun Byun, Kim Ji-Yeon, Sungjoon Choi, Tae-Hyun Oh

We propose CLIP-Actor-X, a text-driven motion generation and neural mesh stylization system for 4D human avatar generation. CLIP-Actor-X generates a detailed 3D human mesh, motion animation, and texture to conform to a given text prompt input from a user. CLIP- Actor-X system mainly consists of two modules. First, for generating realistic human motion, we build a text-driven human motion synthesis module modeled by a retrieval-augmented generative model, powered by a text-to-motion diffusion model. Second, our novel zero-shot neural style optimization module detailizes and texturizes the sampled sequence of a neutral human mesh template, such that the resulting mesh and appearance comply with the input text prompt in a temporally-consistent and pose-agnostic manner. In contrast to the prior arts that use an artist-designed, non-animatable mesh as an input, our output representation is animatable and better aligned between an input text and the generated avatar without additional post-processes, e.g., re-alignment, retargeting, or rigging. We further propose the ways to stabilize the optimization process: spatio-temporal view augmentation and visibility-aware embedding attention, which deals with poorly rendered views. We demonstrate that CLIP-Actor-X produces perceptually plausible and human-recognizable human avatar in motion with detailed geometry and texture solely from a natural language prompt.

我们提出了CLIP-Actor-X,一个文本驱动的运动生成和神经网格风格化系统,用于4D人类化身生成。CLIP-Actor-X生成详细的3D人体网格、运动动画和纹理,以符合来自用户的给定文本提示输入。CLIP- Actor-X系统主要由两个模块组成。首先,为了生成逼真的人体运动,我们构建了一个文本驱动的人体运动合成模块,该模块由检索增强生成模型建模,由文本到运动扩散模型提供支持。其次,我们新颖的零镜头神经风格优化模块对中性人体网格模板的采样序列进行细节化和纹理化处理,从而使生成的网格和外观以时间一致和姿态无关的方式符合输入文本提示。与使用艺术家设计的非动画网格作为输入的现有技术相比,我们的输出表示是可动画的,并且在输入文本和生成的角色之间更好地对齐,而无需额外的后期处理,例如重新对齐,重新定位或操纵。我们进一步提出了稳定优化过程的方法:时空视图增强和可见性感知嵌入注意力,以处理渲染不良的视图。我们证明CLIP-Actor-X仅从自然语言提示中产生具有详细几何和纹理的感知上可信和人类可识别的运动中的人类化身。
{"title":"CLIP-Actor-X: Text-driven 4D Human Avatar Generation via Cross-modal Synthesis-through-Optimization.","authors":"Kim Youwang, Taehyun Byun, Kim Ji-Yeon, Sungjoon Choi, Tae-Hyun Oh","doi":"10.1109/TPAMI.2026.3665111","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3665111","url":null,"abstract":"<p><p>We propose CLIP-Actor-X, a text-driven motion generation and neural mesh stylization system for 4D human avatar generation. CLIP-Actor-X generates a detailed 3D human mesh, motion animation, and texture to conform to a given text prompt input from a user. CLIP- Actor-X system mainly consists of two modules. First, for generating realistic human motion, we build a text-driven human motion synthesis module modeled by a retrieval-augmented generative model, powered by a text-to-motion diffusion model. Second, our novel zero-shot neural style optimization module detailizes and texturizes the sampled sequence of a neutral human mesh template, such that the resulting mesh and appearance comply with the input text prompt in a temporally-consistent and pose-agnostic manner. In contrast to the prior arts that use an artist-designed, non-animatable mesh as an input, our output representation is animatable and better aligned between an input text and the generated avatar without additional post-processes, e.g., re-alignment, retargeting, or rigging. We further propose the ways to stabilize the optimization process: spatio-temporal view augmentation and visibility-aware embedding attention, which deals with poorly rendered views. We demonstrate that CLIP-Actor-X produces perceptually plausible and human-recognizable human avatar in motion with detailed geometry and texture solely from a natural language prompt.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Aligning Few-Step Diffusion Models with Dense Reward Difference Learning. 基于密集奖励差分学习的分级扩散模型。
IF 18.6 Pub Date : 2026-02-23 DOI: 10.1109/TPAMI.2026.3665753
Ziyi Zhang, Li Shen, Sen Zhang, Deheng Ye, Yong Luo, Miaojing Shi, Dongjing Shan, Bo Du, Dacheng Tao

Few-step diffusion models enable efficient high-resolution image synthesis but struggle to align with specific downstream objectives due to limitations of existing reinforcement learning (RL) methods in low-step regimes with limited state spaces and suboptimal sample quality. To address this, we propose Stepwise Diffusion Policy Optimization (SDPO), a novel RL framework tailored for few-step diffusion models. SDPO introduces a dual-state trajectory sampling mechanism, tracking both noisy and predicted clean states at each step to provide dense reward feedback and enable low-variance, mixed-step optimization. For further efficiency, we develop a latent similarity-based dense reward prediction strategy to minimize costly dense reward queries. Leveraging these dense rewards, SDPO optimizes a dense reward difference learning objective that enables more frequent and granular policy updates. Additional refinements, including stepwise advantage estimates, temporal importance weighting, and step-shuffled gradient updates, further enhance long-term dependency, low-step priority, and gradient stability. Our experiments demonstrate that SDPO consistently delivers superior reward-aligned results across diverse few-step settings and tasks.

几步扩散模型能够实现高效的高分辨率图像合成,但由于现有的强化学习(RL)方法在有限状态空间和次优样本质量的低步制度中的局限性,很难与特定的下游目标保持一致。为了解决这个问题,我们提出了逐步扩散策略优化(SDPO),这是一种针对小步扩散模型量身定制的新型强化学习框架。SDPO引入了双状态轨迹采样机制,在每一步都跟踪噪声和预测的干净状态,以提供密集的奖励反馈,并实现低方差的混合步优化。为了提高效率,我们开发了一种基于潜在相似度的密集奖励预测策略,以最小化代价高昂的密集奖励查询。利用这些密集奖励,SDPO优化了密集奖励差异学习目标,从而实现更频繁和更细粒度的策略更新。其他的改进,包括逐步优势估计、时间重要性加权和逐步变换的梯度更新,进一步增强了长期依赖性、低阶优先级和梯度稳定性。我们的实验表明,SDPO在不同的步骤设置和任务中始终如一地提供卓越的奖励对齐结果。
{"title":"Aligning Few-Step Diffusion Models with Dense Reward Difference Learning.","authors":"Ziyi Zhang, Li Shen, Sen Zhang, Deheng Ye, Yong Luo, Miaojing Shi, Dongjing Shan, Bo Du, Dacheng Tao","doi":"10.1109/TPAMI.2026.3665753","DOIUrl":"https://doi.org/10.1109/TPAMI.2026.3665753","url":null,"abstract":"<p><p>Few-step diffusion models enable efficient high-resolution image synthesis but struggle to align with specific downstream objectives due to limitations of existing reinforcement learning (RL) methods in low-step regimes with limited state spaces and suboptimal sample quality. To address this, we propose Stepwise Diffusion Policy Optimization (SDPO), a novel RL framework tailored for few-step diffusion models. SDPO introduces a dual-state trajectory sampling mechanism, tracking both noisy and predicted clean states at each step to provide dense reward feedback and enable low-variance, mixed-step optimization. For further efficiency, we develop a latent similarity-based dense reward prediction strategy to minimize costly dense reward queries. Leveraging these dense rewards, SDPO optimizes a dense reward difference learning objective that enables more frequent and granular policy updates. Additional refinements, including stepwise advantage estimates, temporal importance weighting, and step-shuffled gradient updates, further enhance long-term dependency, low-step priority, and gradient stability. Our experiments demonstrate that SDPO consistently delivers superior reward-aligned results across diverse few-step settings and tasks.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2026-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147277962","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1