首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
The Value of Corrective Feedback in the Online Active Learning Paradigm 纠正反馈在在线主动学习范式中的价值。
IF 18.6 Pub Date : 2025-12-03 DOI: 10.1109/TPAMI.2025.3639522
Mark Lindsey;Francis Kubala;Richard M. Stern
Online Active Learning (OAL) is a powerful tool for classifying evolving data streams using limited annotations from a human operator who is a domain expert. The objective of the OAL learning paradigm is to minimize jointly the classification error rate and the annotation cost across the data stream by posing periodic Active Learning (AL) queries. In this paper, this objective is extended to include identification of classifier errors by the expert during the typical workflow. To this end, Corrective Feedback (CF) is introduced as a second channel of interaction between the expert and the learning algorithm, complementary to the AL channel, that allows the algorithm to obtain additional training labels without disrupting the expert’s workflow. Online Active Learning with Corrective Feedback (OAL-CF) is formally defined as a paradigm, and its efficacy is proven through experimental application to two binary classification tasks, Spoken Language Verification and Voice-Type Discrimination. Finally, the effects of adding CF to the OAL paradigm are analyzed in terms of classification performance, annotation cost, trends over time, and class balance of the collected training data. Overall, the addition of CF results in a 53% relative reduction in cost compared to OAL without CF.
在线主动学习(Online Active Learning, OAL)是一种强大的工具,它利用领域专家操作员提供的有限注释对不断变化的数据流进行分类。主动学习(AL)学习范式的目标是通过提出周期性的主动学习(AL)查询,使数据流中的分类错误率和标注成本共同最小化。在本文中,这一目标被扩展到包括专家在典型工作流程中识别分类器错误。为此,引入纠正反馈(CF)作为专家和学习算法之间的第二个交互通道,补充人工智能通道,使算法能够在不中断专家工作流程的情况下获得额外的训练标签。在线主动学习与纠正反馈(Online Active Learning with Corrective Feedback,简称al - cf)被正式定义为一种范式,并通过实验应用于口语验证和语音类型识别两个二元分类任务,证明了其有效性。最后,从分类性能、注释成本、随时间变化的趋势和收集的训练数据的类平衡等方面分析了将CF添加到OAL范式的影响。总的来说,与不添加CF的OAL相比,添加CF的成本相对降低了53%。
{"title":"The Value of Corrective Feedback in the Online Active Learning Paradigm","authors":"Mark Lindsey;Francis Kubala;Richard M. Stern","doi":"10.1109/TPAMI.2025.3639522","DOIUrl":"10.1109/TPAMI.2025.3639522","url":null,"abstract":"Online Active Learning (OAL) is a powerful tool for classifying evolving data streams using limited annotations from a human operator who is a domain expert. The objective of the OAL learning paradigm is to minimize jointly the classification error rate and the annotation cost across the data stream by posing periodic Active Learning (AL) queries. In this paper, this objective is extended to include identification of classifier errors by the expert during the typical workflow. To this end, Corrective Feedback (CF) is introduced as a second channel of interaction between the expert and the learning algorithm, complementary to the AL channel, that allows the algorithm to obtain additional training labels without disrupting the expert’s workflow. Online Active Learning with Corrective Feedback (OAL-CF) is formally defined as a paradigm, and its efficacy is proven through experimental application to two binary classification tasks, Spoken Language Verification and Voice-Type Discrimination. Finally, the effects of adding CF to the OAL paradigm are analyzed in terms of classification performance, annotation cost, trends over time, and class balance of the collected training data. Overall, the addition of CF results in a 53% relative reduction in cost compared to OAL without CF.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 3","pages":"3885-3898"},"PeriodicalIF":18.6,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11274545","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145663912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FedFask: Fast Sketching Distributed PCA for Large-Scale Federated Data FedFask:快速绘制大规模联邦数据的分布式PCA。
IF 18.6 Pub Date : 2025-12-03 DOI: 10.1109/TPAMI.2025.3639635
Xingcai Zhou;Guang Yang;Haotian Zheng;Linglong Kong;Jinde Cao
We study distributed principal component analysis (PCA) for large-scale federated data when the sample size $n$ and dimension $d$ are both ultra-large. This type of data is currently very common, but faces numerous challenges in PCA learning, such as communication overhead and computational complexity. We develop a new algorithm ${mathsf {FedFask}}$ (Fast Sketching for Federated learning) with lower communication cost $O(dr)$ and lower computational complexity $O(d(np/m+p^{2}+r^{2}))$, where $m$ is the number of workers, $r$ is the rank of matrix, $p$ is the dimension of sketched column space, and $rleq pll d$. In ${mathsf {FedFask}}$, we adopt and develop technologies such as fast sketching, alignments with orthogonal Procrustes Fixing, and matrix Stiefel manifold via Kolmogorov-Nagumo-type average. Thus, ${mathsf {FedFask}}$ has a higher accuracy, lower stochastic variation, and best representation of multiple randomly projected eigenspaces, and avoids the orthogonal ambiguity of eigenspaces. We show that ${mathsf {FedFask}}$ achieves the same rate of learning $Oleft(frac{kappa _{r}r}{lambda _{r}}sqrt{frac{r^{*}}{n}}right)$ as the centralized PCA uses all data, and tolerates more workers to parallel acceleration computation. We conduct extensive experiments to demonstrate the effectiveness of ${mathsf {FedFask}}$.
在样本量$n$和维数$d$都非常大的情况下,研究了大规模联邦数据的分布式主成分分析(PCA)。这种类型的数据目前非常常见,但在PCA学习中面临许多挑战,例如通信开销和计算复杂性。我们开发了一种新的算法${sf FedFask}$ (Fast Sketching for Federated learning),具有更低的通信成本$O(dr)$和更低的计算复杂度$O(d(np/m+p^{2}+r^{2}))$,其中$m$为工人的数量,$r$为矩阵的秩,$p$为草图列空间的维数,$rleq pll d$。在${sf FedFask}$中,我们采用并开发了诸如快速草图,正交Procrustes固定对齐以及通过kolmogorov - nagumo型平均的矩阵Stiefel流形等技术。因此,${sf FedFask}$具有更高的精度,更低的随机变异,以及对多个随机投影特征空间的最佳表示,并且避免了特征空间的正交模糊。我们表明${sf FedFask}$达到了与集中式PCA使用所有数据相同的学习速度$Oleft(frac{kappa _{r}r}{lambda _{r}}sqrt{frac{r^*}{n}}right)$,并且允许更多的工作人员并行加速计算。我们进行了大量的实验来证明${sf FedFask}$的有效性。
{"title":"FedFask: Fast Sketching Distributed PCA for Large-Scale Federated Data","authors":"Xingcai Zhou;Guang Yang;Haotian Zheng;Linglong Kong;Jinde Cao","doi":"10.1109/TPAMI.2025.3639635","DOIUrl":"10.1109/TPAMI.2025.3639635","url":null,"abstract":"We study distributed principal component analysis (PCA) for large-scale federated data when the sample size <inline-formula><tex-math>$n$</tex-math></inline-formula> and dimension <inline-formula><tex-math>$d$</tex-math></inline-formula> are both ultra-large. This type of data is currently very common, but faces numerous challenges in PCA learning, such as communication overhead and computational complexity. We develop a new algorithm <inline-formula><tex-math>${mathsf {FedFask}}$</tex-math></inline-formula> (<b>Fa</b>st <b>Sk</b>etching for <b>Fed</b>erated learning) with lower communication cost <inline-formula><tex-math>$O(dr)$</tex-math></inline-formula> and lower computational complexity <inline-formula><tex-math>$O(d(np/m+p^{2}+r^{2}))$</tex-math></inline-formula>, where <inline-formula><tex-math>$m$</tex-math></inline-formula> is the number of workers, <inline-formula><tex-math>$r$</tex-math></inline-formula> is the rank of matrix, <inline-formula><tex-math>$p$</tex-math></inline-formula> is the dimension of sketched column space, and <inline-formula><tex-math>$rleq pll d$</tex-math></inline-formula>. In <inline-formula><tex-math>${mathsf {FedFask}}$</tex-math></inline-formula>, we adopt and develop technologies such as fast sketching, alignments with orthogonal Procrustes Fixing, and matrix Stiefel manifold via Kolmogorov-Nagumo-type average. Thus, <inline-formula><tex-math>${mathsf {FedFask}}$</tex-math></inline-formula> has a higher accuracy, lower stochastic variation, and best representation of multiple randomly projected eigenspaces, and avoids the orthogonal ambiguity of eigenspaces. We show that <inline-formula><tex-math>${mathsf {FedFask}}$</tex-math></inline-formula> achieves the same rate of learning <inline-formula><tex-math>$Oleft(frac{kappa _{r}r}{lambda _{r}}sqrt{frac{r^{*}}{n}}right)$</tex-math></inline-formula> as the centralized PCA uses all data, and tolerates more workers to parallel acceleration computation. We conduct extensive experiments to demonstrate the effectiveness of <inline-formula><tex-math>${mathsf {FedFask}}$</tex-math></inline-formula>.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 3","pages":"3714-3725"},"PeriodicalIF":18.6,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145664268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Meta-Learning-Based Surrogate Models for Efficient Hyperparameter Optimization 基于元学习的高效超参数优化代理模型
IF 18.6 Pub Date : 2025-11-06 DOI: 10.1109/TPAMI.2025.3630178
Liping Deng;Maziar Raissi;MingQing Xiao
Sequential Model-Based Optimization (SMBO) is a highly effective strategy for hyperparameter search in machine learning. It utilizes a surrogate model that fits previous trials and approximates the hyperparameter response surface (performance). This surrogate model primarily guides the decision-making process for selecting the next set of hyperparameters. Existing classic surrogates, such as Gaussian processes and random forests, focus solely on the current task of interest and cannot incorporate trials from historical tasks. This limitation hinders their efficacy in various applications. Inspired by the state-of-the-art convolutional neural process, this paper proposes a novel meta-learning-based surrogate model for efficient and effective hyperparameter optimization. Our surrogate is trained on the meta-knowledge from a range of historical tasks, enabling it to accurately predict the hyperparameter response surface even with a limited number of trials on a new task. We tested our approach on the hyperparameter selection problem for the well-known support vector machine (SVM), residual neural network (ResNet), and vision transformer (ViT) across hundreds of real-world classification datasets. The empirical results demonstrate its superiority over existing surrogate models, highlighting the effectiveness of meta-learning in hyperparameter optimization.
序列模型优化(SMBO)是机器学习中一种非常有效的超参数搜索策略。它利用了一个拟合先前试验的代理模型,并近似于超参数响应面(性能)。该代理模型主要指导选择下一组超参数的决策过程。现有的经典替代方法,如高斯过程和随机森林,只关注当前感兴趣的任务,而不能纳入历史任务的试验。这一限制阻碍了它们在各种应用中的有效性。受卷积神经过程的启发,本文提出了一种新的基于元学习的代理模型,用于高效的超参数优化。我们的代理使用来自一系列历史任务的元知识进行训练,即使在新任务上进行有限次数的试验,也能准确预测超参数响应面。我们在数百个真实世界的分类数据集上测试了我们的方法,用于众所周知的支持向量机(SVM)、残差神经网络(ResNet)和视觉转换器(ViT)的超参数选择问题。实证结果表明其优于现有的代理模型,突出了元学习在超参数优化中的有效性。
{"title":"Meta-Learning-Based Surrogate Models for Efficient Hyperparameter Optimization","authors":"Liping Deng;Maziar Raissi;MingQing Xiao","doi":"10.1109/TPAMI.2025.3630178","DOIUrl":"10.1109/TPAMI.2025.3630178","url":null,"abstract":"Sequential Model-Based Optimization (SMBO) is a highly effective strategy for hyperparameter search in machine learning. It utilizes a surrogate model that fits previous trials and approximates the hyperparameter response surface (performance). This surrogate model primarily guides the decision-making process for selecting the next set of hyperparameters. Existing classic surrogates, such as Gaussian processes and random forests, focus solely on the current task of interest and cannot incorporate trials from historical tasks. This limitation hinders their efficacy in various applications. Inspired by the state-of-the-art convolutional neural process, this paper proposes a novel meta-learning-based surrogate model for efficient and effective hyperparameter optimization. Our surrogate is trained on the meta-knowledge from a range of historical tasks, enabling it to accurately predict the hyperparameter response surface even with a limited number of trials on a new task. We tested our approach on the hyperparameter selection problem for the well-known support vector machine (SVM), residual neural network (ResNet), and vision transformer (ViT) across hundreds of real-world classification datasets. The empirical results demonstrate its superiority over existing surrogate models, highlighting the effectiveness of meta-learning in hyperparameter optimization.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 3","pages":"3931-3938"},"PeriodicalIF":18.6,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145454731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Affine Correspondences Between Multi-Camera Systems for Relative Pose Estimation 用于相对姿态估计的多相机系统间的仿射对应。
IF 18.6 Pub Date : 2025-10-27 DOI: 10.1109/TPAMI.2025.3626134
Banglei Guan;Ji Zhao
We present a novel method to compute the relative pose of multi-camera systems using two affine correspondences (ACs). Existing solutions to the multi-camera relative pose estimation are either restricted to special cases of motion, have too high computational complexity, or require too many point correspondences (PCs). Thus, these solvers impede an efficient or accurate relative pose estimation when applying RANSAC as a robust estimator. This paper shows that the 6DOF relative pose estimation problem using ACs permits a feasible minimal solution, when exploiting the geometric constraints between ACs and multi-camera systems using a special parameterization. We present a problem formulation based on two ACs that encompass two common types of ACs across two views, i.e., inter-camera and intra-camera. Moreover, we exploit a unified and versatile framework for generating 6DOF solvers. Building upon this foundation, we use this framework to address two categories of practical scenarios. First, for the more challenging 7DOF relative pose estimation problem—where the scale transformation of multi-camera systems is unknown—we propose 7DOF solvers to compute the relative pose and scale using three ACs. Second, leveraging inertial measurement units (IMUs), we introduce several minimal solvers for constrained relative pose estimation problems. These include 5DOF solvers with known relative rotation angle, and 4DOF solver with known vertical direction. Experiments on both virtual and real multi-camera systems prove that the proposed solvers are more efficient than the state-of-the-art algorithms, while resulting in a better relative pose accuracy.
提出了一种利用两个仿射对应(ACs)计算多相机系统相对位姿的新方法。现有的多相机相对姿态估计方法要么局限于运动的特殊情况,要么计算复杂度太高,要么需要太多的点对应(pc)。因此,当使用RANSAC作为鲁棒估计器时,这些解算器阻碍了有效或准确的相对姿态估计。本文表明,当使用特殊的参数化方法利用ACs和多相机系统之间的几何约束时,使用ACs的6自由度相对姿态估计问题允许一个可行的最小解。我们提出了一个基于两个ac的问题公式,该ac包含跨两个视图的两种常见类型的ac,即相机间和相机内。此外,我们还开发了一个统一的通用框架来生成6自由度求解器。在此基础上,我们使用此框架来处理两类实际场景。首先,对于更具挑战性的7DOF相对姿态估计问题(其中多相机系统的尺度变换未知),我们提出了使用三个ac计算相对姿态和尺度的7DOF解算器。其次,利用惯性测量单元(imu),我们引入了约束相对姿态估计问题的几个最小解。这包括已知相对旋转角的5DOF解算器和已知垂直方向的4DOF解算器。在虚拟和真实多相机系统上的实验证明,该算法比现有算法更有效,同时产生了更好的相对姿态精度。源代码可从https://github.com/jizhaox/relpose-mcs-depth获得。
{"title":"Affine Correspondences Between Multi-Camera Systems for Relative Pose Estimation","authors":"Banglei Guan;Ji Zhao","doi":"10.1109/TPAMI.2025.3626134","DOIUrl":"10.1109/TPAMI.2025.3626134","url":null,"abstract":"We present a novel method to compute the relative pose of multi-camera systems using two affine correspondences (ACs). Existing solutions to the multi-camera relative pose estimation are either restricted to special cases of motion, have too high computational complexity, or require too many point correspondences (PCs). Thus, these solvers impede an efficient or accurate relative pose estimation when applying RANSAC as a robust estimator. This paper shows that the 6DOF relative pose estimation problem using ACs permits a feasible minimal solution, when exploiting the geometric constraints between ACs and multi-camera systems using a special parameterization. We present a problem formulation based on two ACs that encompass two common types of ACs across two views, i.e., inter-camera and intra-camera. Moreover, we exploit a unified and versatile framework for generating 6DOF solvers. Building upon this foundation, we use this framework to address two categories of practical scenarios. First, for the more challenging 7DOF relative pose estimation problem—where the scale transformation of multi-camera systems is unknown—we propose 7DOF solvers to compute the relative pose and scale using three ACs. Second, leveraging inertial measurement units (IMUs), we introduce several minimal solvers for constrained relative pose estimation problems. These include 5DOF solvers with known relative rotation angle, and 4DOF solver with known vertical direction. Experiments on both virtual and real multi-camera systems prove that the proposed solvers are more efficient than the state-of-the-art algorithms, while resulting in a better relative pose accuracy.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"2012-2029"},"PeriodicalIF":18.6,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145380504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-Resolution Open-Vocabulary Object 6D Pose Estimation 高分辨率开放词汇对象6D姿态估计。
IF 18.6 Pub Date : 2025-10-23 DOI: 10.1109/TPAMI.2025.3624589
Jaime Corsetti;Davide Boscaini;Francesco Giuliari;Changjae Oh;Andrea Cavallaro;Fabio Poiesi
The generalisation to unseen objects in the 6D pose estimation task is very challenging. While Vision-Language Models (VLMs) enable using natural language descriptions to support 6D pose estimation of unseen objects, these solutions underperform compared to model-based methods. In this work we present Horyon, an open-vocabulary VLM-based architecture that addresses relative pose estimation between two scenes of an unseen object, described by a textual prompt only. We use the textual prompt to identify the unseen object in the scenes and then obtain high-resolution multi-scale features. These features are used to extract cross-scene matches for registration. We evaluate our model on a benchmark with a large variety of unseen objects across four datasets, namely REAL275, Toyota-Light, Linemod, and YCB-Video. Our method achieves state-of-the-art performance on all datasets, outperforming by 12.6 in Average Recall the previous best-performing approach.
在6D姿态估计任务中推广到不可见物体是非常具有挑战性的。虽然视觉语言模型(VLMs)可以使用自然语言描述来支持对未见物体的6D姿态估计,但与基于模型的方法相比,这些解决方案的性能较差。在这项工作中,我们提出了Horyon,一个基于开放词汇表vmm的架构,它解决了一个看不见的物体的两个场景之间的相对姿态估计,仅由文本提示描述。我们利用文本提示来识别场景中看不见的物体,从而获得高分辨率的多尺度特征。这些特征用于提取跨场景匹配进行配准。我们在四个数据集(REAL275、Toyota Light、Linemod和YCB-Video)上对我们的模型进行了基准测试。我们的方法在所有数据集上都达到了最先进的性能,在平均召回率上比之前表现最好的方法高出12.6。
{"title":"High-Resolution Open-Vocabulary Object 6D Pose Estimation","authors":"Jaime Corsetti;Davide Boscaini;Francesco Giuliari;Changjae Oh;Andrea Cavallaro;Fabio Poiesi","doi":"10.1109/TPAMI.2025.3624589","DOIUrl":"10.1109/TPAMI.2025.3624589","url":null,"abstract":"The generalisation to unseen objects in the 6D pose estimation task is very challenging. While Vision-Language Models (VLMs) enable using natural language descriptions to support 6D pose estimation of unseen objects, these solutions underperform compared to model-based methods. In this work we present Horyon, an open-vocabulary VLM-based architecture that addresses relative pose estimation between two scenes of an unseen object, described by a textual prompt only. We use the textual prompt to identify the unseen object in the scenes and then obtain high-resolution multi-scale features. These features are used to extract cross-scene matches for registration. We evaluate our model on a benchmark with a large variety of unseen objects across four datasets, namely REAL275, Toyota-Light, Linemod, and YCB-Video. Our method achieves state-of-the-art performance on <italic>all</i> datasets, outperforming by 12.6 in Average Recall the previous best-performing approach.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"2066-2077"},"PeriodicalIF":18.6,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145357313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SMC++: Masked Learning of Unsupervised Video Semantic Compression 无监督视频语义压缩的掩膜学习。
IF 18.6 Pub Date : 2025-10-23 DOI: 10.1109/TPAMI.2025.3625063
Yuan Tian;Xiaoyue Ling;Cong Geng;Qiang Hu;Guo Lu;Guangtao Zhai
Most video compression methods focus on human visual perception, neglecting semantic preservation. This leads to severe semantic loss during the compression, hampering downstream video analysis tasks. In this paper, we propose a Masked Video Modeling (MVM)-powered compression framework that particularly preserves video semantics, by jointly mining and compressing the semantics in a self-supervised manner. While MVM is proficient at learning generalizable semantics through the masked patch prediction task, it may also encode non-semantic information like trivial textural details, wasting bitcost and bringing semantic noises. To suppress this, we explicitly regularize the non-semantic entropy of the compressed video in the MVM token space. The proposed framework is instantiated as a simple Semantic-Mining-then-Compression (SMC) model. Furthermore, we extend SMC as an advanced SMC++ model from several aspects. First, we equip it with a masked motion prediction objective, leading to better temporal semantic learning ability. Second, we introduce a Transformer-based compression module, to improve the semantic compression efficacy. Considering that directly mining the complex redundancy among heterogeneous features in different coding stages is non-trivial, we introduce a compact blueprint semantic representation to align these features into a similar form, fully unleashing the power of the Transformer-based compression module. Extensive results demonstrate the proposed SMC and SMC++ models show remarkable superiority over previous traditional, learnable, and perceptual quality-oriented video codecs, on three video analysis tasks and seven datasets.
大多数视频压缩方法关注的是人的视觉感知,而忽略了语义的保存。这将导致压缩过程中严重的语义丢失,阻碍后续的视频分析任务。在本文中,我们提出了一个屏蔽视频建模(MVM)驱动的压缩框架,该框架通过以自监督的方式联合挖掘和压缩语义,特别保留了视频语义。虽然MVM精通通过掩码补丁预测任务学习可泛化语义,但它也可能编码非语义信息,如琐碎的纹理细节,浪费比特成本并带来语义噪声。为了抑制这种情况,我们在MVM令牌空间中显式正则化压缩视频的非语义熵。该框架被实例化为一个简单的语义挖掘-压缩(SMC)模型。此外,我们从几个方面将SMC扩展为先进的smc++模型。首先,我们为它配备了一个掩蔽的运动预测目标,从而提高了它的时间语义学习能力。其次,我们引入了一个基于transformer的压缩模块,以提高语义压缩的效率。考虑到在不同编码阶段直接挖掘异构特征之间的复杂冗余是非常重要的,我们引入了一个紧凑的蓝图语义表示来将这些特征对齐到一个相似的形式,充分释放基于transformer的压缩模块的力量。广泛的结果表明,在三个视频分析任务和七个数据集上,所提出的SMC和SMC++模型比以前传统的、可学习的、面向感知质量的视频编解码器具有显著的优势。
{"title":"SMC++: Masked Learning of Unsupervised Video Semantic Compression","authors":"Yuan Tian;Xiaoyue Ling;Cong Geng;Qiang Hu;Guo Lu;Guangtao Zhai","doi":"10.1109/TPAMI.2025.3625063","DOIUrl":"10.1109/TPAMI.2025.3625063","url":null,"abstract":"Most video compression methods focus on human visual perception, neglecting semantic preservation. This leads to severe semantic loss during the compression, hampering downstream video analysis tasks. In this paper, we propose a Masked Video Modeling (MVM)-powered compression framework that particularly preserves video semantics, by jointly mining and compressing the semantics in a self-supervised manner. While MVM is proficient at learning generalizable semantics through the masked patch prediction task, it may also encode non-semantic information like trivial textural details, wasting bitcost and bringing semantic noises. To suppress this, we explicitly regularize the non-semantic entropy of the compressed video in the MVM token space. The proposed framework is instantiated as a simple Semantic-Mining-then-Compression (SMC) model. Furthermore, we extend SMC as an advanced SMC++ model from several aspects. First, we equip it with a masked motion prediction objective, leading to better temporal semantic learning ability. Second, we introduce a Transformer-based compression module, to improve the semantic compression efficacy. Considering that directly mining the complex redundancy among heterogeneous features in different coding stages is non-trivial, we introduce a compact blueprint semantic representation to align these features into a similar form, fully unleashing the power of the Transformer-based compression module. Extensive results demonstrate the proposed SMC and SMC++ models show remarkable superiority over previous traditional, learnable, and perceptual quality-oriented video codecs, on three video analysis tasks and seven datasets.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1992-2011"},"PeriodicalIF":18.6,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145357314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Orthogonal Decoupling Contrastive Regularization: Toward Uncorrelated Feature Decoupling for Unpaired Image Restoration 正交解耦对比正则化:面向非相关特征解耦的非配对图像恢复。
IF 18.6 Pub Date : 2025-10-20 DOI: 10.1109/TPAMI.2025.3620803
Zhongze Wang;Jingchao Peng;Haitao Zhao;Lujian Yao;Kaijie Zhao
Unpaired image restoration (UIR) is a significant task due to the difficulty of acquiring paired degraded/clear images with identical backgrounds. In this paper, we propose a novel UIR method based on the assumption that an image contains both degradation-related features, which affect the level of degradation, and degradation-unrelated features, such as texture and semantic information. Our method aims to ensure that the degradation-related features of the restoration result closely resemble those of the clear image, while the degradation-unrelated features align with the input degraded image. Specifically, we introduce a Feature Orthogonalization Module optimized on Stiefel manifold to decouple image features, ensuring feature uncorrelation. A task-driven Depth-wise Feature Classifier is proposed to assign weights to uncorrelated features based on their relevance to degradation prediction. To avoid the dependence of the training process on the quality of the clear image in a single pair of input data, we propose to maintain several degradation-related proxies describing the degradation level of clear images to enhance the model’s robustness. Finally, a weighted PatchNCE loss is introduced to pull degradation-related features in the output image toward those of clear images, while bringing degradation-unrelated features close to those of the degraded input.
未配对图像恢复(UIR)是一项重要的任务,因为难以获得具有相同背景的成对退化/清晰图像。在本文中,我们提出了一种新的UIR方法,该方法基于图像同时包含影响退化水平的退化相关特征和退化无关特征(如纹理和语义信息)的假设。我们的方法旨在确保恢复结果中与退化相关的特征与清晰图像非常相似,而与退化无关的特征与输入的退化图像对齐。具体来说,我们引入了一个基于Stiefel流形优化的特征正交化模块来解耦图像特征,确保特征不相关。提出了一种任务驱动的深度特征分类器,根据不相关特征与退化预测的相关性为其分配权重。为了避免训练过程对单对输入数据中清晰图像质量的依赖,我们建议保留几个与退化相关的代理来描述清晰图像的退化程度,以增强模型的鲁棒性。最后,引入加权PatchNCE损失,将输出图像中与退化相关的特征拉向清晰图像,同时使与退化无关的特征接近退化的输入图像。
{"title":"Orthogonal Decoupling Contrastive Regularization: Toward Uncorrelated Feature Decoupling for Unpaired Image Restoration","authors":"Zhongze Wang;Jingchao Peng;Haitao Zhao;Lujian Yao;Kaijie Zhao","doi":"10.1109/TPAMI.2025.3620803","DOIUrl":"10.1109/TPAMI.2025.3620803","url":null,"abstract":"Unpaired image restoration (UIR) is a significant task due to the difficulty of acquiring paired degraded/clear images with identical backgrounds. In this paper, we propose a novel UIR method based on the assumption that an image contains both degradation-related features, which affect the level of degradation, and degradation-unrelated features, such as texture and semantic information. Our method aims to ensure that the degradation-related features of the restoration result closely resemble those of the clear image, while the degradation-unrelated features align with the input degraded image. Specifically, we introduce a Feature Orthogonalization Module optimized on Stiefel manifold to decouple image features, ensuring feature uncorrelation. A task-driven Depth-wise Feature Classifier is proposed to assign weights to uncorrelated features based on their relevance to degradation prediction. To avoid the dependence of the training process on the quality of the clear image in a single pair of input data, we propose to maintain several degradation-related proxies describing the degradation level of clear images to enhance the model’s robustness. Finally, a weighted PatchNCE loss is introduced to pull degradation-related features in the output image toward those of clear images, while bringing degradation-unrelated features close to those of the degraded input.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1842-1859"},"PeriodicalIF":18.6,"publicationDate":"2025-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145331646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RealCustom++: Representing Images as Real Textual Word for Real-Time Customization realcustom++:将图像表示为实时定制的真实文本单词
IF 18.6 Pub Date : 2025-10-17 DOI: 10.1109/TPAMI.2025.3623025
Zhendong Mao;Mengqi Huang;Fei Ding;Mingcong Liu;Qian He;Yongdong Zhang
Text-to-image customization aims to generate images that align with both the given text and the subject in the given image. Existing works follow the pseudo-word paradigm, which represents the subject as a non-existent pseudo word and combines it with other text to generate images. However, the pseudo word inherently conflicts and entangles with other real words, resulting in a dual-optimum paradox between the subject similarity and text controllability. To address this, we propose RealCustom++, a novel real-word paradigm that represents the subject with a non-conflicting real word to generate a coherent guidance image and corresponding subject mask, there by disentangling the influence scopes of the text and subject for simultaneous optimization. Specifically, RealCustom++ introduces a train-inference decoupled framework: (1) during training, it learns a general alignment between visual conditions and all real text words; and (2) during inference, a dual-branch architecture is employed, where the Guidance Branch produces the subject guidance mask, and the Generation Branch utilizes this mask to customize the generation of the specific real word exclusively within subject-relevant regions. Extensive experiments validate RealCustom++s superior performance, which improves controllability by 7.48%, similarity by 3.04% and quality by 76.43% simultaneously. Moreover, RealCustom++ further improves controllability by 4.6% and multi-subject similarity by 6.34% for multisubject customization
文本到图像自定义旨在生成与给定文本和给定图像中的主题对齐的图像。现有的作品遵循伪词范式,将主题作为一个不存在的伪词来表现,并与其他文本结合生成图像。然而,伪词与其他真实词之间存在着内在的冲突和纠缠,导致了主体相似性与文本可控性的双重最优悖论。为了解决这个问题,我们提出了realcustom++,这是一种新颖的真实世界范式,它用一个不冲突的真实世界来表示主题,从而生成连贯的指导图像和相应的主题掩码,从而解耦文本和主题的影响范围,同时进行优化。具体来说,realcustom++引入了一个训练-推理解耦框架:(1)在训练过程中,它学习视觉条件与所有真实文本单词之间的一般对齐;(2)在推理过程中,采用双分支架构,引导分支生成主题引导掩码,生成分支利用该掩码在主题相关区域内专门定制生成特定真实词。大量的实验验证了realcustom++的优越性能,同时将可控性提高7.48%,相似度提高3.04%,质量提高76.43%。此外,realcustom++进一步提高了多主题定制的可控性4.6%和多主题相似度6.34%
{"title":"RealCustom++: Representing Images as Real Textual Word for Real-Time Customization","authors":"Zhendong Mao;Mengqi Huang;Fei Ding;Mingcong Liu;Qian He;Yongdong Zhang","doi":"10.1109/TPAMI.2025.3623025","DOIUrl":"10.1109/TPAMI.2025.3623025","url":null,"abstract":"Text-to-image customization aims to generate images that align with both the given text and the subject in the given image. Existing works follow the pseudo-word paradigm, which represents the subject as a non-existent pseudo word and combines it with other text to generate images. However, the pseudo word inherently conflicts and entangles with other real words, resulting in a dual-optimum paradox between the subject similarity and text controllability. To address this, we propose RealCustom++, a novel real-word paradigm that represents the subject with a non-conflicting real word to generate a coherent guidance image and corresponding subject mask, there by disentangling the influence scopes of the text and subject for simultaneous optimization. Specifically, RealCustom++ introduces a train-inference decoupled framework: (1) during training, it learns a general alignment between visual conditions and all real text words; and (2) during inference, a dual-branch architecture is employed, where the Guidance Branch produces the subject guidance mask, and the Generation Branch utilizes this mask to customize the generation of the specific real word exclusively within subject-relevant regions. Extensive experiments validate RealCustom++s superior performance, which improves controllability by 7.48%, similarity by 3.04% and quality by 76.43% simultaneously. Moreover, RealCustom++ further improves controllability by 4.6% and multi-subject similarity by 6.34% for multisubject customization","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"2078-2095"},"PeriodicalIF":18.6,"publicationDate":"2025-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145310795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vicinal Gaussian Transform: Rethinking Source-Free Domain Adaptation Through Source-Informed Label Consistency 邻域高斯变换:基于源知情标签一致性的无源域自适应思考
IF 18.6 Pub Date : 2025-10-15 DOI: 10.1109/TPAMI.2025.3621631
Jing Wang;Yongchao Xu;Jing Tang;Zeyu Gong;Bo Tao;Clarence W. de Silva;Xiang Bai
A central challenge in source-free domain adaptation (SFDA) is the lack of a theoretical framework for explicitly analyzing domain shifts, as the absence of source data prevents direct domain comparisons. In this paper, we introduce the Vicinal Gaussian Transform (VGT), an analytical operator that models source-informed latent vicinities as Gaussians and shows that vicinal prediction divergence is bounded by their covariance. By this formulation, SFDA can be reframed as shrinking covariance to reinforce label consistency. To operationalize this idea, we introduce the Energy-based VGT (EBVGT), a novel SDE that realizes the Gaussian transform by contracting covariance through a denoising mechanism. A recovery-likelihood with a Schrödinger-Bridge smoothness penalty denoises perturbed states, while a BYOL-derived energy function, directly obtained from model predictions, provides the score to guide label-consistent trajectories within the vicinity. This design not only yields noise-suppressed vicinal features for adaptation without source data, but also eliminates the need for additional learnable parameters for score estimation, in contrast to conventional deep SDEs. Our EBVGT is model- and modality-agnostic, efficient for classification, and improves state-of-the-art SFDA methods by 1.3–3.0% (2.0% on average) across both 2D image and 3D point cloud benchmarks.
无源域适应(SFDA)的一个核心挑战是缺乏明确分析域转移的理论框架,因为缺乏源数据阻碍了直接的域比较。在本文中,我们引入了邻域高斯变换(VGT),这是一种分析算子,它将消息源通知的潜在邻域建模为高斯分布,并表明邻域预测散度受其协方差的限制。通过这个公式,SFDA可以重新定义为缩小协方差以加强标签一致性。为了实现这一思想,我们引入了基于能量的VGT (EBVGT),这是一种通过去噪机制收缩协方差来实现高斯变换的新型SDE。带有Schrödinger-Bridge平滑惩罚的恢复可能性去噪了扰动状态,而直接从模型预测中获得的byol衍生能量函数提供了分数,以指导附近的标签一致轨迹。与传统的深度SDEs相比,这种设计不仅可以在没有源数据的情况下产生噪声抑制的邻近特征,而且还消除了对分数估计的额外可学习参数的需要。我们的EBVGT与模型和模态无关,分类效率高,并且在2D图像和3D点云基准测试中比最先进的SFDA方法提高了1.3-3.0%(平均2.0%)。
{"title":"Vicinal Gaussian Transform: Rethinking Source-Free Domain Adaptation Through Source-Informed Label Consistency","authors":"Jing Wang;Yongchao Xu;Jing Tang;Zeyu Gong;Bo Tao;Clarence W. de Silva;Xiang Bai","doi":"10.1109/TPAMI.2025.3621631","DOIUrl":"10.1109/TPAMI.2025.3621631","url":null,"abstract":"A central challenge in source-free domain adaptation (SFDA) is the lack of a theoretical framework for explicitly analyzing domain shifts, as the absence of source data prevents direct domain comparisons. In this paper, we introduce the Vicinal Gaussian Transform (VGT), an analytical operator that models source-informed latent vicinities as Gaussians and shows that vicinal prediction divergence is bounded by their covariance. By this formulation, SFDA can be reframed as shrinking covariance to reinforce label consistency. To operationalize this idea, we introduce the Energy-based VGT (EBVGT), a novel SDE that realizes the Gaussian transform by contracting covariance through a denoising mechanism. A recovery-likelihood with a Schrödinger-Bridge smoothness penalty denoises perturbed states, while a BYOL-derived energy function, directly obtained from model predictions, provides the score to guide label-consistent trajectories within the vicinity. This design not only yields noise-suppressed vicinal features for adaptation without source data, but also eliminates the need for additional learnable parameters for score estimation, in contrast to conventional deep SDEs. Our EBVGT is model- and modality-agnostic, efficient for classification, and improves state-of-the-art SFDA methods by 1.3–3.0% (2.0% on average) across both 2D image and 3D point cloud benchmarks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"2030-2047"},"PeriodicalIF":18.6,"publicationDate":"2025-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145295630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing Unnormalized Statistical Models Through Compositional Optimization 通过组合优化优化非标准化统计模型
IF 18.6 Pub Date : 2025-10-14 DOI: 10.1109/TPAMI.2025.3621320
Wei Jiang;Jiayu Qin;Lingyu Wu;Changyou Chen;Tianbao Yang;Lijun Zhang
Learning unnormalized statistical models (e.g., energy-based models) is computationally challenging due to the complexity of handling the partition function. To eschew this complexity, noise-contrastive estimation (NCE) has been proposed by formulating the objective as the logistic loss between the real data and the artificial noise. However, previous research indicates that NCE may perform poorly in many tasks due to its flat loss landscape and slow convergence. In this paper, we study a direct approach for optimizing the negative log-likelihood of unnormalized models through the lens of compositional optimization. To tackle the partition function, a noise distribution is introduced such that the log partition function can be expressed as a compositional function whose inner function can be estimated using stochastic samples. Consequently, the objective can be optimized via stochastic compositional optimization algorithms. Despite being a simple method, we demonstrate it is more favorable than NCE by (1) establishing a fast convergence rate and quantifying its dependence on the noise distribution through the variance of stochastic estimators; (2) developing better results in Gaussian mean estimation by showing our method has a much favorable loss landscape and enjoys faster convergence; (3) demonstrating better performance on various applications, including density estimation, out-of-distribution detection, and real image generation.
由于处理配分函数的复杂性,学习非规范化统计模型(例如,基于能量的模型)在计算上具有挑战性。为了避免这种复杂性,噪声对比估计(NCE)被提出,它将目标表述为真实数据和人工噪声之间的逻辑损失。然而,先前的研究表明,由于其平坦的损失景观和缓慢的收敛,NCE可能在许多任务中表现不佳。本文从组合优化的角度研究了一种直接优化非归一化模型负对数似然的方法。为了处理配分函数,引入了噪声分布,使得对数配分函数可以表示为一个组合函数,其内部函数可以使用随机样本估计。因此,可以通过随机组合优化算法对目标进行优化。尽管是一种简单的方法,但我们证明了它比NCE更有利,通过(1)建立一个快速的收敛速度,并通过随机估计量的方差量化其对噪声分布的依赖;(2)通过表明我们的方法具有更有利的损失格局和更快的收敛速度,在高斯均值估计中得到更好的结果;(3)在密度估计、离分布检测和真实图像生成等应用中表现出更好的性能。
{"title":"Optimizing Unnormalized Statistical Models Through Compositional Optimization","authors":"Wei Jiang;Jiayu Qin;Lingyu Wu;Changyou Chen;Tianbao Yang;Lijun Zhang","doi":"10.1109/TPAMI.2025.3621320","DOIUrl":"10.1109/TPAMI.2025.3621320","url":null,"abstract":"Learning unnormalized statistical models (e.g., energy-based models) is computationally challenging due to the complexity of handling the partition function. To eschew this complexity, noise-contrastive estimation (NCE) has been proposed by formulating the objective as the logistic loss between the real data and the artificial noise. However, previous research indicates that NCE may perform poorly in many tasks due to its flat loss landscape and slow convergence. In this paper, we study a direct approach for optimizing the negative log-likelihood of unnormalized models through the lens of compositional optimization. To tackle the partition function, a noise distribution is introduced such that the log partition function can be expressed as a compositional function whose inner function can be estimated using stochastic samples. Consequently, the objective can be optimized via stochastic compositional optimization algorithms. Despite being a simple method, we demonstrate it is more favorable than NCE by (1) establishing a fast convergence rate and quantifying its dependence on the noise distribution through the variance of stochastic estimators; (2) developing better results in Gaussian mean estimation by showing our method has a much favorable loss landscape and enjoys faster convergence; (3) demonstrating better performance on various applications, including density estimation, out-of-distribution detection, and real image generation.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1949-1960"},"PeriodicalIF":18.6,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145289305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1