首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
VQ-FedDiff: Federated Learning Algorithm of Diffusion Models With Client-Specific Vector-Quantized Conditioning VQ-FedDiff:具有客户特定矢量量化条件的扩散模型的联邦学习算法
IF 18.6 Pub Date : 2025-08-22 DOI: 10.1109/TPAMI.2025.3602282
Tehrim Yoon;Minyoung Hwang;Eunho Yang
Modern generative models, particularly denoising diffusion probabilistic models (DDPMs), provide high-quality synthetic images, enabling users to generate diverse images and videos that are realistic. However, in a number of situations, edge devices or individual institutions may possess locally collected data that is highly sensitive and should ensure data privacy, such as in the field of healthcare and finance. Under such federated learning (FL) settings, various methods on training generative models have been studied, but most of them assume generative adversarial networks (GANs), and the algorithms are specific to GANs and not other forms of generative models such as DDPM. This paper proposes a new algorithm for training DDPMs under federated learning settings, VQ-FedDiff, which provides a personalized algorithm for training diffusion models that can generate higher-quality images FID while still keeping risk of breaching sensitive information as low as locally-trained secure models. We demonstrate that VQ-FedDiff shows state-of-the-art performance on existing federated learning of diffusion models in both IID and non-IID settings, and in benchmark photorealistic and medical image datasets. Our results show that diffusion models can efficiently learn with decentralized, sensitive data, generating high-quality images while preserving data privacy.
现代生成模型,特别是去噪扩散概率模型(ddpm),提供高质量的合成图像,使用户能够生成逼真的各种图像和视频。然而,在许多情况下,边缘设备或个别机构可能拥有本地收集的高度敏感数据,应确保数据隐私,例如在医疗保健和金融领域。在这种联邦学习(FL)设置下,已经研究了各种训练生成模型的方法,但大多数方法都假设生成对抗网络(GANs),并且算法是针对GANs而不是其他形式的生成模型(如DDPM)的。本文提出了一种在联邦学习设置下训练ddpm的新算法VQ-FedDiff,该算法为训练扩散模型提供了一种个性化的算法,该算法可以生成更高质量的图像FID,同时将泄露敏感信息的风险保持在与本地训练的安全模型一样低的水平。我们证明了VQ-FedDiff在IID和非IID设置中,以及在基准逼真图像和医学图像数据集中,在现有的扩散模型的联邦学习上表现出了最先进的性能。我们的研究结果表明,扩散模型可以有效地学习分散的、敏感的数据,在保护数据隐私的同时生成高质量的图像。
{"title":"VQ-FedDiff: Federated Learning Algorithm of Diffusion Models With Client-Specific Vector-Quantized Conditioning","authors":"Tehrim Yoon;Minyoung Hwang;Eunho Yang","doi":"10.1109/TPAMI.2025.3602282","DOIUrl":"10.1109/TPAMI.2025.3602282","url":null,"abstract":"Modern generative models, particularly denoising diffusion probabilistic models (DDPMs), provide high-quality synthetic images, enabling users to generate diverse images and videos that are realistic. However, in a number of situations, edge devices or individual institutions may possess locally collected data that is highly sensitive and should ensure data privacy, such as in the field of healthcare and finance. Under such federated learning (FL) settings, various methods on training generative models have been studied, but most of them assume generative adversarial networks (GANs), and the algorithms are specific to GANs and not other forms of generative models such as DDPM. This paper proposes a new algorithm for training DDPMs under federated learning settings, VQ-FedDiff, which provides a personalized algorithm for training diffusion models that can generate higher-quality images FID while still keeping risk of breaching sensitive information as low as locally-trained secure models. We demonstrate that VQ-FedDiff shows state-of-the-art performance on existing federated learning of diffusion models in both IID and non-IID settings, and in benchmark photorealistic and medical image datasets. Our results show that diffusion models can efficiently learn with decentralized, sensitive data, generating high-quality images while preserving data privacy.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11863-11873"},"PeriodicalIF":18.6,"publicationDate":"2025-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144900424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unified Modality Separation: A Vision-Language Framework for Unsupervised Domain Adaptation 统一模态分离:无监督领域自适应的视觉语言框架
IF 18.6 Pub Date : 2025-08-22 DOI: 10.1109/TPAMI.2025.3597436
Xinyao Li;Jingjing Li;Zhekai Du;Lei Zhu;Heng Tao Shen
Unsupervised domain adaptation (UDA) enables models trained on a labeled source domain to handle new unlabeled domains. Recently, pre-trained vision-language models (VLMs) have demonstrated promising zero-shot performance by leveraging semantic information to facilitate target tasks. By aligning vision and text embeddings, VLMs have shown notable success in bridging domain gaps. However, inherent differences naturally exist between modalities, which is known as modality gap. Our findings reveal that direct UDA with the presence of modality gap only transfers modality-invariant knowledge, leading to suboptimal target performance. To address this limitation, we propose a unified modality separation framework that accommodates both modality-specific and modality-invariant components. During training, different modality components are disentangled from VLM features then handled separately in a unified manner. At test time, modality-adaptive ensemble weights are automatically determined to maximize the synergy of different components. To evaluate instance-level modality characteristics, we design a modality discrepancy metric to categorize samples into modality-invariant, modality-specific, and uncertain ones. The modality-invariant samples are exploited to facilitate cross-modal alignment, while uncertain ones are annotated to enhance model capabilities. Building upon prompt tuning techniques, our methods achieve up to 9% performance gain with 9 times of computational efficiencies. Extensive experiments and analysis across various backbones, baselines, datasets and adaptation settings demonstrate the efficacy of our design.
无监督域适应(UDA)使在标记源域上训练的模型能够处理新的未标记域。最近,预训练的视觉语言模型(VLMs)通过利用语义信息来促进目标任务,显示出了良好的零射击性能。通过对齐视觉和文本嵌入,vlm在弥合领域差距方面取得了显著的成功。然而,情态之间自然存在着固有的差异,这种差异被称为情态差距。我们的研究结果表明,存在模态差距的直接UDA只传递模态不变的知识,导致次优目标性能。为了解决这个限制,我们提出了一个统一的模态分离框架,它可以容纳模态特定组件和模态不变组件。在训练过程中,将不同的模态分量从VLM特征中分离出来,分别进行统一处理。在测试时,自动确定模态自适应集成权重,以最大化不同组件的协同作用。为了评估实例级模态特征,我们设计了一个模态差异度量,将样本分为模态不变、模态特定和不确定三种。利用模态不变的样本来促进跨模态对齐,而对不确定样本进行注释以增强模型能力。基于即时调优技术,我们的方法实现了高达9%的性能增益和9倍的计算效率。对各种主干、基线、数据集和适应设置的广泛实验和分析证明了我们设计的有效性。
{"title":"Unified Modality Separation: A Vision-Language Framework for Unsupervised Domain Adaptation","authors":"Xinyao Li;Jingjing Li;Zhekai Du;Lei Zhu;Heng Tao Shen","doi":"10.1109/TPAMI.2025.3597436","DOIUrl":"10.1109/TPAMI.2025.3597436","url":null,"abstract":"Unsupervised domain adaptation (UDA) enables models trained on a labeled source domain to handle new unlabeled domains. Recently, pre-trained vision-language models (VLMs) have demonstrated promising zero-shot performance by leveraging semantic information to facilitate target tasks. By aligning vision and text embeddings, VLMs have shown notable success in bridging domain gaps. However, inherent differences naturally exist between modalities, which is known as <italic>modality gap</i>. Our findings reveal that direct UDA with the presence of modality gap only transfers modality-invariant knowledge, leading to suboptimal target performance. To address this limitation, we propose a unified modality separation framework that accommodates both modality-specific and modality-invariant components. During training, different modality components are disentangled from VLM features then handled separately in a unified manner. At test time, modality-adaptive ensemble weights are automatically determined to maximize the synergy of different components. To evaluate instance-level modality characteristics, we design a modality discrepancy metric to categorize samples into modality-invariant, modality-specific, and uncertain ones. The modality-invariant samples are exploited to facilitate cross-modal alignment, while uncertain ones are annotated to enhance model capabilities. Building upon prompt tuning techniques, our methods achieve up to 9% performance gain with 9 times of computational efficiencies. Extensive experiments and analysis across various backbones, baselines, datasets and adaptation settings demonstrate the efficacy of our design.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 11","pages":"10604-10618"},"PeriodicalIF":18.6,"publicationDate":"2025-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144900428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Consistent and Optimal Solution to Camera Motion Estimation 摄像机运动估计的一致性和最优解
IF 18.6 Pub Date : 2025-08-21 DOI: 10.1109/TPAMI.2025.3601430
Guangyang Zeng;Qingcheng Zeng;Xinghan Li;Biqiang Mu;Jiming Chen;Ling Shi;Junfeng Wu
Given 2D point correspondences between an image pair, inferring the camera motion is a fundamental issue in the computer vision community. The existing works generally set out from the epipolar constraint and estimate the essential matrix, which is not optimal in the maximum likelihood (ML) sense. In this paper, we dive into the original measurement model with respect to the rotation matrix and normalized translation vector and formulate the ML problem. We then propose an optimal two-step algorithm to solve it: In the first step, we estimate the variance of measurement noises and devise a consistent estimator based on bias elimination; In the second step, we execute a one-step Gauss-Newton iteration on manifold to refine the consistent estimator. We prove that the proposed estimator achieves the same asymptotic statistical properties as the ML estimator: The first is consistency, i.e., the estimator converges to the ground truth as the point number increases; The second is asymptotic efficiency, i.e., the mean squared error of the estimator converges to the theoretical lower bound — Cramer-Rao bound. In addition, we show that our algorithm has linear time complexity. These appealing characteristics endow our estimator with a great advantage in the case of dense point correspondences. Experiments on both synthetic data and real images demonstrate that when the point number reaches the order of hundreds, our estimator outperforms the state-of-the-art ones in terms of estimation accuracy and CPU time.
给定图像对之间的二维点对应关系,推断相机运动是计算机视觉社区的一个基本问题。现有的工作一般从极外约束出发,估计本质矩阵,这在最大似然(ML)意义上不是最优的。在本文中,我们深入研究了关于旋转矩阵和归一化平移向量的原始测量模型,并制定了ML问题。然后,我们提出了一个最优的两步算法来解决这个问题:第一步,我们估计测量噪声的方差,并设计一个基于偏差消除的一致估计器;在第二步中,我们对流形执行一步高斯-牛顿迭代来改进一致性估计。我们证明了所提出的估计量达到了与ML估计量相同的渐近统计性质:一是一致性,即随着点数的增加,估计量收敛于基真值;二是渐近效率,即估计量的均方误差收敛于理论下界- Cramer-Rao界。此外,我们还证明了我们的算法具有线性时间复杂度。这些吸引人的特性使我们的估计器在密集点对应的情况下具有很大的优势。在合成数据和真实图像上的实验表明,当点数达到数百数量级时,我们的估计器在估计精度和CPU时间方面优于最先进的估计器。
{"title":"Consistent and Optimal Solution to Camera Motion Estimation","authors":"Guangyang Zeng;Qingcheng Zeng;Xinghan Li;Biqiang Mu;Jiming Chen;Ling Shi;Junfeng Wu","doi":"10.1109/TPAMI.2025.3601430","DOIUrl":"10.1109/TPAMI.2025.3601430","url":null,"abstract":"Given 2D point correspondences between an image pair, inferring the camera motion is a fundamental issue in the computer vision community. The existing works generally set out from the epipolar constraint and estimate the essential matrix, which is not optimal in the maximum likelihood (ML) sense. In this paper, we dive into the original measurement model with respect to the rotation matrix and normalized translation vector and formulate the ML problem. We then propose an optimal two-step algorithm to solve it: In the first step, we estimate the variance of measurement noises and devise a consistent estimator based on bias elimination; In the second step, we execute a one-step Gauss-Newton iteration on manifold to refine the consistent estimator. We prove that the proposed estimator achieves the same asymptotic statistical properties as the ML estimator: The first is consistency, i.e., the estimator converges to the ground truth as the point number increases; The second is asymptotic efficiency, i.e., the mean squared error of the estimator converges to the theoretical lower bound — Cramer-Rao bound. In addition, we show that our algorithm has linear time complexity. These appealing characteristics endow our estimator with a great advantage in the case of dense point correspondences. Experiments on both synthetic data and real images demonstrate that when the point number reaches the order of hundreds, our estimator outperforms the state-of-the-art ones in terms of estimation accuracy and CPU time.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"12005-12020"},"PeriodicalIF":18.6,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144898490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FaceTracer: Unveiling Source Identities From Swapped Face Images and Videos for Fraud Prevention FACETRACER:从交换的人脸图像和视频中揭示源身份,以防止欺诈
IF 18.6 Pub Date : 2025-08-20 DOI: 10.1109/TPAMI.2025.3601141
Zhongyi Zhang;Jie Zhang;Wenbo Zhou;Xinghui Zhou;Qing Guo;Weiming Zhang;Tianwei Zhang;Nenghai Yu
Face-swapping techniques have advanced rapidly with the evolution of deep learning, leading to widespread use and growing concerns about potential misuse, especially in cases of fraud. While many efforts have focused on detecting swapped face images or videos, these methods are insufficient for tracing the malicious users behind fraudulent activities. Intrusive watermark-based approaches also fail to trace unmarked identities, limiting their practical utility. To address these challenges, we introduce FaceTracer, the first non-intrusive framework specifically designed to trace the identity of the source person from swapped face images or videos. Specifically, FaceTracer leverages a disentanglement module that effectively suppresses identity information related to the target person while isolating the identity features of the source person. This allows us to extract robust identity information that can directly link the swapped face back to the original individual, aiding in uncovering the actors behind fraudulent activities. Extensive experiments demonstrate FaceTracer’s effectiveness across various face-swapping techniques, successfully identifying the source person in swapped content and enabling the tracing of malicious actors involved in fraudulent activities. Additionally, FaceTracer shows strong transferability to unseen face-swapping methods including commercial applications and robustness against transmission distortions and adaptive attacks.
人脸交换技术随着深度学习的发展而迅速发展,导致广泛使用和越来越多的人担心潜在的滥用,特别是在欺诈案件中。虽然很多努力都集中在检测交换的人脸图像或视频上,但这些方法不足以追踪欺诈活动背后的恶意用户。基于侵入式水印的方法也无法追踪未标记的身份,限制了它们的实际用途。为了应对这些挑战,我们引入了FaceTracer,这是第一个非侵入性框架,专门用于从交换的面部图像或视频中追踪源人员的身份。具体来说,FaceTracer利用了一个解缠模块,可以有效地抑制与目标人员相关的身份信息,同时隔离源人员的身份特征。这使我们能够提取可靠的身份信息,这些信息可以直接将交换后的脸与原始个人联系起来,从而帮助发现欺诈活动背后的参与者。大量的实验证明了FaceTracer在各种面部交换技术中的有效性,成功地识别交换内容中的源人员,并能够跟踪涉及欺诈活动的恶意行为者。此外,FaceTracer显示出强大的可移植性,可用于不可见的面部交换方法,包括商业应用和抗传输失真和自适应攻击的鲁棒性。
{"title":"FaceTracer: Unveiling Source Identities From Swapped Face Images and Videos for Fraud Prevention","authors":"Zhongyi Zhang;Jie Zhang;Wenbo Zhou;Xinghui Zhou;Qing Guo;Weiming Zhang;Tianwei Zhang;Nenghai Yu","doi":"10.1109/TPAMI.2025.3601141","DOIUrl":"10.1109/TPAMI.2025.3601141","url":null,"abstract":"Face-swapping techniques have advanced rapidly with the evolution of deep learning, leading to widespread use and growing concerns about potential misuse, especially in cases of fraud. While many efforts have focused on detecting swapped face images or videos, these methods are insufficient for tracing the malicious users behind fraudulent activities. Intrusive watermark-based approaches also fail to trace unmarked identities, limiting their practical utility. To address these challenges, we introduce <sc>FaceTracer</small>, the first non-intrusive framework specifically designed to trace the identity of the source person from swapped face images or videos. Specifically, <sc>FaceTracer</small> leverages a disentanglement module that effectively suppresses identity information related to the target person while isolating the identity features of the source person. This allows us to extract robust identity information that can directly link the swapped face back to the original individual, aiding in uncovering the actors behind fraudulent activities. Extensive experiments demonstrate <sc>FaceTracer</small>’s effectiveness across various face-swapping techniques, successfully identifying the source person in swapped content and enabling the tracing of malicious actors involved in fraudulent activities. Additionally, <sc>FaceTracer</small> shows strong transferability to unseen face-swapping methods including commercial applications and robustness against transmission distortions and adaptive attacks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"12021-12037"},"PeriodicalIF":18.6,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144898544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LBONet: Supervised Spectral Descriptors for Shape Analysis 用于形状分析的监督谱描述子
IF 18.6 Pub Date : 2025-08-20 DOI: 10.1109/TPAMI.2025.3600873
Oguzhan Yigit;Richard C. Wilson
The Laplace-Beltrami operator has established itself in the field of non-rigid shape analysis due to its many useful properties such as being invariant under isometric transformation, having a countable eigensystem forming an orthonormal basis, and fully characterizing geodesic distances of the manifold. However, this invariancy only applies under isometric deformations, which leads to a performance breakdown in many real-world applications. In recent years emphasis has been placed upon extracting optimal features using deep learning methods, however spectral signatures play a crucial role and still add value. In this paper we take a step back, revisiting the LBO and proposing a supervised way to learn several operators on a manifold. Depending on the task, by applying these functions, we can train the LBO eigenbasis to be more task-specific. The optimization of the LBO leads to enormous improvements to established descriptors such as the heat kernel signature in various tasks such as retrieval, classification, segmentation, and correspondence, proving the adaptation of the LBO eigenbasis to both global and highly local learning settings.
Laplace-Beltrami算子由于其在等距变换下的不变性、具有构成正交基的可数本征系统以及充分表征流形的测地线距离等特性而在非刚性形状分析领域中确立了自己的地位。然而,这种不变性只适用于等距变形,这会导致许多实际应用程序的性能崩溃。近年来,人们把重点放在了使用深度学习方法提取最优特征上,然而光谱特征起着至关重要的作用,仍然具有附加价值。在本文中,我们退后一步,重新审视杠杆收购,并提出一种监督的方法来学习流形上的几个算子。根据任务的不同,通过应用这些函数,我们可以训练LBO特征基,使其更加特定于任务。LBO的优化极大地改进了已有的描述符,如检索、分类、分割和对应等各种任务中的热核签名,证明了LBO特征基对全局和高度局部学习设置的适应性。
{"title":"LBONet: Supervised Spectral Descriptors for Shape Analysis","authors":"Oguzhan Yigit;Richard C. Wilson","doi":"10.1109/TPAMI.2025.3600873","DOIUrl":"10.1109/TPAMI.2025.3600873","url":null,"abstract":"The Laplace-Beltrami operator has established itself in the field of non-rigid shape analysis due to its many useful properties such as being invariant under isometric transformation, having a countable eigensystem forming an orthonormal basis, and fully characterizing geodesic distances of the manifold. However, this invariancy only applies under isometric deformations, which leads to a performance breakdown in many real-world applications. In recent years emphasis has been placed upon extracting optimal features using deep learning methods, however spectral signatures play a crucial role and still add value. In this paper we take a step back, revisiting the LBO and proposing a supervised way to learn several operators on a manifold. Depending on the task, by applying these functions, we can train the LBO eigenbasis to be more task-specific. The optimization of the LBO leads to enormous improvements to established descriptors such as the heat kernel signature in various tasks such as retrieval, classification, segmentation, and correspondence, proving the adaptation of the LBO eigenbasis to both global and highly local learning settings.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"12054-12068"},"PeriodicalIF":18.6,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144898541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reconstructing Satellites in 3D from Amateur Telescope Images. 利用业余望远镜图像重建三维卫星。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3599949
Zhiming Chang, Boyang Liu, Yifei Xia, Youming Guo, Boxin Shi, He Sun

Monitoring space objects is crucial for space situational awareness, yet reconstructing 3D satellite models from ground-based telescope images is super challenging due to atmospheric turbulence, long observation distances, limited viewpoints, and low signal-to-noise ratios. In this paper, we propose a novel computational imaging framework that overcomes these obstacles by integrating a hybrid image pre-processing pipeline with a joint pose estimation and 3D reconstruction module based on controlled Gaussian Splatting (GS) and Branch-and-Bound (BnB) search. We validate our approach on both synthetic satellite datasets and on-sky observations of China's Tiangong Space Station and the International Space Station, achieving robust 3D reconstructions of low-Earth orbit satellites from ground-based data. Quantitative evaluations using SSIM, PSNR, LPIPS, and Chamfer Distance demonstrate that our method outperforms state-of-the-art NeRF-based approaches, and ablation studies confirm the critical role of each component. Our framework enables high-fidelity 3D satellite monitoring from Earth, offering a cost-effective alternative for space situational awareness. Project page: https://ai4scientificimaging.org/3DSatellites.

监测空间物体对于空间态势感知至关重要,然而,由于大气湍流、观测距离长、视点有限和信噪比低,从地面望远镜图像重建3D卫星模型极具挑战性。在本文中,我们提出了一种新的计算成像框架,通过将混合图像预处理管道与基于可控高斯飞溅(GS)和分支边界(BnB)搜索的联合姿态估计和三维重建模块集成在一起,克服了这些障碍。我们在中国天宫空间站和国际空间站的合成卫星数据集和天空观测数据上验证了我们的方法,从地面数据实现了低地球轨道卫星的强大3D重建。使用SSIM、PSNR、LPIPS和Chamfer Distance进行的定量评估表明,我们的方法优于最先进的基于nerf的方法,并且消融研究证实了每个组件的关键作用。我们的框架能够实现地球上的高保真3D卫星监测,为空间态势感知提供了一种具有成本效益的替代方案。项目页面:https://ai4scientificimaging.org/3DSatellites。
{"title":"Reconstructing Satellites in 3D from Amateur Telescope Images.","authors":"Zhiming Chang, Boyang Liu, Yifei Xia, Youming Guo, Boxin Shi, He Sun","doi":"10.1109/TPAMI.2025.3599949","DOIUrl":"10.1109/TPAMI.2025.3599949","url":null,"abstract":"<p><p>Monitoring space objects is crucial for space situational awareness, yet reconstructing 3D satellite models from ground-based telescope images is super challenging due to atmospheric turbulence, long observation distances, limited viewpoints, and low signal-to-noise ratios. In this paper, we propose a novel computational imaging framework that overcomes these obstacles by integrating a hybrid image pre-processing pipeline with a joint pose estimation and 3D reconstruction module based on controlled Gaussian Splatting (GS) and Branch-and-Bound (BnB) search. We validate our approach on both synthetic satellite datasets and on-sky observations of China's Tiangong Space Station and the International Space Station, achieving robust 3D reconstructions of low-Earth orbit satellites from ground-based data. Quantitative evaluations using SSIM, PSNR, LPIPS, and Chamfer Distance demonstrate that our method outperforms state-of-the-art NeRF-based approaches, and ablation studies confirm the critical role of each component. Our framework enables high-fidelity 3D satellite monitoring from Earth, offering a cost-effective alternative for space situational awareness. Project page: https://ai4scientificimaging.org/3DSatellites.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Frequency-Based Comprehensive Prompt Learning for Vision-Language Models 基于频率的视觉语言模型综合提示学习。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3599830
Liangchen Liu;Nannan Wang;Chen Chen;Decheng Liu;Xi Yang;Xinbo Gao;Tongliang Liu
This paper targets to learn multiple comprehensive text prompts that can describe the visual concepts from coarse to fine, thereby endowing pre-trained VLMs with better transfer ability to various downstream tasks. We focus on exploring this idea on transformer-based VLMs since this kind of architecture achieves more compelling performances than CNN-based ones. Unfortunately, unlike CNNs, the transformer-based visual encoder of pre-trained VLMs cannot naturally provide discriminative and representative local visual information. To solve this problem, we propose Frequency-based Comprehensive Prompt Learning (FCPrompt) to excavate representative local visual information from the redundant output features of the visual encoder. FCPrompt transforms these features into frequency domain via Discrete Cosine Transform (DCT). Taking the advantages of energy concentration and information orthogonality of DCT, we can obtain compact, informative and disentangled local visual information by leveraging specific frequency components of the transformed frequency features. To better fit with transformer architectures, FCPrompt further adopts and optimizes different text prompts to respectively align with the global and frequency-based local visual information via a dual-branch framework. Finally, the learned text prompts can thus describe the entire visual concepts from coarse to fine comprehensively. Extensive experiments indicate that FCPrompt achieves the state-of-the-art performances on various benchmarks.
本文的目标是学习多个能够从粗到细描述视觉概念的综合文本提示,从而使预训练的vlm具有更好的向各种下游任务迁移的能力。我们专注于在基于变压器的vlm上探索这个想法,因为这种架构比基于cnn的架构实现了更引人注目的性能。不幸的是,与cnn不同,预训练vlm的基于变压器的视觉编码器不能自然地提供判别性和代表性的局部视觉信息。为了解决这个问题,我们提出了基于频率的综合提示学习(FCPrompt),从视觉编码器的冗余输出特征中挖掘具有代表性的局部视觉信息。FCPrompt通过离散余弦变换(DCT)将这些特征转换到频域。利用DCT的能量集中和信息正交性,利用变换后的频率特征的特定频率分量,可以得到紧凑、信息量大、解纠缠的局部视觉信息。为了更好地适应变压器架构,FCPrompt进一步采用并优化了不同的文本提示,通过双分支框架分别与全局和基于频率的局部视觉信息对齐。最后,学习到的文本提示可以从粗到细全面地描述整个视觉概念。大量的实验表明,FCPrompt在各种基准测试中都达到了最先进的性能。代码可从https://github.com/llcllc1997/FCPrompt获得。
{"title":"Frequency-Based Comprehensive Prompt Learning for Vision-Language Models","authors":"Liangchen Liu;Nannan Wang;Chen Chen;Decheng Liu;Xi Yang;Xinbo Gao;Tongliang Liu","doi":"10.1109/TPAMI.2025.3599830","DOIUrl":"10.1109/TPAMI.2025.3599830","url":null,"abstract":"This paper targets to learn multiple comprehensive text prompts that can describe the visual concepts from coarse to fine, thereby endowing pre-trained VLMs with better transfer ability to various downstream tasks. We focus on exploring this idea on transformer-based VLMs since this kind of architecture achieves more compelling performances than CNN-based ones. Unfortunately, unlike CNNs, the transformer-based visual encoder of pre-trained VLMs cannot naturally provide discriminative and representative local visual information. To solve this problem, we propose Frequency-based Comprehensive Prompt Learning (FCPrompt) to excavate representative local visual information from the redundant output features of the visual encoder. FCPrompt transforms these features into frequency domain via Discrete Cosine Transform (DCT). Taking the advantages of energy concentration and information orthogonality of DCT, we can obtain compact, informative and disentangled local visual information by leveraging specific frequency components of the transformed frequency features. To better fit with transformer architectures, FCPrompt further adopts and optimizes different text prompts to respectively align with the global and frequency-based local visual information via a dual-branch framework. Finally, the learned text prompts can thus describe the entire visual concepts from coarse to fine comprehensively. Extensive experiments indicate that FCPrompt achieves the state-of-the-art performances on various benchmarks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11974-11989"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Constrained Clustering Ensemble 自约束聚类集成。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600256
Wei Wei;Jianguo Wu;Xinyao Guo;Jing Yan;Jiye Liang
Existing clustering ensemble methods typically fuse all base clusterings in one shot under unsupervised settings, making it difficult to distinguish the quality of individual base clusterings and to exploit latent prior knowledge; consequently, their adaptability to data distributions and overall performance are limited. To address these issues, this paper proposes the Self-Constrained Clustering Ensemble (SCCE) algorithm. SCCE treats the pseudolabels automatically generated from current clustering results as selfsupervised signals and performs metric learning to obtain a linear transformation that enlarges interclass distances while compressing intraclass distances. The base clusterings are then reclustered in the new metric space to enhance separability and consistency. Afterward, ensemble updating is iteratively applied, forming a self-driven closed loop that continuously improves model performance. Theoretical analysis shows that the model converges efficiently via alternating optimization, with computational complexity on the same order as mainstream methods. Experiments on public datasets demonstrate that the proposed algorithm significantly outperforms representative clustering ensemble approaches, validating its effectiveness and robustness in scenarios lacking external supervision.
现有的聚类集成方法通常在无监督的情况下将所有的碱基聚类一次性融合在一起,难以区分单个碱基聚类的质量和利用潜在的先验知识;因此,它们对数据分布和整体性能的适应性受到限制。为了解决这些问题,本文提出了自约束聚类集成(SCCE)算法。SCCE将当前聚类结果自动生成的伪标签作为自监督信号,并进行度量学习,得到在压缩类内距离的同时增大类间距离的线性变换。然后在新的度量空间中对基本聚类进行重新聚类,以增强可分离性和一致性。然后,迭代地应用集成更新,形成一个不断提高模型性能的自驱动闭环。理论分析表明,通过交替优化,该模型收敛效率高,计算复杂度与主流方法在同一数量级。在公共数据集上的实验表明,该算法显著优于代表性聚类集成方法,验证了其在缺乏外部监督的场景下的有效性和鲁棒性。
{"title":"Self-Constrained Clustering Ensemble","authors":"Wei Wei;Jianguo Wu;Xinyao Guo;Jing Yan;Jiye Liang","doi":"10.1109/TPAMI.2025.3600256","DOIUrl":"10.1109/TPAMI.2025.3600256","url":null,"abstract":"Existing clustering ensemble methods typically fuse all base clusterings in one shot under unsupervised settings, making it difficult to distinguish the quality of individual base clusterings and to exploit latent prior knowledge; consequently, their adaptability to data distributions and overall performance are limited. To address these issues, this paper proposes the Self-Constrained Clustering Ensemble (SCCE) algorithm. SCCE treats the pseudolabels automatically generated from current clustering results as selfsupervised signals and performs metric learning to obtain a linear transformation that enlarges interclass distances while compressing intraclass distances. The base clusterings are then reclustered in the new metric space to enhance separability and consistency. Afterward, ensemble updating is iteratively applied, forming a self-driven closed loop that continuously improves model performance. Theoretical analysis shows that the model converges efficiently via alternating optimization, with computational complexity on the same order as mainstream methods. Experiments on public datasets demonstrate that the proposed algorithm significantly outperforms representative clustering ensemble approaches, validating its effectiveness and robustness in scenarios lacking external supervision.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11946-11960"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MicroDreamer: Efficient 3D Generation in $sim$20 Seconds by Score-based Iterative Reconstruction. MicroDreamer:通过基于分数的迭代重建,在$sim$20秒内实现高效3D生成。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600494
Luxi Chen, Zhengyi Wang, Zihan Zhou, Tingting Gao, Hang Su, Jun Zhu, Chongxuan Li

Optimization-based approaches, such as score distillation sampling (SDS), show promise in zero-shot 3D generation but suffer from low efficiency, primarily due to the high number of function evaluations (NFEs) required for each sample and the limitation of optimization confined to latent space. This paper introduces score-based iterative reconstruction (SIR), an efficient and general algorithm mimicking a differentiable 3D reconstruction process to reduce the NFEs and enable optimization in pixel space. Given a single set of images sampled from a multi-view score-based diffusion model, SIR repeatedly optimizes 3D parameters, unlike the single-step optimization in SDS. With other improvements in training, we present an efficient approach called MicroDreamer that generally applies to various 3D representations and 3D generation tasks. In particular, MicroDreamer is 5-20 times faster than SDS in generating neural radiance field while retaining a comparable performance and takes about 20 seconds to create meshes from 3D Gaussian splatting on a single A100 GPU, halving the time of the fastest optimization-based baseline DreamGaussian with significantly superior performance compared to the measurement standard deviation. Our code is available at https://github.com/ML-GSAI/MicroDreamer.

基于优化的方法,如分数蒸馏采样(SDS),在零射击3D生成中表现出希望,但效率较低,主要是由于每个样本需要大量的功能评估(nfe)以及优化局限于潜在空间。本文介绍了基于分数的迭代重建(SIR),这是一种模拟可微三维重建过程的高效通用算法,可以减少nfe并实现像素空间的优化。给定从基于分数的多视图扩散模型中采样的一组图像,SIR会重复优化3D参数,而不像SDS中的单步优化。随着训练的其他改进,我们提出了一种称为MicroDreamer的有效方法,该方法通常适用于各种3D表示和3D生成任务。特别是,microdream在生成神经辐射场方面比SDS快5-20倍,同时保持了相当的性能,并且在单个A100 GPU上从3D高斯飞溅创建网格大约需要20秒,将基于最快优化的基线DreamGaussian的时间缩短了一半,与测量标准偏差相比,性能显着优于SDS。我们的代码可在https://github.com/ML-GSAI/MicroDreamer上获得。
{"title":"MicroDreamer: Efficient 3D Generation in $sim$20 Seconds by Score-based Iterative Reconstruction.","authors":"Luxi Chen, Zhengyi Wang, Zihan Zhou, Tingting Gao, Hang Su, Jun Zhu, Chongxuan Li","doi":"10.1109/TPAMI.2025.3600494","DOIUrl":"10.1109/TPAMI.2025.3600494","url":null,"abstract":"<p><p>Optimization-based approaches, such as score distillation sampling (SDS), show promise in zero-shot 3D generation but suffer from low efficiency, primarily due to the high number of function evaluations (NFEs) required for each sample and the limitation of optimization confined to latent space. This paper introduces score-based iterative reconstruction (SIR), an efficient and general algorithm mimicking a differentiable 3D reconstruction process to reduce the NFEs and enable optimization in pixel space. Given a single set of images sampled from a multi-view score-based diffusion model, SIR repeatedly optimizes 3D parameters, unlike the single-step optimization in SDS. With other improvements in training, we present an efficient approach called MicroDreamer that generally applies to various 3D representations and 3D generation tasks. In particular, MicroDreamer is 5-20 times faster than SDS in generating neural radiance field while retaining a comparable performance and takes about 20 seconds to create meshes from 3D Gaussian splatting on a single A100 GPU, halving the time of the fastest optimization-based baseline DreamGaussian with significantly superior performance compared to the measurement standard deviation. Our code is available at https://github.com/ML-GSAI/MicroDreamer.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SKDF: A Simple Knowledge Distillation Framework for Distilling Open-Vocabulary Knowledge to Open-World Object Detector SKDF:将开放词汇知识提炼到开放世界对象检测器的简单知识蒸馏框架。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600435
Shuailei Ma;Yuefeng Wang;Ying Wei;Enming Zhang;Jiaqi Fan;Xinyu Sun;Peihao Chen
Open World Object Detection (OWOD) is a novel computer vision task with a considerable challenge, bridging the gap between classic object detection (OD) and real-world object detection. In addition to detecting and classifying seen/known objects, OWOD algorithms are expected to localize all potential unseen/unknown objects and incrementally learn them. The large pre-trained vision-language grounding models (VLM, e.g., GLIP) have rich knowledge about the open world, but are limited by text prompts and cannot localize indescribable objects. However, there are many detection scenarios in which pre-defined language descriptions are unavailable during inference. In this paper, we attempt to specialize the VLM model for OWOD tasks by distilling its open-world knowledge into a language-agnostic detector. Surprisingly, we observe that the simple knowledge distillation approach leads to unexpected performance for unknown object detection, even with a small amount of data. Unfortunately, knowledge distillation for unknown objects severely affects the learning of detectors with conventional structures, leading to catastrophic damage to the model’s ability to learn about known objects. To alleviate these problems, we propose the down-weight training strategy for knowledge distillation from vision-language model to single visual modality one. Meanwhile, we propose the cascade decoupled decoders that decouple the learning of localization and recognition to reduce the impact of category interactions of known and unknown objects on the localization learning process. Ablation experiments demonstrate that both of them are effective in mitigating the impact of open-world knowledge distillation on the learning of known objects. Additionally, to alleviate the current lack of comprehensive benchmarks for evaluating the ability of the open-world detector to detect unknown objects in the open world, we refine the benchmark for evaluating the performance of unknown object detection by augmenting annotations for unknown objects which we name“IntensiveSet$scriptstylespadesuit$”. Comprehensive experiments performed on OWOD, MS-COCO, and our proposed benchmarks demonstrate the effectiveness of our methods.
开放世界目标检测(OWOD)是一项具有相当大挑战的新型计算机视觉任务,它弥合了经典目标检测(OD)和现实世界目标检测之间的差距。除了检测和分类可见/已知对象外,OWOD算法还有望定位所有潜在的未见/未知对象并逐步学习它们。大型预训练的视觉语言基础模型(VLM,如GLIP)具有丰富的开放世界知识,但受文本提示的限制,无法定位不可描述的对象。然而,在许多检测场景中,预定义的语言描述在推理期间不可用。在本文中,我们试图通过将其开放世界知识提炼成语言不可知检测器来专门化用于OWOD任务的VLM模型。令人惊讶的是,我们观察到简单的知识蒸馏方法即使在少量数据的情况下也会导致未知对象检测的意想不到的性能。不幸的是,未知对象的知识蒸馏严重影响了传统结构检测器的学习,导致模型学习已知对象的能力受到灾难性的破坏。为了解决这些问题,我们提出了从视觉语言模型到单一视觉模态模型的知识升华的降权训练策略。同时,我们提出了级联解耦解码器,将定位和识别的学习解耦,以减少已知和未知对象的类别交互对定位学习过程的影响。烧蚀实验表明,这两种方法都能有效缓解开放世界知识蒸馏对已知对象学习的影响。此外,为了缓解目前缺乏评估开放世界检测器在开放世界中检测未知物体能力的综合基准的问题,我们通过增加未知物体的注释来改进评估未知物体检测性能的基准,我们将其命名为“intenveset $spadesuit$”。在OWOD, MS-COCO和我们提出的基准上进行的综合实验证明了我们方法的有效性。
{"title":"SKDF: A Simple Knowledge Distillation Framework for Distilling Open-Vocabulary Knowledge to Open-World Object Detector","authors":"Shuailei Ma;Yuefeng Wang;Ying Wei;Enming Zhang;Jiaqi Fan;Xinyu Sun;Peihao Chen","doi":"10.1109/TPAMI.2025.3600435","DOIUrl":"10.1109/TPAMI.2025.3600435","url":null,"abstract":"Open World Object Detection (OWOD) is a novel computer vision task with a considerable challenge, bridging the gap between classic object detection (OD) and real-world object detection. In addition to detecting and classifying seen/known objects, OWOD algorithms are expected to localize all potential unseen/unknown objects and incrementally learn them. The large pre-trained vision-language grounding models (VLM, e.g., GLIP) have rich knowledge about the open world, but are limited by text prompts and cannot localize indescribable objects. However, there are many detection scenarios in which pre-defined language descriptions are unavailable during inference. In this paper, we attempt to specialize the VLM model for OWOD tasks by distilling its open-world knowledge into a language-agnostic detector. Surprisingly, we observe that the simple <b>knowledge distillation</b> approach leads to unexpected performance for unknown object detection, even with a small amount of data. Unfortunately, knowledge distillation for unknown objects severely affects the learning of detectors with conventional structures, leading to catastrophic damage to the model’s ability to learn about known objects. To alleviate these problems, we propose the <b>down-weight training strategy</b> for knowledge distillation from vision-language model to single visual modality one. Meanwhile, we propose the <b>cascade decoupled decoders</b> that decouple the learning of localization and recognition to reduce the impact of category interactions of known and unknown objects on the localization learning process. Ablation experiments demonstrate that both of them are effective in mitigating the impact of open-world knowledge distillation on the learning of known objects. Additionally, to alleviate the current lack of comprehensive benchmarks for evaluating the ability of the open-world detector to detect unknown objects in the open world, we refine the benchmark for evaluating the performance of unknown object detection by augmenting annotations for unknown objects which we name“<b>IntensiveSet</b><inline-formula><tex-math>$scriptstylespadesuit$</tex-math></inline-formula>”. Comprehensive experiments performed on OWOD, MS-COCO, and our proposed benchmarks demonstrate the effectiveness of our methods.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11738-11752"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1