首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
FaceTracer: Unveiling Source Identities From Swapped Face Images and Videos for Fraud Prevention FACETRACER:从交换的人脸图像和视频中揭示源身份,以防止欺诈
IF 18.6 Pub Date : 2025-08-20 DOI: 10.1109/TPAMI.2025.3601141
Zhongyi Zhang;Jie Zhang;Wenbo Zhou;Xinghui Zhou;Qing Guo;Weiming Zhang;Tianwei Zhang;Nenghai Yu
Face-swapping techniques have advanced rapidly with the evolution of deep learning, leading to widespread use and growing concerns about potential misuse, especially in cases of fraud. While many efforts have focused on detecting swapped face images or videos, these methods are insufficient for tracing the malicious users behind fraudulent activities. Intrusive watermark-based approaches also fail to trace unmarked identities, limiting their practical utility. To address these challenges, we introduce FaceTracer, the first non-intrusive framework specifically designed to trace the identity of the source person from swapped face images or videos. Specifically, FaceTracer leverages a disentanglement module that effectively suppresses identity information related to the target person while isolating the identity features of the source person. This allows us to extract robust identity information that can directly link the swapped face back to the original individual, aiding in uncovering the actors behind fraudulent activities. Extensive experiments demonstrate FaceTracer’s effectiveness across various face-swapping techniques, successfully identifying the source person in swapped content and enabling the tracing of malicious actors involved in fraudulent activities. Additionally, FaceTracer shows strong transferability to unseen face-swapping methods including commercial applications and robustness against transmission distortions and adaptive attacks.
人脸交换技术随着深度学习的发展而迅速发展,导致广泛使用和越来越多的人担心潜在的滥用,特别是在欺诈案件中。虽然很多努力都集中在检测交换的人脸图像或视频上,但这些方法不足以追踪欺诈活动背后的恶意用户。基于侵入式水印的方法也无法追踪未标记的身份,限制了它们的实际用途。为了应对这些挑战,我们引入了FaceTracer,这是第一个非侵入性框架,专门用于从交换的面部图像或视频中追踪源人员的身份。具体来说,FaceTracer利用了一个解缠模块,可以有效地抑制与目标人员相关的身份信息,同时隔离源人员的身份特征。这使我们能够提取可靠的身份信息,这些信息可以直接将交换后的脸与原始个人联系起来,从而帮助发现欺诈活动背后的参与者。大量的实验证明了FaceTracer在各种面部交换技术中的有效性,成功地识别交换内容中的源人员,并能够跟踪涉及欺诈活动的恶意行为者。此外,FaceTracer显示出强大的可移植性,可用于不可见的面部交换方法,包括商业应用和抗传输失真和自适应攻击的鲁棒性。
{"title":"FaceTracer: Unveiling Source Identities From Swapped Face Images and Videos for Fraud Prevention","authors":"Zhongyi Zhang;Jie Zhang;Wenbo Zhou;Xinghui Zhou;Qing Guo;Weiming Zhang;Tianwei Zhang;Nenghai Yu","doi":"10.1109/TPAMI.2025.3601141","DOIUrl":"10.1109/TPAMI.2025.3601141","url":null,"abstract":"Face-swapping techniques have advanced rapidly with the evolution of deep learning, leading to widespread use and growing concerns about potential misuse, especially in cases of fraud. While many efforts have focused on detecting swapped face images or videos, these methods are insufficient for tracing the malicious users behind fraudulent activities. Intrusive watermark-based approaches also fail to trace unmarked identities, limiting their practical utility. To address these challenges, we introduce <sc>FaceTracer</small>, the first non-intrusive framework specifically designed to trace the identity of the source person from swapped face images or videos. Specifically, <sc>FaceTracer</small> leverages a disentanglement module that effectively suppresses identity information related to the target person while isolating the identity features of the source person. This allows us to extract robust identity information that can directly link the swapped face back to the original individual, aiding in uncovering the actors behind fraudulent activities. Extensive experiments demonstrate <sc>FaceTracer</small>’s effectiveness across various face-swapping techniques, successfully identifying the source person in swapped content and enabling the tracing of malicious actors involved in fraudulent activities. Additionally, <sc>FaceTracer</small> shows strong transferability to unseen face-swapping methods including commercial applications and robustness against transmission distortions and adaptive attacks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"12021-12037"},"PeriodicalIF":18.6,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144898544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LBONet: Supervised Spectral Descriptors for Shape Analysis 用于形状分析的监督谱描述子
IF 18.6 Pub Date : 2025-08-20 DOI: 10.1109/TPAMI.2025.3600873
Oguzhan Yigit;Richard C. Wilson
The Laplace-Beltrami operator has established itself in the field of non-rigid shape analysis due to its many useful properties such as being invariant under isometric transformation, having a countable eigensystem forming an orthonormal basis, and fully characterizing geodesic distances of the manifold. However, this invariancy only applies under isometric deformations, which leads to a performance breakdown in many real-world applications. In recent years emphasis has been placed upon extracting optimal features using deep learning methods, however spectral signatures play a crucial role and still add value. In this paper we take a step back, revisiting the LBO and proposing a supervised way to learn several operators on a manifold. Depending on the task, by applying these functions, we can train the LBO eigenbasis to be more task-specific. The optimization of the LBO leads to enormous improvements to established descriptors such as the heat kernel signature in various tasks such as retrieval, classification, segmentation, and correspondence, proving the adaptation of the LBO eigenbasis to both global and highly local learning settings.
Laplace-Beltrami算子由于其在等距变换下的不变性、具有构成正交基的可数本征系统以及充分表征流形的测地线距离等特性而在非刚性形状分析领域中确立了自己的地位。然而,这种不变性只适用于等距变形,这会导致许多实际应用程序的性能崩溃。近年来,人们把重点放在了使用深度学习方法提取最优特征上,然而光谱特征起着至关重要的作用,仍然具有附加价值。在本文中,我们退后一步,重新审视杠杆收购,并提出一种监督的方法来学习流形上的几个算子。根据任务的不同,通过应用这些函数,我们可以训练LBO特征基,使其更加特定于任务。LBO的优化极大地改进了已有的描述符,如检索、分类、分割和对应等各种任务中的热核签名,证明了LBO特征基对全局和高度局部学习设置的适应性。
{"title":"LBONet: Supervised Spectral Descriptors for Shape Analysis","authors":"Oguzhan Yigit;Richard C. Wilson","doi":"10.1109/TPAMI.2025.3600873","DOIUrl":"10.1109/TPAMI.2025.3600873","url":null,"abstract":"The Laplace-Beltrami operator has established itself in the field of non-rigid shape analysis due to its many useful properties such as being invariant under isometric transformation, having a countable eigensystem forming an orthonormal basis, and fully characterizing geodesic distances of the manifold. However, this invariancy only applies under isometric deformations, which leads to a performance breakdown in many real-world applications. In recent years emphasis has been placed upon extracting optimal features using deep learning methods, however spectral signatures play a crucial role and still add value. In this paper we take a step back, revisiting the LBO and proposing a supervised way to learn several operators on a manifold. Depending on the task, by applying these functions, we can train the LBO eigenbasis to be more task-specific. The optimization of the LBO leads to enormous improvements to established descriptors such as the heat kernel signature in various tasks such as retrieval, classification, segmentation, and correspondence, proving the adaptation of the LBO eigenbasis to both global and highly local learning settings.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"12054-12068"},"PeriodicalIF":18.6,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144898541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reconstructing Satellites in 3D from Amateur Telescope Images. 利用业余望远镜图像重建三维卫星。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3599949
Zhiming Chang, Boyang Liu, Yifei Xia, Youming Guo, Boxin Shi, He Sun

Monitoring space objects is crucial for space situational awareness, yet reconstructing 3D satellite models from ground-based telescope images is super challenging due to atmospheric turbulence, long observation distances, limited viewpoints, and low signal-to-noise ratios. In this paper, we propose a novel computational imaging framework that overcomes these obstacles by integrating a hybrid image pre-processing pipeline with a joint pose estimation and 3D reconstruction module based on controlled Gaussian Splatting (GS) and Branch-and-Bound (BnB) search. We validate our approach on both synthetic satellite datasets and on-sky observations of China's Tiangong Space Station and the International Space Station, achieving robust 3D reconstructions of low-Earth orbit satellites from ground-based data. Quantitative evaluations using SSIM, PSNR, LPIPS, and Chamfer Distance demonstrate that our method outperforms state-of-the-art NeRF-based approaches, and ablation studies confirm the critical role of each component. Our framework enables high-fidelity 3D satellite monitoring from Earth, offering a cost-effective alternative for space situational awareness. Project page: https://ai4scientificimaging.org/3DSatellites.

监测空间物体对于空间态势感知至关重要,然而,由于大气湍流、观测距离长、视点有限和信噪比低,从地面望远镜图像重建3D卫星模型极具挑战性。在本文中,我们提出了一种新的计算成像框架,通过将混合图像预处理管道与基于可控高斯飞溅(GS)和分支边界(BnB)搜索的联合姿态估计和三维重建模块集成在一起,克服了这些障碍。我们在中国天宫空间站和国际空间站的合成卫星数据集和天空观测数据上验证了我们的方法,从地面数据实现了低地球轨道卫星的强大3D重建。使用SSIM、PSNR、LPIPS和Chamfer Distance进行的定量评估表明,我们的方法优于最先进的基于nerf的方法,并且消融研究证实了每个组件的关键作用。我们的框架能够实现地球上的高保真3D卫星监测,为空间态势感知提供了一种具有成本效益的替代方案。项目页面:https://ai4scientificimaging.org/3DSatellites。
{"title":"Reconstructing Satellites in 3D from Amateur Telescope Images.","authors":"Zhiming Chang, Boyang Liu, Yifei Xia, Youming Guo, Boxin Shi, He Sun","doi":"10.1109/TPAMI.2025.3599949","DOIUrl":"10.1109/TPAMI.2025.3599949","url":null,"abstract":"<p><p>Monitoring space objects is crucial for space situational awareness, yet reconstructing 3D satellite models from ground-based telescope images is super challenging due to atmospheric turbulence, long observation distances, limited viewpoints, and low signal-to-noise ratios. In this paper, we propose a novel computational imaging framework that overcomes these obstacles by integrating a hybrid image pre-processing pipeline with a joint pose estimation and 3D reconstruction module based on controlled Gaussian Splatting (GS) and Branch-and-Bound (BnB) search. We validate our approach on both synthetic satellite datasets and on-sky observations of China's Tiangong Space Station and the International Space Station, achieving robust 3D reconstructions of low-Earth orbit satellites from ground-based data. Quantitative evaluations using SSIM, PSNR, LPIPS, and Chamfer Distance demonstrate that our method outperforms state-of-the-art NeRF-based approaches, and ablation studies confirm the critical role of each component. Our framework enables high-fidelity 3D satellite monitoring from Earth, offering a cost-effective alternative for space situational awareness. Project page: https://ai4scientificimaging.org/3DSatellites.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Frequency-Based Comprehensive Prompt Learning for Vision-Language Models 基于频率的视觉语言模型综合提示学习。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3599830
Liangchen Liu;Nannan Wang;Chen Chen;Decheng Liu;Xi Yang;Xinbo Gao;Tongliang Liu
This paper targets to learn multiple comprehensive text prompts that can describe the visual concepts from coarse to fine, thereby endowing pre-trained VLMs with better transfer ability to various downstream tasks. We focus on exploring this idea on transformer-based VLMs since this kind of architecture achieves more compelling performances than CNN-based ones. Unfortunately, unlike CNNs, the transformer-based visual encoder of pre-trained VLMs cannot naturally provide discriminative and representative local visual information. To solve this problem, we propose Frequency-based Comprehensive Prompt Learning (FCPrompt) to excavate representative local visual information from the redundant output features of the visual encoder. FCPrompt transforms these features into frequency domain via Discrete Cosine Transform (DCT). Taking the advantages of energy concentration and information orthogonality of DCT, we can obtain compact, informative and disentangled local visual information by leveraging specific frequency components of the transformed frequency features. To better fit with transformer architectures, FCPrompt further adopts and optimizes different text prompts to respectively align with the global and frequency-based local visual information via a dual-branch framework. Finally, the learned text prompts can thus describe the entire visual concepts from coarse to fine comprehensively. Extensive experiments indicate that FCPrompt achieves the state-of-the-art performances on various benchmarks.
本文的目标是学习多个能够从粗到细描述视觉概念的综合文本提示,从而使预训练的vlm具有更好的向各种下游任务迁移的能力。我们专注于在基于变压器的vlm上探索这个想法,因为这种架构比基于cnn的架构实现了更引人注目的性能。不幸的是,与cnn不同,预训练vlm的基于变压器的视觉编码器不能自然地提供判别性和代表性的局部视觉信息。为了解决这个问题,我们提出了基于频率的综合提示学习(FCPrompt),从视觉编码器的冗余输出特征中挖掘具有代表性的局部视觉信息。FCPrompt通过离散余弦变换(DCT)将这些特征转换到频域。利用DCT的能量集中和信息正交性,利用变换后的频率特征的特定频率分量,可以得到紧凑、信息量大、解纠缠的局部视觉信息。为了更好地适应变压器架构,FCPrompt进一步采用并优化了不同的文本提示,通过双分支框架分别与全局和基于频率的局部视觉信息对齐。最后,学习到的文本提示可以从粗到细全面地描述整个视觉概念。大量的实验表明,FCPrompt在各种基准测试中都达到了最先进的性能。代码可从https://github.com/llcllc1997/FCPrompt获得。
{"title":"Frequency-Based Comprehensive Prompt Learning for Vision-Language Models","authors":"Liangchen Liu;Nannan Wang;Chen Chen;Decheng Liu;Xi Yang;Xinbo Gao;Tongliang Liu","doi":"10.1109/TPAMI.2025.3599830","DOIUrl":"10.1109/TPAMI.2025.3599830","url":null,"abstract":"This paper targets to learn multiple comprehensive text prompts that can describe the visual concepts from coarse to fine, thereby endowing pre-trained VLMs with better transfer ability to various downstream tasks. We focus on exploring this idea on transformer-based VLMs since this kind of architecture achieves more compelling performances than CNN-based ones. Unfortunately, unlike CNNs, the transformer-based visual encoder of pre-trained VLMs cannot naturally provide discriminative and representative local visual information. To solve this problem, we propose Frequency-based Comprehensive Prompt Learning (FCPrompt) to excavate representative local visual information from the redundant output features of the visual encoder. FCPrompt transforms these features into frequency domain via Discrete Cosine Transform (DCT). Taking the advantages of energy concentration and information orthogonality of DCT, we can obtain compact, informative and disentangled local visual information by leveraging specific frequency components of the transformed frequency features. To better fit with transformer architectures, FCPrompt further adopts and optimizes different text prompts to respectively align with the global and frequency-based local visual information via a dual-branch framework. Finally, the learned text prompts can thus describe the entire visual concepts from coarse to fine comprehensively. Extensive experiments indicate that FCPrompt achieves the state-of-the-art performances on various benchmarks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11974-11989"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Constrained Clustering Ensemble 自约束聚类集成。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600256
Wei Wei;Jianguo Wu;Xinyao Guo;Jing Yan;Jiye Liang
Existing clustering ensemble methods typically fuse all base clusterings in one shot under unsupervised settings, making it difficult to distinguish the quality of individual base clusterings and to exploit latent prior knowledge; consequently, their adaptability to data distributions and overall performance are limited. To address these issues, this paper proposes the Self-Constrained Clustering Ensemble (SCCE) algorithm. SCCE treats the pseudolabels automatically generated from current clustering results as selfsupervised signals and performs metric learning to obtain a linear transformation that enlarges interclass distances while compressing intraclass distances. The base clusterings are then reclustered in the new metric space to enhance separability and consistency. Afterward, ensemble updating is iteratively applied, forming a self-driven closed loop that continuously improves model performance. Theoretical analysis shows that the model converges efficiently via alternating optimization, with computational complexity on the same order as mainstream methods. Experiments on public datasets demonstrate that the proposed algorithm significantly outperforms representative clustering ensemble approaches, validating its effectiveness and robustness in scenarios lacking external supervision.
现有的聚类集成方法通常在无监督的情况下将所有的碱基聚类一次性融合在一起,难以区分单个碱基聚类的质量和利用潜在的先验知识;因此,它们对数据分布和整体性能的适应性受到限制。为了解决这些问题,本文提出了自约束聚类集成(SCCE)算法。SCCE将当前聚类结果自动生成的伪标签作为自监督信号,并进行度量学习,得到在压缩类内距离的同时增大类间距离的线性变换。然后在新的度量空间中对基本聚类进行重新聚类,以增强可分离性和一致性。然后,迭代地应用集成更新,形成一个不断提高模型性能的自驱动闭环。理论分析表明,通过交替优化,该模型收敛效率高,计算复杂度与主流方法在同一数量级。在公共数据集上的实验表明,该算法显著优于代表性聚类集成方法,验证了其在缺乏外部监督的场景下的有效性和鲁棒性。
{"title":"Self-Constrained Clustering Ensemble","authors":"Wei Wei;Jianguo Wu;Xinyao Guo;Jing Yan;Jiye Liang","doi":"10.1109/TPAMI.2025.3600256","DOIUrl":"10.1109/TPAMI.2025.3600256","url":null,"abstract":"Existing clustering ensemble methods typically fuse all base clusterings in one shot under unsupervised settings, making it difficult to distinguish the quality of individual base clusterings and to exploit latent prior knowledge; consequently, their adaptability to data distributions and overall performance are limited. To address these issues, this paper proposes the Self-Constrained Clustering Ensemble (SCCE) algorithm. SCCE treats the pseudolabels automatically generated from current clustering results as selfsupervised signals and performs metric learning to obtain a linear transformation that enlarges interclass distances while compressing intraclass distances. The base clusterings are then reclustered in the new metric space to enhance separability and consistency. Afterward, ensemble updating is iteratively applied, forming a self-driven closed loop that continuously improves model performance. Theoretical analysis shows that the model converges efficiently via alternating optimization, with computational complexity on the same order as mainstream methods. Experiments on public datasets demonstrate that the proposed algorithm significantly outperforms representative clustering ensemble approaches, validating its effectiveness and robustness in scenarios lacking external supervision.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11946-11960"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MicroDreamer: Efficient 3D Generation in $sim$20 Seconds by Score-based Iterative Reconstruction. MicroDreamer:通过基于分数的迭代重建,在$sim$20秒内实现高效3D生成。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600494
Luxi Chen, Zhengyi Wang, Zihan Zhou, Tingting Gao, Hang Su, Jun Zhu, Chongxuan Li

Optimization-based approaches, such as score distillation sampling (SDS), show promise in zero-shot 3D generation but suffer from low efficiency, primarily due to the high number of function evaluations (NFEs) required for each sample and the limitation of optimization confined to latent space. This paper introduces score-based iterative reconstruction (SIR), an efficient and general algorithm mimicking a differentiable 3D reconstruction process to reduce the NFEs and enable optimization in pixel space. Given a single set of images sampled from a multi-view score-based diffusion model, SIR repeatedly optimizes 3D parameters, unlike the single-step optimization in SDS. With other improvements in training, we present an efficient approach called MicroDreamer that generally applies to various 3D representations and 3D generation tasks. In particular, MicroDreamer is 5-20 times faster than SDS in generating neural radiance field while retaining a comparable performance and takes about 20 seconds to create meshes from 3D Gaussian splatting on a single A100 GPU, halving the time of the fastest optimization-based baseline DreamGaussian with significantly superior performance compared to the measurement standard deviation. Our code is available at https://github.com/ML-GSAI/MicroDreamer.

基于优化的方法,如分数蒸馏采样(SDS),在零射击3D生成中表现出希望,但效率较低,主要是由于每个样本需要大量的功能评估(nfe)以及优化局限于潜在空间。本文介绍了基于分数的迭代重建(SIR),这是一种模拟可微三维重建过程的高效通用算法,可以减少nfe并实现像素空间的优化。给定从基于分数的多视图扩散模型中采样的一组图像,SIR会重复优化3D参数,而不像SDS中的单步优化。随着训练的其他改进,我们提出了一种称为MicroDreamer的有效方法,该方法通常适用于各种3D表示和3D生成任务。特别是,microdream在生成神经辐射场方面比SDS快5-20倍,同时保持了相当的性能,并且在单个A100 GPU上从3D高斯飞溅创建网格大约需要20秒,将基于最快优化的基线DreamGaussian的时间缩短了一半,与测量标准偏差相比,性能显着优于SDS。我们的代码可在https://github.com/ML-GSAI/MicroDreamer上获得。
{"title":"MicroDreamer: Efficient 3D Generation in $sim$20 Seconds by Score-based Iterative Reconstruction.","authors":"Luxi Chen, Zhengyi Wang, Zihan Zhou, Tingting Gao, Hang Su, Jun Zhu, Chongxuan Li","doi":"10.1109/TPAMI.2025.3600494","DOIUrl":"10.1109/TPAMI.2025.3600494","url":null,"abstract":"<p><p>Optimization-based approaches, such as score distillation sampling (SDS), show promise in zero-shot 3D generation but suffer from low efficiency, primarily due to the high number of function evaluations (NFEs) required for each sample and the limitation of optimization confined to latent space. This paper introduces score-based iterative reconstruction (SIR), an efficient and general algorithm mimicking a differentiable 3D reconstruction process to reduce the NFEs and enable optimization in pixel space. Given a single set of images sampled from a multi-view score-based diffusion model, SIR repeatedly optimizes 3D parameters, unlike the single-step optimization in SDS. With other improvements in training, we present an efficient approach called MicroDreamer that generally applies to various 3D representations and 3D generation tasks. In particular, MicroDreamer is 5-20 times faster than SDS in generating neural radiance field while retaining a comparable performance and takes about 20 seconds to create meshes from 3D Gaussian splatting on a single A100 GPU, halving the time of the fastest optimization-based baseline DreamGaussian with significantly superior performance compared to the measurement standard deviation. Our code is available at https://github.com/ML-GSAI/MicroDreamer.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SKDF: A Simple Knowledge Distillation Framework for Distilling Open-Vocabulary Knowledge to Open-World Object Detector SKDF:将开放词汇知识提炼到开放世界对象检测器的简单知识蒸馏框架。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600435
Shuailei Ma;Yuefeng Wang;Ying Wei;Enming Zhang;Jiaqi Fan;Xinyu Sun;Peihao Chen
Open World Object Detection (OWOD) is a novel computer vision task with a considerable challenge, bridging the gap between classic object detection (OD) and real-world object detection. In addition to detecting and classifying seen/known objects, OWOD algorithms are expected to localize all potential unseen/unknown objects and incrementally learn them. The large pre-trained vision-language grounding models (VLM, e.g., GLIP) have rich knowledge about the open world, but are limited by text prompts and cannot localize indescribable objects. However, there are many detection scenarios in which pre-defined language descriptions are unavailable during inference. In this paper, we attempt to specialize the VLM model for OWOD tasks by distilling its open-world knowledge into a language-agnostic detector. Surprisingly, we observe that the simple knowledge distillation approach leads to unexpected performance for unknown object detection, even with a small amount of data. Unfortunately, knowledge distillation for unknown objects severely affects the learning of detectors with conventional structures, leading to catastrophic damage to the model’s ability to learn about known objects. To alleviate these problems, we propose the down-weight training strategy for knowledge distillation from vision-language model to single visual modality one. Meanwhile, we propose the cascade decoupled decoders that decouple the learning of localization and recognition to reduce the impact of category interactions of known and unknown objects on the localization learning process. Ablation experiments demonstrate that both of them are effective in mitigating the impact of open-world knowledge distillation on the learning of known objects. Additionally, to alleviate the current lack of comprehensive benchmarks for evaluating the ability of the open-world detector to detect unknown objects in the open world, we refine the benchmark for evaluating the performance of unknown object detection by augmenting annotations for unknown objects which we name“IntensiveSet$scriptstylespadesuit$”. Comprehensive experiments performed on OWOD, MS-COCO, and our proposed benchmarks demonstrate the effectiveness of our methods.
开放世界目标检测(OWOD)是一项具有相当大挑战的新型计算机视觉任务,它弥合了经典目标检测(OD)和现实世界目标检测之间的差距。除了检测和分类可见/已知对象外,OWOD算法还有望定位所有潜在的未见/未知对象并逐步学习它们。大型预训练的视觉语言基础模型(VLM,如GLIP)具有丰富的开放世界知识,但受文本提示的限制,无法定位不可描述的对象。然而,在许多检测场景中,预定义的语言描述在推理期间不可用。在本文中,我们试图通过将其开放世界知识提炼成语言不可知检测器来专门化用于OWOD任务的VLM模型。令人惊讶的是,我们观察到简单的知识蒸馏方法即使在少量数据的情况下也会导致未知对象检测的意想不到的性能。不幸的是,未知对象的知识蒸馏严重影响了传统结构检测器的学习,导致模型学习已知对象的能力受到灾难性的破坏。为了解决这些问题,我们提出了从视觉语言模型到单一视觉模态模型的知识升华的降权训练策略。同时,我们提出了级联解耦解码器,将定位和识别的学习解耦,以减少已知和未知对象的类别交互对定位学习过程的影响。烧蚀实验表明,这两种方法都能有效缓解开放世界知识蒸馏对已知对象学习的影响。此外,为了缓解目前缺乏评估开放世界检测器在开放世界中检测未知物体能力的综合基准的问题,我们通过增加未知物体的注释来改进评估未知物体检测性能的基准,我们将其命名为“intenveset $spadesuit$”。在OWOD, MS-COCO和我们提出的基准上进行的综合实验证明了我们方法的有效性。
{"title":"SKDF: A Simple Knowledge Distillation Framework for Distilling Open-Vocabulary Knowledge to Open-World Object Detector","authors":"Shuailei Ma;Yuefeng Wang;Ying Wei;Enming Zhang;Jiaqi Fan;Xinyu Sun;Peihao Chen","doi":"10.1109/TPAMI.2025.3600435","DOIUrl":"10.1109/TPAMI.2025.3600435","url":null,"abstract":"Open World Object Detection (OWOD) is a novel computer vision task with a considerable challenge, bridging the gap between classic object detection (OD) and real-world object detection. In addition to detecting and classifying seen/known objects, OWOD algorithms are expected to localize all potential unseen/unknown objects and incrementally learn them. The large pre-trained vision-language grounding models (VLM, e.g., GLIP) have rich knowledge about the open world, but are limited by text prompts and cannot localize indescribable objects. However, there are many detection scenarios in which pre-defined language descriptions are unavailable during inference. In this paper, we attempt to specialize the VLM model for OWOD tasks by distilling its open-world knowledge into a language-agnostic detector. Surprisingly, we observe that the simple <b>knowledge distillation</b> approach leads to unexpected performance for unknown object detection, even with a small amount of data. Unfortunately, knowledge distillation for unknown objects severely affects the learning of detectors with conventional structures, leading to catastrophic damage to the model’s ability to learn about known objects. To alleviate these problems, we propose the <b>down-weight training strategy</b> for knowledge distillation from vision-language model to single visual modality one. Meanwhile, we propose the <b>cascade decoupled decoders</b> that decouple the learning of localization and recognition to reduce the impact of category interactions of known and unknown objects on the localization learning process. Ablation experiments demonstrate that both of them are effective in mitigating the impact of open-world knowledge distillation on the learning of known objects. Additionally, to alleviate the current lack of comprehensive benchmarks for evaluating the ability of the open-world detector to detect unknown objects in the open world, we refine the benchmark for evaluating the performance of unknown object detection by augmenting annotations for unknown objects which we name“<b>IntensiveSet</b><inline-formula><tex-math>$scriptstylespadesuit$</tex-math></inline-formula>”. Comprehensive experiments performed on OWOD, MS-COCO, and our proposed benchmarks demonstrate the effectiveness of our methods.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11738-11752"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MinD-3D++: Advancing fMRI-Based 3D Reconstruction With High-Quality Textured Mesh Generation and a Comprehensive Dataset MinD-3D++:使用高质量纹理网格生成和综合数据集推进基于fmri的3D重建。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3599860
Jianxiong Gao;Yanwei Fu;Yuqian Fu;Yun Wang;Xuelin Qian;Jianfeng Feng
Reconstructing 3D visuals from functional Magnetic Resonance Imaging (fMRI) data, introduced as Recon3DMind, is of significant interest to both cognitive neuroscience and computer vision. To advance this task, we present the fMRI-3D dataset, which includes data from 15 participants and showcases a total of 4,768 3D objects. The dataset consists of two components: fMRI-Shape, previously introduced and available at https://huggingface.co/datasets/Fudan-fMRI/fMRI-Shape, and fMRI-Objaverse, proposed in this paper and available at https://huggingface.co/datasets/Fudan-fMRI/fMRI-Objaverse. fMRI-Objaverse includes data from 5 subjects, 4 of whom are also part of the core set in fMRI-Shape. Each subject views 3,142 3D objects across 117 categories, all accompanied by text captions. This significantly enhances the diversity and potential applications of the dataset. Moreover, we propose MinD-3D++, a novel framework for decoding textured 3D visual information from fMRI signals. The framework evaluates the feasibility of not only reconstructing 3D objects from the human mind but also generating, for the first time, 3D textured meshes with detailed textures from fMRI data. We establish new benchmarks by designing metrics at the semantic, structural, and textured levels to evaluate model performance. Furthermore, we assess the model’s effectiveness in out-of-distribution settings and analyze the attribution of the proposed 3D pari fMRI dataset in visual regions of interest (ROIs) in fMRI signals. Our experiments demonstrate that MinD-3D++ not only reconstructs 3D objects with high semantic and spatial accuracy but also provides deeper insights into how the human brain processes 3D visual information.
从功能磁共振成像(fMRI)数据重建3D视觉,称为Recon3DMind,是认知神经科学和计算机视觉的重要兴趣。为了推进这项任务,我们展示了fMRI-3D数据集,其中包括来自15名参与者的数据,共展示了4,768个3D物体。该数据集由两个部分组成:fMRI-Shape(之前介绍过,可在https://huggingface.co/datasets/Fudan-fMRI/fMRI-Shape上获得)和fMRI-Objaverse(本文提出,可在https://huggingface.co/datasets/Fudan-fMRI/fMRI-Objaverse上获得)。fMRI-Objaverse包含了5个被试的数据,其中4个被试也是fMRI-Shape核心集的一部分。每个主题查看117个类别中的3,142个3D对象,所有对象都附有文字说明。这大大增强了数据集的多样性和潜在的应用。此外,我们提出了MinD-3D++,这是一个从fMRI信号中解码纹理三维视觉信息的新框架。该框架不仅评估了从人类思维中重建3D物体的可行性,而且还首次从fMRI数据中生成具有详细纹理的3D纹理网格。我们通过在语义、结构和纹理级别设计度量来评估模型性能,从而建立新的基准。此外,我们评估了该模型在非分布环境下的有效性,并分析了在fMRI信号的视觉感兴趣区域(roi)中提出的3D pari fMRI数据集的归属。我们的实验表明,MinD-3D++不仅重建了具有高语义和空间精度的3D物体,而且还为人类大脑如何处理3D视觉信息提供了更深入的见解。项目页面:https://jianxgao.github.io/MinD-3D。
{"title":"MinD-3D++: Advancing fMRI-Based 3D Reconstruction With High-Quality Textured Mesh Generation and a Comprehensive Dataset","authors":"Jianxiong Gao;Yanwei Fu;Yuqian Fu;Yun Wang;Xuelin Qian;Jianfeng Feng","doi":"10.1109/TPAMI.2025.3599860","DOIUrl":"10.1109/TPAMI.2025.3599860","url":null,"abstract":"Reconstructing 3D visuals from functional Magnetic Resonance Imaging (fMRI) data, introduced as Recon3DMind, is of significant interest to both cognitive neuroscience and computer vision. To advance this task, we present the fMRI-3D dataset, which includes data from 15 participants and showcases a total of 4,768 3D objects. The dataset consists of two components: fMRI-Shape, previously introduced and available at <uri>https://huggingface.co/datasets/Fudan-fMRI/fMRI-Shape</uri>, and fMRI-Objaverse, proposed in this paper and available at <uri>https://huggingface.co/datasets/Fudan-fMRI/fMRI-Objaverse</uri>. fMRI-Objaverse includes data from 5 subjects, 4 of whom are also part of the core set in fMRI-Shape. Each subject views 3,142 3D objects across 117 categories, all accompanied by text captions. This significantly enhances the diversity and potential applications of the dataset. Moreover, we propose <italic>MinD-3D++</i>, a novel framework for decoding textured 3D visual information from fMRI signals. The framework evaluates the feasibility of not only reconstructing 3D objects from the human mind but also generating, for the first time, 3D textured meshes with detailed textures from fMRI data. We establish new benchmarks by designing metrics at the semantic, structural, and textured levels to evaluate model performance. Furthermore, we assess the model’s effectiveness in out-of-distribution settings and analyze the attribution of the proposed 3D pari fMRI dataset in visual regions of interest (ROIs) in fMRI signals. Our experiments demonstrate that MinD-3D++ not only reconstructs 3D objects with high semantic and spatial accuracy but also provides deeper insights into how the human brain processes 3D visual information.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"11802-11816"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph Memory Learning: Imitating Lifelong Remembering and Forgetting of Brain Networks 图形记忆学习:模仿大脑网络的终身记忆和遗忘。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3599898
Jiaxing Miao;Liang Hu;Qi Zhang;Longbing Cao
Graph data in real-world scenarios undergo rapid and frequent changes, making it challenging for existing graph models to effectively handle the continuous influx of new data and accommodate data withdrawal requests. The approach to frequently retraining graph models is resource intensive and impractical. To address this pressing challenge, this paper introduces a new concept of graph memory learning. Its core idea is to enable a graph model to selectively remember new knowledge but forget old knowledge. Building on this approach, the paper presents a novel graph memory learning framework - Brain-inspired Graph Memory Learning (BGML), inspired by brain network dynamics and function-structure coupling strategies. BGML incorporates a multi-granular hierarchical progressive learning mechanism rooted in feature graph grain learning to mitigate potential conflict between memorization and forgetting in graph memory learning. This mechanism allows for a comprehensive and multi-level perception of local details within evolving graphs. In addition, to tackle the issue of unreliable structures in newly added incremental information, the paper introduces an information self-assessment ownership mechanism. This mechanism not only facilitates the propagation of incremental information within the model but also effectively preserves the integrity of past experiences. We design five types of graph memory learning tasks: regular, memory, unlearning, data-incremental, and class-incremental to evaluate BGML. Its excellent performance is confirmed through extensive experiments on multiple node classification datasets.
现实场景中的图数据变化迅速且频繁,这使得现有的图模型难以有效地处理不断涌入的新数据和适应数据提取请求。频繁重新训练图模型的方法是资源密集型的,而且不切实际。为了解决这一紧迫的挑战,本文引入了图记忆学习的新概念。其核心思想是使图模型能够选择性地记住新知识而忘记旧知识。在此基础上,本文提出了一种新的图记忆学习框架——脑启发图记忆学习(BGML),该框架受脑网络动力学和功能结构耦合策略的启发。BGML结合了基于特征图颗粒学习的多颗粒分层渐进学习机制,以缓解图记忆学习中记忆和遗忘之间的潜在冲突。这种机制允许对不断发展的图中的局部细节进行全面和多层次的感知。此外,为了解决新增增量信息中结构不可靠的问题,本文引入了信息自评估所有权机制。这种机制不仅有利于增量信息在模型内的传播,而且有效地保持了过去经验的完整性。我们设计了五种类型的图记忆学习任务:常规、记忆、遗忘、数据增量和类增量来评估BGML。通过在多节点分类数据集上的大量实验,验证了该算法的优异性能。
{"title":"Graph Memory Learning: Imitating Lifelong Remembering and Forgetting of Brain Networks","authors":"Jiaxing Miao;Liang Hu;Qi Zhang;Longbing Cao","doi":"10.1109/TPAMI.2025.3599898","DOIUrl":"10.1109/TPAMI.2025.3599898","url":null,"abstract":"Graph data in real-world scenarios undergo rapid and frequent changes, making it challenging for existing graph models to effectively handle the continuous influx of new data and accommodate data withdrawal requests. The approach to frequently retraining graph models is resource intensive and impractical. To address this pressing challenge, this paper introduces a new concept of <italic>graph memory learning</i>. Its core idea is to enable a graph model to selectively remember new knowledge but forget old knowledge. Building on this approach, the paper presents a novel graph memory learning framework - <bold>B</b>rain-inspired <bold>G</b>raph <bold>M</b>emory <bold>L</b>earning (BGML), inspired by brain network dynamics and function-structure coupling strategies. BGML incorporates a multi-granular hierarchical progressive learning mechanism rooted in feature graph grain learning to mitigate potential conflict between memorization and forgetting in graph memory learning. This mechanism allows for a comprehensive and multi-level perception of local details within evolving graphs. In addition, to tackle the issue of unreliable structures in newly added incremental information, the paper introduces an information self-assessment ownership mechanism. This mechanism not only facilitates the propagation of incremental information within the model but also effectively preserves the integrity of past experiences. We design five types of graph memory learning tasks: regular, memory, unlearning, data-incremental, and class-incremental to evaluate BGML. Its excellent performance is confirmed through extensive experiments on multiple node classification datasets.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 12","pages":"12038-12053"},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UrbanGen: Urban Generation with Compositional and Controllable Neural Fields. UrbanGen:具有组成和可控神经场的城市生成。
IF 18.6 Pub Date : 2025-08-19 DOI: 10.1109/TPAMI.2025.3600440
Yuanbo Yang, Yujun Shen, Yue Wang, Andreas Geiger, Yiyi Liao

Despite the rapid progress in generative radiance fields, most existing methods focus on object-centric applications and are not able to generate complex urban scenes. In this paper, we propose UrbanGen, a solution for the challenging task of generating urban radiance fields with photorealistic rendering, accurate geometry, high controllability, and diverse city styles. Our key idea is to leverage a coarse 3D panoptic prior, represented by a semantic voxel grid for stuff and bounding boxes for countable objects, to condition a compositional generative radiance field. This panoptic prior simplifies the task of learning complex urban geometry, enables disentanglement of stuff and objects, and provides versatile control over both. Moreover, by combining semantic and geometry losses with adversarial training, our method faithfully adheres to the input conditions, allowing for joint rendering of semantic and depth maps alongside RGB images. In addition, we collect a unified dataset with images and their panoptic priors in the same format from 3 diverse real-world datasets: KITTI-360, nuScenes, and Waymo, and train a city style-aware model on this data. Our systematic study shows that UrbanGen outperforms state-of-the-art generative radiance field baselines in terms of image fidelity and geometry accuracy for urban scene generation. Furthermore, UrbenGen brings a new set of controllability features, including large camera movements, stuff editing, and city style control.

尽管生成辐射场的研究进展迅速,但大多数现有方法都侧重于以物体为中心的应用,无法生成复杂的城市场景。在本文中,我们提出了urban bangen,这是一个具有挑战性的任务,具有逼真的渲染,精确的几何形状,高度可控制性和多样化的城市风格。我们的关键思想是利用粗糙的3D全景先验,用语义体素网格表示物体,用边界框表示可数物体,来调节合成生成辐射场。这种全景先验简化了学习复杂城市几何的任务,使材料和物体能够解开纠缠,并提供对两者的通用控制。此外,通过将语义和几何损失与对抗训练相结合,我们的方法忠实地遵循输入条件,允许与RGB图像一起联合渲染语义和深度图。此外,我们从KITTI-360、nuScenes和Waymo 3个不同的现实世界数据集中收集了一个统一的数据集,其中包含相同格式的图像及其全景先验,并在此数据上训练了一个城市风格感知模型。我们的系统研究表明,UrbanGen在城市场景生成的图像保真度和几何精度方面优于最先进的生成辐光场基线。此外,UrbenGen带来了一系列新的可控性功能,包括大镜头移动、素材编辑和城市风格控制。
{"title":"UrbanGen: Urban Generation with Compositional and Controllable Neural Fields.","authors":"Yuanbo Yang, Yujun Shen, Yue Wang, Andreas Geiger, Yiyi Liao","doi":"10.1109/TPAMI.2025.3600440","DOIUrl":"10.1109/TPAMI.2025.3600440","url":null,"abstract":"<p><p>Despite the rapid progress in generative radiance fields, most existing methods focus on object-centric applications and are not able to generate complex urban scenes. In this paper, we propose UrbanGen, a solution for the challenging task of generating urban radiance fields with photorealistic rendering, accurate geometry, high controllability, and diverse city styles. Our key idea is to leverage a coarse 3D panoptic prior, represented by a semantic voxel grid for stuff and bounding boxes for countable objects, to condition a compositional generative radiance field. This panoptic prior simplifies the task of learning complex urban geometry, enables disentanglement of stuff and objects, and provides versatile control over both. Moreover, by combining semantic and geometry losses with adversarial training, our method faithfully adheres to the input conditions, allowing for joint rendering of semantic and depth maps alongside RGB images. In addition, we collect a unified dataset with images and their panoptic priors in the same format from 3 diverse real-world datasets: KITTI-360, nuScenes, and Waymo, and train a city style-aware model on this data. Our systematic study shows that UrbanGen outperforms state-of-the-art generative radiance field baselines in terms of image fidelity and geometry accuracy for urban scene generation. Furthermore, UrbenGen brings a new set of controllability features, including large camera movements, stuff editing, and city style control.</p>","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"PP ","pages":""},"PeriodicalIF":18.6,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1