首页 > 最新文献

Pattern Recognition最新文献

英文 中文
Data-efficient generalization for zero-shot composed image retrieval 零镜头合成图像检索的数据高效泛化
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-28 DOI: 10.1016/j.patcog.2026.113187
Zining Chen , Zhicheng Zhao , Fei Su , Shijian Lu
Zero-shot Composed Image Retrieval (ZS-CIR) aims to retrieve the target image based on a reference image and a text description without requiring in-distribution triplets for training. One prevalent approach follows the vision-language pretraining paradigm that employs a mapping network to transfer the image embedding to a pseudo-word token in the text embedding space. However, this approach tends to impede network generalization due to modality discrepancy and distribution shift between training and inference. To this end, we propose a Data-efficient Generalization (DeG) framework, including two novel designs, namely, Textual Supplement (TS) module and Semantic Sample Pool (SSP) module. The TS module exploits compositional textual semantics during training, enhancing the pseudo-word token with more linguistic semantics and thus mitigating the modality discrepancy effectively. The SSP module exploits the zero-shot capability of pretrained Vision-Language Models (VLMs), alleviating the distribution shift and mitigating the overfitting issue from the redundancy of the large-scale image-text data. Extensive experiments over four ZS-CIR benchmarks show that DeG outperforms the state-of-the-art (SOTA) methods with much less training data, and saves substantial training and inference time for practical usage.
Zero-shot组合图像检索(ZS-CIR)旨在基于参考图像和文本描述检索目标图像,而不需要在分布中三元组进行训练。一种流行的方法遵循视觉语言预训练范式,使用映射网络将图像嵌入转移到文本嵌入空间中的伪词标记。然而,由于训练和推理之间的模态差异和分布转移,这种方法容易阻碍网络的泛化。为此,我们提出了一个数据高效泛化(DeG)框架,其中包括两个新颖的设计,即文本补充(TS)模块和语义样本池(SSP)模块。TS模块在训练过程中利用组合文本语义,使伪词标记具有更多的语言语义,从而有效地缓解情态差异。SSP模块利用了预训练视觉语言模型(VLMs)的零射击能力,减轻了大规模图像文本数据冗余带来的分布偏移和过拟合问题。在四个ZS-CIR基准上进行的广泛实验表明,DeG在训练数据少得多的情况下优于最先进的(SOTA)方法,并为实际使用节省了大量的训练和推理时间。
{"title":"Data-efficient generalization for zero-shot composed image retrieval","authors":"Zining Chen ,&nbsp;Zhicheng Zhao ,&nbsp;Fei Su ,&nbsp;Shijian Lu","doi":"10.1016/j.patcog.2026.113187","DOIUrl":"10.1016/j.patcog.2026.113187","url":null,"abstract":"<div><div>Zero-shot Composed Image Retrieval (ZS-CIR) aims to retrieve the target image based on a reference image and a text description without requiring in-distribution triplets for training. One prevalent approach follows the vision-language pretraining paradigm that employs a mapping network to transfer the image embedding to a pseudo-word token in the text embedding space. However, this approach tends to impede network generalization due to modality discrepancy and distribution shift between training and inference. To this end, we propose a Data-efficient Generalization (DeG) framework, including two novel designs, namely, Textual Supplement (TS) module and Semantic Sample Pool (SSP) module. The TS module exploits compositional textual semantics during training, enhancing the pseudo-word token with more linguistic semantics and thus mitigating the modality discrepancy effectively. The SSP module exploits the zero-shot capability of pretrained Vision-Language Models (VLMs), alleviating the distribution shift and mitigating the overfitting issue from the redundancy of the large-scale image-text data. Extensive experiments over four ZS-CIR benchmarks show that DeG outperforms the state-of-the-art (SOTA) methods with much less training data, and saves substantial training and inference time for practical usage.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113187"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prioritized scanning: Combining spatial information multiple instance learning for computational pathology 优先扫描:结合空间信息多实例学习的计算病理学
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-24 DOI: 10.1016/j.patcog.2026.113151
Yuqi Zhang , Jiakai Wang , Baoyu Liang , Yuancheng Yang , Siyang Wu , Chao Tong
Multiple instance learning (MIL) has emerged as a reliable paradigm that has propelled the integration of computational pathology (CPath) into clinical histopathology. However, despite significant advancements, current MIL approaches continue to face challenges due to inadequate spatial information representation resulting from the disorder of the original whole slide images (WSIs). To address this limitation, we first demonstrate the importance of prioritized scanning within the structured state space models (SSM). We introduce a MIL framework that incorporates spatial information, termed Prioritized Scanning MIL (PSMIL). PSMIL primarily comprises two branches and a fusion block. The first branch, known as the spatial branch, incorporates potential spatial information into the patch sequence using the original 2D positions and employs SSM to model the spatial features of the WSI. The second branch, referred to as the cross-spatial branch, utilizes a significance scoring block along with SSM to harness feature relationships among similar instances across spatial locations. Finally, a lightweight feature fusion block integrates the outputs of both branches, facilitating more comprehensive feature utilization. Extensive experiments on 5 popular datasets and 3 downstream tasks strongly demonstrate that PSMIL surpasses the state-of-the-art MIL methods significantly, up to 5.26% ACC improvements for cancer sub-typing. Our code is available at https://github.com/YuqiZhang-Buaa/PSMIL.
多实例学习(MIL)已经成为一种可靠的范式,它推动了计算病理学(CPath)与临床组织病理学的整合。然而,尽管取得了重大进展,目前的MIL方法仍然面临着挑战,因为原始整个幻灯片图像(wsi)的无序导致空间信息表示不足。为了解决这一限制,我们首先展示了在结构化状态空间模型(SSM)中优先扫描的重要性。我们引入了一个包含空间信息的MIL框架,称为优先扫描MIL (PSMIL)。PSMIL主要由两个分支和一个融合块组成。第一个分支称为空间分支,利用原始二维位置将潜在的空间信息整合到patch序列中,并使用SSM对WSI的空间特征进行建模。第二个分支称为跨空间分支,它利用显著性评分块和SSM来利用跨空间位置的类似实例之间的特征关系。最后,一个轻量级的特征融合块集成了两个分支的输出,便于更全面地利用特征。在5个流行数据集和3个下游任务上进行的大量实验表明,PSMIL显著优于最先进的MIL方法,在癌症亚型分型方面的ACC提高高达5.26%。我们的代码可在https://github.com/YuqiZhang-Buaa/PSMIL上获得。
{"title":"Prioritized scanning: Combining spatial information multiple instance learning for computational pathology","authors":"Yuqi Zhang ,&nbsp;Jiakai Wang ,&nbsp;Baoyu Liang ,&nbsp;Yuancheng Yang ,&nbsp;Siyang Wu ,&nbsp;Chao Tong","doi":"10.1016/j.patcog.2026.113151","DOIUrl":"10.1016/j.patcog.2026.113151","url":null,"abstract":"<div><div>Multiple instance learning (MIL) has emerged as a reliable paradigm that has propelled the integration of computational pathology (CPath) into clinical histopathology. However, despite significant advancements, current MIL approaches continue to face challenges due to inadequate spatial information representation resulting from the disorder of the original whole slide images (WSIs). To address this limitation, we first demonstrate the importance of prioritized scanning within the structured state space models (SSM). We introduce a MIL framework that incorporates spatial information, termed <strong>Prioritized Scanning MIL (PSMIL)</strong>. PSMIL primarily comprises two branches and a fusion block. The first branch, known as the spatial branch, incorporates potential spatial information into the patch sequence using the original 2D positions and employs SSM to model the spatial features of the WSI. The second branch, referred to as the cross-spatial branch, utilizes a significance scoring block along with SSM to harness feature relationships among similar instances across spatial locations. Finally, a lightweight feature fusion block integrates the outputs of both branches, facilitating more comprehensive feature utilization. Extensive experiments on 5 popular datasets and 3 downstream tasks strongly demonstrate that PSMIL surpasses the state-of-the-art MIL methods significantly, up to 5.26% ACC improvements for cancer sub-typing. Our code is available at <span><span>https://github.com/YuqiZhang-Buaa/PSMIL</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113151"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TranSAC: An unsupervised transferability metric based on task speciality and domain commonality TranSAC:基于任务特殊性和领域共性的无监督可转移性度量
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-29 DOI: 10.1016/j.patcog.2026.113137
Qianshan Zhan , Xiao-Jun Zeng , Qian Wang
In transfer learning, one fundamental problem is transferability estimation, where a metric measures transfer performance without training. Existing metrics face two issues: 1) requiring target domain labels, and 2) only focusing on task speciality but ignoring equally important domain commonality. To overcome these limitations, we propose TranSAC, a Transferability metric based on task Speciality And domain Commonality, capturing the separation between classes and the similarity between domains. Its main advantages are: 1) unsupervised, 2) fine-tuning free, and 3) applicable to source-dependent and source-free transfer scenarios. To achieve this, we investigate the upper and lower bounds of transfer performance based on fixed representations extracted from the pre-trained model. Theoretical results reveal that unsupervised transfer performance is characterized by entropy-based quantities, naturally reflecting task specificity and domain commonality. These insights motivate the design of TranSAC, which integrates both factors to enhance transferability. Extensive experiments are performed across 12 target datasets with 36 pre-trained models, including supervised CNNs, self-supervised CNNs, and ViTs. Results demonstrate the importance of domain commonality and task speciality, allowing TranSAC as superior to state-of-the-art metrics for pre-trained model ranking, target domain ranking, and source domain ranking.
在迁移学习中,一个基本问题是可迁移性估计,其中度量是在未经训练的情况下度量迁移性能。现有的度量标准面临两个问题:1)需要目标领域标签;2)只关注任务的特殊性,而忽略了同样重要的领域共性。为了克服这些限制,我们提出了TranSAC,一种基于任务特殊性和领域共性的可转移性度量,捕获类之间的分离和领域之间的相似性。它的主要优点是:1)无监督,2)无微调,以及3)适用于依赖源和无源的传输场景。为了实现这一点,我们研究了基于从预训练模型中提取的固定表示的传输性能的上界和下界。理论结果表明,无监督迁移性能具有基于熵的特征量,自然地反映了任务的特殊性和领域的共性。这些见解激发了TranSAC的设计,它整合了这两个因素以增强可转移性。在12个目标数据集和36个预训练模型上进行了广泛的实验,包括监督cnn、自监督cnn和vit。结果证明了领域共性和任务特殊性的重要性,使得TranSAC在预训练模型排名、目标领域排名和源领域排名方面优于最先进的指标。
{"title":"TranSAC: An unsupervised transferability metric based on task speciality and domain commonality","authors":"Qianshan Zhan ,&nbsp;Xiao-Jun Zeng ,&nbsp;Qian Wang","doi":"10.1016/j.patcog.2026.113137","DOIUrl":"10.1016/j.patcog.2026.113137","url":null,"abstract":"<div><div>In transfer learning, one fundamental problem is transferability estimation, where a metric measures transfer performance without training. Existing metrics face two issues: 1) requiring target domain labels, and 2) only focusing on task speciality but ignoring equally important domain commonality. To overcome these limitations, we propose TranSAC, a <strong>Tran</strong>sferability metric based on task <strong>S</strong>peciality <strong>A</strong>nd domain <strong>C</strong>ommonality, capturing the separation between classes and the similarity between domains. Its main advantages are: 1) unsupervised, 2) fine-tuning free, and 3) applicable to source-dependent and source-free transfer scenarios. To achieve this, we investigate the upper and lower bounds of transfer performance based on fixed representations extracted from the pre-trained model. Theoretical results reveal that unsupervised transfer performance is characterized by entropy-based quantities, naturally reflecting task specificity and domain commonality. These insights motivate the design of TranSAC, which integrates both factors to enhance transferability. Extensive experiments are performed across 12 target datasets with 36 pre-trained models, including supervised CNNs, self-supervised CNNs, and ViTs. Results demonstrate the importance of domain commonality and task speciality, allowing TranSAC as superior to state-of-the-art metrics for pre-trained model ranking, target domain ranking, and source domain ranking.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113137"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Audio-visual perceptual quality measurement via multi-perspective spatio-temporal EEG analysis 基于多视角时空脑电图分析的视听感知质量测量
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-24 DOI: 10.1016/j.patcog.2026.113156
Shuzhan Hu , Mingyu Li , Yang Liu , Weiwei Jiang , Bingrui Geng , Wei Zhong , Long Ye
In human-centered communication systems, establishing human perception-aligned audio-visual quality assessment methods is crucial for enhancing multimedia system performance and service quality. However, conventional subjective evaluation methods based on user ratings are susceptible to biases induced by high-level cognitive processes. To address this limitation, we propose an electroencephalography (EEG) feature fusion approach to establish correlations between audio-visual distortions and perceptual experiences. Specifically, we construct an audio-visual degradation-EEG dataset by recording neural responses from subjects exposed to progressively degraded stimuli. Leveraging this dataset, we extract event-related potential (ERP) features to quantify variations in subjects’ perception of audio-visual quality, demonstrating the feasibility of EEG-based perceptual experience assessment. Capitalizing on EEG’s sensitivity to dynamic multimodal perceptual changes, we develop a multi-perspective feature fusion framework, incorporating a spatio-temporal feature fusion architecture and a diffusion-driven EEG augmentation strategy. This framework enables the extraction of experience-related features from single-trial EEG signals, establishing an EEG-based classifier to detect whether distortions induce perceptual experience alterations. Experimental results validate that EEG signals effectively reflect perception changes induced by quality degradation, while the proposed model achieves efficient and dynamic detection of perception alterations from single-trial EEG data.
在以人为中心的通信系统中,建立符合人的感知的视听质量评价方法是提高多媒体系统性能和服务质量的关键。然而,传统的基于用户评分的主观评价方法容易受到高层次认知过程的影响。为了解决这一限制,我们提出了一种脑电图(EEG)特征融合方法来建立视听扭曲和感知体验之间的相关性。具体来说,我们通过记录暴露于逐渐退化的刺激的受试者的神经反应,构建了一个视听退化-脑电图数据集。利用该数据集,我们提取事件相关电位(ERP)特征来量化受试者对视听质量感知的变化,证明了基于脑电图的感知体验评估的可行性。利用脑电对动态多模态感知变化的敏感性,我们开发了一个多视角特征融合框架,将时空特征融合架构和扩散驱动的脑电增强策略相结合。该框架能够从单次脑电图信号中提取与经验相关的特征,建立基于脑电图的分类器来检测扭曲是否会引起感知经验的改变。实验结果表明,脑电信号能够有效地反映质量退化引起的感知变化,该模型能够实现对单次脑电信号感知变化的高效、动态检测。
{"title":"Audio-visual perceptual quality measurement via multi-perspective spatio-temporal EEG analysis","authors":"Shuzhan Hu ,&nbsp;Mingyu Li ,&nbsp;Yang Liu ,&nbsp;Weiwei Jiang ,&nbsp;Bingrui Geng ,&nbsp;Wei Zhong ,&nbsp;Long Ye","doi":"10.1016/j.patcog.2026.113156","DOIUrl":"10.1016/j.patcog.2026.113156","url":null,"abstract":"<div><div>In human-centered communication systems, establishing human perception-aligned audio-visual quality assessment methods is crucial for enhancing multimedia system performance and service quality. However, conventional subjective evaluation methods based on user ratings are susceptible to biases induced by high-level cognitive processes. To address this limitation, we propose an electroencephalography (EEG) feature fusion approach to establish correlations between audio-visual distortions and perceptual experiences. Specifically, we construct an audio-visual degradation-EEG dataset by recording neural responses from subjects exposed to progressively degraded stimuli. Leveraging this dataset, we extract event-related potential (ERP) features to quantify variations in subjects’ perception of audio-visual quality, demonstrating the feasibility of EEG-based perceptual experience assessment. Capitalizing on EEG’s sensitivity to dynamic multimodal perceptual changes, we develop a multi-perspective feature fusion framework, incorporating a spatio-temporal feature fusion architecture and a diffusion-driven EEG augmentation strategy. This framework enables the extraction of experience-related features from single-trial EEG signals, establishing an EEG-based classifier to detect whether distortions induce perceptual experience alterations. Experimental results validate that EEG signals effectively reflect perception changes induced by quality degradation, while the proposed model achieves efficient and dynamic detection of perception alterations from single-trial EEG data.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113156"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
One-step multi-view graph clustering via bottom-up structural learning 基于自底向上结构学习的一步多视图聚类
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-29 DOI: 10.1016/j.patcog.2026.113175
Wenzhe Liu , Li Jiang , Huibing Wang , Yong Zhang
In recent years, tensor-based methods have seen considerable success in multi-view clustering. However, the current approach has several limitations: 1) Insufficient exploration of underlying similarity information (i.e. latent representation); 2) Insufficient exploration of higher-order structure information of both inter-view and intra-view; 3) Treating clustering learning independently from tensor learning and the overall learning framework. To address these issues, we propose a unified framework called Bottom-up Structural Exploration for One-step Multi-view Graph Clustering (BSE_OMGC). Specifically, we first employ an anchor strategy to build similarity graphs, reducing the complexity of graph learning. To deeply represent the underlying similar information of the data and mitigate the influence of noise on similar structures in the original space, BSE_OMGC adaptively separates the noise matrix from the similarity graphs to learn high-quality enhanced graphs. Subsequently, from the bottom up, the enhanced graphs serve as the foundation for constructing high-order tensors. We rotate the constructed tensors and apply the t-TNN to preserve the low-rank properties and to better capture higher-order structure information of both inter-view and intra-view. Finally, we introduce a symmetric non-negative matrix factorization-based graph partitioning technique, which learns non-negative embeddings during dynamic optimization to reveal clustering results. This approach unifies clustering learning within the entire learning framework. Extensive experiments on multiple real-world multi-view datasets, along with comparisons to state-of-the-art methods, demonstrate the effectiveness and robustness of the proposed approach.
近年来,基于张量的聚类方法在多视图聚类中取得了相当大的成功。然而,目前的方法存在一些局限性:1)对潜在相似信息(即潜在表示)的探索不足;2)对视图间和视图内高阶结构信息的挖掘不足;3)将聚类学习独立于张量学习和整体学习框架。为了解决这些问题,我们提出了一个统一的框架,称为自下而上的结构探索一步多视图图聚类(BSE_OMGC)。具体来说,我们首先采用锚点策略来构建相似图,降低了图学习的复杂性。为了深度表示数据的潜在相似信息,减轻噪声对原始空间相似结构的影响,BSE_OMGC自适应地将噪声矩阵从相似图中分离出来,学习高质量的增强图。随后,从下往上,增强图作为构造高阶张量的基础。我们旋转构造张量并应用t-TNN来保持低秩性质,并更好地捕获视图间和视图内的高阶结构信息。最后,我们介绍了一种基于对称非负矩阵分解的图划分技术,该技术在动态优化过程中学习非负嵌入以显示聚类结果。这种方法在整个学习框架内统一了聚类学习。在多个真实世界的多视图数据集上进行的大量实验,以及与最先进方法的比较,证明了所提出方法的有效性和鲁棒性。
{"title":"One-step multi-view graph clustering via bottom-up structural learning","authors":"Wenzhe Liu ,&nbsp;Li Jiang ,&nbsp;Huibing Wang ,&nbsp;Yong Zhang","doi":"10.1016/j.patcog.2026.113175","DOIUrl":"10.1016/j.patcog.2026.113175","url":null,"abstract":"<div><div>In recent years, tensor-based methods have seen considerable success in multi-view clustering. However, the current approach has several limitations: 1) Insufficient exploration of underlying similarity information (i.e. latent representation); 2) Insufficient exploration of higher-order structure information of both inter-view and intra-view; 3) Treating clustering learning independently from tensor learning and the overall learning framework. To address these issues, we propose a unified framework called Bottom-up Structural Exploration for One-step Multi-view Graph Clustering (BSE_OMGC). Specifically, we first employ an anchor strategy to build similarity graphs, reducing the complexity of graph learning. To deeply represent the underlying similar information of the data and mitigate the influence of noise on similar structures in the original space, BSE_OMGC adaptively separates the noise matrix from the similarity graphs to learn high-quality enhanced graphs. Subsequently, from the bottom up, the enhanced graphs serve as the foundation for constructing high-order tensors. We rotate the constructed tensors and apply the t-TNN to preserve the low-rank properties and to better capture higher-order structure information of both inter-view and intra-view. Finally, we introduce a symmetric non-negative matrix factorization-based graph partitioning technique, which learns non-negative embeddings during dynamic optimization to reveal clustering results. This approach unifies clustering learning within the entire learning framework. Extensive experiments on multiple real-world multi-view datasets, along with comparisons to state-of-the-art methods, demonstrate the effectiveness and robustness of the proposed approach.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113175"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning generalizable visual representations with causal diffusion model for controllable editing 基于因果扩散模型的可控编辑可泛化视觉表征学习
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-29 DOI: 10.1016/j.patcog.2026.113162
Shanshan Huang , Lei Wang , Haoxuan Chen , Yuxuan Liang , Li Liu
Representation learning has been widely employed to learn low-dimensional representations that consist of multiple independent and interpretable generative factors like visual attributes in images, enabling controllable image editing by manipulating specific attributes in the learned representation space. However, in real-world scenarios, generative factors with semantic meanings are often causally related rather than independent. Previous methods with independence assumption are failed to capture such causal relationships, even in the supervised settings. To this end, we propose a diffusion model-based causal representation learning framework, named CausalDiffuser, which models causal prior distributions by the structural causal models (SCMs) to explicitly characterize the causal relations among the underlying generative factors. Such modelling scheme encourages the framework to learn the latent representations of causality for generative factors. Furthermore, a composite loss function is introduced to ensure causal disentanglement of latent representations by combining supervision information from the ground truth factors (i.e., image labels). Empirical evaluations on one synthetic dataset and two real-world benchmark datasets suggest our approach significantly outperforms the state-of-the-art methods. CausalDiffuser effectively edits image attributes by restoring causal relationships among generative factors and generates counterfactual images through intervention operation.
表征学习已被广泛用于学习由多个独立且可解释的生成因素(如图像中的视觉属性)组成的低维表征,通过对学习到的表征空间中的特定属性进行操作,实现对图像的可控编辑。然而,在现实场景中,具有语义意义的生成因素往往是因果相关的,而不是独立的。以前的方法与独立性假设未能捕获这样的因果关系,即使在监督设置。为此,我们提出了一个基于扩散模型的因果表示学习框架,名为CausalDiffuser,它通过结构因果模型(scm)对因果先验分布进行建模,以明确表征潜在生成因素之间的因果关系。这种建模方案鼓励框架学习生成因素因果关系的潜在表征。此外,引入了一个复合损失函数,通过结合来自地面真值因素(即图像标签)的监督信息来确保潜在表示的因果解纠缠。对一个合成数据集和两个真实世界基准数据集的实证评估表明,我们的方法明显优于最先进的方法。CausalDiffuser通过还原生成因素之间的因果关系,有效地编辑图像属性,并通过干预操作生成反事实图像。
{"title":"Learning generalizable visual representations with causal diffusion model for controllable editing","authors":"Shanshan Huang ,&nbsp;Lei Wang ,&nbsp;Haoxuan Chen ,&nbsp;Yuxuan Liang ,&nbsp;Li Liu","doi":"10.1016/j.patcog.2026.113162","DOIUrl":"10.1016/j.patcog.2026.113162","url":null,"abstract":"<div><div>Representation learning has been widely employed to learn low-dimensional representations that consist of multiple independent and interpretable generative factors like visual attributes in images, enabling controllable image editing by manipulating specific attributes in the learned representation space. However, in real-world scenarios, generative factors with semantic meanings are often causally related rather than independent. Previous methods with independence assumption are failed to capture such causal relationships, even in the supervised settings. To this end, we propose a diffusion model-based causal representation learning framework, named CausalDiffuser, which models causal prior distributions by the structural causal models (SCMs) to explicitly characterize the causal relations among the underlying generative factors. Such modelling scheme encourages the framework to learn the latent representations of causality for generative factors. Furthermore, a composite loss function is introduced to ensure causal disentanglement of latent representations by combining supervision information from the ground truth factors (i.e., image labels). Empirical evaluations on one synthetic dataset and two real-world benchmark datasets suggest our approach significantly outperforms the state-of-the-art methods. CausalDiffuser effectively edits image attributes by restoring causal relationships among generative factors and generates counterfactual images through intervention operation.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113162"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Frequency-aligned supervision for few-shot neural rendering 基于频率对齐的少镜头神经渲染监督
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-28 DOI: 10.1016/j.patcog.2026.113183
Su-Ji Jang, Ue-Hwan Kim
Neural rendering has shown significant potential in generating high-quality 3D scenes from sparse inputs. However, existing methods struggle to simultaneously capture both low-frequency global structures and high-frequency fine details, leading to suboptimal scene representations. To overcome this limitation, we propose a frequency-aligned supervision framework that explicitly separates the learning process into low-frequency and full-spectrum components. By introducing two sub-networks and aligning supervision signals at appropriate layers, our method enhances the formation of global structures while preserving fine details. Specifically, the low-frequency network (LFN) is supervised with low-pass targets (Gaussian-filtered images) to form global structures, while the full-spectrum network (FSN) is supervised with the original images to refine high-frequency details. The proposed approach is broadly applicable to MLP-based NeRF architectures without requiring major architectural modifications. Extensive experiments demonstrate that our method consistently improves PSNR, SSIM, and LPIPS across multiple NeRF variants and datasets, confirming its robustness in sparse input scenarios.
神经渲染在从稀疏输入生成高质量3D场景方面显示出巨大的潜力。然而,现有的方法很难同时捕获低频全局结构和高频精细细节,导致次优的场景表示。为了克服这一限制,我们提出了一个频率一致的监督框架,明确地将学习过程分为低频和全频谱组件。该方法通过引入两个子网络并在适当的层上对齐监督信号,增强了全局结构的形成,同时保留了精细的细节。其中,低频网络(LFN)采用低通目标(高斯滤波图像)进行监督,形成全局结构;全谱网络(FSN)采用原始图像进行监督,细化高频细节。所提出的方法广泛适用于基于mlp的NeRF体系结构,而不需要对体系结构进行重大修改。大量实验表明,我们的方法在多个NeRF变量和数据集上持续提高PSNR、SSIM和LPIPS,证实了其在稀疏输入场景下的鲁棒性。
{"title":"Frequency-aligned supervision for few-shot neural rendering","authors":"Su-Ji Jang,&nbsp;Ue-Hwan Kim","doi":"10.1016/j.patcog.2026.113183","DOIUrl":"10.1016/j.patcog.2026.113183","url":null,"abstract":"<div><div>Neural rendering has shown significant potential in generating high-quality 3D scenes from sparse inputs. However, existing methods struggle to simultaneously capture both low-frequency global structures and high-frequency fine details, leading to suboptimal scene representations. To overcome this limitation, we propose a frequency-aligned supervision framework that explicitly separates the learning process into low-frequency and full-spectrum components. By introducing two sub-networks and aligning supervision signals at appropriate layers, our method enhances the formation of global structures while preserving fine details. Specifically, the low-frequency network (LFN) is supervised with low-pass targets (Gaussian-filtered images) to form global structures, while the full-spectrum network (FSN) is supervised with the original images to refine high-frequency details. The proposed approach is broadly applicable to MLP-based NeRF architectures without requiring major architectural modifications. Extensive experiments demonstrate that our method consistently improves PSNR, SSIM, and LPIPS across multiple NeRF variants and datasets, confirming its robustness in sparse input scenarios.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113183"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint asymmetric discrete hashing for cross-modal retrieval 跨模态检索的联合非对称离散哈希
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-28 DOI: 10.1016/j.patcog.2026.113180
Jiaxing Li , Lin Jiang , Zuopeng Yang , Xiaozhao Fang , Shengli Xie , Yong Xu
Cross-modal hashing is one of the promising practical applications in information retrieval for multimedia data. However, there exist some technical hurdles, e.g., how to further reduce the heterogeneous gaps for cross-modal data semantically, how to extract cross-modal knowledge by jointly training data from different modality and how to better leverage the label information to generate more discriminative hash codes, etc. To overcome the above-mentioned challenges, this paper proposes a joint asymmetric discrete hashing (JADH for short) for cross-modal retrieval. By leveraging kernel mapping operation, JADH extracts the non-linear features of cross-modal data to better preserve the semantic information in the latent common space learning. Then, a joint asymmetric hash codes learning term is customized to learn hash codes for data from different modalities jointly. As such, more cross-modal information can be preserved, which can effectively reduce the heterogeneous semantic gaps. Finally, a log-likelihood similarity preserving term is proposed to boost hash codes learning from the similarity matrix, while a classifier learning term is proposed to further improve the quality of the learned hash codes. In addition, an alternative algorithm is derived to solve the optimization problem in JADH efficiently. Experimental results on four widely used datasets show that, JADH outperforms some state-of-the-art baseline methods in hashing-based cross-modal retrieval, on accuracy and efficiency.
跨模态哈希是多媒体数据信息检索中很有前途的实际应用之一。然而,存在一些技术障碍,如如何在语义上进一步减少跨模态数据的异构差距,如何通过联合训练不同模态的数据来提取跨模态知识,如何更好地利用标签信息生成更具判别性的哈希码等。为了克服上述挑战,本文提出了一种联合非对称离散哈希(joint asymmetric discrete hash,简称JADH)的跨模态检索方法。JADH通过核映射操作提取跨模态数据的非线性特征,更好地保留潜在公共空间学习中的语义信息。然后,自定义一个联合非对称哈希码学习项,用于联合学习不同模态数据的哈希码。这样可以保留更多的跨模态信息,有效减少异构语义间隙。最后,提出了一个对数似然相似保持项来增强哈希码从相似矩阵中学习的能力,同时提出了一个分类器学习项来进一步提高学习到的哈希码的质量。在此基础上,提出了一种有效解决JADH优化问题的替代算法。在四个广泛使用的数据集上的实验结果表明,在基于哈希的跨模态检索中,JADH在准确性和效率上都优于一些最先进的基线方法。
{"title":"Joint asymmetric discrete hashing for cross-modal retrieval","authors":"Jiaxing Li ,&nbsp;Lin Jiang ,&nbsp;Zuopeng Yang ,&nbsp;Xiaozhao Fang ,&nbsp;Shengli Xie ,&nbsp;Yong Xu","doi":"10.1016/j.patcog.2026.113180","DOIUrl":"10.1016/j.patcog.2026.113180","url":null,"abstract":"<div><div>Cross-modal hashing is one of the promising practical applications in information retrieval for multimedia data. However, there exist some technical hurdles, e.g., how to further reduce the heterogeneous gaps for cross-modal data semantically, how to extract cross-modal knowledge by jointly training data from different modality and how to better leverage the label information to generate more discriminative hash codes, etc. To overcome the above-mentioned challenges, this paper proposes a joint asymmetric discrete hashing (JADH for short) for cross-modal retrieval. By leveraging kernel mapping operation, JADH extracts the non-linear features of cross-modal data to better preserve the semantic information in the latent common space learning. Then, a joint asymmetric hash codes learning term is customized to learn hash codes for data from different modalities jointly. As such, more cross-modal information can be preserved, which can effectively reduce the heterogeneous semantic gaps. Finally, a log-likelihood similarity preserving term is proposed to boost hash codes learning from the similarity matrix, while a classifier learning term is proposed to further improve the quality of the learned hash codes. In addition, an alternative algorithm is derived to solve the optimization problem in JADH efficiently. Experimental results on four widely used datasets show that, JADH outperforms some state-of-the-art baseline methods in hashing-based cross-modal retrieval, on accuracy and efficiency.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113180"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
S2I-DiT: Unlocking the semantic-to-image transferability by fine-tuning large diffusion transformer models S2I-DiT:通过微调大型扩散变压器模型解锁语义到图像的可转移性
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-25 DOI: 10.1016/j.patcog.2026.113158
Gang Li , Enze Xie , Chongjian Ge , Xiang Li , Lingyu Si , Changwen Zheng , Zhenguo Li
Denoising Diffusion Probabilistic Models (DDPMs) have made significant progress in image generation. Recent works in semantic-to-image (S2I) synthesis have also shifted from the previously de facto GAN-based methods to DDPMs, yielding better results. However, these works mostly employ a U-Net structure and vanilla training-from-scratch scheme for S2I, unconsciously neglecting the potential benefits offered by task-related pre-training. In this work, we introduce a Transformer-based architecture, namely S2I-DiT, and reconsider the merits of a pre-trained large diffusion model for cross-task adaptation (i.e., from the class-conditional generation to S2I). In S2I-DiT, we propose the integration of semantic embedders within Diffusion Transformers (DiTs) to maximize the utilization of semantic information. The semantic embedder densely encodes semantic layouts to guide the adaptive normalization process. We configure semantic embedders in a layer-wise manner to learn pixel-level correspondence, enabling finer-grained semantic-to-image control. Besides, to fully unleash the cross-task transferability of DDPMs, we introduce a two-stage fine-tuning strategy, which involves initially adapting the semantic embedders in the pixel-level space, followed by fine-tuning the partial/entire model for cross-task adaptation. Notably, S2I-DiT pioneers the application of Large Diffusion Transformers to cross-task fine-tuning. Extensive experiments on four benchmark datasets demonstrate S2I-DiT’s effectiveness, as it achieves state-of-the-art performance in terms of quality (FID) and diversity (LPIPS), while consuming fewer training iterations. This work establishes a new state-of-the-art for semantic-to-image generation and provides valuable insights into cross-task transferability of large generative models.
消噪扩散概率模型(ddpm)在图像生成方面取得了重大进展。最近在语义到图像(S2I)合成方面的工作也从以前事实上的基于gan的方法转向了ddpm,产生了更好的结果。然而,这些作品大多采用U-Net结构和香草的S2I从头开始训练方案,无意识地忽略了与任务相关的预训练提供的潜在好处。在这项工作中,我们引入了一个基于transformer的架构,即S2I- dit,并重新考虑了用于跨任务适应的预训练大型扩散模型的优点(即从类条件生成到S2I)。在S2I-DiT中,我们提出在扩散转换器(dit)中集成语义嵌入器,以最大限度地利用语义信息。语义嵌入器对语义布局进行密集编码,引导自适应归一化过程。我们以分层方式配置语义嵌入器以学习像素级对应,从而实现更细粒度的语义到图像控制。此外,为了充分释放ddpm的跨任务可移植性,我们引入了一种两阶段的微调策略,即首先在像素级空间调整语义嵌入器,然后对部分/整个模型进行微调以进行跨任务适应。值得注意的是,S2I-DiT率先将大型扩散变压器应用于跨任务微调。在四个基准数据集上进行的大量实验证明了S2I-DiT的有效性,因为它在质量(FID)和多样性(LPIPS)方面达到了最先进的性能,同时消耗了更少的训练迭代。这项工作建立了语义到图像生成的新技术,并为大型生成模型的跨任务可移植性提供了有价值的见解。
{"title":"S2I-DiT: Unlocking the semantic-to-image transferability by fine-tuning large diffusion transformer models","authors":"Gang Li ,&nbsp;Enze Xie ,&nbsp;Chongjian Ge ,&nbsp;Xiang Li ,&nbsp;Lingyu Si ,&nbsp;Changwen Zheng ,&nbsp;Zhenguo Li","doi":"10.1016/j.patcog.2026.113158","DOIUrl":"10.1016/j.patcog.2026.113158","url":null,"abstract":"<div><div>Denoising Diffusion Probabilistic Models (DDPMs) have made significant progress in image generation. Recent works in semantic-to-image (S2I) synthesis have also shifted from the previously <em>de facto</em> GAN-based methods to DDPMs, yielding better results. However, these works mostly employ a U-Net structure and vanilla training-from-scratch scheme for S2I, unconsciously neglecting the potential benefits offered by task-related pre-training. In this work, we introduce a Transformer-based architecture, namely S2I-DiT, and reconsider the merits of a pre-trained large diffusion model for cross-task adaptation (i.e., from the class-conditional generation to S2I). In S2I-DiT, we propose the integration of semantic embedders within Diffusion Transformers (DiTs) to maximize the utilization of semantic information. The semantic embedder densely encodes semantic layouts to guide the adaptive normalization process. We configure semantic embedders in a layer-wise manner to learn pixel-level correspondence, enabling finer-grained semantic-to-image control. Besides, to fully unleash the cross-task transferability of DDPMs, we introduce a two-stage fine-tuning strategy, which involves initially adapting the semantic embedders in the pixel-level space, followed by fine-tuning the partial/entire model for cross-task adaptation. Notably, S2I-DiT pioneers the application of Large Diffusion Transformers to cross-task fine-tuning. Extensive experiments on four benchmark datasets demonstrate S2I-DiT’s effectiveness, as it achieves state-of-the-art performance in terms of quality (FID) and diversity (LPIPS), while consuming fewer training iterations. This work establishes a new state-of-the-art for semantic-to-image generation and provides valuable insights into cross-task transferability of large generative models.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113158"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalizable face forgery detection via mining single-step reconstruction difference 基于单步重建差分挖掘的人脸伪造检测方法
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-10 DOI: 10.1016/j.patcog.2026.113265
Kai Zhou , Guanglu Sun , Linsen Yu , Jun Wang
Existing face forgery detection methods mainly focus on capturing specific artifacts. While achieving high accuracy on in-distribution data, they often generalize poorly to unseen manipulation techniques due to strong correlation between the learned feature representation and training set. To mitigate this strong correlation and move towards a more generalizable feature representation, we propose a novel face forgery detection framework based on Single-step Reconstruction Difference (SRD). Our approach explores more generalizable features by mining differences between the original and single-step reconstructed features of both real and fake faces. More specifically, we design a feature enhancement module that processes and refines the single-step reconstruction difference, which progressively integrates forgery-related clues into the features of neural networks through attention mechanism. In addition, we design a Frequency-Constrained Contrastive Loss (FCC Loss) to learn discriminative and robust features by contrasting real and fake faces using frequency-domain information. Experimental results demonstrate that the proposed method not only exhibits excellent generalization performance on different datasets but also shows strong robustness of the detection method against various image attacks. Our code is released at: https://github.com/zhouk369/SRD.
现有的人脸伪造检测方法主要集中在捕获特定的伪迹。虽然在分布内数据上获得了很高的准确性,但由于学习到的特征表示与训练集之间存在很强的相关性,它们往往不能很好地推广到看不见的操作技术。为了减轻这种强相关性并向更一般化的特征表示方向发展,我们提出了一种基于单步重建差分(SRD)的人脸伪造检测框架。我们的方法通过挖掘真实人脸和假人脸的原始特征和单步重建特征之间的差异来探索更多可推广的特征。具体而言,我们设计了一个特征增强模块,对单步重建差异进行处理和细化,通过注意机制逐步将伪造相关线索整合到神经网络特征中。此外,我们设计了一个频率约束的对比损耗(FCC Loss),通过使用频域信息对比真实和虚假的人脸来学习判别性和鲁棒性特征。实验结果表明,该方法不仅在不同的数据集上表现出优异的泛化性能,而且对各种图像攻击具有较强的鲁棒性。我们的代码发布在:https://github.com/zhouk369/SRD。
{"title":"Generalizable face forgery detection via mining single-step reconstruction difference","authors":"Kai Zhou ,&nbsp;Guanglu Sun ,&nbsp;Linsen Yu ,&nbsp;Jun Wang","doi":"10.1016/j.patcog.2026.113265","DOIUrl":"10.1016/j.patcog.2026.113265","url":null,"abstract":"<div><div>Existing face forgery detection methods mainly focus on capturing specific artifacts. While achieving high accuracy on in-distribution data, they often generalize poorly to unseen manipulation techniques due to strong correlation between the learned feature representation and training set. To mitigate this strong correlation and move towards a more generalizable feature representation, we propose a novel face forgery detection framework based on Single-step Reconstruction Difference (SRD). Our approach explores more generalizable features by mining differences between the original and single-step reconstructed features of both real and fake faces. More specifically, we design a feature enhancement module that processes and refines the single-step reconstruction difference, which progressively integrates forgery-related clues into the features of neural networks through attention mechanism. In addition, we design a Frequency-Constrained Contrastive Loss (FCC Loss) to learn discriminative and robust features by contrasting real and fake faces using frequency-domain information. Experimental results demonstrate that the proposed method not only exhibits excellent generalization performance on different datasets but also shows strong robustness of the detection method against various image attacks. Our code is released at: <span><span>https://github.com/zhouk369/SRD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113265"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1