首页 > 最新文献

Pattern Recognition最新文献

英文 中文
Tab2Visual: Deep learning for limited tabular data via visual representations and augmentation Tab2Visual:通过可视化表示和增强对有限表格数据进行深度学习
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-29 DOI: 10.1016/j.patcog.2026.113173
Ahmed Mamdouh , Moumen El-Melegy , Samia Ali , Ron Kikinis
This research addresses the challenge of limited data in tabular data classification, particularly prevalent in domains with constraints like healthcare. We propose Tab2Visual, a novel approach that transforms heterogeneous tabular data into visual representations, enabling the application of powerful deep learning models. Tab2Visual effectively addresses data scarcity by incorporating novel image augmentation techniques and facilitating transfer learning. We extensively evaluate the proposed approach on diverse tabular datasets, comparing its performance against a wide range of machine learning algorithms, including classical methods, tree-based ensembles, and state-of-the-art deep learning models specifically designed for tabular data. We also perform an in-depth analysis of factors influencing Tab2Visual’s performance. Our experimental results demonstrate that Tab2Visual outperforms other methods in classification problems with limited tabular data.
本研究解决了表格数据分类中有限数据的挑战,特别是在医疗保健等有限制的领域。我们提出了Tab2Visual,这是一种将异构表格数据转换为视觉表示的新方法,使强大的深度学习模型得以应用。Tab2Visual通过结合新的图像增强技术和促进迁移学习,有效地解决了数据稀缺问题。我们在不同的表格数据集上广泛评估了所提出的方法,并将其与各种机器学习算法(包括经典方法、基于树的集成和专门为表格数据设计的最先进的深度学习模型)的性能进行了比较。我们还对影响Tab2Visual性能的因素进行了深入的分析。我们的实验结果表明,Tab2Visual在有限表格数据的分类问题上优于其他方法。
{"title":"Tab2Visual: Deep learning for limited tabular data via visual representations and augmentation","authors":"Ahmed Mamdouh ,&nbsp;Moumen El-Melegy ,&nbsp;Samia Ali ,&nbsp;Ron Kikinis","doi":"10.1016/j.patcog.2026.113173","DOIUrl":"10.1016/j.patcog.2026.113173","url":null,"abstract":"<div><div>This research addresses the challenge of limited data in tabular data classification, particularly prevalent in domains with constraints like healthcare. We propose Tab2Visual, a novel approach that transforms heterogeneous tabular data into visual representations, enabling the application of powerful deep learning models. Tab2Visual effectively addresses data scarcity by incorporating novel image augmentation techniques and facilitating transfer learning. We extensively evaluate the proposed approach on diverse tabular datasets, comparing its performance against a wide range of machine learning algorithms, including classical methods, tree-based ensembles, and state-of-the-art deep learning models specifically designed for tabular data. We also perform an in-depth analysis of factors influencing Tab2Visual’s performance. Our experimental results demonstrate that Tab2Visual outperforms other methods in classification problems with limited tabular data.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113173"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Amplitude-guided deep reinforcement learning for semi-supervised layer segmentation 半监督层分割的幅度引导深度强化学习
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-30 DOI: 10.1016/j.patcog.2026.113204
Enting Gao , Zian Zha , Yonggang Li , Junhui Zhu , Yong Wang , Xinjian Chen , Naihui Zhou , Dehui Xiang
Accurate segmentation of scalp tissue layers is essential for mechanistic studies and staging of androgenetic alopecia (AGA), a common form of hair loss that impacts quality of life and mental health. High-resolution magnetic resonance imaging (HR-MR) offers a promising assessment tool. However, accurate segmentation remains challenging due to the lack of large-scale annotated datasets, structural deformation, and low image quality. To address these issues, an Amplitude-guided Deep Reinforcement Learning (ADRL) framework is designed to decouple the data distribution of images and adaptively fuse into the distribution of unlabeled images. This enables effective feature learning of lamellar and asymmetrically thickened structures from both labeled and unlabeled data. Then, phase component alignment (PHA) is imposed to mitigate the adverse impacts of noise or artifacts. To further enhance the discriminative capability of this network, a Cross-Power Spectrum Correlation (CPSC) module is proposed to mitigate inaccurate segmentation of layer structures. Comprehensive experiments on a scalp HR-MR image dataset and a publicly available retinal OCT image dataset demonstrate that our method significantly outperforms state-of-the-art methods in semi-supervised layer segmentation.
雄激素性脱发(AGA)是一种影响生活质量和心理健康的常见脱发,准确分割头皮组织层对于雄激素性脱发的机制研究和分期至关重要。高分辨率磁共振成像(HR-MR)是一种很有前途的评估工具。然而,由于缺乏大规模的注释数据集、结构变形和低图像质量,准确分割仍然具有挑战性。为了解决这些问题,设计了一个幅度引导深度强化学习(ADRL)框架来解耦图像的数据分布,并自适应地融合到未标记图像的分布中。这使得从标记和未标记数据中有效地学习层状和不对称加厚结构的特征成为可能。然后,相位分量对准(PHA)被施加以减轻噪声或伪影的不利影响。为了进一步提高该网络的判别能力,提出了一个交叉功率谱相关(CPSC)模块,以减轻层结构的不准确分割。在头皮HR-MR图像数据集和公开可用的视网膜OCT图像数据集上的综合实验表明,我们的方法在半监督层分割方面明显优于最先进的方法。
{"title":"Amplitude-guided deep reinforcement learning for semi-supervised layer segmentation","authors":"Enting Gao ,&nbsp;Zian Zha ,&nbsp;Yonggang Li ,&nbsp;Junhui Zhu ,&nbsp;Yong Wang ,&nbsp;Xinjian Chen ,&nbsp;Naihui Zhou ,&nbsp;Dehui Xiang","doi":"10.1016/j.patcog.2026.113204","DOIUrl":"10.1016/j.patcog.2026.113204","url":null,"abstract":"<div><div>Accurate segmentation of scalp tissue layers is essential for mechanistic studies and staging of androgenetic alopecia (AGA), a common form of hair loss that impacts quality of life and mental health. High-resolution magnetic resonance imaging (HR-MR) offers a promising assessment tool. However, accurate segmentation remains challenging due to the lack of large-scale annotated datasets, structural deformation, and low image quality. To address these issues, an Amplitude-guided Deep Reinforcement Learning (ADRL) framework is designed to decouple the data distribution of images and adaptively fuse into the distribution of unlabeled images. This enables effective feature learning of lamellar and asymmetrically thickened structures from both labeled and unlabeled data. Then, phase component alignment (PHA) is imposed to mitigate the adverse impacts of noise or artifacts. To further enhance the discriminative capability of this network, a Cross-Power Spectrum Correlation (CPSC) module is proposed to mitigate inaccurate segmentation of layer structures. Comprehensive experiments on a scalp HR-MR image dataset and a publicly available retinal OCT image dataset demonstrate that our method significantly outperforms state-of-the-art methods in semi-supervised layer segmentation.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113204"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-shot incremental food recognition via cross-domain guided pseudo-targets 基于跨域制导伪目标的少弹增量食物识别
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-07 DOI: 10.1016/j.patcog.2026.113280
Minkang Chai , Lu Wei , Zheng Qian , Ran Zhang , Ye Zhu
The explosive growth of global food culture has expanded the application scope of visual recognition; however, it has introduced complex challenges arising from high intra-class variability and inter-class similarity. However, existing systems struggle to address fine-grained confusion and the trade-off between retaining old knowledge and adapting to new information. Traditional methods are constrained by a heavy reliance on large-scale datasets, whereas emerging zero-shot techniques are prone to semantic hallucination when encountering unseen dishes, thereby posing a severe challenge to precise recognition. To address these challenges, we propose the Cross-domain Guided Food Pseudo-Target Estimation (CFPE) framework, establishing a novel paradigm that is vision-led and semantically enhanced. First, to tackle the scarcity of incremental data, we utilize cross-domain adversarial training and an adaptive mask generator to synthesize high-quality pseudo-targets, thus establishing stable geometric anchors within the feature space. Second, by integrating Bessel Estimation Loss of Hypersphere (BELH) and Perturbation Margin Enhanced Prototype Regularization (PMEPR), we geometrically reconstruct the hyperspherical manifold distribution of features, effectively correcting estimation biases induced by few-shot samples. Crucially, we introduce a Food Factor-based Visual Semantic Consistency (FVSC) constraint, which explicitly decouples fine-grained visual confusion by injecting structured semantics. This is complemented by a depth-aware feature decoupling strategy to dynamically balance the plasticity and stability of the model. Experimental results demonstrate that CFPE achieves state-of-the-art performance across multiple benchmark datasets. It not only significantly improves incremental learning accuracy but also exhibits exceptional robustness in recognizing high-entropy food images.
全球饮食文化的爆发式增长,扩大了视觉识别的应用范围;然而,它带来了复杂的挑战,这是由于阶级内部的高度变异性和阶级之间的相似性。然而,现有的系统很难处理细粒度的混淆,以及保留旧知识和适应新信息之间的权衡。传统方法受限于对大规模数据集的依赖,而新兴的零射击技术在遇到看不见的菜肴时容易产生语义幻觉,从而对精确识别提出了严峻的挑战。为了解决这些挑战,我们提出了跨域引导食品伪目标估计(CFPE)框架,建立了一个视觉主导和语义增强的新范式。首先,为了解决增量数据的稀缺性,我们利用跨域对抗训练和自适应掩码生成器来合成高质量的伪目标,从而在特征空间内建立稳定的几何锚点。其次,通过整合超球贝塞尔估计损失(BELH)和微扰边际增强原型正则化(PMEPR),从几何上重构了特征的超球流形分布,有效地纠正了由少量样本引起的估计偏差。关键是,我们引入了基于食物因子的视觉语义一致性(FVSC)约束,该约束通过注入结构化语义显式解耦细粒度视觉混淆。辅以深度感知特征解耦策略,动态平衡模型的可塑性和稳定性。实验结果表明,CFPE在多个基准数据集上实现了最先进的性能。它不仅显著提高了增量学习的准确性,而且在识别高熵食物图像方面表现出优异的鲁棒性。
{"title":"Few-shot incremental food recognition via cross-domain guided pseudo-targets","authors":"Minkang Chai ,&nbsp;Lu Wei ,&nbsp;Zheng Qian ,&nbsp;Ran Zhang ,&nbsp;Ye Zhu","doi":"10.1016/j.patcog.2026.113280","DOIUrl":"10.1016/j.patcog.2026.113280","url":null,"abstract":"<div><div>The explosive growth of global food culture has expanded the application scope of visual recognition; however, it has introduced complex challenges arising from high intra-class variability and inter-class similarity. However, existing systems struggle to address fine-grained confusion and the trade-off between retaining old knowledge and adapting to new information. Traditional methods are constrained by a heavy reliance on large-scale datasets, whereas emerging zero-shot techniques are prone to semantic hallucination when encountering unseen dishes, thereby posing a severe challenge to precise recognition. To address these challenges, we propose the Cross-domain Guided Food Pseudo-Target Estimation (CFPE) framework, establishing a novel paradigm that is vision-led and semantically enhanced. First, to tackle the scarcity of incremental data, we utilize cross-domain adversarial training and an adaptive mask generator to synthesize high-quality pseudo-targets, thus establishing stable geometric anchors within the feature space. Second, by integrating Bessel Estimation Loss of Hypersphere (BELH) and Perturbation Margin Enhanced Prototype Regularization (PMEPR), we geometrically reconstruct the hyperspherical manifold distribution of features, effectively correcting estimation biases induced by few-shot samples. Crucially, we introduce a Food Factor-based Visual Semantic Consistency (FVSC) constraint, which explicitly decouples fine-grained visual confusion by injecting structured semantics. This is complemented by a depth-aware feature decoupling strategy to dynamically balance the plasticity and stability of the model. Experimental results demonstrate that CFPE achieves state-of-the-art performance across multiple benchmark datasets. It not only significantly improves incremental learning accuracy but also exhibits exceptional robustness in recognizing high-entropy food images.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113280"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adversarial supervised contrastive feature learning for cross-modal retrieval 跨模态检索的对抗性监督对比特征学习
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-10 DOI: 10.1016/j.patcog.2026.113256
Xin Shu, Yikang Guo, Shou Gang Ren
Cross-modal hashing methods have attracted substantial interest in information retrieval because of their efficiency and low memory costs. Recent advancements in contrastive learning have greatly improved the retrieval performance of these hashing techniques. However, these approaches still encounter two significant drawbacks: (1) most current methods transform multimodal data into a common Hamming space to reduce the semantic gap, which may fail to capture the strong feature correlations across modalities; and (2) semantic similarity is represented as a binary value, neglecting the semantic relationships among multiple labels. To address these issues, we propose a novel adversarial supervised contrastive feature learning approach for cross-modal hashing. Specifically, we utilize a pre-trained CLIP model to extract multimodal features and apply contrastive learning to integrate these features effectively. Additionally, we introduce an adversarial feature learning mechanism to enhance the correlation between features from different modalities. Furthermore, we employ a graph convolutional network to model label correlations. Experimental results on benchmark datasets demonstrate the effectiveness and efficiency of our proposed method.
跨模态哈希方法由于其高效和低内存成本而引起了人们对信息检索的极大兴趣。对比学习的最新进展极大地提高了这些哈希技术的检索性能。然而,这些方法仍然存在两个明显的缺陷:(1)目前大多数方法将多模态数据转换为一个公共的Hamming空间,以减少语义差距,可能无法捕获模态之间的强特征相关性;(2)语义相似度用二值表示,忽略了多个标签之间的语义关系。为了解决这些问题,我们提出了一种新的对抗性监督对比特征学习方法用于跨模态哈希。具体而言,我们利用预训练的CLIP模型提取多模态特征,并应用对比学习有效地整合这些特征。此外,我们引入了一种对抗特征学习机制来增强不同模态特征之间的相关性。此外,我们使用图卷积网络来建模标签相关性。在基准数据集上的实验结果证明了该方法的有效性和高效性。
{"title":"Adversarial supervised contrastive feature learning for cross-modal retrieval","authors":"Xin Shu,&nbsp;Yikang Guo,&nbsp;Shou Gang Ren","doi":"10.1016/j.patcog.2026.113256","DOIUrl":"10.1016/j.patcog.2026.113256","url":null,"abstract":"<div><div>Cross-modal hashing methods have attracted substantial interest in information retrieval because of their efficiency and low memory costs. Recent advancements in contrastive learning have greatly improved the retrieval performance of these hashing techniques. However, these approaches still encounter two significant drawbacks: (1) most current methods transform multimodal data into a common Hamming space to reduce the semantic gap, which may fail to capture the strong feature correlations across modalities; and (2) semantic similarity is represented as a binary value, neglecting the semantic relationships among multiple labels. To address these issues, we propose a novel adversarial supervised contrastive feature learning approach for cross-modal hashing. Specifically, we utilize a pre-trained CLIP model to extract multimodal features and apply contrastive learning to integrate these features effectively. Additionally, we introduce an adversarial feature learning mechanism to enhance the correlation between features from different modalities. Furthermore, we employ a graph convolutional network to model label correlations. Experimental results on benchmark datasets demonstrate the effectiveness and efficiency of our proposed method.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113256"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DuoNet: Joint optimization of representation learning and prototype classifier for unbiased scene graph generation DuoNet:用于无偏场景图生成的表示学习和原型分类器联合优化
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-27 DOI: 10.1016/j.patcog.2026.113152
Zhaodi Wang , Biao Leng , Shuo Zhang
Unbiased Scene Graph Generation (SGG) aims to parse visual scenes into highly informative graphs under the long-tail challenge. While prototype-based methods have shown promise in unbiased SGG, they highlight the importance of learning discriminative features that are intra-class compact and inter-class separable. In this paper, we revisit prototype-based methods and analyze critical roles of representation learning and prototype classifier in driving unbiased SGG, and accordingly propose a novel framework DuoNet. To enhance intra-class compactness, we introduce a Bi-Directional Representation Refinement (BiDR2) module that captures relation-sensitive visual variability and within-relation visual consistency of entities. This module adopts relation-to-entity-to-relation refinement by integrating dual-level relation pattern modeling with a relation-specific entity constraint. Furthermore, a Knowledge-Guided Prototype Learning (KGPL) module is devised to strengthen inter-class separability by constructing an equidistributed prototypical classifier with maximum inter-class margins. The equidistributed prototype classifier is frozen during SGG training to mitigate long-tail bias, thus a knowledge-driven triplet loss is developed to strengthen the learning of BiDR2, enhancing relation-prototype matching. Extensive experiments demonstrate the effectiveness of our method, which sets new state-of-the-art performance on Visual Genome, GQA and Open Images datasets.
无偏场景图生成(Unbiased Scene Graph Generation, SGG)的目标是在长尾挑战下将视觉场景解析成高信息量的图。虽然基于原型的方法在无偏SGG中显示出了希望,但它们强调了学习类内紧凑和类间可分离的判别特征的重要性。本文回顾了基于原型的方法,分析了表征学习和原型分类器在驱动无偏SGG中的关键作用,并据此提出了一个新的框架DuoNet。为了增强类内的紧凑性,我们引入了双向表示细化(BiDR2)模块,该模块捕获关系敏感的视觉可变性和实体的关系内视觉一致性。该模块通过将双级关系模式建模与特定于关系的实体约束集成,采用关系到实体到关系的细化。在此基础上,设计了知识引导原型学习(knowledge guided Prototype Learning, KGPL)模块,通过构造类间边界最大的等分布原型分类器来增强类间可分性。在SGG训练过程中冻结等分布原型分类器以减轻长尾偏差,因此开发了知识驱动的三重损失来加强BiDR2的学习,增强关系原型匹配。大量的实验证明了我们的方法的有效性,它在视觉基因组,GQA和开放图像数据集上设置了新的最先进的性能。
{"title":"DuoNet: Joint optimization of representation learning and prototype classifier for unbiased scene graph generation","authors":"Zhaodi Wang ,&nbsp;Biao Leng ,&nbsp;Shuo Zhang","doi":"10.1016/j.patcog.2026.113152","DOIUrl":"10.1016/j.patcog.2026.113152","url":null,"abstract":"<div><div>Unbiased Scene Graph Generation (SGG) aims to parse visual scenes into highly informative graphs under the long-tail challenge. While prototype-based methods have shown promise in unbiased SGG, they highlight the importance of learning discriminative features that are intra-class compact and inter-class separable. In this paper, we revisit prototype-based methods and analyze critical roles of representation learning and prototype classifier in driving unbiased SGG, and accordingly propose a novel framework DuoNet. To enhance intra-class compactness, we introduce a Bi-Directional Representation Refinement (BiDR<sup>2</sup>) module that captures relation-sensitive visual variability and within-relation visual consistency of entities. This module adopts relation-to-entity-to-relation refinement by integrating dual-level relation pattern modeling with a relation-specific entity constraint. Furthermore, a Knowledge-Guided Prototype Learning (KGPL) module is devised to strengthen inter-class separability by constructing an equidistributed prototypical classifier with maximum inter-class margins. The equidistributed prototype classifier is frozen during SGG training to mitigate long-tail bias, thus a knowledge-driven triplet loss is developed to strengthen the learning of BiDR<sup>2</sup>, enhancing relation-prototype matching. Extensive experiments demonstrate the effectiveness of our method, which sets new state-of-the-art performance on Visual Genome, GQA and Open Images datasets.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113152"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SCALAR: Spatial-concept alignment for robust vision in harsh open world 标量:在严酷的开放世界中实现健壮视觉的空间概念对齐
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-03 DOI: 10.1016/j.patcog.2026.113203
Xiaoyu Yang , Lijian Xu , Xingyu Zeng , Xiaosong Wang , Hongsheng Li , Shaoting Zhang
Foundation models have recently transformed visual-linguistic representation learning, yet their robustness under adverse imaging conditions of open worlds remains insufficiently understood. In this work, we introduce SCALAR, a scene-aware framework that endows multi-modal large language models with enhanced capability for robust spatial-concept alignment in degraded visual environments of open worlds. SCALAR proceeds in two complementary stages. The supervised alignment stage reconstructs hierarchical concept chains from visual-linguistic corpora, thereby enabling efficient spatial relationship decoding. The subsequent reinforced fine-tuning stage dispenses with annotations and leverages a consistency-driven reward to facilitate open-world self-evolution, yielding improved adaptability across diverse degraded domains. Crucially, SCALAR jointly optimizes multi-dimensional spatial representations and heterogeneous knowledge structures, thereby fostering resilience and generalization beyond canonical benchmarks. Extensive evaluations across five tasks and eight large-scale datasets demonstrate the efficacy of SCALAR in advancing state-of-the-art performance on visual grounding and complex scene understanding, even under challenging open-world environments with harsh visual conditions. Comprehensive ablation studies further elucidate the contributions of reinforced fine-tuning and multi-task joint optimization. Finally, to encourage future research, we provide a new multi-task visual grounding dataset emphasizing fine-grained scene-object relations under degradation, along with code: https://github.com/AnonymGiant/SCALAR.
基础模型最近改变了视觉语言表征学习,但它们在开放世界的不利成像条件下的鲁棒性仍然没有得到充分的理解。在这项工作中,我们引入了一个场景感知框架SCALAR,它赋予多模态大型语言模型在开放世界的退化视觉环境中具有增强的鲁棒空间概念对齐能力。SCALAR分两个互补的阶段进行。监督对齐阶段从视觉语言语料库中重构层次概念链,从而实现高效的空间关系解码。随后的强化微调阶段省去了注释,并利用一致性驱动的奖励来促进开放世界的自我进化,从而提高了对不同退化领域的适应性。至关重要的是,SCALAR联合优化了多维空间表示和异构知识结构,从而促进了超越规范基准的弹性和泛化。对五个任务和八个大规模数据集的广泛评估表明,即使在具有恶劣视觉条件的具有挑战性的开放世界环境中,SCALAR在提高视觉基础和复杂场景理解方面的最先进性能方面的有效性。综合消融研究进一步阐明了强化微调和多任务关节优化的贡献。最后,为了鼓励未来的研究,我们提供了一个新的多任务视觉基础数据集,强调在退化下的细粒度场景-对象关系,以及代码:https://github.com/AnonymGiant/SCALAR。
{"title":"SCALAR: Spatial-concept alignment for robust vision in harsh open world","authors":"Xiaoyu Yang ,&nbsp;Lijian Xu ,&nbsp;Xingyu Zeng ,&nbsp;Xiaosong Wang ,&nbsp;Hongsheng Li ,&nbsp;Shaoting Zhang","doi":"10.1016/j.patcog.2026.113203","DOIUrl":"10.1016/j.patcog.2026.113203","url":null,"abstract":"<div><div>Foundation models have recently transformed visual-linguistic representation learning, yet their robustness under adverse imaging conditions of open worlds remains insufficiently understood. In this work, we introduce SCALAR, a scene-aware framework that endows multi-modal large language models with enhanced capability for robust spatial-concept alignment in degraded visual environments of open worlds. SCALAR proceeds in two complementary stages. The supervised alignment stage reconstructs hierarchical concept chains from visual-linguistic corpora, thereby enabling efficient spatial relationship decoding. The subsequent reinforced fine-tuning stage dispenses with annotations and leverages a consistency-driven reward to facilitate open-world self-evolution, yielding improved adaptability across diverse degraded domains. Crucially, SCALAR jointly optimizes multi-dimensional spatial representations and heterogeneous knowledge structures, thereby fostering resilience and generalization beyond canonical benchmarks. Extensive evaluations across five tasks and eight large-scale datasets demonstrate the efficacy of SCALAR in advancing state-of-the-art performance on visual grounding and complex scene understanding, even under challenging open-world environments with harsh visual conditions. Comprehensive ablation studies further elucidate the contributions of reinforced fine-tuning and multi-task joint optimization. Finally, to encourage future research, we provide a new multi-task visual grounding dataset emphasizing fine-grained scene-object relations under degradation, along with code: <span><span>https://github.com/AnonymGiant/SCALAR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113203"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CSAFNet: Cross-modal spatial alignment and fusion network for RGB-T crowd counting CSAFNet: RGB-T人群计数的跨模态空间对齐与融合网络
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-07 DOI: 10.1016/j.patcog.2026.113250
Yongjie Zhao , Liuru Pu , Huaibo Song , Bo Jiang
Crowd counting is critical for public safety and urban management in smart cities, yet faces challenges in complex scenarios. While RGB-Thermal (RGB-T) fusion helps address information loss in low-light conditions, current methods still suffer from two key limitations. (a) Existing RGB-T crowd counting methods fail to address the spatial misalignment between RGB and thermal features caused by different capturing devices, which diminishes fusion performance and impedes improvements in crowd counting accuracy. (b) Current methods fail to adequately distinguish between specific and common features of RGB and thermal modalities, leading to redundant feature fusion that compromises feature representation and results in suboptimal counting performance. To address the aforementioned challenges, the Cross-modal Spatial Alignment and Fusion Network (CSAFNet) is proposed. CSAFNet integrates three novel modules: the Cross-modal Feature Space Alignment (CFSA), Multiscale Spatia l Displacement Compensation (MSDC) and the Cross-modal Feature Decoupling Fusion (CFDF) modules. The CFSA module performs precise spatial alignment via feature windows and achieves wide spatial consistency through the MSDC module. The CFDF module employs Kullback-Leibler divergence and Jensen-Shannon divergence to perform decoupled fusion of cross-modal features, preserving modality-specific details, enhancing cross-modal commonalities, reducing redundant features, and strengthening discriminative feature representation. Extensive experiments demonstrate that the proposed CSAFNet achieves competitive performance on the RGBT-CC dataset, reducing GAME(0) to 10.75 and RMSE to 17.91. These results validate the effectiveness and promising potential of CSAFNet for cross-modal crowd counting tasks. Code is released at https://github.com/Zyjer888/CSAFNet.
人群统计对智慧城市的公共安全和城市管理至关重要,但在复杂的场景下面临挑战。虽然RGB-Thermal (RGB-T)融合有助于解决弱光条件下的信息丢失问题,但目前的方法仍然存在两个关键限制。(a)现有的RGB- t人群计数方法未能解决由于不同的捕获设备导致的RGB与热特征之间的空间不对准问题,从而降低了融合性能,阻碍了人群计数精度的提高。(b)目前的方法不能充分区分RGB和热模态的特定特征和共同特征,导致冗余特征融合,损害特征表示并导致次优计数性能。为了解决上述挑战,提出了跨模态空间对齐和融合网络(CSAFNet)。CSAFNet集成了三个新模块:跨模态特征空间对齐(CFSA)、多尺度空间位移补偿(MSDC)和跨模态特征解耦融合(CFDF)模块。CFSA模块通过特征窗口实现精确的空间对准,并通过MSDC模块实现广泛的空间一致性。CFDF模块采用Kullback-Leibler散度和Jensen-Shannon散度对跨模态特征进行解耦融合,保留模态特定细节,增强跨模态共性,减少冗余特征,加强判别特征表示。大量的实验表明,所提出的CSAFNet在RGBT-CC数据集上取得了具有竞争力的性能,将GAME(0)降低到10.75,RMSE降低到17.91。这些结果验证了CSAFNet在跨模式人群计数任务中的有效性和潜力。代码发布在https://github.com/Zyjer888/CSAFNet。
{"title":"CSAFNet: Cross-modal spatial alignment and fusion network for RGB-T crowd counting","authors":"Yongjie Zhao ,&nbsp;Liuru Pu ,&nbsp;Huaibo Song ,&nbsp;Bo Jiang","doi":"10.1016/j.patcog.2026.113250","DOIUrl":"10.1016/j.patcog.2026.113250","url":null,"abstract":"<div><div>Crowd counting is critical for public safety and urban management in smart cities, yet faces challenges in complex scenarios. While RGB-Thermal (RGB-T) fusion helps address information loss in low-light conditions, current methods still suffer from two key limitations. (a) Existing RGB-T crowd counting methods fail to address the spatial misalignment between RGB and thermal features caused by different capturing devices, which diminishes fusion performance and impedes improvements in crowd counting accuracy. (b) Current methods fail to adequately distinguish between specific and common features of RGB and thermal modalities, leading to redundant feature fusion that compromises feature representation and results in suboptimal counting performance. To address the aforementioned challenges, the Cross-modal Spatial Alignment and Fusion Network (CSAFNet) is proposed. CSAFNet integrates three novel modules: the Cross-modal Feature Space Alignment (CFSA), Multiscale Spatia l Displacement Compensation (MSDC) and the Cross-modal Feature Decoupling Fusion (CFDF) modules. The CFSA module performs precise spatial alignment via feature windows and achieves wide spatial consistency through the MSDC module. The CFDF module employs Kullback-Leibler divergence and Jensen-Shannon divergence to perform decoupled fusion of cross-modal features, preserving modality-specific details, enhancing cross-modal commonalities, reducing redundant features, and strengthening discriminative feature representation. Extensive experiments demonstrate that the proposed CSAFNet achieves competitive performance on the RGBT-CC dataset, reducing GAME(0) to 10.75 and RMSE to 17.91. These results validate the effectiveness and promising potential of CSAFNet for cross-modal crowd counting tasks. <em><strong>Code is released at</strong></em> <span><span>https://github.com/Zyjer888/CSAFNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113250"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MG-TVMF: Multi-grained text-video matching and fusing for weakly supervised video anomaly detection MG-TVMF:用于弱监督视频异常检测的多粒度文本视频匹配和融合
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-03 DOI: 10.1016/j.patcog.2026.113201
Ping He, Xiaonan Gao, Huibin Li
Weakly supervised video anomaly detection (WS-VAD) often suffers from false alarms and incomplete localization due to the lack of precise temporal annotations. To address these limitations, we propose a novel method, multi-grained text-video matching and fusing (MG-TVMF), which leverages semantic cues from anomaly category text labels to enhance both the accuracy and completeness of anomaly localization. MG-TVMF integrates two complementary branches: the MG-TVM branch improves localization accuracy through a hierarchical structure comprising a coarse-grained classification module and two fine-grained matching modules, including a video-text matching (VTM) module for global semantic alignment and a segment-text matching (STM) module for local video (i.e. segment) text alignment via optimal transport algorithm. Meanwhile, the MG-TVF branch enhances localization completeness by prepending a global video-level text prompt to each segment-level caption for multi-grained textual fusion, and reconstructing the masked anomaly-related caption of the top-scoring segment using video segment features and anomaly scores. Extensive experiments on the UCF-Crime and XD-Violence datasets demonstrate the effectiveness of the proposed VTM and STM modules as well as the MG-TVF branch, and the proposed MG-TVMF method achieves state-of-the-art performance on UCF-Crime, XD-Violence, and ShanghaiTech datasets.
弱监督视频异常检测(WS-VAD)由于缺乏精确的时间标注,常常存在误报和定位不完整的问题。为了解决这些限制,我们提出了一种新的方法,多粒度文本视频匹配和融合(MG-TVMF),该方法利用异常类别文本标签的语义线索来提高异常定位的准确性和完整性。MG-TVMF集成了两个互补的分支:MG-TVM分支通过一个由粗粒度分类模块和两个细粒度匹配模块组成的层次结构来提高定位精度,其中包括一个用于全局语义对齐的视频文本匹配(VTM)模块和一个通过最优传输算法用于局部视频(即片段)文本对齐的段文本匹配(STM)模块。同时,MG-TVF分支通过在每个片段级标题前添加全局视频级文本提示进行多粒度文本融合,增强定位完整性,并利用视频片段特征和异常分数重构得分最高的片段的屏蔽异常相关标题。在UCF-Crime和XD-Violence数据集上的大量实验证明了所提出的VTM和STM模块以及MG-TVF分支的有效性,所提出的MG-TVMF方法在UCF-Crime、XD-Violence和ShanghaiTech数据集上实现了最先进的性能。
{"title":"MG-TVMF: Multi-grained text-video matching and fusing for weakly supervised video anomaly detection","authors":"Ping He,&nbsp;Xiaonan Gao,&nbsp;Huibin Li","doi":"10.1016/j.patcog.2026.113201","DOIUrl":"10.1016/j.patcog.2026.113201","url":null,"abstract":"<div><div>Weakly supervised video anomaly detection (WS-VAD) often suffers from false alarms and incomplete localization due to the lack of precise temporal annotations. To address these limitations, we propose a novel method, multi-grained text-video matching and fusing (MG-TVMF), which leverages semantic cues from anomaly category text labels to enhance both the accuracy and completeness of anomaly localization. MG-TVMF integrates two complementary branches: the MG-TVM branch improves localization accuracy through a hierarchical structure comprising a coarse-grained classification module and two fine-grained matching modules, including a video-text matching (VTM) module for global semantic alignment and a segment-text matching (STM) module for local video (i.e. segment) text alignment via optimal transport algorithm. Meanwhile, the MG-TVF branch enhances localization completeness by prepending a global video-level text prompt to each segment-level caption for multi-grained textual fusion, and reconstructing the masked anomaly-related caption of the top-scoring segment using video segment features and anomaly scores. Extensive experiments on the UCF-Crime and XD-Violence datasets demonstrate the effectiveness of the proposed VTM and STM modules as well as the MG-TVF branch, and the proposed MG-TVMF method achieves state-of-the-art performance on UCF-Crime, XD-Violence, and ShanghaiTech datasets.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113201"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain generalization via domain uncertainty shrinkage 通过域不确定性收缩进行域泛化
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-23 DOI: 10.1016/j.patcog.2026.113118
Jun-Zheng Chu , Bin Pan , Tian-Yang Shi , Zhen-Wei Shi
Ensuring model robustness against distributional shifts still presents a significant challenge in many machine learning applications. To address this issue, a wide range of domain generalization (DG) methods have been developed. However, these approaches mainly focus on invariant representations by leveraging multiple source domain data, which ignore the uncertainty presented from different domains. In this paper, we establish a novel DG framework in form of evidential deep learning (EDL-DG). To reach DG objective under finite given domains, we propose a new Domain Uncertainty Shrinkage (DUS) regularization scheme on the output Dirichlet distribution parameters, which achieves better generalization across unseen domains without introducing additional structures. Theoretically, we analyze the convergence of EDL-DG, and provide a generalization bound in the framework of PAC-Bayesian learning. We show that our proposed method reduce the PAC-Bayesian bound under certain conditions, and thus achieve better generalization across unseen domains. In our experiments, we validate the effectiveness our proposed method on DomainBed benchmark in multiple real-world datasets.
在许多机器学习应用中,确保模型对分布变化的鲁棒性仍然是一个重大挑战。为了解决这个问题,已经开发了各种领域泛化(DG)方法。然而,这些方法主要侧重于利用多源领域数据的不变表示,而忽略了来自不同领域的不确定性。在本文中,我们以证据深度学习(EDL-DG)的形式建立了一个新的DG框架。为了在有限给定域下达到DG目标,我们对输出的Dirichlet分布参数提出了一种新的域不确定性收缩(DUS)正则化方案,该方案在不引入额外结构的情况下实现了更好的跨未知域的泛化。从理论上分析了EDL-DG的收敛性,并给出了PAC-Bayesian学习框架下的泛化界。在一定条件下,我们的方法减小了PAC-Bayesian边界,从而实现了更好的跨未知域的泛化。在我们的实验中,我们在多个真实数据集的DomainBed基准上验证了我们提出的方法的有效性。
{"title":"Domain generalization via domain uncertainty shrinkage","authors":"Jun-Zheng Chu ,&nbsp;Bin Pan ,&nbsp;Tian-Yang Shi ,&nbsp;Zhen-Wei Shi","doi":"10.1016/j.patcog.2026.113118","DOIUrl":"10.1016/j.patcog.2026.113118","url":null,"abstract":"<div><div>Ensuring model robustness against distributional shifts still presents a significant challenge in many machine learning applications. To address this issue, a wide range of domain generalization (DG) methods have been developed. However, these approaches mainly focus on invariant representations by leveraging multiple source domain data, which ignore the uncertainty presented from different domains. In this paper, we establish a novel DG framework in form of evidential deep learning (EDL-DG). To reach DG objective under finite given domains, we propose a new <em>Domain Uncertainty Shrinkage</em> (<strong>DUS</strong>) regularization scheme on the output Dirichlet distribution parameters, which achieves better generalization across unseen domains without introducing additional structures. Theoretically, we analyze the convergence of EDL-DG, and provide a generalization bound in the framework of PAC-Bayesian learning. We show that our proposed method reduce the PAC-Bayesian bound under certain conditions, and thus achieve better generalization across unseen domains. In our experiments, we validate the effectiveness our proposed method on DomainBed benchmark in multiple real-world datasets.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113118"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
YOLO-PICO: Lightweight object recognition in remote sensing images using expansion attention modules YOLO-PICO:使用扩展注意力模块的遥感图像中的轻量级目标识别
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-17 DOI: 10.1016/j.patcog.2026.113114
Mohamad Ebrahim Aghili, Hassan Ghassemian, Maryam Imani
Recognizing small objects in remote sensing imagery remains a significant challenge. This paper introduces YOLO-PICO, a novel and highly efficient object detector designed for small object recognition. At its core is the Expansion Attention (EA) Module, a new operator for spatial-channel feature fusion that enhances fine-grained details with minimal computational cost. This allows YOLO-PICO to achieve competitive performance with significantly fewer parameters than existing models, as demonstrated by our new parameter efficiency metric, Size-Normalized Average Precision (SNAP). Furthermore, we show that YOLO-PICO's efficiency makes it an ideal foundation for an Ensemble of Specialists (EoS) framework, a decision-level fusion strategy that substantially boosts detection accuracy with a modest increase in inference time. Our results demonstrate that this combination of an efficient core model and an advanced fusion strategy offers a compelling solution for high-performance recognition on resource-constrained platforms. The code will be made available at: https://github.com/MohamadEbrahimAghili/YOLO-PICO.
在遥感图像中识别小目标仍然是一个重大挑战。本文介绍了一种用于小目标识别的新型高效目标检测器YOLO-PICO。其核心是扩展注意力(EA)模块,这是一种用于空间信道特征融合的新算子,可以以最小的计算成本增强细粒度细节。这使得YOLO-PICO能够以比现有模型少得多的参数实现具有竞争力的性能,正如我们新的参数效率指标——尺寸归一化平均精度(SNAP)所证明的那样。此外,我们表明YOLO-PICO的效率使其成为专家集成(EoS)框架的理想基础,这是一种决策级融合策略,可以在适度增加推理时间的情况下大幅提高检测精度。我们的研究结果表明,这种高效核心模型和先进融合策略的结合为资源受限平台上的高性能识别提供了令人信服的解决方案。代码将在https://github.com/MohamadEbrahimAghili/YOLO-PICO上提供。
{"title":"YOLO-PICO: Lightweight object recognition in remote sensing images using expansion attention modules","authors":"Mohamad Ebrahim Aghili,&nbsp;Hassan Ghassemian,&nbsp;Maryam Imani","doi":"10.1016/j.patcog.2026.113114","DOIUrl":"10.1016/j.patcog.2026.113114","url":null,"abstract":"<div><div>Recognizing small objects in remote sensing imagery remains a significant challenge. This paper introduces YOLO-PICO, a novel and highly efficient object detector designed for small object recognition. At its core is the Expansion Attention (EA) Module, a new operator for spatial-channel feature fusion that enhances fine-grained details with minimal computational cost. This allows YOLO-PICO to achieve competitive performance with significantly fewer parameters than existing models, as demonstrated by our new parameter efficiency metric, Size-Normalized Average Precision (SNAP). Furthermore, we show that YOLO-PICO's efficiency makes it an ideal foundation for an Ensemble of Specialists (EoS) framework, a decision-level fusion strategy that substantially boosts detection accuracy with a modest increase in inference time. Our results demonstrate that this combination of an efficient core model and an advanced fusion strategy offers a compelling solution for high-performance recognition on resource-constrained platforms. The code will be made available at: <span><span>https://github.com/MohamadEbrahimAghili/YOLO-PICO</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113114"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1