首页 > 最新文献

Pattern Recognition最新文献

英文 中文
Adaptive centroid guided hashing for cross-modal retrieval 跨模态检索的自适应质心引导哈希
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-29 DOI: 10.1016/j.patcog.2026.113186
Zhenqiu Shu, Julong Zhang, Zhengtao Yu
Deep hashing technology is widely used in cross-modal retrieval tasks due to its low storage costs and high computational efficiency. However, most existing supervised hashing methods suffer from the following challenges: (1) Relying on manually labeled semantic affinity levels as supervisory information for hash learning may ignore the underlying structure of semantic information, potentially resulting in semantic structure degradation. (2) They fail to consider both the semantic relationships among labels and the relative significance of each label to individual samples. To address these challenges, we propose a novel adaptive centroid guided hashing (ACGH) method for cross-modal retrieval. Specifically, we extract global and local features using Transformer models, and then fuse them to obtain fine-grained feature representations of multimodal data. Subsequently, the hash centroid generation module leverages the category semantic embedding to construct category hash centers and combine them with learnable Label-Affinity Coefficients (LAC) memory banks to learn adaptive hash centroids. Furthermore, we design a hash centroid guidance module, which employs the hash centroids to guide hash code learning and then updates the hash centers and LAC memory banks through the newly learned hash codes. Extensive experimental results on several benchmark multimodal datasets demonstrate that the proposed ACGH method significantly outperforms other state-of-the-art methods in cross-modal retrieval tasks.
深度哈希技术以其存储成本低、计算效率高的特点被广泛应用于跨模态检索任务中。然而,现有的大多数监督哈希方法都面临以下挑战:(1)依赖人工标记的语义亲和度作为哈希学习的监督信息,可能会忽略语义信息的底层结构,可能导致语义结构退化。(2)他们没有考虑标签之间的语义关系和每个标签对单个样本的相对重要性。为了解决这些挑战,我们提出了一种新的自适应质心引导哈希(ACGH)方法用于跨模态检索。具体来说,我们使用Transformer模型提取全局和局部特征,然后融合它们以获得多模态数据的细粒度特征表示。随后,哈希质心生成模块利用类别语义嵌入构造类别哈希中心,并将其与可学习的标签亲和系数(LAC)记忆库结合学习自适应哈希质心。此外,我们设计了一个哈希质心引导模块,利用哈希质心引导哈希码学习,然后通过新学习的哈希码更新哈希中心和LAC内存库。在多个基准多模态数据集上的大量实验结果表明,所提出的ACGH方法在跨模态检索任务中显著优于其他最先进的方法。
{"title":"Adaptive centroid guided hashing for cross-modal retrieval","authors":"Zhenqiu Shu,&nbsp;Julong Zhang,&nbsp;Zhengtao Yu","doi":"10.1016/j.patcog.2026.113186","DOIUrl":"10.1016/j.patcog.2026.113186","url":null,"abstract":"<div><div>Deep hashing technology is widely used in cross-modal retrieval tasks due to its low storage costs and high computational efficiency. However, most existing supervised hashing methods suffer from the following challenges: (1) Relying on manually labeled semantic affinity levels as supervisory information for hash learning may ignore the underlying structure of semantic information, potentially resulting in semantic structure degradation. (2) They fail to consider both the semantic relationships among labels and the relative significance of each label to individual samples. To address these challenges, we propose a novel adaptive centroid guided hashing (ACGH) method for cross-modal retrieval. Specifically, we extract global and local features using Transformer models, and then fuse them to obtain fine-grained feature representations of multimodal data. Subsequently, the hash centroid generation module leverages the category semantic embedding to construct category hash centers and combine them with learnable Label-Affinity Coefficients (LAC) memory banks to learn adaptive hash centroids. Furthermore, we design a hash centroid guidance module, which employs the hash centroids to guide hash code learning and then updates the hash centers and LAC memory banks through the newly learned hash codes. Extensive experimental results on several benchmark multimodal datasets demonstrate that the proposed ACGH method significantly outperforms other state-of-the-art methods in cross-modal retrieval tasks.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113186"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interaction-aware adaptive network for drug-drug interaction prediction 药物-药物相互作用预测的相互作用感知自适应网络
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-01-29 DOI: 10.1016/j.patcog.2026.113172
Dongjiang Niu , Xiaofeng Wang , Zengqian Deng , Bowen Tang , Zhen Li
The prediction of drug-drug interactions (DDI) is crucial for drug safety and combination therapies. However, existing computational approaches face significant challenges in modeling drug interactions and effectively integrating multi-view information. To this end, AMIE-DDI, an Adaptive Multi-view Integration framework is proposed. First, Interaction-Enhanced Graph Transformer is designed to model complex relationships between drugs and capture the underlying interaction mechanisms. Second, a Multi-Channel Adaptive Fusion Module (MAF) is introduced to dynamically integrate information from different representations, enhancing feature learning and ensuring efficient multi-view feature integration. Finally, a Dynamic Interaction Scaling Prediction Module (DIS) is developed to adaptively adjust interaction intensity, thus improving both predictive accuracy and stability. Experimental results on multiple datasets demonstrate that AMIE-DDI outperforms state-of-the-art baselines in both warm-start and cold-start scenarios. Moreover, ablation studies and visualization experiments validate its capability to capture key motifs and enhance DDI prediction accuracy.
药物相互作用(DDI)的预测对药物安全和联合治疗至关重要。然而,现有的计算方法在模拟药物相互作用和有效整合多视图信息方面面临重大挑战。为此,提出了自适应多视图集成框架mie - ddi。首先,交互增强图转换器设计用于模拟药物之间的复杂关系并捕获潜在的交互机制。其次,引入多通道自适应融合模块(MAF),对不同表征的信息进行动态集成,增强特征学习能力,保证多视图特征的高效集成;最后,开发了动态交互尺度预测模块(DIS),自适应调整交互强度,提高了预测精度和稳定性。在多个数据集上的实验结果表明,ami - ddi在热启动和冷启动场景下都优于最先进的基线。此外,消融研究和可视化实验验证了其捕获关键基序和提高DDI预测精度的能力。
{"title":"Interaction-aware adaptive network for drug-drug interaction prediction","authors":"Dongjiang Niu ,&nbsp;Xiaofeng Wang ,&nbsp;Zengqian Deng ,&nbsp;Bowen Tang ,&nbsp;Zhen Li","doi":"10.1016/j.patcog.2026.113172","DOIUrl":"10.1016/j.patcog.2026.113172","url":null,"abstract":"<div><div>The prediction of drug-drug interactions (DDI) is crucial for drug safety and combination therapies. However, existing computational approaches face significant challenges in modeling drug interactions and effectively integrating multi-view information. To this end, AMIE-DDI, an Adaptive Multi-view Integration framework is proposed. First, Interaction-Enhanced Graph Transformer is designed to model complex relationships between drugs and capture the underlying interaction mechanisms. Second, a Multi-Channel Adaptive Fusion Module (MAF) is introduced to dynamically integrate information from different representations, enhancing feature learning and ensuring efficient multi-view feature integration. Finally, a Dynamic Interaction Scaling Prediction Module (DIS) is developed to adaptively adjust interaction intensity, thus improving both predictive accuracy and stability. Experimental results on multiple datasets demonstrate that AMIE-DDI outperforms state-of-the-art baselines in both warm-start and cold-start scenarios. Moreover, ablation studies and visualization experiments validate its capability to capture key motifs and enhance DDI prediction accuracy.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113172"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Panoptic-VSNet: Visual-semantic prior knowledge-driven multimodal 3D panoptic segmentation panoptic - vsnet:视觉语义先验知识驱动的多模态三维全景分割
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-05 DOI: 10.1016/j.patcog.2026.113239
Xiao Li , Hui Li , Xiangzhen Kong , Yuang Ji , Zhiyu Liu , Hao Liu
Precise and robust perception is critical for ensuring the safe operation of autonomous vehicles. However, current methods are constrained by sparse image-LiDAR alignment, insufficient annotations, and ineffective structural discrepancy modeling, causing semantic degradation and generalization deficiency. Therefore, we propose Panoptic-VSNet, a visual-semantic prior knowledge-driven multimodal 3D panoptic segmentation network. Firstly, we design a progressive fusion semantic alignment module that effectively aggregates visual prior features obtained from the large Visual-Language model, establishing a point-semantic region association, thereby enhancing semantic awareness. Secondly, we propose an instance-aware superpixel cross-modal fusion module that incorporates instance prior knowledge, forming a unified representation with spatial precision and class consistency. Finally, we introduce a correlation-aware adaptive panoptic segmentation network that reduces parameter count while dynamically capturing contextual information and enhancing local details, thereby improving panoptic perception capabilities. Experimental evaluations on benchmark datasets show that Panoptic-VSNet outperforms state-of-the-art methods. Code is available at https://github.com/lixiao0125/panoptic-vsnet.git.
精确而稳健的感知对于确保自动驾驶汽车的安全运行至关重要。然而,目前的方法受到稀疏图像- lidar对齐、注释不足和结构差异建模无效的限制,导致语义退化和泛化不足。因此,我们提出了一种视觉语义先验知识驱动的多模态三维全景分割网络panoptic - vsnet。首先,我们设计了一个递进式融合语义对齐模块,该模块有效地聚合了从大型视觉语言模型中获得的视觉先验特征,建立了点-语义区域关联,从而增强了语义感知;其次,我们提出了一个实例感知的超像素跨模态融合模块,该模块融合了实例先验知识,形成了具有空间精度和类一致性的统一表示。最后,我们引入了一种相关感知的自适应泛光分割网络,该网络在动态捕获上下文信息和增强局部细节的同时减少了参数计数,从而提高了泛光感知能力。对基准数据集的实验评估表明,Panoptic-VSNet优于最先进的方法。代码可从https://github.com/lixiao0125/panoptic-vsnet.git获得。
{"title":"Panoptic-VSNet: Visual-semantic prior knowledge-driven multimodal 3D panoptic segmentation","authors":"Xiao Li ,&nbsp;Hui Li ,&nbsp;Xiangzhen Kong ,&nbsp;Yuang Ji ,&nbsp;Zhiyu Liu ,&nbsp;Hao Liu","doi":"10.1016/j.patcog.2026.113239","DOIUrl":"10.1016/j.patcog.2026.113239","url":null,"abstract":"<div><div>Precise and robust perception is critical for ensuring the safe operation of autonomous vehicles. However, current methods are constrained by sparse image-LiDAR alignment, insufficient annotations, and ineffective structural discrepancy modeling, causing semantic degradation and generalization deficiency. Therefore, we propose Panoptic-VSNet, a visual-semantic prior knowledge-driven multimodal 3D panoptic segmentation network. Firstly, we design a progressive fusion semantic alignment module that effectively aggregates visual prior features obtained from the large Visual-Language model, establishing a point-semantic region association, thereby enhancing semantic awareness. Secondly, we propose an instance-aware superpixel cross-modal fusion module that incorporates instance prior knowledge, forming a unified representation with spatial precision and class consistency. Finally, we introduce a correlation-aware adaptive panoptic segmentation network that reduces parameter count while dynamically capturing contextual information and enhancing local details, thereby improving panoptic perception capabilities. Experimental evaluations on benchmark datasets show that Panoptic-VSNet outperforms state-of-the-art methods. Code is available at <span><span>https://github.com/lixiao0125/panoptic-vsnet.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113239"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SWFTI: Facial template inversion via StyleSwin mapping 通过StyleSwin映射实现面部模板反转
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-01 DOI: 10.1016/j.patcog.2026.113190
Zixuan Shen, Zhihua Xia, Kaikai Gan, Peipeng Yu, Xiaoyu Zhou
Face recognition systems storing facial templates for identity authentication are widely deployed in real-world applications. However, these systems are inherently vulnerable to facial template inversion attacks. To expose such risks, we propose SWFTI, a novel StyleSwin-based facial template inverter explicitly designed for high-fidelity identity-preserving face image reconstruction. Specifically, we train an optimized mapping network to transform the facial templates into intermediate latent codes, which are then fed into StyleSwin’s synthesis network for face generation. The mapping network is trained with a carefully designed suite of latent-level and image-level losses. Our latent-level losses include discrimination loss, numerical loss, and directional loss. They ensure that the transformed intermediate latent codes conform to the distribution of the original codes in StyleSwin. Furthermore, to mitigate misalignment between the generated latent codes and the underlying subject identity, we introduce image-level losses which include template loss and attribute loss to optimize mapping network training. Without relying on complex training strategies, SWFTI effectively reconstructs face images from templates with strong identity preservation. Experiments on the LFW and AgeDB datasets demonstrate that SWFTI outperforms state-of-the-art methods by a significant margin. It improves the TAR by over 20% under Type-II attacks. Ablation studies further validate the effectiveness of each core component in our proposed framework.
存储用于身份认证的人脸模板的人脸识别系统被广泛应用于实际应用中。然而,这些系统本身就容易受到面部模板反转攻击。为了暴露这些风险,我们提出了SWFTI,一种新颖的基于styleswin5的人脸模板逆变器,专门用于高保真的保持身份的人脸图像重建。具体来说,我们训练了一个优化的映射网络,将面部模板转换为中间潜在代码,然后将其馈送到StyleSwin的合成网络中进行面部生成。映射网络是用一套精心设计的潜在级和图像级损失来训练的。我们的潜在级损失包括辨别损失、数值损失和方向损失。它们确保转换后的中间潜在代码符合StyleSwin中原始代码的分布。此外,为了减轻生成的潜在代码与潜在主题身份之间的不一致,我们引入了包括模板损失和属性损失在内的图像级损失来优化映射网络训练。在不依赖复杂训练策略的情况下,SWFTI可以有效地从具有强身份保留的模板中重建人脸图像。在LFW和AgeDB数据集上的实验表明,SWFTI的性能明显优于最先进的方法。在ii型攻击下,它将TAR提高了20%以上。消融研究进一步验证了我们提出的框架中每个核心组件的有效性。
{"title":"SWFTI: Facial template inversion via StyleSwin mapping","authors":"Zixuan Shen,&nbsp;Zhihua Xia,&nbsp;Kaikai Gan,&nbsp;Peipeng Yu,&nbsp;Xiaoyu Zhou","doi":"10.1016/j.patcog.2026.113190","DOIUrl":"10.1016/j.patcog.2026.113190","url":null,"abstract":"<div><div>Face recognition systems storing facial templates for identity authentication are widely deployed in real-world applications. However, these systems are inherently vulnerable to facial template inversion attacks. To expose such risks, we propose SWFTI, a novel StyleSwin-based facial template inverter explicitly designed for high-fidelity identity-preserving face image reconstruction. Specifically, we train an optimized mapping network to transform the facial templates into intermediate latent codes, which are then fed into StyleSwin’s synthesis network for face generation. The mapping network is trained with a carefully designed suite of latent-level and image-level losses. Our latent-level losses include discrimination loss, numerical loss, and directional loss. They ensure that the transformed intermediate latent codes conform to the distribution of the original codes in StyleSwin. Furthermore, to mitigate misalignment between the generated latent codes and the underlying subject identity, we introduce image-level losses which include template loss and attribute loss to optimize mapping network training. Without relying on complex training strategies, SWFTI effectively reconstructs face images from templates with strong identity preservation. Experiments on the LFW and AgeDB datasets demonstrate that SWFTI outperforms state-of-the-art methods by a significant margin. It improves the TAR by over 20% under Type-II attacks. Ablation studies further validate the effectiveness of each core component in our proposed framework.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113190"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond similarity: Mutual information-guided retrieval for in-context learning in VQA 超越相似性:VQA中上下文学习的相互信息引导检索
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-03 DOI: 10.1016/j.patcog.2026.113214
Jun Zhang , Zezhong Lv , Jian Zhao , Yan Wang , Tianle Zhang , Yuchen Yuan , Yuchu Jiang , Chi Zhang , Wenqi Ren , Xuelong Li
Visual Question Answering (VQA) is a challenging multi-modal task. In-context Learning (ICL) has shown promise in improving the generalization of pre-trained models on VQA by retrieving image-text pairs that are similar to the given query. However, existing approaches overlook two critical issues: i) The effectiveness of the In-context Demonstration (ICD) in prompting a pre-trained model is not strictly correlated with the feature similarity. ii) As a multi-modal task involving both vision and language, VQA requires a joint understanding of visual and textual modalities, which is difficult to achieve when retrieval is based on a single modality. To address these limitations, we propose a novel Mutual Information-Guided Retrieval (MIGR) model. Specifically, we annotate a small subset of data (5% of the dataset) with ICD quality scores based on VQA performance, and train our model to maximize the multi-modal mutual information between each query and its corresponding high-quality ICDs. This enables the model to capture more complex relationships beyond feature-level similarity, leading to improved generalization in ICL. Extensive experiments demonstrate that our mutual information-based retrieval strategy significantly outperforms conventional similarity-based retrieval methods in VQA tasks.
可视化问答(VQA)是一项具有挑战性的多模态任务。上下文学习(ICL)通过检索与给定查询相似的图像-文本对,在提高VQA预训练模型的泛化方面显示出了希望。然而,现有的方法忽略了两个关键问题:i)上下文演示(ICD)在提示预训练模型方面的有效性与特征相似度并不严格相关。ii) VQA是一项涉及视觉和语言的多模态任务,需要对视觉模态和文本模态进行联合理解,当检索基于单一模态时难以实现。为了解决这些限制,我们提出了一种新的互信息引导检索(MIGR)模型。具体来说,我们用基于VQA性能的ICD质量分数注释了一小部分数据(数据集的5%),并训练我们的模型最大化每个查询与其相应的高质量ICD之间的多模态互信息。这使得模型能够捕获超越特征级相似性的更复杂的关系,从而改进ICL中的泛化。大量的实验表明,我们的基于互信息的检索策略在VQA任务中显著优于传统的基于相似性的检索方法。
{"title":"Beyond similarity: Mutual information-guided retrieval for in-context learning in VQA","authors":"Jun Zhang ,&nbsp;Zezhong Lv ,&nbsp;Jian Zhao ,&nbsp;Yan Wang ,&nbsp;Tianle Zhang ,&nbsp;Yuchen Yuan ,&nbsp;Yuchu Jiang ,&nbsp;Chi Zhang ,&nbsp;Wenqi Ren ,&nbsp;Xuelong Li","doi":"10.1016/j.patcog.2026.113214","DOIUrl":"10.1016/j.patcog.2026.113214","url":null,"abstract":"<div><div>Visual Question Answering (VQA) is a challenging multi-modal task. In-context Learning (ICL) has shown promise in improving the generalization of pre-trained models on VQA by retrieving image-text pairs that are similar to the given query. However, existing approaches overlook two critical issues: i) The effectiveness of the In-context Demonstration (ICD) in prompting a pre-trained model is not strictly correlated with the feature similarity. ii) As a multi-modal task involving both vision and language, VQA requires a joint understanding of visual and textual modalities, which is difficult to achieve when retrieval is based on a single modality. To address these limitations, we propose a novel Mutual Information-Guided Retrieval (MIGR) model. Specifically, we annotate a small subset of data (5% of the dataset) with ICD quality scores based on VQA performance, and train our model to maximize the multi-modal mutual information between each query and its corresponding high-quality ICDs. This enables the model to capture more complex relationships beyond feature-level similarity, leading to improved generalization in ICL. Extensive experiments demonstrate that our mutual information-based retrieval strategy significantly outperforms conventional similarity-based retrieval methods in VQA tasks.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113214"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards invisible backdoor attacks on multi-object tracking via suppressed feature learning 基于抑制特征学习的多目标跟踪隐形后门攻击研究
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-07 DOI: 10.1016/j.patcog.2026.113248
Yilang Zhang , Bo Lang
In current practice, training efficient multi-object tracking (MOT) models often requires collecting large-scale third-party datasets. However, directly incorporating these unverified datasets introduces new security threats to MOT. In this paper, we reveal such a threat, where an adversary can implant a hidden backdoor into an MOT tracker by poisoning only a small portion of the dataset. Specifically, we propose a feature-map-suppression-based poison-only backdoor attack, which adopts a sample-specific trigger paradigm and optimizes the trigger based on multi-scale feature maps of video frames. In addition, we introduce an inter-frame motion analysis method for selecting poisoned frames. In our attack, once the tracker is embedded with a backdoor, the object with the trigger will evade tracking. Extensive experiments under various settings demonstrate that our attack significantly degrades the performance of both Tracking-by-Detection and Joint-Detection-and-Tracking MOT trackers. Furthermore, we validate the robustness of our attack against several potential backdoor defense methods. The code will be available at https://github.com/Magic0825/MOT-BA.
在目前的实践中,训练高效的多目标跟踪(MOT)模型通常需要收集大规模的第三方数据集。然而,直接合并这些未经验证的数据集会给MOT带来新的安全威胁。在本文中,我们揭示了这样一种威胁,攻击者可以通过毒害一小部分数据集来将隐藏的后门植入到MOT跟踪器中。具体而言,我们提出了一种基于特征映射抑制的纯毒后门攻击,该攻击采用特定于样本的触发器范式,并基于视频帧的多尺度特征映射优化触发器。此外,我们还介绍了一种帧间运动分析方法来选择有毒帧。在我们的攻击中,一旦追踪器嵌入了后门,带有触发器的物体就会逃避追踪。在各种设置下的大量实验表明,我们的攻击显著降低了检测跟踪和联合检测跟踪MOT跟踪器的性能。此外,我们验证了针对几种潜在后门防御方法的攻击的鲁棒性。代码可在https://github.com/Magic0825/MOT-BA上获得。
{"title":"Towards invisible backdoor attacks on multi-object tracking via suppressed feature learning","authors":"Yilang Zhang ,&nbsp;Bo Lang","doi":"10.1016/j.patcog.2026.113248","DOIUrl":"10.1016/j.patcog.2026.113248","url":null,"abstract":"<div><div>In current practice, training efficient multi-object tracking (MOT) models often requires collecting large-scale third-party datasets. However, directly incorporating these unverified datasets introduces new security threats to MOT. In this paper, we reveal such a threat, where an adversary can implant a hidden backdoor into an MOT tracker by poisoning only a small portion of the dataset. Specifically, we propose a feature-map-suppression-based poison-only backdoor attack, which adopts a sample-specific trigger paradigm and optimizes the trigger based on multi-scale feature maps of video frames. In addition, we introduce an inter-frame motion analysis method for selecting poisoned frames. In our attack, once the tracker is embedded with a backdoor, the object with the trigger will evade tracking. Extensive experiments under various settings demonstrate that our attack significantly degrades the performance of both Tracking-by-Detection and Joint-Detection-and-Tracking MOT trackers. Furthermore, we validate the robustness of our attack against several potential backdoor defense methods. The code will be available at <span><span>https://github.com/Magic0825/MOT-BA</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113248"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VLCounting: Taming zero-shot counting via language-driven exemplar grounding VLCounting:通过语言驱动的范例基础来训练零射击计数
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-05 DOI: 10.1016/j.patcog.2026.113205
Mingjie Wang , Zhuohang Li , Yong Dai , Eric Buys , Minglun Gong
Recently, Class-Agnostic Counting (CAC) problem has garnered increasing attention owing to its broad applicability and superior efficiency compared to Class-Specific Counting (CSC). This paper proposes a novel framework, VLCounting, designed to enhance zero-shot object counting by deeply integrating language-guided exemplar grounding. Specifically, VLCounting consists of an innovative Language-oriented Exemplar Perceptron and a downstream visual Zero-shot Counting pipeline. The perceptron focuses on extracting accurate exemplar cues from collaborative language-vision signals by leveraging rich semantic priors from state-of-the-art pre-trained Large Language Models (LLMs). Meanwhile, the counting pipeline excels in extracting fine-grained features through dual-branch and cross-attention schemes, contributing to the high-quality similarity learning. In addition to bridging the gap between LLMs and visual counting tasks, expression-guided exemplar estimation significantly advances zero-shot learning capabilities for counting instances with arbitrary classes. Moreover, the development of FSC-147-Express, with meticulously annotated linguistic expressions, opens up new avenues for developing and validating language-based counting models. Extensive experiments demonstrate VLCounting’s superior performance, outperforming state-of-the-art general exemplar learning approaches with 35.6% lower MAE and 39.1% lower RMSE on the validation set, respectively. Moreover, it attains accuracy comparable to that of several category-specific counting models, further demonstrating its competitive advantage. The implementation code and trained models are publicly available at https://github.com/ZSTU-CV-Lab/VLCounting.
近年来,类不可知论计数(CAC)由于其广泛的适用性和相对于类特定计数(CSC)的优越性而受到越来越多的关注。本文提出了一个新的框架,VLCounting,旨在通过深度集成语言引导的范例基础来增强零射击目标计数。具体来说,VLCounting由一个创新的面向语言的范例感知器和一个下游的视觉零射击计数管道组成。感知器专注于通过利用来自最先进的预训练大型语言模型(LLMs)的丰富语义先验,从协作语言视觉信号中提取准确的示例线索。同时,计数管道通过双分支和交叉关注方案提取细粒度特征,有助于实现高质量的相似学习。除了弥合llm和视觉计数任务之间的差距之外,表达式引导的范例估计显着提高了使用任意类计数实例的零射击学习能力。此外,带有精心注释的语言表达式的FSC-147-Express的开发为开发和验证基于语言的计数模型开辟了新的途径。大量的实验证明了VLCounting的优越性能,在验证集上的MAE和RMSE分别降低了35.6%和39.1%,优于最先进的一般范例学习方法。此外,它达到了与几种特定类别计数模型相当的准确性,进一步证明了它的竞争优势。实现代码和经过训练的模型可在https://github.com/ZSTU-CV-Lab/VLCounting上公开获得。
{"title":"VLCounting: Taming zero-shot counting via language-driven exemplar grounding","authors":"Mingjie Wang ,&nbsp;Zhuohang Li ,&nbsp;Yong Dai ,&nbsp;Eric Buys ,&nbsp;Minglun Gong","doi":"10.1016/j.patcog.2026.113205","DOIUrl":"10.1016/j.patcog.2026.113205","url":null,"abstract":"<div><div>Recently, Class-Agnostic Counting (CAC) problem has garnered increasing attention owing to its broad applicability and superior efficiency compared to Class-Specific Counting (CSC). This paper proposes a novel framework, VLCounting, designed to enhance zero-shot object counting by deeply integrating language-guided exemplar grounding. Specifically, VLCounting consists of an innovative Language-oriented Exemplar Perceptron and a downstream visual Zero-shot Counting pipeline. The perceptron focuses on extracting accurate exemplar cues from collaborative language-vision signals by leveraging rich semantic priors from state-of-the-art pre-trained Large Language Models (LLMs). Meanwhile, the counting pipeline excels in extracting fine-grained features through dual-branch and cross-attention schemes, contributing to the high-quality similarity learning. In addition to bridging the gap between LLMs and visual counting tasks, expression-guided exemplar estimation significantly advances zero-shot learning capabilities for counting instances with arbitrary classes. Moreover, the development of FSC-147-Express, with meticulously annotated linguistic expressions, opens up new avenues for developing and validating language-based counting models. Extensive experiments demonstrate VLCounting’s superior performance, outperforming state-of-the-art general exemplar learning approaches with 35.6% lower MAE and 39.1% lower RMSE on the validation set, respectively. Moreover, it attains accuracy comparable to that of several category-specific counting models, further demonstrating its competitive advantage. The implementation code and trained models are publicly available at <span><span>https://github.com/ZSTU-CV-Lab/VLCounting</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113205"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Disco: Disentangled identity-action extraction and spatiotemporal context modeling for LLM-based identity-aware basketball video captioning 基于llm的身份感知篮球视频字幕的解纠缠身份-动作提取和时空上下文建模
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-09 DOI: 10.1016/j.patcog.2026.113188
Zeyu Xi, Ya Jing, Haoying Sun, Haoran Zhang, Lifang Wu
Identity-aware sports video captioning (IAVC) is a challenging task that involves recognizing players and describing fine-grained actions. Existing methods focus heavily on player identification, often overlooking the substantial potential of mining action information in enhancing captioning performance. In this paper, we propose a novel identity-aware basketball video captioning network featuring disentangled identity-action extraction and spatiotemporal context modeling (Disco). Specifically, key players are selected based on their interactions with the ball. And a pretrained identity-action disentanglement network (IADN) is used for the joint extraction of player identities and action semantics from key players. To enhance contextual understanding, our adaptive spatiotemporal context modeling (ASCM) module employs learnable query vectors to capture scene-level visual cues. A player-scene interaction (PSI) module is designed to associate players with the scene context. The outputs of the above components are concatenated into a prompt containing identity and action information, guiding the large language model (LLM) to generate accurate captions. Extensive experiments on VC-NBA-2022 and NBA-Identity datasets demonstrate that Disco achieves impressive results, significantly outperforming existing advanced methods. Code is publicly available at https://github.com/Zeyu1226-mt/Disco.
身份感知体育视频字幕(IAVC)是一项具有挑战性的任务,涉及识别球员和描述细粒度的动作。现有的方法主要集中在玩家识别上,往往忽略了挖掘动作信息在提高字幕性能方面的巨大潜力。在本文中,我们提出了一种新的身份感知篮球视频字幕网络,该网络具有解纠缠的身份-动作提取和时空上下文建模(Disco)。具体来说,关键球员是根据他们与球的互动来选择的。并利用预训练的身份-动作解纠缠网络(IADN)对关键参与者的身份和动作语义进行联合提取。为了增强上下文理解,我们的自适应时空上下文建模(ASCM)模块采用可学习的查询向量来捕获场景级视觉线索。玩家-场景交互(PSI)模块旨在将玩家与场景上下文联系起来。上述组件的输出被连接到包含身份和操作信息的提示符中,指导大型语言模型(LLM)生成准确的标题。在VC-NBA-2022和nba identity数据集上的大量实验表明,Disco取得了令人印象深刻的结果,显著优于现有的先进方法。代码可在https://github.com/Zeyu1226-mt/Disco上公开获取。
{"title":"Disco: Disentangled identity-action extraction and spatiotemporal context modeling for LLM-based identity-aware basketball video captioning","authors":"Zeyu Xi,&nbsp;Ya Jing,&nbsp;Haoying Sun,&nbsp;Haoran Zhang,&nbsp;Lifang Wu","doi":"10.1016/j.patcog.2026.113188","DOIUrl":"10.1016/j.patcog.2026.113188","url":null,"abstract":"<div><div>Identity-aware sports video captioning (IAVC) is a challenging task that involves recognizing players and describing fine-grained actions. Existing methods focus heavily on player identification, often overlooking the substantial potential of mining action information in enhancing captioning performance. In this paper, we propose a novel identity-aware basketball video captioning network featuring <strong>d</strong>isentangled <strong>i</strong>dentity-action extraction and <strong>s</strong>patiotemporal <strong>co</strong>ntext modeling (Disco). Specifically, key players are selected based on their interactions with the ball. And a pretrained identity-action disentanglement network (IADN) is used for the joint extraction of player identities and action semantics from key players. To enhance contextual understanding, our adaptive spatiotemporal context modeling (ASCM) module employs learnable query vectors to capture scene-level visual cues. A player-scene interaction (PSI) module is designed to associate players with the scene context. The outputs of the above components are concatenated into a prompt containing identity and action information, guiding the large language model (LLM) to generate accurate captions. Extensive experiments on VC-NBA-2022 and NBA-Identity datasets demonstrate that Disco achieves impressive results, significantly outperforming existing advanced methods. Code is publicly available at <span><span>https://github.com/Zeyu1226-mt/Disco</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113188"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal emotion recognition via unified granularity contrastive learning and similar negative discrimination 基于统一粒度对比学习和相似负性辨别的多模态情绪识别
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-05 DOI: 10.1016/j.patcog.2026.113224
Yongwei Li , Wei Gao , Jianwu Li
Audio-visual emotion recognition plays a crucial role in advancing human-computer interaction by enabling systems to perceive users’ emotional states. While recent advances have primarily focused on audio-visual feature fusion and alignment, existing approaches often overlook two critical challenges: (1) the alignment of audio-visual features across varying levels of granularity, and (2) the effective discrimination of hard negative sample with highly similar feature representations but belonging to different emotional categories. To address these limitations, we propose a novel audio-visual emotion recognition framework. First, we introduce a unified granularity contrastive learning strategy, which employs a shared vector space to harmonize features of different granularities, thereby enabling more consistent cross-modal alignment. Second, to improve class discrimination, particularly in the presence of hard negative samples, we propose a similar negative discrimination module that utilizes an auxiliary classification head to explicitly separate semantically similar but class-distinct samples across modalities. Extensive experiments conducted on two widely used benchmark datasets, CREMA-D and IEMOCAP, demonstrate that our method achieves state-of-the-art performance, validating the effectiveness of the proposed method. Our source code is available at https://github.com/gaoweibit/multi-modal_emotion_recognition.
视听情感识别通过使系统感知用户的情绪状态,在推进人机交互中起着至关重要的作用。虽然最近的进展主要集中在视听特征的融合和对齐上,但现有的方法往往忽视了两个关键的挑战:(1)跨不同粒度级别的视听特征的对齐,以及(2)具有高度相似特征表示但属于不同情感类别的硬负样本的有效区分。为了解决这些限制,我们提出了一个新的视听情感识别框架。首先,我们引入了统一粒度对比学习策略,该策略采用共享向量空间来协调不同粒度的特征,从而实现更一致的跨模态对齐。其次,为了提高类别识别,特别是在存在硬负样本的情况下,我们提出了一个类似的负识别模块,该模块利用辅助分类头在模态中明确分离语义相似但类别不同的样本。在两个广泛使用的基准数据集CREMA-D和IEMOCAP上进行的大量实验表明,我们的方法达到了最先进的性能,验证了所提出方法的有效性。我们的源代码可从https://github.com/gaoweibit/multi-modal_emotion_recognition获得。
{"title":"Multimodal emotion recognition via unified granularity contrastive learning and similar negative discrimination","authors":"Yongwei Li ,&nbsp;Wei Gao ,&nbsp;Jianwu Li","doi":"10.1016/j.patcog.2026.113224","DOIUrl":"10.1016/j.patcog.2026.113224","url":null,"abstract":"<div><div>Audio-visual emotion recognition plays a crucial role in advancing human-computer interaction by enabling systems to perceive users’ emotional states. While recent advances have primarily focused on audio-visual feature fusion and alignment, existing approaches often overlook two critical challenges: (1) the alignment of audio-visual features across varying levels of granularity, and (2) the effective discrimination of hard negative sample with highly similar feature representations but belonging to different emotional categories. To address these limitations, we propose a novel audio-visual emotion recognition framework. First, we introduce a unified granularity contrastive learning strategy, which employs a shared vector space to harmonize features of different granularities, thereby enabling more consistent cross-modal alignment. Second, to improve class discrimination, particularly in the presence of hard negative samples, we propose a similar negative discrimination module that utilizes an auxiliary classification head to explicitly separate semantically similar but class-distinct samples across modalities. Extensive experiments conducted on two widely used benchmark datasets, CREMA-D and IEMOCAP, demonstrate that our method achieves state-of-the-art performance, validating the effectiveness of the proposed method. Our source code is available at <span><span>https://github.com/gaoweibit/multi-modal_emotion_recognition</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113224"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dual-level modality debiasing learning for unsupervised visible-infrared person re-identification 无监督可见红外人再识别的双阶模态去偏学习
IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-08-01 Epub Date: 2026-02-09 DOI: 10.1016/j.patcog.2026.113257
Jiaze Li , Yan Lu , Bin Liu , Guojun Yin , Mang Ye
Two-stage learning pipeline has achieved promising results in unsupervised visible-infrared person re-identification (USL-VI-ReID). It first performs single-modality learning and then operates cross-modality learning to tackle the modality discrepancy. Although promising, this pipeline inevitably introduces modality bias: modality-specific cues learned in the single-modality training naturally propagate into the following cross-modality learning, impairing identity discrimination and generalization. To address this issue, we propose a Dual-level Modality Debiasing Learning (DMDL) framework that implements debiasing at both the model and optimization levels. At the model level, we propose a Causality-inspired Adjustment Intervention (CAI) module that replaces likelihood-based modeling with causal modeling, preventing modality-induced spurious patterns from being introduced, leading to a low-biased model. At the optimization level, a Collaborative Bias-free Training (CBT) strategy is introduced to interrupt the propagation of modality bias across data, labels, and features by integrating modality-specific augmentation, label refinement, and feature alignment. Extensive experiments on benchmark datasets demonstrate that DMDL could enable modality-invariant feature learning and a more generalized model.
两阶段学习管道在无监督可见红外人员再识别(USL-VI-ReID)中取得了可喜的成果。它首先进行单模态学习,然后进行跨模态学习,以解决模态差异。虽然有希望,但这种管道不可避免地引入了模态偏见:在单模态训练中学习到的模态特定线索自然会传播到后续的跨模态学习中,从而损害身份歧视和泛化。为了解决这个问题,我们提出了一个双级模态去偏学习(DMDL)框架,该框架在模型和优化级别实现去偏。在模型层面,我们提出了一个因果关系启发的调整干预(CAI)模块,该模块用因果模型取代基于似然的模型,防止引入模态诱导的虚假模式,从而导致低偏差模型。在优化层面,引入了协作无偏差训练(CBT)策略,通过集成特定于模态的增强、标签细化和特征对齐来中断模态偏差在数据、标签和特征之间的传播。在基准数据集上的大量实验表明,DMDL可以实现模态不变的特征学习和更广义的模型。
{"title":"Dual-level modality debiasing learning for unsupervised visible-infrared person re-identification","authors":"Jiaze Li ,&nbsp;Yan Lu ,&nbsp;Bin Liu ,&nbsp;Guojun Yin ,&nbsp;Mang Ye","doi":"10.1016/j.patcog.2026.113257","DOIUrl":"10.1016/j.patcog.2026.113257","url":null,"abstract":"<div><div>Two-stage learning pipeline has achieved promising results in unsupervised visible-infrared person re-identification (USL-VI-ReID). It first performs single-modality learning and then operates cross-modality learning to tackle the modality discrepancy. Although promising, this pipeline inevitably introduces modality bias: modality-specific cues learned in the single-modality training naturally propagate into the following cross-modality learning, impairing identity discrimination and generalization. To address this issue, we propose a Dual-level Modality Debiasing Learning (DMDL) framework that implements debiasing at both the model and optimization levels. At the model level, we propose a Causality-inspired Adjustment Intervention (CAI) module that replaces likelihood-based modeling with causal modeling, preventing modality-induced spurious patterns from being introduced, leading to a low-biased model. At the optimization level, a Collaborative Bias-free Training (CBT) strategy is introduced to interrupt the propagation of modality bias across data, labels, and features by integrating modality-specific augmentation, label refinement, and feature alignment. Extensive experiments on benchmark datasets demonstrate that DMDL could enable modality-invariant feature learning and a more generalized model.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"176 ","pages":"Article 113257"},"PeriodicalIF":7.6,"publicationDate":"2026-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1