首页 > 最新文献

IEEE transactions on pattern analysis and machine intelligence最新文献

英文 中文
Causal Interventional Prompt Tuning for Few-Shot Out-of-Distribution Generalization 少弹分布外泛化的因果介入提示调谐。
IF 18.6 Pub Date : 2025-10-14 DOI: 10.1109/TPAMI.2025.3621250
Jie Wen;Yicheng Liu;Chao Huang;Chengliang Liu;Yong Xu;Xiaochun Cao
Fine-tuning pre-trained vision-language models (VLMs) has shown substantial benefits in a wide range of downstream tasks, often achieving impressive performance with minimal labeled data. Parameter-efficient fine-tuning techniques, in particular, have demonstrated their effectiveness in enhancing downstream task performance. However, these methods frequently struggle to generalize to out-of-distribution (OOD) data due to their reliance on non-causal representations, which can introduce biases and spurious correlations that negatively impact decision-making. Such spurious factors hinder the model’s generalization ability beyond the training distribution. To address these challenges, in this paper, we propose a novel causal intervention-based prompt tuning method to adapt VLMs to few-shot OOD generalization. Specifically, we leverage the front-door adjustment technique from causal inference to mitigate the effects of spurious correlations and enhance the model’s focus on causal relationships. Built upon VLMs, our approach begins by decoupling causal and non-causal representations in the vision-language alignment process. The causal representation that captures only essential semantically relevant information can serve as a mediator variable between the input image and output label, mitigating the biases from the latent confounder. To further enrich this causal representation, we propose a novel text-based diversity augmentation technique that uses textual features to provide additional semantic context. This augmentation technique can enhance the diversity of the causal representation, making it more robust and generalizable to various OOD scenarios. Experimental results across multiple OOD datasets demonstrate that our method significantly outperforms existing approaches, achieving state-of-the-art generalization performance.
微调预训练的视觉语言模型(VLMs)在广泛的下游任务中显示出了实质性的好处,通常可以用最少的标记数据实现令人印象深刻的性能。参数高效微调技术,特别是,已经证明了它们在提高下游任务性能方面的有效性。然而,这些方法往往难以推广到分布外(OOD)数据,因为它们依赖于非因果表示,这可能会引入对决策产生负面影响的偏差和虚假相关性。这些虚假因素阻碍了模型在训练分布之外的泛化能力。为了解决这些挑战,在本文中,我们提出了一种新的基于因果干预的提示调谐方法,以使vlm适应少弹OOD泛化。具体来说,我们利用因果推理的前门调整技术来减轻虚假相关性的影响,并增强模型对因果关系的关注。基于vlm,我们的方法首先解耦视觉语言对齐过程中的因果和非因果表示。仅捕获基本语义相关信息的因果表示可以作为输入图像和输出标签之间的中介变量,减轻潜在混杂因素的偏差。为了进一步丰富这种因果表示,我们提出了一种新的基于文本的多样性增强技术,该技术使用文本特征提供额外的语义上下文。这种增强技术可以增强因果表示的多样性,使其更健壮,并可推广到各种OOD场景。跨多个OOD数据集的实验结果表明,我们的方法显著优于现有方法,实现了最先进的泛化性能。
{"title":"Causal Interventional Prompt Tuning for Few-Shot Out-of-Distribution Generalization","authors":"Jie Wen;Yicheng Liu;Chao Huang;Chengliang Liu;Yong Xu;Xiaochun Cao","doi":"10.1109/TPAMI.2025.3621250","DOIUrl":"10.1109/TPAMI.2025.3621250","url":null,"abstract":"Fine-tuning pre-trained vision-language models (VLMs) has shown substantial benefits in a wide range of downstream tasks, often achieving impressive performance with minimal labeled data. Parameter-efficient fine-tuning techniques, in particular, have demonstrated their effectiveness in enhancing downstream task performance. However, these methods frequently struggle to generalize to out-of-distribution (OOD) data due to their reliance on non-causal representations, which can introduce biases and spurious correlations that negatively impact decision-making. Such spurious factors hinder the model’s generalization ability beyond the training distribution. To address these challenges, in this paper, we propose a novel causal intervention-based prompt tuning method to adapt VLMs to few-shot OOD generalization. Specifically, we leverage the front-door adjustment technique from causal inference to mitigate the effects of spurious correlations and enhance the model’s focus on causal relationships. Built upon VLMs, our approach begins by decoupling causal and non-causal representations in the vision-language alignment process. The causal representation that captures only essential semantically relevant information can serve as a mediator variable between the input image and output label, mitigating the biases from the latent confounder. To further enrich this causal representation, we propose a novel text-based diversity augmentation technique that uses textual features to provide additional semantic context. This augmentation technique can enhance the diversity of the causal representation, making it more robust and generalizable to various OOD scenarios. Experimental results across multiple OOD datasets demonstrate that our method significantly outperforms existing approaches, achieving state-of-the-art generalization performance.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1978-1991"},"PeriodicalIF":18.6,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145288451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploiting the Benefits of Temporal Information in the Realm of LiDAR Panoptic Segmentation 利用时间信息在激光雷达全景分割领域的优势。
IF 18.6 Pub Date : 2025-10-14 DOI: 10.1109/TPAMI.2025.3621650
Ngoc-Quan Ha-Phan;Myungsik Yoo
LiDAR perception for autonomous driving applications offers highly accurate scene depiction in three-dimensional (3D) spaces, whose most representative task is LiDAR panoptic segmentation (LPS), as it offers exhibition of both instance- and semantic-level segmentation in a holistic manner. Although previous approaches have achieved mature performance, no research has explored temporal information for enhancing LPS performance. As multi-frame processing can assist in better predictions in terms of feature representation and recursive forecasting, which has been proven in other LiDAR perception challenges, this study proposes an effective and temporal-aware panoptic segmentation method for LiDAR point clouds. Specifically, we introduce two modules: convolution-based cross-frame fusion attention (CFFA) and adjacent shifted feature encoder (ASFE) modules. The CFFA module can fuse multi-frame features on the basis of the idea of convolution-based attention, whereas the ASFE module leverages adjacent model outputs and serves as an intermediate guide for final segmentation predictions. Consequent to our extensive experiments, the two modules have been reaffirmed in terms of their productivity in the realm of the LPS. The proposed LPS model achieves impressive panoptic-quality metric scores that are evaluated on different popular benchmarks (63.36% under SemanticKITTI and 78.54% under Panoptic nuScenes), outperforming previous state-of-the-art methods by a significant margin. Further quantitative and qualitative analyses provide evidence of the advantages of multi-frame processing for the LPS together with demonstrations of its particular behavior under different settings.
用于自动驾驶应用的激光雷达感知提供了三维(3D)空间的高精度场景描述,其中最具代表性的任务是激光雷达全景分割(LPS),因为它以整体的方式展示了实例和语义级分割。虽然以前的方法已经获得了成熟的性能,但没有研究探索时间信息来提高LPS性能。由于多帧处理可以在特征表示和递归预测方面帮助更好地预测,这在其他LiDAR感知挑战中得到了证明,因此本研究提出了一种有效的、时间感知的LiDAR点云全光分割方法。具体来说,我们介绍了两个模块:基于卷积的跨帧融合注意(CFFA)和相邻移位特征编码器(ASFE)模块。CFFA模块可以基于基于卷积的注意力思想融合多帧特征,而ASFE模块利用相邻模型输出并作为最终分割预测的中间指南。经过我们广泛的实验,这两个模块在LPS领域的生产力方面得到了重申。提出的LPS模型在不同的流行基准上获得了令人印象深刻的全景质量度量分数(在SemanticKITTI下为63.36%,在Panoptic nuScenes下为78.54%),显著优于之前的最先进方法。进一步的定量和定性分析证明了多帧处理对LPS的优势,并展示了其在不同设置下的特殊行为。
{"title":"Exploiting the Benefits of Temporal Information in the Realm of LiDAR Panoptic Segmentation","authors":"Ngoc-Quan Ha-Phan;Myungsik Yoo","doi":"10.1109/TPAMI.2025.3621650","DOIUrl":"10.1109/TPAMI.2025.3621650","url":null,"abstract":"LiDAR perception for autonomous driving applications offers highly accurate scene depiction in three-dimensional (3D) spaces, whose most representative task is LiDAR panoptic segmentation (LPS), as it offers exhibition of both instance- and semantic-level segmentation in a holistic manner. Although previous approaches have achieved mature performance, no research has explored temporal information for enhancing LPS performance. As multi-frame processing can assist in better predictions in terms of feature representation and recursive forecasting, which has been proven in other LiDAR perception challenges, this study proposes an effective and temporal-aware panoptic segmentation method for LiDAR point clouds. Specifically, we introduce two modules: convolution-based cross-frame fusion attention (CFFA) and adjacent shifted feature encoder (ASFE) modules. The CFFA module can fuse multi-frame features on the basis of the idea of convolution-based attention, whereas the ASFE module leverages adjacent model outputs and serves as an intermediate guide for final segmentation predictions. Consequent to our extensive experiments, the two modules have been reaffirmed in terms of their productivity in the realm of the LPS. The proposed LPS model achieves impressive panoptic-quality metric scores that are evaluated on different popular benchmarks (63.36% under SemanticKITTI and 78.54% under Panoptic nuScenes), outperforming previous state-of-the-art methods by a significant margin. Further quantitative and qualitative analyses provide evidence of the advantages of multi-frame processing for the LPS together with demonstrations of its particular behavior under different settings.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"2048-2065"},"PeriodicalIF":18.6,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145288448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hier-EgoPack: Hierarchical Egocentric Video Understanding With Diverse Task Perspectives Hier-EgoPack:基于不同任务视角的分层自我中心视频理解
IF 18.6 Pub Date : 2025-10-14 DOI: 10.1109/TPAMI.2025.3621326
Simone Alberto Peirone;Francesca Pistilli;Antonio Alliegro;Tatiana Tommasi;Giuseppe Averta
Our comprehension of video streams depicting human activities is naturally multifaceted: in just a few moments, we can grasp what is happening, identify the relevance and interactions of objects in the scene, and forecast what will happen soon, everything all at once. To endow autonomous systems with such a holistic perception, learning how to correlate concepts, abstract knowledge across diverse tasks, and leverage tasks synergies when learning novel skills is essential. A significant step in this direction is EgoPack, a unified framework for understanding human activities across diverse tasks with minimal overhead. EgoPack promotes information sharing and collaboration among downstream tasks, essential for efficiently learning new skills. In this paper, we introduce Hier-EgoPack, which advances EgoPack by enabling reasoning also across diverse temporal granularities, which expands its applicability to a broader range of downstream tasks. To achieve this, we propose a novel hierarchical architecture for temporal reasoning equipped with a GNN layer specifically designed to tackle the challenges of multi-granularity reasoning effectively. We evaluate our approach on multiple Ego4D benchmarks involving both clip-level and frame-level reasoning, demonstrating how our hierarchical unified architecture effectively solves these diverse tasks simultaneously.
我们对描绘人类活动的视频流的理解自然是多方面的:在短短几分钟内,我们就能掌握正在发生的事情,识别场景中物体的相关性和相互作用,并预测即将发生的事情,所有这些都是一次性的。要赋予自主系统这样的整体感知能力,学习如何关联概念、跨不同任务的抽象知识,以及在学习新技能时利用任务协同作用至关重要。EgoPack是朝这个方向迈出的重要一步,它是一个统一的框架,用于以最小的开销理解跨不同任务的人类活动。EgoPack促进下游任务之间的信息共享和协作,这对于有效学习新技能至关重要。在本文中,我们介绍了her -EgoPack,它通过支持跨不同时间粒度的推理来改进EgoPack,从而扩展了其对更广泛的下游任务的适用性。为了实现这一目标,我们提出了一种新的时间推理层次结构,配备了一个专门设计用于有效解决多粒度推理挑战的GNN层。我们对涉及剪辑级和框架级推理的多个Ego4D基准评估我们的方法,展示了我们的分层统一架构如何同时有效地解决这些不同的任务。
{"title":"Hier-EgoPack: Hierarchical Egocentric Video Understanding With Diverse Task Perspectives","authors":"Simone Alberto Peirone;Francesca Pistilli;Antonio Alliegro;Tatiana Tommasi;Giuseppe Averta","doi":"10.1109/TPAMI.2025.3621326","DOIUrl":"10.1109/TPAMI.2025.3621326","url":null,"abstract":"Our comprehension of video streams depicting human activities is naturally multifaceted: in just a few moments, we can grasp what is happening, identify the relevance and interactions of objects in the scene, and forecast what will happen soon, everything all at once. To endow autonomous systems with such a holistic perception, learning how to correlate concepts, abstract knowledge across diverse tasks, and leverage tasks synergies when learning novel skills is essential. A significant step in this direction is EgoPack, a unified framework for understanding human activities across diverse tasks with minimal overhead. EgoPack promotes information sharing and collaboration among downstream tasks, essential for efficiently learning new skills. In this paper, we introduce Hier-EgoPack, which advances EgoPack by enabling reasoning also across diverse temporal granularities, which expands its applicability to a broader range of downstream tasks. To achieve this, we propose a novel hierarchical architecture for temporal reasoning equipped with a GNN layer specifically designed to tackle the challenges of multi-granularity reasoning effectively. We evaluate our approach on multiple Ego4D benchmarks involving both clip-level and frame-level reasoning, demonstrating how our hierarchical unified architecture effectively solves these diverse tasks simultaneously.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1917-1931"},"PeriodicalIF":18.6,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11202655","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145289306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Unified Masked Jigsaw Puzzle Framework for Vision and Language Models 视觉和语言模型的统一蒙面拼图框架。
IF 18.6 Pub Date : 2025-10-14 DOI: 10.1109/TPAMI.2025.3621246
Weixin Ye;Wei Wang;Yahui Liu;Yue Song;Bin Ren;Wei Bi;Rita Cucchiara;Nicu Sebe
In federated learning, Transformer, as a popular architecture, faces critical challenges in defending against gradient attacks and improving model performance in both Computer Vision (CV) and Natural Language Processing (NLP) tasks. It has been revealed that the gradient of Position Embeddings (PEs) in Transformer contains sufficient information, which can be used to reconstruct the input data. To mitigate this issue, we introduce a Masked Jigsaw Puzzle (MJP) framework. MJP starts with random token shuffling to break the token order, and then a learnable unknown (unk) position embedding is used to mask out the PEs of the shuffled tokens. In this manner, the local spatial information which is encoded in the position embeddings is disrupted, and the models are forced to learn feature representations that are less reliant on the local spatial information. Notably, with the careful use of MJP, we can not only improve models’ robustness against gradient attacks, but also boost their performance in both vision and text application scenarios, such as classification for images (e.g., ImageNet-1 K) and sentiment analysis for text (e.g., Yelp and Amazon). Experimental results suggest that MJP is a unified framework for different Transformer-based models in both vision and language tasks.
在联邦学习中,Transformer作为一种流行的架构,在防御梯度攻击和提高计算机视觉(CV)和自然语言处理(NLP)任务中的模型性能方面面临着严峻的挑战。结果表明,变压器中位置嵌入(PEs)的梯度包含了足够的信息,可以用来重建输入数据。为了缓解这个问题,我们引入了一个蒙面拼图(MJP)框架。MJP从随机标记洗牌开始打破标记顺序,然后使用可学习的未知(unk)位置嵌入来掩盖洗牌标记的pe。通过这种方式,位置嵌入中编码的局部空间信息被破坏,模型被迫学习对局部空间信息依赖程度较低的特征表示。值得注意的是,通过仔细使用MJP,我们不仅可以提高模型对梯度攻击的鲁棒性,还可以提高它们在视觉和文本应用场景中的性能,例如图像分类(例如ImageNet-1K)和文本情感分析(例如Yelp和Amazon)。实验结果表明,MJP是一种统一的框架,适用于视觉和语言任务中基于transformer的不同模型。代码可通过https://github.com/ywxsuperstar/transformerattack公开获取。
{"title":"A Unified Masked Jigsaw Puzzle Framework for Vision and Language Models","authors":"Weixin Ye;Wei Wang;Yahui Liu;Yue Song;Bin Ren;Wei Bi;Rita Cucchiara;Nicu Sebe","doi":"10.1109/TPAMI.2025.3621246","DOIUrl":"10.1109/TPAMI.2025.3621246","url":null,"abstract":"In federated learning, Transformer, as a popular architecture, faces critical challenges in defending against gradient attacks and improving model performance in both Computer Vision (CV) and Natural Language Processing (NLP) tasks. It has been revealed that the gradient of Position Embeddings (PEs) in Transformer contains sufficient information, which can be used to reconstruct the input data. To mitigate this issue, we introduce a Masked Jigsaw Puzzle (MJP) framework. MJP starts with random token shuffling to break the token order, and then a learnable <italic>unknown (unk)</i> position embedding is used to mask out the PEs of the shuffled tokens. In this manner, the local spatial information which is encoded in the position embeddings is disrupted, and the models are forced to learn feature representations that are less reliant on the local spatial information. Notably, with the careful use of MJP, we can not only improve models’ robustness against gradient attacks, but also boost their performance in both vision and text application scenarios, such as classification for images (e.g., ImageNet-1 K) and sentiment analysis for text (e.g., Yelp and Amazon). Experimental results suggest that MJP is a unified framework for different Transformer-based models in both vision and language tasks.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1873-1887"},"PeriodicalIF":18.6,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145288485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Discourse-Aware Language Representation 话语感知语言表征
IF 18.6 Pub Date : 2025-10-14 DOI: 10.1109/TPAMI.2025.3621229
Zhuosheng Zhang;Siru Ouyang;Hai Zhao
Recent Transformer-based language representation techniques have commonly adopted a straightforward approach to modeling textual context as a linear sequence of successive tokens. However, this sequential modeling strategy falls short in actively exploring intermediate structures present in natural languages and does not account for the rich interactive relationships between sentences. To overcome these limitations, we propose a discourse-aware framework that bridges the gap between sequential contextualization and the interactive nature of conversational reading comprehension. Concretely, we first divide the context into elementary discourse units (EDUs), ensuring that each unit contains precisely one condition. Then, we systematically explore three instantiations for modeling discourse features: sequential EDU encoding, discourse-aware masking, and discourse graph network. These techniques allow us to capture the nuanced interactions within the discourse. To assess the efficacy of our methodologies, we perform experiments on three conversational reading comprehension tasks: multi-turn response selection, conversational question answering, and conversational machine reading. Experimental results demonstrate the superiority of our proposed approach. Moreover, analysis reveals that the discourse-aware approach enables the model to effectively capture intricate relationships within the context and fosters reasoning interpretability. Additionally, our method exhibits efficacy across various backbone PLMs and diverse domains.
最近基于transformer的语言表示技术通常采用一种直接的方法来将文本上下文建模为连续标记的线性序列。然而,这种顺序建模策略在积极探索自然语言中存在的中间结构方面存在不足,并且没有考虑到句子之间丰富的交互关系。为了克服这些限制,我们提出了一个话语感知框架,该框架在顺序语境化和会话阅读理解的互动性之间架起了桥梁。具体来说,我们首先将上下文划分为基本话语单元(edu),确保每个单元只包含一个条件。然后,我们系统地探索了三种话语特征建模的实例:顺序EDU编码、话语感知掩蔽和话语图网络。这些技术使我们能够捕捉到话语中微妙的相互作用。为了评估我们的方法的有效性,我们在三个会话阅读理解任务上进行了实验:多回合反应选择、会话问答和会话机器阅读。实验结果证明了该方法的优越性。此外,分析表明,话语感知方法使模型能够有效地捕捉上下文中的复杂关系,并促进推理的可解释性。此外,我们的方法在不同的主干plm和不同的领域显示出有效性。
{"title":"Discourse-Aware Language Representation","authors":"Zhuosheng Zhang;Siru Ouyang;Hai Zhao","doi":"10.1109/TPAMI.2025.3621229","DOIUrl":"10.1109/TPAMI.2025.3621229","url":null,"abstract":"Recent Transformer-based language representation techniques have commonly adopted a straightforward approach to modeling textual context as a linear sequence of successive tokens. However, this sequential modeling strategy falls short in actively exploring intermediate structures present in natural languages and does not account for the rich interactive relationships between sentences. To overcome these limitations, we propose a discourse-aware framework that bridges the gap between sequential contextualization and the interactive nature of conversational reading comprehension. Concretely, we first divide the context into elementary discourse units (EDUs), ensuring that each unit contains precisely one condition. Then, we systematically explore three instantiations for modeling discourse features: sequential EDU encoding, discourse-aware masking, and discourse graph network. These techniques allow us to capture the nuanced interactions within the discourse. To assess the efficacy of our methodologies, we perform experiments on three conversational reading comprehension tasks: multi-turn response selection, conversational question answering, and conversational machine reading. Experimental results demonstrate the superiority of our proposed approach. Moreover, analysis reveals that the discourse-aware approach enables the model to effectively capture intricate relationships within the context and fosters reasoning interpretability. Additionally, our method exhibits efficacy across various backbone PLMs and diverse domains.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1888-1903"},"PeriodicalIF":18.6,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145289473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IRNet: Iterative Refinement Network for Noisy Partial Label Learning IRNet:用于有噪声部分标签学习的迭代改进网络
IF 18.6 Pub Date : 2025-10-13 DOI: 10.1109/TPAMI.2025.3620388
Zheng Lian;Mingyu Xu;Lan Chen;Licai Sun;Bin Liu;Lei Feng;Jianhua Tao
Partial label learning (PLL) is a typical weakly supervised learning, where each sample is associated with a set of candidate labels. Its basic assumption is that the ground-truth label must be in the candidate set, but this assumption may not be satisfied due to the unprofessional judgment of annotators. Therefore, we relax this assumption and focus on a more general task, noisy PLL, where the ground-truth label may not exist in the candidate set. To address this challenging task, we propose a novel framework called “Iterative Refinement Network (IRNet)”, aiming to purify noisy samples through two key modules (i.e., noisy sample detection and label correction). To achieve better performance, we exploit smoothness constraints to reduce prediction errors in these modules. Through theoretical analysis, we prove that IRNet is able to reduce the noise level of the dataset and eventually approximate the Bayes optimal classifier. Meanwhile, IRNet is a plug-in strategy that can be integrated with existing PLL approaches. Experimental results on multiple benchmark datasets show that IRNet outperforms state-of-the-art approaches on noisy PLL.
部分标签学习(PLL)是一种典型的弱监督学习,其中每个样本与一组候选标签相关联。它的基本假设是基本真值标签必须在候选集中,但由于注释者的不专业判断,这个假设可能无法满足。因此,我们放宽这一假设,并将重点放在更一般的任务上,即噪声锁相环,其中候选集中可能不存在基真值标签。为了解决这一具有挑战性的任务,我们提出了一个名为“迭代细化网络(IRNet)”的新框架,旨在通过两个关键模块(即噪声样本检测和标签校正)来净化噪声样本。为了获得更好的性能,我们利用平滑约束来减少这些模块中的预测误差。通过理论分析,我们证明了IRNet能够降低数据集的噪声水平,最终逼近贝叶斯最优分类器。同时,IRNet是一种可以与现有PLL方法集成的插件策略。在多个基准数据集上的实验结果表明,IRNet在噪声锁相环上优于最先进的方法。
{"title":"IRNet: Iterative Refinement Network for Noisy Partial Label Learning","authors":"Zheng Lian;Mingyu Xu;Lan Chen;Licai Sun;Bin Liu;Lei Feng;Jianhua Tao","doi":"10.1109/TPAMI.2025.3620388","DOIUrl":"10.1109/TPAMI.2025.3620388","url":null,"abstract":"Partial label learning (PLL) is a typical weakly supervised learning, where each sample is associated with a set of candidate labels. Its basic assumption is that the ground-truth label must be in the candidate set, but this assumption may not be satisfied due to the unprofessional judgment of annotators. Therefore, we relax this assumption and focus on a more general task, noisy PLL, where the ground-truth label may not exist in the candidate set. To address this challenging task, we propose a novel framework called “Iterative Refinement Network (IRNet)”, aiming to purify noisy samples through two key modules (i.e., noisy sample detection and label correction). To achieve better performance, we exploit smoothness constraints to reduce prediction errors in these modules. Through theoretical analysis, we prove that IRNet is able to reduce the noise level of the dataset and eventually approximate the Bayes optimal classifier. Meanwhile, IRNet is a plug-in strategy that can be integrated with existing PLL approaches. Experimental results on multiple benchmark datasets show that IRNet outperforms state-of-the-art approaches on noisy PLL.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1932-1948"},"PeriodicalIF":18.6,"publicationDate":"2025-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145282990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multilingual Text-to-Image Person Retrieval via Bidirectional Relation Reasoning and Aligning 基于双向关系推理和对齐的多语言文本-图像人物检索
IF 18.6 Pub Date : 2025-10-10 DOI: 10.1109/TPAMI.2025.3620139
Min Cao;Xinyu Zhou;Ding Jiang;Bo Du;Mang Ye;Min Zhang
Text-to-image person retrieval (TIPR) aims to identify the target person using textual descriptions, facing challenge in modality heterogeneity. Prior works have attempted to address it by developing cross-modal global or local alignment strategies. However, global methods typically overlook fine-grained cross-modal differences, whereas local methods require prior information to explore explicit part alignments. Additionally, current methods are English-centric, restricting their application in multilingual contexts. To alleviate these issues, we pioneer a multilingual TIPR task by developing a multilingual TIPR benchmark, for which we leverage large language models for initial translations and refine them by integrating domain-specific knowledge. Correspondingly, we propose Bi-IRRA: a Bidirectional Implicit Relation Reasoning and Aligning framework to learn alignment across languages and modalities. Within Bi-IRRA, a bidirectional implicit relation reasoning module enables bidirectional prediction of masked image and text, implicitly enhancing the modeling of local relations across languages and modalities, a multi-dimensional global alignment module is integrated to bridge the modality heterogeneity. The proposed method achieves new state-of-the-art results on all multilingual TIPR datasets.
文本到图像人物检索(TIPR)旨在利用文本描述识别目标人物,但面临着模态异质性的挑战。先前的工作试图通过开发跨模式的全球或本地对齐策略来解决这个问题。然而,全局方法通常忽略了细粒度的跨模态差异,而局部方法需要先验信息来探索显式部件对齐。此外,目前的方法以英语为中心,限制了它们在多语言环境中的应用。为了缓解这些问题,我们通过开发一个多语言TIPR基准,开创了一个多语言TIPR任务,为此,我们利用大型语言模型进行初始翻译,并通过集成特定领域的知识来改进它们。相应地,我们提出了双向隐式关系推理和对齐框架,以学习跨语言和模态的对齐。在Bi-IRRA中,双向隐式关系推理模块可以对被屏蔽的图像和文本进行双向预测,隐式地增强了跨语言和模态的局部关系建模,集成了多维全局对齐模块以弥合模态异质性。所提出的方法在所有多语言TIPR数据集上取得了最新的结果。
{"title":"Multilingual Text-to-Image Person Retrieval via Bidirectional Relation Reasoning and Aligning","authors":"Min Cao;Xinyu Zhou;Ding Jiang;Bo Du;Mang Ye;Min Zhang","doi":"10.1109/TPAMI.2025.3620139","DOIUrl":"10.1109/TPAMI.2025.3620139","url":null,"abstract":"Text-to-image person retrieval (TIPR) aims to identify the target person using textual descriptions, facing challenge in modality heterogeneity. Prior works have attempted to address it by developing cross-modal global or local alignment strategies. However, global methods typically overlook fine-grained cross-modal differences, whereas local methods require prior information to explore explicit part alignments. Additionally, current methods are English-centric, restricting their application in multilingual contexts. To alleviate these issues, we pioneer a multilingual TIPR task by developing a multilingual TIPR benchmark, for which we leverage large language models for initial translations and refine them by integrating domain-specific knowledge. Correspondingly, we propose Bi-IRRA: a Bidirectional Implicit Relation Reasoning and Aligning framework to learn alignment across languages and modalities. Within Bi-IRRA, a bidirectional implicit relation reasoning module enables bidirectional prediction of masked image and text, implicitly enhancing the modeling of local relations across languages and modalities, a multi-dimensional global alignment module is integrated to bridge the modality heterogeneity. The proposed method achieves new state-of-the-art results on all multilingual TIPR datasets.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1961-1977"},"PeriodicalIF":18.6,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145260880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Probabilistically Aligned View-Unaligned Clustering With Adaptive Template Selection 具有自适应模板选择的概率对齐视图不对齐聚类
IF 18.6 Pub Date : 2025-10-07 DOI: 10.1109/TPAMI.2025.3618984
Wenhua Dong;Xiao-Jun Wu;Zhenhua Feng;Sara Atito;Muhammad Awais;Josef Kittler
In most existing multi-view modeling scenarios, cross-view correspondence (CVC) between instances of the same target from different views, like paired image-text data, is a crucial prerequisite for effortlessly deriving a consistent representation. Nevertheless, this premise is frequently compromised in certain applications, where each view is organized and transmitted independently, resulting in the view-unaligned problem (VuP). Restoring CVC of unaligned multi-view data is a challenging and highly demanding task that has received limited attention from the research community. To tackle this practical challenge, we propose to integrate the permutation derivation procedure into the bipartite graph paradigm for view-unaligned clustering, termed Probabilistically Aligned View-unaligned Clustering with Adaptive Template Selection (PAVuC-ATS). Specifically, we learn consistent anchors and view-specific graphs by the bipartite graph, and derive permutations applied to the unaligned graphs by reformulating the alignment between two latent representations as a 2-step transition of a Markov chain with adaptive template selection, thereby achieving the probabilistic alignment. The convergence of the resultant optimization problem is validated both experimentally and theoretically. Extensive experiments on six benchmark datasets demonstrate the superiority of the proposed PAVuC-ATS over the baseline methods.
在大多数现有的多视图建模场景中,来自不同视图的相同目标实例之间的跨视图通信(CVC),如成对的图像-文本数据,是毫不费力地获得一致表示的关键先决条件。然而,在某些应用程序中,这个前提经常被破坏,其中每个视图都是独立组织和传输的,从而导致视图未对齐问题(VuP)。恢复未对齐多视图数据的CVC是一项具有挑战性和高要求的任务,但受到研究界的关注有限。为了解决这一实际挑战,我们提出将排列派生过程集成到视图不对齐聚类的二部图范式中,称为概率对齐视图不对齐聚类与自适应模板选择(PAVuC-ATS)。具体而言,我们通过二部图学习一致锚点和特定视图图,并通过自适应模板选择将两个潜在表示之间的对齐重新表述为马尔可夫链的两步转移,从而获得应用于未对齐图的排列,从而实现概率对齐。实验和理论验证了所得优化问题的收敛性。在六个基准数据集上的大量实验证明了所提出的PAVuC-ATS方法优于基线方法。
{"title":"Probabilistically Aligned View-Unaligned Clustering With Adaptive Template Selection","authors":"Wenhua Dong;Xiao-Jun Wu;Zhenhua Feng;Sara Atito;Muhammad Awais;Josef Kittler","doi":"10.1109/TPAMI.2025.3618984","DOIUrl":"10.1109/TPAMI.2025.3618984","url":null,"abstract":"In most existing multi-view modeling scenarios, cross-view correspondence (CVC) between instances of the same target from different views, like paired image-text data, is a crucial prerequisite for effortlessly deriving a consistent representation. Nevertheless, this premise is frequently compromised in certain applications, where each view is organized and transmitted independently, resulting in the view-unaligned problem (VuP). Restoring CVC of unaligned multi-view data is a challenging and highly demanding task that has received limited attention from the research community. To tackle this practical challenge, we propose to integrate the permutation derivation procedure into the bipartite graph paradigm for view-unaligned clustering, termed Probabilistically Aligned View-unaligned Clustering with Adaptive Template Selection (PAVuC-ATS). Specifically, we learn consistent anchors and view-specific graphs by the bipartite graph, and derive permutations applied to the unaligned graphs by reformulating the alignment between two latent representations as a 2-step transition of a Markov chain with adaptive template selection, thereby achieving the probabilistic alignment. The convergence of the resultant optimization problem is validated both experimentally and theoretically. Extensive experiments on six benchmark datasets demonstrate the superiority of the proposed PAVuC-ATS over the baseline methods.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1904-1916"},"PeriodicalIF":18.6,"publicationDate":"2025-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145241237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Neural Network Parameter Selection via Dataset Similarity Under Meta-Learning Framework 元学习框架下基于数据集相似度的深度神经网络参数选择。
IF 18.6 Pub Date : 2025-10-07 DOI: 10.1109/TPAMI.2025.3618991
Liping Deng;Maziar Raissi;MingQing Xiao
Optimizing the performance of deep neural networks (DNNs) remains a significant challenge due to the sensitivity of models to both hyperparameter selection and weight initialization. Existing approaches typically address these two factors independently, which often leads to limiting adaptability and overall effectiveness. In this paper, we present a novel meta-learning framework that jointly recommends hyperparameters and initial weights by leveraging dataset similarity. Our method begins by extracting meta-features from a collection of historical datasets. For a given query dataset, similarity is computed based on distances in the meta-feature space, and the most similar historical datasets are used to recommend the underlying parameter configurations. To capture the diverse characteristics of image datasets, we introduce two complementary types of meta-features. The first, referred to as shallow or visible meta-features, comprises five groups of statistical measures that summarize color and texture information. The second, termed deep or invisible meta-features, consists of 512 descriptors extracted from a convolutional neural network pre-trained on ImageNet. We evaluated our framework in 105 real-world image classification tasks, using 75 datasets for historical modeling and 30 for querying. Experimental results with both vision transformers and convolutional neural networks demonstrate that our approach consistently outperforms state-of-the-art baselines, underscoring the effectiveness of dataset-driven parameter recommendation in deep learning.
由于模型对超参数选择和权值初始化的敏感性,优化深度神经网络(dnn)的性能仍然是一个重大挑战。现有的方法通常独立地处理这两个因素,这常常导致限制适应性和总体有效性。在本文中,我们提出了一个新的元学习框架,通过利用数据集相似度来联合推荐超参数和初始权重。我们的方法首先从历史数据集集合中提取元特征。对于给定的查询数据集,基于元特征空间中的距离计算相似性,并使用最相似的历史数据集来推荐底层参数配置。为了捕捉图像数据集的不同特征,我们引入了两种互补类型的元特征。第一种被称为浅层或可见元特征,包括五组汇总颜色和纹理信息的统计度量。第二种被称为深度或不可见元特征,由512个描述符组成,这些描述符是从ImageNet上预训练的卷积神经网络中提取的。我们在105个真实世界的图像分类任务中评估了我们的框架,使用75个数据集进行历史建模,30个数据集进行查询。视觉变压器和卷积神经网络的实验结果表明,我们的方法始终优于最先进的基线,强调了数据集驱动的参数推荐在深度学习中的有效性。
{"title":"Deep Neural Network Parameter Selection via Dataset Similarity Under Meta-Learning Framework","authors":"Liping Deng;Maziar Raissi;MingQing Xiao","doi":"10.1109/TPAMI.2025.3618991","DOIUrl":"10.1109/TPAMI.2025.3618991","url":null,"abstract":"Optimizing the performance of deep neural networks (DNNs) remains a significant challenge due to the sensitivity of models to both hyperparameter selection and weight initialization. Existing approaches typically address these two factors independently, which often leads to limiting adaptability and overall effectiveness. In this paper, we present a novel meta-learning framework that jointly recommends hyperparameters and initial weights by leveraging dataset similarity. Our method begins by extracting meta-features from a collection of historical datasets. For a given query dataset, similarity is computed based on distances in the meta-feature space, and the most similar historical datasets are used to recommend the underlying parameter configurations. To capture the diverse characteristics of image datasets, we introduce two complementary types of meta-features. The first, referred to as shallow or visible meta-features, comprises five groups of statistical measures that summarize color and texture information. The second, termed deep or invisible meta-features, consists of 512 descriptors extracted from a convolutional neural network pre-trained on ImageNet. We evaluated our framework in 105 real-world image classification tasks, using 75 datasets for historical modeling and 30 for querying. Experimental results with both vision transformers and convolutional neural networks demonstrate that our approach consistently outperforms state-of-the-art baselines, underscoring the effectiveness of dataset-driven parameter recommendation in deep learning.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"1860-1872"},"PeriodicalIF":18.6,"publicationDate":"2025-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145240879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Dynamic Graph Embeddings With Neural Controlled Differential Equations 用神经控制微分方程学习动态图嵌入
IF 18.6 Pub Date : 2025-10-03 DOI: 10.1109/TPAMI.2025.3617660
Tiexin Qin;Benjamin Walker;Terry Lyons;Hong Yan;Haoliang Li
This paper focuses on representation learning for dynamic graphs with temporal interactions. A fundamental issue is that both the graph structure and the nodes own their own dynamics, and their blending induces intractable complexity in the temporal evolution over graphs. Drawing inspiration from the recent progress of physical dynamic models in deep neural networks, we propose Graph Neural Controlled Differential Equations (GN-CDEs), a continuous-time framework that jointly models node embeddings and structural dynamics by incorporating a graph enhanced neural network vector field with a time-varying graph path as the control signal. Our framework exhibits several desirable characteristics, including the ability to express dynamics on evolving graphs without piecewise integration, the capability to calibrate trajectories with subsequent data, and robustness to missing observations. Empirical evaluation on a range of dynamic graph representation learning tasks demonstrates the effectiveness of our proposed approach in capturing the complex dynamics of dynamic graphs.
本文主要研究具有时间交互的动态图的表示学习。一个基本的问题是,图结构和节点都有自己的动态,它们的混合在图的时间演化中引起了难以处理的复杂性。从深度神经网络物理动力学模型的最新进展中获得灵感,我们提出了图神经控制微分方程(GN-CDEs),这是一个连续时间框架,通过将具有时变图路径的图增强神经网络向量场作为控制信号,联合建模节点嵌入和结构动力学。我们的框架展示了几个令人满意的特性,包括在演化图上表达动态而不需要分段集成的能力,使用后续数据校准轨迹的能力,以及对缺失观测的鲁棒性。对一系列动态图表示学习任务的经验评估表明,我们提出的方法在捕获动态图的复杂动态方面是有效的。
{"title":"Learning Dynamic Graph Embeddings With Neural Controlled Differential Equations","authors":"Tiexin Qin;Benjamin Walker;Terry Lyons;Hong Yan;Haoliang Li","doi":"10.1109/TPAMI.2025.3617660","DOIUrl":"10.1109/TPAMI.2025.3617660","url":null,"abstract":"This paper focuses on representation learning for dynamic graphs with temporal interactions. A fundamental issue is that both the graph structure and the nodes own their own dynamics, and their blending induces intractable complexity in the temporal evolution over graphs. Drawing inspiration from the recent progress of physical dynamic models in deep neural networks, we propose <italic>Graph Neural Controlled Differential Equations</i> (GN-CDEs), a continuous-time framework that jointly models node embeddings and structural dynamics by incorporating a graph enhanced neural network vector field with a time-varying graph path as the control signal. Our framework exhibits several desirable characteristics, including the ability to express dynamics on evolving graphs without piecewise integration, the capability to calibrate trajectories with subsequent data, and robustness to missing observations. Empirical evaluation on a range of dynamic graph representation learning tasks demonstrates the effectiveness of our proposed approach in capturing the complex dynamics of dynamic graphs.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"48 2","pages":"2096-2103"},"PeriodicalIF":18.6,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145215644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on pattern analysis and machine intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1