首页 > 最新文献

IEEE Transactions on Multimedia最新文献

英文 中文
Deep Mutual Distillation for Unsupervised Domain Adaptation Person Re-identification 用于无监督领域适应性人员再识别的深度相互提炼技术
IF 7.3 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-12 DOI: 10.1109/tmm.2024.3459637
Xingyu Gao, Zhenyu Chen, Jianze Wei, Rubo Wang, Zhijun Zhao
{"title":"Deep Mutual Distillation for Unsupervised Domain Adaptation Person Re-identification","authors":"Xingyu Gao, Zhenyu Chen, Jianze Wei, Rubo Wang, Zhijun Zhao","doi":"10.1109/tmm.2024.3459637","DOIUrl":"https://doi.org/10.1109/tmm.2024.3459637","url":null,"abstract":"","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"1 1","pages":""},"PeriodicalIF":7.3,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaborative License Plate Recognition via Association Enhancement Network With Auxiliary Learning and a Unified Benchmark 借助辅助学习和统一基准,通过关联增强网络实现协作式车牌识别
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-10 DOI: 10.1109/TMM.2024.3452982
Yifei Deng;Guohao Wang;Chenglong Li;Wei Wang;Cheng Zhang;Jin Tang
Since the standard license plate of large vehicle is easily affected by occlusion and stain, the traffic management department introduces the enlarged license plate at the rear of the large vehicle to assist license plate recognition. However, current researches regards standard license plate recognition and enlarged license plate recognition as independent tasks, and do not take advantage of the complementary benefits from the two types of license plates. In this work, we propose a new computer vision task called collaborative license plate recognition, aiming to leverage the complementary advantages of standard and enlarged license plates for achieving more accurate license plate recognition. To achieve this goal, we propose an Association Enhancement Network (AENet), which achieves robust collaborative licence plate recognition by capturing the correlations between characters within a single licence plate and enhancing the associations between two license plates. In particular, we design an association enhancement branch, which supervises the fusion of two licence plate information using the complete licence plate number to mine the association between them. To enhance the representation ability of each type of licence plates, we design an auxiliary learning branch in the training stage, which supervises the learning of individual license plates in the association enhancement between two license plates. In addition, we contribute a comprehensive benchmark dataset called CLPR, which consists of a total of 19,782 standard and enlarged licence plates from 24 provinces in China and covers most of the challenges in real scenarios, for collaborative license plate recognition. Extensive experiments on the proposed CLPR dataset demonstrate the effectiveness of the proposed AENet against several state-of-the-art methods.
由于大型车辆的标准车牌容易受到遮挡和污渍的影响,交通管理部门在大型车辆尾部引入了放大车牌来辅助车牌识别。然而,目前的研究将标准车牌识别和放大车牌识别视为独立的任务,没有利用两种车牌的互补优势。在这项工作中,我们提出了一种新的计算机视觉任务--协同车牌识别,旨在利用标准车牌和放大车牌的互补优势,实现更准确的车牌识别。为实现这一目标,我们提出了关联增强网络(AENet),通过捕捉单个车牌内字符之间的关联,增强两个车牌之间的关联,从而实现稳健的协同车牌识别。具体来说,我们设计了一个关联增强分支,它利用完整的车牌号码来挖掘两个车牌之间的关联,从而监督两个车牌信息的融合。为了提高各类车牌的表示能力,我们在训练阶段设计了一个辅助学习分支,在两个车牌的关联增强中监督单个车牌的学习。此外,我们还提供了一个名为 CLPR 的综合基准数据集,该数据集由来自中国 24 个省份的 19782 个标准车牌和放大车牌组成,涵盖了真实场景中协同车牌识别所面临的大部分挑战。在拟议的 CLPR 数据集上进行的大量实验证明,拟议的 AENet 与几种最先进的方法相比非常有效。
{"title":"Collaborative License Plate Recognition via Association Enhancement Network With Auxiliary Learning and a Unified Benchmark","authors":"Yifei Deng;Guohao Wang;Chenglong Li;Wei Wang;Cheng Zhang;Jin Tang","doi":"10.1109/TMM.2024.3452982","DOIUrl":"10.1109/TMM.2024.3452982","url":null,"abstract":"Since the standard license plate of large vehicle is easily affected by occlusion and stain, the traffic management department introduces the enlarged license plate at the rear of the large vehicle to assist license plate recognition. However, current researches regards standard license plate recognition and enlarged license plate recognition as independent tasks, and do not take advantage of the complementary benefits from the two types of license plates. In this work, we propose a new computer vision task called collaborative license plate recognition, aiming to leverage the complementary advantages of standard and enlarged license plates for achieving more accurate license plate recognition. To achieve this goal, we propose an Association Enhancement Network (AENet), which achieves robust collaborative licence plate recognition by capturing the correlations between characters within a single licence plate and enhancing the associations between two license plates. In particular, we design an association enhancement branch, which supervises the fusion of two licence plate information using the complete licence plate number to mine the association between them. To enhance the representation ability of each type of licence plates, we design an auxiliary learning branch in the training stage, which supervises the learning of individual license plates in the association enhancement between two license plates. In addition, we contribute a comprehensive benchmark dataset called CLPR, which consists of a total of 19,782 standard and enlarged licence plates from 24 provinces in China and covers most of the challenges in real scenarios, for collaborative license plate recognition. Extensive experiments on the proposed CLPR dataset demonstrate the effectiveness of the proposed AENet against several state-of-the-art methods.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11402-11414"},"PeriodicalIF":8.4,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VLDadaptor: Domain Adaptive Object Detection With Vision-Language Model Distillation VLDadaptor:通过视觉语言模型提炼实现领域自适应目标检测
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-06 DOI: 10.1109/TMM.2024.3453061
Junjie Ke;Lihuo He;Bo Han;Jie Li;Di Wang;Xinbo Gao
Domain adaptive object detection (DAOD) aims to develop a detector trained on labeled source domains to identify objects in unlabeled target domains. A primary challenge in DAOD is the domain shift problem. Most existing methods learn domain-invariant features within single domain embedding space, often resulting in heavy model biases due to the intrinsic data properties of source domains. To mitigate the model biases, this paper proposes VLDadaptor, a domain adaptive object detector based on vision-language models (VLMs) distillation. Firstly, the proposed method integrates domain-mixed contrastive knowledge distillation between the visual encoder of CLIP and the detector by transferring category-level instance features, which guarantees the detector can extract domain-invariant visual instance features across domains. Then, VLDadaptor employs domain-mixed consistency distillation between the text encoder of CLIP and detector by aligning text prompt embeddings with visual instance features, which helps to maintain the category-level feature consistency among the detector, text encoder and the visual encoder of VLMs. Finally, the proposed method further promotes the adaptation ability by adopting a prompt-based memory bank to generate semantic-complete features for graph matching. These contributions enable VLDadaptor to extract visual features into the visual-language embedding space without any evident model bias towards specific domains. Extensive experimental results demonstrate that the proposed method achieves state-of-the-art performance on Pascal VOC to Clipart adaptation tasks and exhibits high accuracy on driving scenario tasks with significantly less training time.
域自适应物体检测(DAOD)旨在开发一种在已标注源域上经过训练的检测器,以识别未标注目标域中的物体。DAOD 面临的一个主要挑战是域偏移问题。大多数现有方法都是在单域嵌入空间内学习域不变特征,由于源域的内在数据属性,往往会导致严重的模型偏差。为了减轻模型偏差,本文提出了一种基于视觉语言模型(VLMs)提炼的域自适应物体检测器--VLDadaptor。首先,本文提出的方法在 CLIP 视觉编码器和检测器之间集成了领域混合对比知识蒸馏,通过转移类别级实例特征,保证检测器能够跨领域提取领域不变的视觉实例特征。然后,VLDadaptor 通过将文本提示嵌入与视觉实例特征对齐,在 CLIP 文本编码器和检测器之间进行域混合一致性提炼,这有助于保持检测器、文本编码器和 VLM 视觉编码器之间的类别级特征一致性。最后,通过采用基于提示的记忆库来生成用于图匹配的语义完整特征,所提出的方法进一步提高了适应能力。这些贡献使 VLDadaptor 能够在视觉语言嵌入空间中提取视觉特征,而不会对特定领域产生明显的模型偏差。广泛的实验结果表明,所提出的方法在 Pascal VOC 到剪贴画的适配任务中取得了最先进的性能,并在驾驶场景任务中表现出较高的准确性,同时大大减少了训练时间。
{"title":"VLDadaptor: Domain Adaptive Object Detection With Vision-Language Model Distillation","authors":"Junjie Ke;Lihuo He;Bo Han;Jie Li;Di Wang;Xinbo Gao","doi":"10.1109/TMM.2024.3453061","DOIUrl":"10.1109/TMM.2024.3453061","url":null,"abstract":"Domain adaptive object detection (DAOD) aims to develop a detector trained on labeled source domains to identify objects in unlabeled target domains. A primary challenge in DAOD is the domain shift problem. Most existing methods learn domain-invariant features within single domain embedding space, often resulting in heavy model biases due to the intrinsic data properties of source domains. To mitigate the model biases, this paper proposes VLDadaptor, a domain adaptive object detector based on vision-language models (VLMs) distillation. Firstly, the proposed method integrates domain-mixed contrastive knowledge distillation between the visual encoder of CLIP and the detector by transferring category-level instance features, which guarantees the detector can extract domain-invariant visual instance features across domains. Then, VLDadaptor employs domain-mixed consistency distillation between the text encoder of CLIP and detector by aligning text prompt embeddings with visual instance features, which helps to maintain the category-level feature consistency among the detector, text encoder and the visual encoder of VLMs. Finally, the proposed method further promotes the adaptation ability by adopting a prompt-based memory bank to generate semantic-complete features for graph matching. These contributions enable VLDadaptor to extract visual features into the visual-language embedding space without any evident model bias towards specific domains. Extensive experimental results demonstrate that the proposed method achieves state-of-the-art performance on Pascal VOC to Clipart adaptation tasks and exhibits high accuracy on driving scenario tasks with significantly less training time.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11316-11331"},"PeriodicalIF":8.4,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Camera-Incremental Object Re-Identification With Identity Knowledge Evolution 利用身份知识演进进行相机增量物体再识别
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-05 DOI: 10.1109/TMM.2024.3453045
Hantao Yao;Jifei Luo;Lu Yu;Changsheng Xu
Object Re-identification (ReID) is a task focused on retrieving a probe object from a multitude of gallery images using a ReID model trained on a stationary, camera-free dataset. This training involves associating and aggregating identities across various camera views. However, when deploying ReID algorithms in real-world scenarios, several challenges, such as storage constraints, privacy considerations, and dynamic changes in camera setups, can hinder their generalizability and practicality. To address these challenges, we introduce a novel ReID task called Camera-Incremental Object Re-identification (CIOR). In CIOR, we treat each camera's data as a separate source and continually optimize the ReID model as new data streams come from various cameras. By associating and consolidating the knowledge of common identities, our aim is to enhance discrimination capabilities and mitigate the problem of catastrophic forgetting. Therefore, we propose a novel Identity Knowledge Evolution (IKE) framework for CIOR, consisting of Identity Knowledge Association (IKA), Identity Knowledge Distillation (IKD), and Identity Knowledge Update (IKU). IKA is proposed to discover common identities between the current identity and historical identities, facilitating the integration of previously acquired knowledge. IKD involves distilling historical identity knowledge from common identities, enabling rapid adaptation of the historical model to the current camera view. After each camera has been trained, IKU is applied to continually expand identity knowledge by combining historical and current identity memories. Market-CL and Veri-CL evaluations show the effectiveness of Identity Knowledge Evolution (IKE) for CIOR.Code: https://github.com/htyao89/Camera-Incremental-Object-ReID
物体再识别(ReID)是一项任务,其重点是利用在静态、无摄像头数据集上训练的 ReID 模型,从大量图库图像中检索探测物体。这种训练包括在不同的相机视图中关联和汇总身份。然而,在真实世界场景中部署 ReID 算法时,存储限制、隐私考虑和摄像头设置的动态变化等一些挑战会阻碍算法的通用性和实用性。为了应对这些挑战,我们引入了一种名为 "摄像头增量对象再识别(CIOR)"的新型再识别任务。在 CIOR 中,我们将每台摄像机的数据视为一个单独的数据源,并随着来自不同摄像机的新数据流不断优化 ReID 模型。通过关联和整合共同身份的知识,我们的目标是提高识别能力,减少灾难性遗忘的问题。因此,我们为CIOR提出了一个新颖的身份知识演进(IKE)框架,由身份知识关联(IKA)、身份知识提炼(IKD)和身份知识更新(IKU)组成。IKA的目的是发现当前身份和历史身份之间的共同点,从而促进先前所获知识的整合。IKD 包括从共同身份中提炼出历史身份知识,使历史模型快速适应当前的摄像机视图。在每个摄像头经过训练后,IKU 将结合历史和当前身份记忆,不断扩展身份知识。Market-CL和Veri-CL评估显示了身份知识进化(IKE)对CIOR的有效性。代码:https://github.com/htyao89/Camera-Incremental-Object-ReID
{"title":"Camera-Incremental Object Re-Identification With Identity Knowledge Evolution","authors":"Hantao Yao;Jifei Luo;Lu Yu;Changsheng Xu","doi":"10.1109/TMM.2024.3453045","DOIUrl":"10.1109/TMM.2024.3453045","url":null,"abstract":"Object Re-identification (ReID) is a task focused on retrieving a probe object from a multitude of gallery images using a ReID model trained on a stationary, camera-free dataset. This training involves associating and aggregating identities across various camera views. However, when deploying ReID algorithms in real-world scenarios, several challenges, such as storage constraints, privacy considerations, and dynamic changes in camera setups, can hinder their generalizability and practicality. To address these challenges, we introduce a novel ReID task called Camera-Incremental Object Re-identification (CIOR). In CIOR, we treat each camera's data as a separate source and continually optimize the ReID model as new data streams come from various cameras. By associating and consolidating the knowledge of common identities, our aim is to enhance discrimination capabilities and mitigate the problem of catastrophic forgetting. Therefore, we propose a novel Identity Knowledge Evolution (IKE) framework for CIOR, consisting of Identity Knowledge Association (IKA), Identity Knowledge Distillation (IKD), and Identity Knowledge Update (IKU). IKA is proposed to discover common identities between the current identity and historical identities, facilitating the integration of previously acquired knowledge. IKD involves distilling historical identity knowledge from common identities, enabling rapid adaptation of the historical model to the current camera view. After each camera has been trained, IKU is applied to continually expand identity knowledge by combining historical and current identity memories. Market-CL and Veri-CL evaluations show the effectiveness of Identity Knowledge Evolution (IKE) for CIOR.Code: \u0000<uri>https://github.com/htyao89/Camera-Incremental-Object-ReID</uri>","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11246-11260"},"PeriodicalIF":8.4,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dual-View Data Hallucination With Semantic Relation Guidance for Few-Shot Image Recognition 利用语义关系指导双视图数据幻象,实现少镜头图像识别
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TMM.2024.3453055
Hefeng Wu;Guangzhi Ye;Ziyang Zhou;Ling Tian;Qing Wang;Liang Lin
Learning to recognize novel concepts from just a few image samples is very challenging as the learned model is easily overfitted on the few data and results in poor generalizability. One promising but underexplored solution is to compensate for the novel classes by generating plausible samples. However, most existing works of this line exploit visual information only, rendering the generated data easy to be distracted by some challenging factors contained in the few available samples. Being aware of the semantic information in the textual modality that reflects human concepts, this work proposes a novel framework that exploits semantic relations to guide dual-view data hallucination for few-shot image recognition. The proposed framework enables generating more diverse and reasonable data samples for novel classes through effective information transfer from base classes. Specifically, an instance-view data hallucination module hallucinates each sample of a novel class to generate new data by employing local semantic correlated attention and global semantic feature fusion derived from base classes. Meanwhile, a prototype-view data hallucination module exploits semantic-aware measure to estimate the prototype of a novel class and the associated distribution from the few samples, which thereby harvests the prototype as a more stable sample and enables resampling a large number of samples. We conduct extensive experiments and comparisons with state-of-the-art methods on several popular few-shot benchmarks to verify the effectiveness of the proposed framework.
从少量图像样本中学习识别新概念非常具有挑战性,因为学习到的模型很容易对少量数据过度拟合,导致普适性差。一种前景广阔但尚未得到充分探索的解决方案是通过生成可信样本来补偿新类别。然而,大多数现有的相关工作都只利用了视觉信息,使得生成的数据很容易被少数可用样本中包含的一些挑战性因素所干扰。考虑到文本模式中的语义信息反映了人类的概念,这项工作提出了一个新颖的框架,利用语义关系来指导双视角数据幻化,从而实现少镜头图像识别。所提出的框架能通过有效的基础类信息转移,为新类别生成更多样、更合理的数据样本。具体来说,实例视图数据幻化模块通过局部语义相关注意和全局语义特征融合,对新类别的每个样本进行幻化,生成新数据。同时,原型视图数据幻象模块利用语义感知措施,从少量样本中估计出新类别的原型和相关分布,从而获得作为更稳定样本的原型,并实现对大量样本的重新采样。我们在几个流行的少量样本基准上进行了大量实验,并与最先进的方法进行了比较,以验证所提框架的有效性。
{"title":"Dual-View Data Hallucination With Semantic Relation Guidance for Few-Shot Image Recognition","authors":"Hefeng Wu;Guangzhi Ye;Ziyang Zhou;Ling Tian;Qing Wang;Liang Lin","doi":"10.1109/TMM.2024.3453055","DOIUrl":"10.1109/TMM.2024.3453055","url":null,"abstract":"Learning to recognize novel concepts from just a few image samples is very challenging as the learned model is easily overfitted on the few data and results in poor generalizability. One promising but underexplored solution is to compensate for the novel classes by generating plausible samples. However, most existing works of this line exploit visual information only, rendering the generated data easy to be distracted by some challenging factors contained in the few available samples. Being aware of the semantic information in the textual modality that reflects human concepts, this work proposes a novel framework that exploits semantic relations to guide dual-view data hallucination for few-shot image recognition. The proposed framework enables generating more diverse and reasonable data samples for novel classes through effective information transfer from base classes. Specifically, an instance-view data hallucination module hallucinates each sample of a novel class to generate new data by employing local semantic correlated attention and global semantic feature fusion derived from base classes. Meanwhile, a prototype-view data hallucination module exploits semantic-aware measure to estimate the prototype of a novel class and the associated distribution from the few samples, which thereby harvests the prototype as a more stable sample and enables resampling a large number of samples. We conduct extensive experiments and comparisons with state-of-the-art methods on several popular few-shot benchmarks to verify the effectiveness of the proposed framework.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11302-11315"},"PeriodicalIF":8.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEIRNet: Inconsistency Exploiting Based Identity Rectification for Face Forgery Detection IEIRNet:基于不一致性开发的人脸伪造检测身份校正技术
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TMM.2024.3453066
Mingqi Fang;Lingyun Yu;Yun Song;Yongdong Zhang;Hongtao Xie
Face forgery detection has attracted much attention due to the ever-increasing social concerns caused by facial manipulation techniques. Recently, identity-based detection methods have made considerable progress, which is especially suitable in the celebrity protection scenario. However, they still suffer from two main limitations: (a) generic identity extractor is not specifically designed for forgery detection, leading to nonnegligible Identity Representation Bias to forged images. (b) existing methods only analyze the identity representation of each image individually, but ignores the query-reference interaction for inconsistency exploiting. To address these issues, a novel Inconsistency Exploiting based Identity Rectification Network (IEIRNet) is proposed in this paper. Firstly, for the identity bias rectification, the IEIRNet follows an effective two-branches structure. Besides the Generic Identity Extractor (GIE) branch, an essential Bias Diminishing Module (BDM) branch is proposed to eliminate the identity bias through a novel Attention-based Bias Rectification (ABR) component, accordingly acquiring the ultimate discriminative identity representation. Secondly, for query-reference inconsistency exploiting, an Inconsistency Exploiting Module (IEM) is applied in IEIRNet to comprehensively exploit the inconsistency clues from both spatial and channel perspectives. In the spatial aspect, an innovative region-aware kernel is derived to activate the local region inconsistency with deep spatial interaction. Afterward in the channel aspect, a coattention mechanism is utilized to model the channel interaction meticulously, and accordingly highlight the channel-wise inconsistency with adaptive weight assignment and channel-wise dropout. Our IEIRNet has shown effectiveness and superiority in various generalization and robustness experiments.
由于人脸伪造技术引发的社会关注与日俱增,人脸伪造检测备受关注。最近,基于身份的检测方法取得了长足的进步,尤其适用于名人保护场景。然而,这些方法仍然存在两大局限性:(a) 通用身份提取器并非专为伪造检测而设计,导致伪造图像存在不可忽略的身份表征偏差。(b) 现有方法只能单独分析每幅图像的身份表征,而忽略了利用查询-参考交互进行不一致利用。针对这些问题,本文提出了一种新型的基于不一致性利用的身份校正网络(IEIRNet)。首先,为了纠正身份偏差,IEIRNet 采用了有效的双分支结构。除了通用身份提取器(GIE)分支外,还提出了一个重要的偏差消除模块(BDM)分支,通过一个新颖的基于注意力的偏差纠正(ABR)组件消除身份偏差,从而获得最终的鉴别性身份表示。其次,在查询-参考不一致利用方面,IEIRNet 采用了不一致利用模块 (IEM),从空间和信道两个角度全面利用不一致线索。在空间方面,一个创新的区域感知内核通过深度空间交互来激活局部区域的不一致性。然后,在信道方面,利用协同关注机制对信道交互进行细致建模,并相应地通过自适应权重分配和信道剔除来突出信道方面的不一致性。我们的 IEIRNet 在各种泛化和鲁棒性实验中显示出了有效性和优越性。
{"title":"IEIRNet: Inconsistency Exploiting Based Identity Rectification for Face Forgery Detection","authors":"Mingqi Fang;Lingyun Yu;Yun Song;Yongdong Zhang;Hongtao Xie","doi":"10.1109/TMM.2024.3453066","DOIUrl":"10.1109/TMM.2024.3453066","url":null,"abstract":"Face forgery detection has attracted much attention due to the ever-increasing social concerns caused by facial manipulation techniques. Recently, identity-based detection methods have made considerable progress, which is especially suitable in the celebrity protection scenario. However, they still suffer from two main limitations: (a) generic identity extractor is not specifically designed for forgery detection, leading to nonnegligible \u0000<italic>Identity Representation Bias</i>\u0000 to forged images. (b) existing methods only analyze the identity representation of each image individually, but ignores the query-reference interaction for inconsistency exploiting. To address these issues, a novel \u0000<italic>Inconsistency Exploiting based Identity Rectification Network</i>\u0000 (IEIRNet) is proposed in this paper. Firstly, for the identity bias rectification, the IEIRNet follows an effective two-branches structure. Besides the \u0000<italic>Generic Identity Extractor</i>\u0000 (GIE) branch, an essential \u0000<italic>Bias Diminishing Module</i>\u0000 (BDM) branch is proposed to eliminate the identity bias through a novel \u0000<italic>Attention-based Bias Rectification</i>\u0000 (ABR) component, accordingly acquiring the ultimate discriminative identity representation. Secondly, for query-reference inconsistency exploiting, an \u0000<italic>Inconsistency Exploiting Module</i>\u0000 (IEM) is applied in IEIRNet to comprehensively exploit the inconsistency clues from both spatial and channel perspectives. In the spatial aspect, an innovative region-aware kernel is derived to activate the local region inconsistency with deep spatial interaction. Afterward in the channel aspect, a coattention mechanism is utilized to model the channel interaction meticulously, and accordingly highlight the channel-wise inconsistency with adaptive weight assignment and channel-wise dropout. Our IEIRNet has shown effectiveness and superiority in various generalization and robustness experiments.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11232-11245"},"PeriodicalIF":8.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pixel-Learnable 3DLUT With Saturation-Aware Compensation for Image Enhancement 具有饱和度补偿功能的像素可学习 3DLUT 图像增强技术
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TMM.2024.3453064
Jing Liu;Qingying Li;Xiongkuo Min;Yuting Su;Guangtao Zhai;Xiaokang Yang
The 3D Lookup Table (3DLUT)-based methods are gaining popularity due to their satisfactory and stable performance in achieving automatic and adaptive real time image enhancement. In this paper, we present a new solution to the intractability in handling continuous color transformations of 3DLUT due to the lookup via three independent color channel coordinates in RGB space. Inspired by the inherent merits of the HSV color space, we separately enhance image intensity and color composition. The Transformer-based Pixel-Learnable 3D Lookup Table is proposed to undermine contouring artifacts, which enhances images in a pixel-wise manner with non-local information to emphasize the diverse spatially variant context. In addition, noticing the underestimation of composition color component, we develop the Saturation-Aware Compensation (SAC) module to enhance the under-saturated region determined by an adaptive SA map with Saturation-Interaction block, achieving well balance between preserving details and color rendition. Our approach can be applied to image retouching and tone mapping tasks with fairly good generality, especially in restoring localized regions with weak visibility. The performance in both theoretical analysis and comparative experiments manifests that the proposed solution is effective and robust.
基于三维查找表(3DLUT)的方法在实现自动和自适应实时图像增强方面具有令人满意的稳定性能,因而越来越受欢迎。在本文中,我们提出了一种新的解决方案,以解决 3DLUT 由于通过 RGB 空间中的三个独立颜色通道坐标进行查找而导致的处理连续颜色变换的困难性。受 HSV 色彩空间固有优点的启发,我们分别增强了图像的强度和色彩构成。为了消除轮廓伪影,我们提出了基于变换器的像素可学习三维查找表,它利用非本地信息以像素方式增强图像,以强调多样化的空间变化背景。此外,我们还注意到成分色彩部分被低估,因此开发了饱和度感知补偿(SAC)模块,以增强由饱和度-交互块自适应 SA 映射确定的饱和度不足区域,从而在保留细节和色彩呈现之间实现良好平衡。我们的方法可用于图像修饰和色调映射任务,具有相当好的通用性,尤其是在恢复能见度较弱的局部区域时。理论分析和对比实验的结果表明,所提出的解决方案既有效又稳健。
{"title":"Pixel-Learnable 3DLUT With Saturation-Aware Compensation for Image Enhancement","authors":"Jing Liu;Qingying Li;Xiongkuo Min;Yuting Su;Guangtao Zhai;Xiaokang Yang","doi":"10.1109/TMM.2024.3453064","DOIUrl":"10.1109/TMM.2024.3453064","url":null,"abstract":"The 3D Lookup Table (3DLUT)-based methods are gaining popularity due to their satisfactory and stable performance in achieving automatic and adaptive real time image enhancement. In this paper, we present a new solution to the intractability in handling continuous color transformations of 3DLUT due to the lookup via three independent color channel coordinates in RGB space. Inspired by the inherent merits of the HSV color space, we separately enhance image intensity and color composition. The Transformer-based Pixel-Learnable 3D Lookup Table is proposed to undermine contouring artifacts, which enhances images in a pixel-wise manner with non-local information to emphasize the diverse spatially variant context. In addition, noticing the underestimation of composition color component, we develop the Saturation-Aware Compensation (SAC) module to enhance the under-saturated region determined by an adaptive SA map with Saturation-Interaction block, achieving well balance between preserving details and color rendition. Our approach can be applied to image retouching and tone mapping tasks with fairly good generality, especially in restoring localized regions with weak visibility. The performance in both theoretical analysis and comparative experiments manifests that the proposed solution is effective and robust.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11219-11231"},"PeriodicalIF":8.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
End-to-End Image Colorization With Multiscale Pyramid Transformer 利用多尺度金字塔变换器实现端到端图像着色
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TMM.2024.3453035
Tongtong Zhao;Gehui Li;Shanshan Zhao
Image colorization is a challenging task due to its ill-posed and multimodal nature, leading to unsatisfactory results in traditional approaches that rely on reference images or user guides. Although deep learning-based methods have been proposed, they may not be sufficient due to the lack of semantic understanding. To overcome this limitation, we present an innovative end-to-end automatic colorization method that does not require any color reference images and achieves superior quantitative and qualitative results compared to state-of-the-art methods. Our approach incorporates a Multiscale Pyramid Transformer that captures both local and global contextual information and a novel attention module called Dual-Attention, which replaces the traditional Window Attention and Channel Attention with faster and lighter Separable Dilated Attention and Factorized Channel Attention. Additionally, we introduce a new color decoder called Color-Attention, which learns colorization patterns from grayscale images and color images of the current training set, resulting in improved generalizability and eliminating the need for constructing color priors. Experimental results demonstrate the effectiveness of our approach in various benchmark datasets, including high-level computer vision tasks such as classification, segmentation, and detection. Our method offers robustness, generalization ability, and improved colorization quality, making it a valuable contribution to the field of image colorization.
图像着色是一项具有挑战性的任务,因为它具有不确定性和多模态性,导致依赖参考图像或用户指南的传统方法无法取得令人满意的结果。虽然已经提出了基于深度学习的方法,但由于缺乏语义理解,这些方法可能还不够充分。为了克服这一局限性,我们提出了一种创新的端到端自动着色方法,这种方法不需要任何色彩参考图像,与最先进的方法相比,在定量和定性方面都取得了卓越的效果。我们的方法采用了多尺度金字塔变换器(Multiscale Pyramid Transformer),可捕捉局部和全局上下文信息;还采用了名为 "双注意力"(Dual-Attention)的新型注意力模块,用更快、更轻的可分离式稀释注意力(Separable Dilated Attention)和因子化通道注意力(Factorized Channel Attention)取代了传统的窗口注意力(Window Attention)和通道注意力(Channel Attention)。此外,我们还引入了一种名为 "色彩注意力"(Color-Attention)的新色彩解码器,它能从当前训练集的灰度图像和彩色图像中学习着色模式,从而提高了泛化能力,并且无需构建色彩先验。实验结果证明了我们的方法在各种基准数据集上的有效性,包括分类、分割和检测等高级计算机视觉任务。我们的方法具有鲁棒性、泛化能力和更高的着色质量,是对图像着色领域的宝贵贡献。
{"title":"End-to-End Image Colorization With Multiscale Pyramid Transformer","authors":"Tongtong Zhao;Gehui Li;Shanshan Zhao","doi":"10.1109/TMM.2024.3453035","DOIUrl":"10.1109/TMM.2024.3453035","url":null,"abstract":"Image colorization is a challenging task due to its ill-posed and multimodal nature, leading to unsatisfactory results in traditional approaches that rely on reference images or user guides. Although deep learning-based methods have been proposed, they may not be sufficient due to the lack of semantic understanding. To overcome this limitation, we present an innovative end-to-end automatic colorization method that does not require any color reference images and achieves superior quantitative and qualitative results compared to state-of-the-art methods. Our approach incorporates a Multiscale Pyramid Transformer that captures both local and global contextual information and a novel attention module called Dual-Attention, which replaces the traditional Window Attention and Channel Attention with faster and lighter Separable Dilated Attention and Factorized Channel Attention. Additionally, we introduce a new color decoder called Color-Attention, which learns colorization patterns from grayscale images and color images of the current training set, resulting in improved generalizability and eliminating the need for constructing color priors. Experimental results demonstrate the effectiveness of our approach in various benchmark datasets, including high-level computer vision tasks such as classification, segmentation, and detection. Our method offers robustness, generalization ability, and improved colorization quality, making it a valuable contribution to the field of image colorization.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11332-11344"},"PeriodicalIF":8.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coarse-to-Fine Target Detection for HFSWR With Spatial-Frequency Analysis and Subnet Structure 利用空间频率分析和子网结构进行从粗到细的 HFSWR 目标检测
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TMM.2024.3453044
Wandong Zhang;Yimin Yang;Tianlong Liu
High-frequency surface wave radar (HFSWR) is a powerful tool for ship detection and surveillance. blackHowever, the use of pre-trained deep learning (DL) networks for ship detection is challenging due to the limited training samples in HFSWR and the substantial differences between remote sensing images and everyday images. To tackle these issues, this paper proposes a coarse-to-fine target detection approach that combines traditional methods with DL, resulting in improved performance. The contributions of this work include: 1) a two-stage learning pipeline that integrates spatial-frequency analysis (SFA) with subnet-based neural networks, 2) an automatic linear thresholding algorithm for plausible target region (PTR) detection, and 3) a robust subnet neural network for fine target detection. The advantage of using SFA and subnet network is that the SFA reduces the need for extensive training data, while the subnet neural network excels at localizing ships even with limited training data. Experimental results on the HFSWR-RD dataset affirm the model's superior performance compared to rival algorithms.
然而,由于高频面波雷达的训练样本有限,而且遥感图像与日常图像之间存在巨大差异,使用预先训练好的深度学习(DL)网络进行船舶探测具有挑战性。为了解决这些问题,本文提出了一种从粗到细的目标检测方法,将传统方法与深度学习相结合,从而提高了性能。这项工作的贡献包括1)将空间频率分析(SFA)与基于子网的神经网络相结合的两阶段学习管道;2)用于可信目标区域(PTR)检测的自动线性阈值算法;3)用于精细目标检测的鲁棒子网神经网络。使用 SFA 和子网神经网络的优势在于,SFA 减少了对大量训练数据的需求,而子网神经网络即使在训练数据有限的情况下也能出色地定位舰船。在 HFSWR-RD 数据集上的实验结果表明,与竞争对手的算法相比,该模型具有更优越的性能。
{"title":"Coarse-to-Fine Target Detection for HFSWR With Spatial-Frequency Analysis and Subnet Structure","authors":"Wandong Zhang;Yimin Yang;Tianlong Liu","doi":"10.1109/TMM.2024.3453044","DOIUrl":"10.1109/TMM.2024.3453044","url":null,"abstract":"High-frequency surface wave radar (HFSWR) is a powerful tool for ship detection and surveillance. blackHowever, the use of pre-trained deep learning (DL) networks for ship detection is challenging due to the limited training samples in HFSWR and the substantial differences between remote sensing images and everyday images. To tackle these issues, this paper proposes a coarse-to-fine target detection approach that combines traditional methods with DL, resulting in improved performance. The contributions of this work include: 1) a two-stage learning pipeline that integrates spatial-frequency analysis (SFA) with subnet-based neural networks, 2) an automatic linear thresholding algorithm for plausible target region (PTR) detection, and 3) a robust subnet neural network for fine target detection. The advantage of using SFA and subnet network is that the SFA reduces the need for extensive training data, while the subnet neural network excels at localizing ships even with limited training data. Experimental results on the HFSWR-RD dataset affirm the model's superior performance compared to rival algorithms.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11290-11301"},"PeriodicalIF":8.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142223657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rethinking Video Sentence Grounding From a Tracking Perspective With Memory Network and Masked Attention 用记忆网络和掩蔽注意力从跟踪角度反思视频句子接地问题
IF 8.4 1区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-09-02 DOI: 10.1109/TMM.2024.3453062
Zeyu Xiong;Daizong Liu;Xiang Fang;Xiaoye Qu;Jianfeng Dong;Jiahao Zhu;Keke Tang;Pan Zhou
Video sentence grounding (VSG) is the task of identifying the segment of an untrimmed video that semantically corresponds to a given natural language query. While many existing methods extract frame-grained features using pre-trained 2D or 3D convolution networks, often fail to capture subtle differences between ambiguous adjacent frames. Although some recent approaches incorporate object-grained features using Faster R-CNN to capture more fine-grained details, they are still primarily based on feature enhancement and lack spatio-temporal modeling to explore the semantics of the core persons/objects. To solve the problem of modeling the core target's behavior, in this paper, we propose a new perspective for addressing the VSG task by tracking pivotal objects and activities to learn more fine-grained spatio-temporal features. Specifically, we introduce the Video Sentence Tracker with Memory Network and Masked Attention (VSTMM), which comprises a cross-modal targets generator for producing multi-modal templates and search space, a memory-based tracker for dynamically tracking multi-modal targets using a memory network to record targets' behaviors, a masked attention localizer which learns local shared features between frames and eliminates interference from long-term dependencies, resulting in improved accuracy when localizing the moment. To evaluate the performance of our VSTMM, we conducted extensive experiments and comparisons with state-of-the-art methods on three challenging benchmarks, including Charades-STA, ActivityNet Captions, and TACoS. Without bells and whistles, our VSTMM achieves leading performance with a considerable real-time speed.
视频句子接地(VSG)是指识别未剪辑视频中与给定自然语言查询语义对应的片段。现有的许多方法都是使用预先训练好的二维或三维卷积网络来提取帧粒度特征,但往往无法捕捉到模糊相邻帧之间的细微差别。虽然最近的一些方法使用 Faster R-CNN 结合了对象粒度特征,以捕捉更多细粒度细节,但它们仍然主要基于特征增强,缺乏时空建模来探索核心人物/对象的语义。为了解决核心目标行为建模的问题,本文提出了一种解决 VSG 任务的新视角,即通过跟踪关键对象和活动来学习更精细的时空特征。具体来说,我们介绍了具有记忆网络和掩码注意力的视频句子跟踪器(VSTMM),它包括一个用于生成多模态模板和搜索空间的跨模态目标生成器、一个用于使用记忆网络记录目标行为动态跟踪多模态目标的基于记忆的跟踪器、一个用于学习帧间局部共享特征并消除长期依赖性干扰的掩码注意力定位器,从而提高了时刻定位的准确性。为了评估 VSTMM 的性能,我们进行了大量实验,并在三个具有挑战性的基准测试中与最先进的方法进行了比较,包括猜字谜-STA、ActivityNet Captions 和 TACoS。我们的 VSTMM 在没有任何附加功能的情况下实现了领先的性能和相当快的实时速度。
{"title":"Rethinking Video Sentence Grounding From a Tracking Perspective With Memory Network and Masked Attention","authors":"Zeyu Xiong;Daizong Liu;Xiang Fang;Xiaoye Qu;Jianfeng Dong;Jiahao Zhu;Keke Tang;Pan Zhou","doi":"10.1109/TMM.2024.3453062","DOIUrl":"10.1109/TMM.2024.3453062","url":null,"abstract":"Video sentence grounding (VSG) is the task of identifying the segment of an untrimmed video that semantically corresponds to a given natural language query. While many existing methods extract frame-grained features using pre-trained 2D or 3D convolution networks, often fail to capture subtle differences between ambiguous adjacent frames. Although some recent approaches incorporate object-grained features using Faster R-CNN to capture more fine-grained details, they are still primarily based on feature enhancement and lack spatio-temporal modeling to explore the semantics of the core persons/objects. To solve the problem of modeling the core target's behavior, in this paper, we propose a new perspective for addressing the VSG task by tracking pivotal objects and activities to learn more fine-grained spatio-temporal features. Specifically, we introduce the Video Sentence Tracker with Memory Network and Masked Attention (VSTMM), which comprises a cross-modal targets generator for producing multi-modal templates and search space, a memory-based tracker for dynamically tracking multi-modal targets using a memory network to record targets' behaviors, a masked attention localizer which learns local shared features between frames and eliminates interference from long-term dependencies, resulting in improved accuracy when localizing the moment. To evaluate the performance of our VSTMM, we conducted extensive experiments and comparisons with state-of-the-art methods on three challenging benchmarks, including Charades-STA, ActivityNet Captions, and TACoS. Without bells and whistles, our VSTMM achieves leading performance with a considerable real-time speed.","PeriodicalId":13273,"journal":{"name":"IEEE Transactions on Multimedia","volume":"26 ","pages":"11204-11218"},"PeriodicalIF":8.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Multimedia
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1