首页 > 最新文献

Journal of Visual Communication and Image Representation最新文献

英文 中文
Multimodal prompt-guided vision transformer for precise image manipulation localization 用于精确图像处理定位的多模态提示引导视觉转换器
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-27 DOI: 10.1016/j.jvcir.2026.104736
Yafang Xiao , Wei Jiang , Shihua Zhou , Bin Wang , Pengfei Wang , Pan Zheng
With the rise of generative AI and advanced image editing technologies, image manipulation localization has become more challenging. Existing methods often struggle with limited semantic understanding and insufficient spatial detail capture, especially in complex scenarios. To address these issues, we propose a novel multimodal text-guided framework for image manipulation localization. By fusing textual prompts with image features, our approach enhances the model’s ability to identify manipulated regions. We introduce a Multimodal Interaction Prompt Module (MIPM) that uses cross-modal attention mechanisms to align visual and textual information. Guided by multimodal prompts, our Vision Transformer-based model accurately localizes forged areas in images. Extensive experiments on public datasets, including CASIAv1 and Columbia, show that our method outperforms existing approaches. Specifically, on the CASIAv1 dataset, our approach achieves an F1 score of 0.734, surpassing the second-best method by 1.3%. These results demonstrate the effectiveness of our multimodal fusion strategy. The code is available at https://github.com/Makabaka613/MPG-ViT.
随着生成式人工智能和先进的图像编辑技术的兴起,图像处理本地化变得更具挑战性。现有的方法往往在有限的语义理解和不足的空间细节捕获方面挣扎,特别是在复杂的场景中。为了解决这些问题,我们提出了一种新的多模态文本引导框架用于图像处理定位。通过将文本提示与图像特征融合,我们的方法增强了模型识别被操纵区域的能力。我们介绍了一个多模态交互提示模块(MIPM),它使用跨模态注意机制来对齐视觉和文本信息。在多模态提示的引导下,我们基于视觉转换器的模型可以准确地定位图像中的伪造区域。在包括CASIAv1和Columbia在内的公共数据集上进行的大量实验表明,我们的方法优于现有方法。具体来说,在CASIAv1数据集上,我们的方法获得了0.734的F1分数,比第二好的方法高出1.3%。这些结果证明了我们的多模态融合策略的有效性。代码可在https://github.com/Makabaka613/MPG-ViT上获得。
{"title":"Multimodal prompt-guided vision transformer for precise image manipulation localization","authors":"Yafang Xiao ,&nbsp;Wei Jiang ,&nbsp;Shihua Zhou ,&nbsp;Bin Wang ,&nbsp;Pengfei Wang ,&nbsp;Pan Zheng","doi":"10.1016/j.jvcir.2026.104736","DOIUrl":"10.1016/j.jvcir.2026.104736","url":null,"abstract":"<div><div>With the rise of generative AI and advanced image editing technologies, image manipulation localization has become more challenging. Existing methods often struggle with limited semantic understanding and insufficient spatial detail capture, especially in complex scenarios. To address these issues, we propose a novel multimodal text-guided framework for image manipulation localization. By fusing textual prompts with image features, our approach enhances the model’s ability to identify manipulated regions. We introduce a Multimodal Interaction Prompt Module (MIPM) that uses cross-modal attention mechanisms to align visual and textual information. Guided by multimodal prompts, our Vision Transformer-based model accurately localizes forged areas in images. Extensive experiments on public datasets, including CASIAv1 and Columbia, show that our method outperforms existing approaches. Specifically, on the CASIAv1 dataset, our approach achieves an F1 score of 0.734, surpassing the second-best method by 1.3%. These results demonstrate the effectiveness of our multimodal fusion strategy. The code is available at <span><span>https://github.com/Makabaka613/MPG-ViT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104736"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146079535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing temporal action localization through cross-modal and cross-structural knowledge distillation 通过跨模态和跨结构的知识提炼来增强时间动作定位
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-02-02 DOI: 10.1016/j.jvcir.2026.104734
Yue Yu, Cheng Wang, Yuxin Shi
This paper proposes Cross-Modal and Cross-Structure distillation for rgb-based temporal action detection(C2MS-Net), a novel fully supervised approach for enhancing temporal action localization by leveraging cross-modal and cross-structural distillation techniques. By integrating information from multiple modalities and structural representations, C2MS-Net significantly improves the discriminative power of action proposals. A distillation framework is introduced, which transfers knowledge from a teacher model trained on rich multi-modal data to a more efficient student model. This approach not only enhances temporal localization accuracy but also improves the robustness of action detection against visual content variations. Extensive experiments on benchmark datasets demonstrate that the proposed C2MS-Net performs competitively with or surpasses state-of-the-art methods, particularly at lower and mid-range tIoU thresholds, while offering substantial gains in computational efficiency. By eliminating the need for optical flow extraction, the proposed method substantially reduces computational complexity, achieving faster inference speeds and smaller model sizes without compromising accuracy. Code, dataset and models are available at: https://github.com/wangcheng666/ActionFormer.
本文提出了基于rgb的时间动作检测(C2MS-Net)的跨模态和跨结构蒸馏,这是一种利用跨模态和跨结构蒸馏技术增强时间动作定位的新型全监督方法。C2MS-Net通过整合来自多种模式和结构表征的信息,显著提高了行动建议的判别能力。介绍了一种精馏框架,该框架将基于丰富多模态数据训练的教师模型中的知识转化为更高效的学生模型。该方法不仅提高了时间定位精度,而且提高了动作检测对视觉内容变化的鲁棒性。在基准数据集上进行的大量实验表明,所提出的C2MS-Net与最先进的方法相比具有竞争力或优于最先进的方法,特别是在较低和中等范围的tIoU阈值时,同时在计算效率方面提供了可观的收益。通过消除光流提取的需要,该方法大大降低了计算复杂度,在不影响精度的情况下实现了更快的推理速度和更小的模型尺寸。代码、数据集和模型可在https://github.com/wangcheng666/ActionFormer上获得。
{"title":"Enhancing temporal action localization through cross-modal and cross-structural knowledge distillation","authors":"Yue Yu,&nbsp;Cheng Wang,&nbsp;Yuxin Shi","doi":"10.1016/j.jvcir.2026.104734","DOIUrl":"10.1016/j.jvcir.2026.104734","url":null,"abstract":"<div><div>This paper proposes Cross-Modal and Cross-Structure distillation for rgb-based temporal action detection(C2MS-Net), a novel fully supervised approach for enhancing temporal action localization by leveraging cross-modal and cross-structural distillation techniques. By integrating information from multiple modalities and structural representations, C2MS-Net significantly improves the discriminative power of action proposals. A distillation framework is introduced, which transfers knowledge from a teacher model trained on rich multi-modal data to a more efficient student model. This approach not only enhances temporal localization accuracy but also improves the robustness of action detection against visual content variations. Extensive experiments on benchmark datasets demonstrate that the proposed C2MS-Net performs competitively with or surpasses state-of-the-art methods, particularly at lower and mid-range tIoU thresholds, while offering substantial gains in computational efficiency. By eliminating the need for optical flow extraction, the proposed method substantially reduces computational complexity, achieving faster inference speeds and smaller model sizes without compromising accuracy. Code, dataset and models are available at: <span><span>https://github.com/wangcheng666/ActionFormer</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104734"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Progressively multi-scale feature fusion for semantic segmentation 渐进式多尺度特征融合语义分割
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-28 DOI: 10.1016/j.jvcir.2026.104739
Guoqing Zhang , Shichao Kan , Yigang Cen , Yi Cen , Qi Cao , Yansen Huang , Ming Zeng
A fundamental challenge in semantic segmentation is the discriminative learning of pixel-level features. Various semantic segmentation methods and decoders in the literature have been reported to address this challenge. These methods involve directly upsampling feature maps of different sizes and then concatenating them along the channel dimension to generate pixel-level features. However, direct upsampling of feature maps can result in the misalignment of information at the pixel level, leading to suboptimal performance. In this paper, we introduce a novel solution called the Progressive Multi-Scale Feature Fusion (PMSFF) decoder to overcome this issue. Specifically, we develop a lightweight feed-forward network and atrous convolution layer, that are combined as a fusion module to fuse feature maps from adjacent layers. This fusion module is applied to different segments of a network to aggregate all feature maps for semantic segmentation. The fusion module is characterized by a simple and convenient structure with fewer parameters, which can be flexibly embedded into both Convolutional Neural Networks (CNNs) and Transformers to achieve progressive multi-scale pixel-level feature fusion. Extensive experiments on benchmark datasets have been conducted. The results illustrate the effectiveness and efficiency of the proposed module.
语义分割的一个基本挑战是像素级特征的判别学习。文献中已经报道了各种语义分割方法和解码器来解决这一挑战。这些方法包括直接对不同大小的特征图进行上采样,然后沿着通道维度将它们连接起来以生成像素级特征。然而,直接对特征映射进行上采样可能会导致像素级信息的不对齐,从而导致次优性能。在本文中,我们提出了一种新的解决方案,称为渐进式多尺度特征融合(PMSFF)解码器来克服这个问题。具体来说,我们开发了一个轻量级的前馈网络和亚光卷积层,它们被组合成一个融合模块来融合来自相邻层的特征映射。该融合模块应用于网络的不同段,聚合所有特征映射进行语义分割。该融合模块具有结构简单方便、参数少的特点,可灵活嵌入卷积神经网络(cnn)和变压器中,实现多尺度像素级特征的渐进融合。在基准数据集上进行了大量的实验。实验结果表明了该模块的有效性和高效性。
{"title":"Progressively multi-scale feature fusion for semantic segmentation","authors":"Guoqing Zhang ,&nbsp;Shichao Kan ,&nbsp;Yigang Cen ,&nbsp;Yi Cen ,&nbsp;Qi Cao ,&nbsp;Yansen Huang ,&nbsp;Ming Zeng","doi":"10.1016/j.jvcir.2026.104739","DOIUrl":"10.1016/j.jvcir.2026.104739","url":null,"abstract":"<div><div>A fundamental challenge in semantic segmentation is the discriminative learning of pixel-level features. Various semantic segmentation methods and decoders in the literature have been reported to address this challenge. These methods involve directly upsampling feature maps of different sizes and then concatenating them along the channel dimension to generate pixel-level features. However, direct upsampling of feature maps can result in the misalignment of information at the pixel level, leading to suboptimal performance. In this paper, we introduce a novel solution called the Progressive Multi-Scale Feature Fusion (PMSFF) decoder to overcome this issue. Specifically, we develop a lightweight feed-forward network and atrous convolution layer, that are combined as a fusion module to fuse feature maps from adjacent layers. This fusion module is applied to different segments of a network to aggregate all feature maps for semantic segmentation. The fusion module is characterized by a simple and convenient structure with fewer parameters, which can be flexibly embedded into both Convolutional Neural Networks (CNNs) and Transformers to achieve progressive multi-scale pixel-level feature fusion. Extensive experiments on benchmark datasets have been conducted. The results illustrate the effectiveness and efficiency of the proposed module.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104739"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MGLA-DSNet: Multi-head global-local attention-enabled dual-stream network for weakly supervised video anomaly detection MGLA-DSNet:用于弱监督视频异常检测的多头全局本地关注双流网络
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-02-12 DOI: 10.1016/j.jvcir.2026.104744
Rashmiranjan Nayak, Umesh Chandra Pati, Santos Kumar Das
Video Anomaly Detection (VAD) is the process of identifying anomalous events by analyzing spatiotemporal patterns in video. Furthermore, VAD is a complex task due to difficulties in obtaining frame-level annotations, data imbalance issues, and the equivocal and context-dependent nature of video anomalies. To address these issues, this article presents a weakly supervised learning-based Multi-head Global-Local Attention-enabled Dual-Stream Network (MGLA-DSNet) that effectively utilizes spatial (appearance) and temporal (motion) features, with an emphasis on context dependency. The proposed model uses two streams to extract RGB and optical flow features corresponding to appearance (spatial) and motion (temporal) properties, respectively. Subsequently, multi-head global and location attention with adaptive gating and head-wise specialization is applied to the concatenated RGB and Flow features to efficiently model global and local contexts, respectively, using multiple instance learning Finally, the proposed MGLA-DSNet model outperforms state-of-the-art methods across three benchmark datasets, including CUHK Avenue, ShanghaiTech Campus, and UCF-Crime.
视频异常检测(VAD)是通过分析视频中的时空模式来识别异常事件的过程。此外,由于难以获得帧级注释、数据不平衡问题以及视频异常的模糊性和上下文依赖性,VAD是一项复杂的任务。为了解决这些问题,本文提出了一种基于弱监督学习的多头全局局部注意双流网络(MGLA-DSNet),该网络有效地利用了空间(外观)和时间(运动)特征,强调了上下文依赖性。该模型使用两个流分别提取与外观(空间)和运动(时间)属性相对应的RGB和光流特征。随后,将具有自适应门控和头部专门化的多头部全局和位置关注应用于连接的RGB和Flow特征,分别使用多实例学习有效地建模全局和局部上下文。最后,所提出的MGLA-DSNet模型在三个基准数据集(包括中大大道、上海科技园区和UCF-Crime)上优于最先进的方法。
{"title":"MGLA-DSNet: Multi-head global-local attention-enabled dual-stream network for weakly supervised video anomaly detection","authors":"Rashmiranjan Nayak,&nbsp;Umesh Chandra Pati,&nbsp;Santos Kumar Das","doi":"10.1016/j.jvcir.2026.104744","DOIUrl":"10.1016/j.jvcir.2026.104744","url":null,"abstract":"<div><div>Video Anomaly Detection (VAD) is the process of identifying anomalous events by analyzing spatiotemporal patterns in video. Furthermore, VAD is a complex task due to difficulties in obtaining frame-level annotations, data imbalance issues, and the equivocal and context-dependent nature of video anomalies. To address these issues, this article presents a weakly supervised learning-based Multi-head Global-Local Attention-enabled Dual-Stream Network (MGLA-DSNet) that effectively utilizes spatial (appearance) and temporal (motion) features, with an emphasis on context dependency. The proposed model uses two streams to extract RGB and optical flow features corresponding to appearance (spatial) and motion (temporal) properties, respectively. Subsequently, multi-head global and location attention with adaptive gating and head-wise specialization is applied to the concatenated RGB and Flow features to efficiently model global and local contexts, respectively, using multiple instance learning Finally, the proposed MGLA-DSNet model outperforms state-of-the-art methods across three benchmark datasets, including CUHK Avenue, ShanghaiTech Campus, and UCF-Crime.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104744"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CKCR: Context-aware knowledge construction and retrieval for knowledge-based visual question answering 基于知识的视觉问答的语境感知知识构建与检索
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-13 DOI: 10.1016/j.jvcir.2026.104711
Fengjuan Wang , Jiayi Liu , Ruonan Zhang, Zhengxue Li, Feng Zhang, Gaoyun An
Knowledge-based Visual Question Answering (KB-VQA) requires models to integrate visual content with external knowledge to answer questions, which is crucial for building intelligent systems capable of real-world understanding. However, effectively incorporating external knowledge into visual reasoning faces three major challenges: the incompleteness of external knowledge bases leads to missing knowledge for many specific visual scenarios, semantic gaps exist between retrieved textual knowledge and visual content making alignment difficult, and effective mechanisms for fusing heterogeneous knowledge sources are lacking. While Multimodal Large Language Models(MLLMs) have demonstrated strong performance in visual understanding tasks, but face notable challenges in KB-VQA, particularly in knowledge utilization efficiency and semantic alignment, which seriously limits the reasoning depth and robustness. To address these problems, a Context-aware Knowledge Construction and Retrieval (CKCR) method is proposed for knowledge-based VQA, which includes the following three modules. The multi-granularity knowledge retrieval module constructs joint query vector based on the multi-dimensional embedding representation of images and questions, accurately obtaining explicit knowledge that is highly matched with the context. The vision-to-knowledge generation module supplements fine-grained semantic clues from the perspective of visual content, generating visual knowledge closely related to the image and making up for the expression limitations of general knowledge. To achieve deep alignment of knowledge representation, the knowledge adaptive learning module accurately embeds multi-source knowledge into the semantic space of MLLM by introducing a learnable knowledge mapping mechanism. Experimental evaluation on OK-VQA and A-OKVQA dataset shows the CKCR outperforms state-of-the-art methods of the same-scale. Ablation experiments and visualization analysis demonstrate the superiority of CKCR in its perception of fine-grained visual information and its ability to align knowledge semantics. Our code will be released on GitHub: https://github.com/fjwang3/CKCR.
基于知识的视觉问答(knowledge -based Visual Question answer, KB-VQA)要求模型集成视觉内容和外部知识来回答问题,这对于构建能够理解现实世界的智能系统至关重要。然而,将外部知识有效地整合到视觉推理中面临着三个主要挑战:外部知识库的不完整性导致许多特定视觉场景的知识缺失;检索的文本知识与视觉内容之间存在语义差距导致对齐困难;缺乏有效的融合异构知识来源的机制。虽然多模态大型语言模型(Multimodal Large Language Models, mllm)在视觉理解任务中表现出了较强的性能,但在知识利用效率和语义对齐方面面临着显著的挑战,严重限制了推理深度和鲁棒性。针对这些问题,本文提出了一种基于知识的VQA的情境感知知识构建与检索方法,该方法包括以下三个模块。多粒度知识检索模块基于图像和问题的多维嵌入表示构建联合查询向量,准确获取与上下文高度匹配的显式知识。视觉到知识生成模块从视觉内容的角度补充细粒度的语义线索,生成与图像密切相关的视觉知识,弥补一般知识的表达局限性。为实现知识表示的深度对齐,知识自适应学习模块通过引入可学习的知识映射机制,将多源知识精确嵌入到MLLM的语义空间中。在OK-VQA和A-OKVQA数据集上的实验评估表明,CKCR优于同尺度的最先进方法。消融实验和可视化分析证明了CKCR在细粒度视觉信息感知和知识语义对齐能力方面的优势。我们的代码将在GitHub上发布:https://github.com/fjwang3/CKCR。
{"title":"CKCR: Context-aware knowledge construction and retrieval for knowledge-based visual question answering","authors":"Fengjuan Wang ,&nbsp;Jiayi Liu ,&nbsp;Ruonan Zhang,&nbsp;Zhengxue Li,&nbsp;Feng Zhang,&nbsp;Gaoyun An","doi":"10.1016/j.jvcir.2026.104711","DOIUrl":"10.1016/j.jvcir.2026.104711","url":null,"abstract":"<div><div>Knowledge-based Visual Question Answering (KB-VQA) requires models to integrate visual content with external knowledge to answer questions, which is crucial for building intelligent systems capable of real-world understanding. However, effectively incorporating external knowledge into visual reasoning faces three major challenges: the incompleteness of external knowledge bases leads to missing knowledge for many specific visual scenarios, semantic gaps exist between retrieved textual knowledge and visual content making alignment difficult, and effective mechanisms for fusing heterogeneous knowledge sources are lacking. While Multimodal Large Language Models(MLLMs) have demonstrated strong performance in visual understanding tasks, but face notable challenges in KB-VQA, particularly in knowledge utilization efficiency and semantic alignment, which seriously limits the reasoning depth and robustness. To address these problems, a Context-aware Knowledge Construction and Retrieval (CKCR) method is proposed for knowledge-based VQA, which includes the following three modules. The multi-granularity knowledge retrieval module constructs joint query vector based on the multi-dimensional embedding representation of images and questions, accurately obtaining explicit knowledge that is highly matched with the context. The vision-to-knowledge generation module supplements fine-grained semantic clues from the perspective of visual content, generating visual knowledge closely related to the image and making up for the expression limitations of general knowledge. To achieve deep alignment of knowledge representation, the knowledge adaptive learning module accurately embeds multi-source knowledge into the semantic space of MLLM by introducing a learnable knowledge mapping mechanism. Experimental evaluation on OK-VQA and A-OKVQA dataset shows the CKCR outperforms state-of-the-art methods of the same-scale. Ablation experiments and visualization analysis demonstrate the superiority of CKCR in its perception of fine-grained visual information and its ability to align knowledge semantics. Our code will be released on GitHub: <span><span>https://github.com/fjwang3/CKCR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104711"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image copy-move forgery detection using three-stage matching with constraints 基于约束的三级匹配图像复制-移动伪造检测
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-08 DOI: 10.1016/j.jvcir.2026.104709
Panpan Niu, Hongxin Wang, Xingqi Wang
Copy-move forgery is one of the most commonly used manipulations for tampering digital images. In recent years, keypoint-based detection methods have achieved encouraging results, but there are still several shortcomings that can be improved. First, unability to generate sufficient keypoints in small or smooth regions, causing detection failure. Second, lack of robust and discriminative descriptors for image keypoints, resulting in false matches. Third, high computational cost of image keypoints matching. To tackle this challenge, we present a new keypoint-based image copy-move forgery detection (CMFD) using three-stage matching with constraints. In keypoint extraction, we extract sufficient SIFT keypoints by adaptively enlarging image and enhancing image contrast. In feature description, we adopt the combination of complex and real values of Polar Harmonic Fourier Moments (PHFMs) as the PHFMs-based hybrid feature vector of each keypoint, which substantially enhances the differentiation of the features. In feature matching, we present a fast stratification approach based on SLIC and locally optimal orientation pattern (LOOP), and utilize the stratification results as the constraints of matching, which can reduce the search space. Then a high-precision three-stage matching strategy based on amplitude information, phase information and distance information is executed. In post-processing, the location of the tampered regions is finally determined by one-step filtering and one-step clustering. Extensive experimental results show the superiority of the proposed method over the existing representative CMFD techniques.
复制-移动伪造是篡改数字图像最常用的手法之一。近年来,基于关键点的检测方法取得了令人鼓舞的成果,但仍有一些不足之处需要改进。首先,无法在小区域或光滑区域生成足够的关键点,导致检测失败。其次,对图像关键点缺乏鲁棒性和判别性的描述符,导致匹配错误。第三,图像关键点匹配的计算成本高。为了解决这一挑战,我们提出了一种新的基于关键点的图像复制-移动伪造检测(CMFD),该检测使用带有约束的三阶段匹配。在关键点提取方面,我们通过自适应放大图像和增强图像对比度来提取足够的SIFT关键点。在特征描述中,我们采用极调和傅里叶矩(PHFMs)的复值与实值的组合作为每个关键点的基于PHFMs的混合特征向量,大大增强了特征的差异性。在特征匹配中,我们提出了一种基于SLIC和局部最优方向模式(LOOP)的快速分层方法,并利用分层结果作为匹配约束,减小了搜索空间。然后执行基于幅值信息、相位信息和距离信息的高精度三级匹配策略。在后处理中,通过一步滤波和一步聚类最终确定篡改区域的位置。大量的实验结果表明,该方法优于现有的具有代表性的CMFD技术。
{"title":"Image copy-move forgery detection using three-stage matching with constraints","authors":"Panpan Niu,&nbsp;Hongxin Wang,&nbsp;Xingqi Wang","doi":"10.1016/j.jvcir.2026.104709","DOIUrl":"10.1016/j.jvcir.2026.104709","url":null,"abstract":"<div><div>Copy-move forgery is one of the most commonly used manipulations for tampering digital images. In recent years, keypoint-based detection methods have achieved encouraging results, but there are still several shortcomings that can be improved. First, unability to generate sufficient keypoints in small or smooth regions, causing detection failure. Second, lack of robust and discriminative descriptors for image keypoints, resulting in false matches. Third, high computational cost of image keypoints matching. To tackle this challenge, we present a new keypoint-based image copy-move forgery detection (CMFD) using three-stage matching with constraints. In keypoint extraction, we extract sufficient SIFT keypoints by adaptively enlarging image and enhancing image contrast. In feature description, we adopt the combination of complex and real values of Polar Harmonic Fourier Moments (PHFMs) as the PHFMs-based hybrid feature vector of each keypoint, which substantially enhances the differentiation of the features. In feature matching, we present a fast stratification approach based on SLIC and locally optimal orientation pattern (LOOP), and utilize the stratification results as the constraints of matching, which can reduce the search space. Then a high-precision three-stage matching strategy based on amplitude information, phase information and distance information is executed. In post-processing, the location of the tampered regions is finally determined by one-step filtering and one-step clustering. Extensive experimental results show the superiority of the proposed method over the existing representative CMFD techniques.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104709"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145950136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
eGoRG: GPU-accelerated depth estimation for immersive video applications based on graph cuts 基于图形切割的沉浸式视频应用的gpu加速深度估计
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-19 DOI: 10.1016/j.jvcir.2026.104727
Jaime Sancho , Manuel Villa , Miguel Chavarrias , Rubén Salvador , Eduardo Juarez , César Sanz
Immersive video is gaining relevance across various fields, but its integration into real applications remains limited due to the technical challenges of depth estimation. Generating accurate depth maps is essential for 3D rendering, yet high-quality algorithms can require hundreds of seconds to produce a single frame. While real-time depth estimation solutions exist — particularly monocular deep learning-based methods and active sensors such as time-of-flight or plenoptic cameras — their depth accuracy and multiview consistency are often insufficient for depth image-based rendering (DIBR) and immersive video applications. This highlights the persistent challenge of jointly achieving real-time performance and high-quality, correlated depth across views. This paper introduces eGoRG, a GPU-accelerated depth estimation algorithm based on MPEG DERS, which employs graph cuts to achieve high-quality results. eGoRG contributes a novel GPU-based graph cuts stage, integrating block-based push-relabel acceleration and a simplified alpha expansion method. These optimizations deliver quality comparable to leading graph-cut approaches while greatly improving speed. Evaluation on an MPEG multiview dataset and a static NeRF dataset demonstrates the algorithm’s effectiveness across different scenarios.
沉浸式视频正在各个领域获得相关性,但由于深度估计的技术挑战,其与实际应用的集成仍然有限。生成精确的深度图对于3D渲染至关重要,然而高质量的算法可能需要数百秒才能生成单个帧。虽然存在实时深度估计解决方案,特别是基于单目深度学习的方法和主动传感器,如飞行时间或全光学相机,但它们的深度精度和多视图一致性通常不足以用于深度图像渲染(DIBR)和沉浸式视频应用。这突出了共同实现实时性能和高质量、跨视图相关深度的持续挑战。本文介绍了一种基于MPEG - DERS的gpu加速深度估计算法eGoRG,该算法利用图切来获得高质量的深度估计结果。eGoRG提供了一种新的基于gpu的图切割阶段,集成了基于块的推标签加速和简化的alpha展开方法。这些优化提供了与领先的图切割方法相当的质量,同时大大提高了速度。对MPEG多视图数据集和静态NeRF数据集的评估证明了该算法在不同场景下的有效性。
{"title":"eGoRG: GPU-accelerated depth estimation for immersive video applications based on graph cuts","authors":"Jaime Sancho ,&nbsp;Manuel Villa ,&nbsp;Miguel Chavarrias ,&nbsp;Rubén Salvador ,&nbsp;Eduardo Juarez ,&nbsp;César Sanz","doi":"10.1016/j.jvcir.2026.104727","DOIUrl":"10.1016/j.jvcir.2026.104727","url":null,"abstract":"<div><div>Immersive video is gaining relevance across various fields, but its integration into real applications remains limited due to the technical challenges of depth estimation. Generating accurate depth maps is essential for 3D rendering, yet high-quality algorithms can require hundreds of seconds to produce a single frame. While real-time depth estimation solutions exist — particularly monocular deep learning-based methods and active sensors such as time-of-flight or plenoptic cameras — their depth accuracy and multiview consistency are often insufficient for depth image-based rendering (DIBR) and immersive video applications. This highlights the persistent challenge of jointly achieving real-time performance and high-quality, correlated depth across views. This paper introduces eGoRG, a GPU-accelerated depth estimation algorithm based on MPEG DERS, which employs graph cuts to achieve high-quality results. eGoRG contributes a novel GPU-based graph cuts stage, integrating block-based push-relabel acceleration and a simplified alpha expansion method. These optimizations deliver quality comparable to leading graph-cut approaches while greatly improving speed. Evaluation on an MPEG multiview dataset and a static NeRF dataset demonstrates the algorithm’s effectiveness across different scenarios.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104727"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146024843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NCC-FDM: Frequency-domain diffusion model driven by non-physical-domain color correction for underwater image enhancement NCC-FDM:非物理域色彩校正驱动的频域扩散模型用于水下图像增强
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-02-09 DOI: 10.1016/j.jvcir.2026.104740
Guanglin Qu, Xiuman Liang, Zhendong Liu, Haifeng Yu
Underwater environments have the characteristics of light absorption and scattering. Images captured in such environments commonly suffer from multiple degradation issues such as color bias, haze, detail loss and low contrast, which further severely interfere with downstream underwater vision tasks. To addressing the challenges of underwater image enhancement, we propose a non-physical color correction driven frequency domain diffusion model (NCC-FDM). The model combines non-physical color correction with a conditional diffusion model. Firstly, we design a non-physical color correction stage (NCCS) to rapidly address severe color shifts in underwater datasets. Image color deviations are corrected through the combined application of rapid optical wave compensation and the gray scale world method. Secondly, we consider the different degrees of degradation between high frequency images and low-frequency images in underwater images. We design a frequency domain conditional diffusion model based on discrete wavelet transform to process the low-frequency components of color-corrected images. A hybrid high-frequency enhancement module (HHEM) is proposed to restore detail and structural information in our images. The module separately enhances the high-frequency components of different spatial dimensions. The enhancement is based on the principle that the proportion of information and noise contained in the high-frequency components varies across different spatial dimensions. Finally, we design a joint loss function to optimize the frequency domain diffusion model. The joint loss function includes noise loss, reconstruction loss and high-frequency loss. Comprehensive evaluation across four public underwater datasets demonstrates that the proposed NCC-FDM algorithm achieves outstanding performance in both visual quality and evaluation metrics.
水下环境具有光的吸收和散射特性。在这种环境中捕获的图像通常会遭受多种退化问题,如色彩偏差、雾霾、细节丢失和低对比度,这进一步严重干扰下游的水下视觉任务。为了解决水下图像增强的挑战,我们提出了一种非物理色彩校正驱动的频域扩散模型(NCC-FDM)。该模型结合了非物理色彩校正和条件扩散模型。首先,我们设计了一个非物理色彩校正阶段(NCCS)来快速解决水下数据集中严重的色彩偏移。采用快速光波补偿和灰度世界法相结合的方法对图像颜色偏差进行了校正。其次,我们考虑了水下图像中高频图像和低频图像的不同退化程度。设计了一种基于离散小波变换的频域条件扩散模型来处理彩色校正图像的低频分量。提出了一种混合高频增强模块(HHEM)来恢复图像中的细节和结构信息。该模块分别增强不同空间维度的高频分量。增强的原理是高频分量中包含的信息和噪声的比例在不同的空间维度上是不同的。最后,我们设计了一个联合损失函数来优化频域扩散模型。联合损失函数包括噪声损失、重构损失和高频损失。对四个公开水下数据集的综合评估表明,所提出的NCC-FDM算法在视觉质量和评估指标方面都取得了出色的性能。
{"title":"NCC-FDM: Frequency-domain diffusion model driven by non-physical-domain color correction for underwater image enhancement","authors":"Guanglin Qu,&nbsp;Xiuman Liang,&nbsp;Zhendong Liu,&nbsp;Haifeng Yu","doi":"10.1016/j.jvcir.2026.104740","DOIUrl":"10.1016/j.jvcir.2026.104740","url":null,"abstract":"<div><div>Underwater environments have the characteristics of light absorption and scattering. Images captured in such environments commonly suffer from multiple degradation issues such as color bias, haze, detail loss and low contrast, which further severely interfere with downstream underwater vision tasks. To addressing the challenges of underwater image enhancement, we propose a non-physical color correction driven frequency domain diffusion model (NCC-FDM). The model combines non-physical color correction with a conditional diffusion model. Firstly, we design a non-physical color correction stage (NCCS) to rapidly address severe color shifts in underwater datasets. Image color deviations are corrected through the combined application of rapid optical wave compensation and the gray scale world method. Secondly, we consider the different degrees of degradation between high frequency images and low-frequency images in underwater images. We design a frequency domain conditional diffusion model based on discrete wavelet transform to process the low-frequency components of color-corrected images. A hybrid high-frequency enhancement module (HHEM) is proposed to restore detail and structural information in our images. The module separately enhances the high-frequency components of different spatial dimensions. The enhancement is based on the principle that the proportion of information and noise contained in the high-frequency components varies across different spatial dimensions. Finally, we design a joint loss function to optimize the frequency domain diffusion model. The joint loss function includes noise loss, reconstruction loss and high-frequency loss. Comprehensive evaluation across four public underwater datasets demonstrates that the proposed NCC-FDM algorithm achieves outstanding performance in both visual quality and evaluation metrics.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104740"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Global–local dual-branch network with local feature enhancement for visual tracking 基于局部特征增强的全局-局部双分支网络视觉跟踪
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-23 DOI: 10.1016/j.jvcir.2026.104725
Yuanyun Wang, Lingtao Zhou, Zhuo An, Lei Sun, Min Hu, Jun Wang
Vision Transformers (ViT) have been widely applied due to their excellent performance. Compared with CNN models, ViT models are more difficult to train and require more training samples because they cannot effectively utilize high-frequency local information. In this paper we propose an efficient tracking framework based on global and local feature extraction, and an enhancement module. To address the high-frequency local information neglected by general ViT-based trackers, we design an effective local branch architecture to capture the information. For local feature extraction and enhancement, we design a local branch, which aggregates local information by using shared weights; it utilizes the optimized context-aware weights to enhance the local features. The integration of the attention mechanism in the global and local branches enables the tracker to perceive both high-frequency local information and low-frequency global information simultaneously. Experimental comparisons show that the tracker achieves superior results and proves the generalization ability and effectiveness. Code will be available at https://github.com/WangJun-CV/GLDTrack.
视觉变压器以其优异的性能得到了广泛的应用。与CNN模型相比,ViT模型由于不能有效利用高频局部信息,训练难度更大,需要更多的训练样本。本文提出了一种基于全局和局部特征提取的高效跟踪框架和增强模块。为了解决一般基于vit的跟踪器所忽略的高频局部信息,我们设计了一个有效的局部分支架构来捕获这些信息。在局部特征提取和增强方面,设计了局部分支,利用共享权值对局部信息进行聚合;它利用优化的上下文感知权重来增强局部特征。全局分支和局部分支的注意机制集成,使跟踪器能够同时感知高频局部信息和低频全局信息。实验对比表明,该跟踪器取得了较好的效果,证明了该跟踪器的泛化能力和有效性。代码将在https://github.com/WangJun-CV/GLDTrack上提供。
{"title":"Global–local dual-branch network with local feature enhancement for visual tracking","authors":"Yuanyun Wang,&nbsp;Lingtao Zhou,&nbsp;Zhuo An,&nbsp;Lei Sun,&nbsp;Min Hu,&nbsp;Jun Wang","doi":"10.1016/j.jvcir.2026.104725","DOIUrl":"10.1016/j.jvcir.2026.104725","url":null,"abstract":"<div><div>Vision Transformers (ViT) have been widely applied due to their excellent performance. Compared with CNN models, ViT models are more difficult to train and require more training samples because they cannot effectively utilize high-frequency local information. In this paper we propose an efficient tracking framework based on global and local feature extraction, and an enhancement module. To address the high-frequency local information neglected by general ViT-based trackers, we design an effective local branch architecture to capture the information. For local feature extraction and enhancement, we design a local branch, which aggregates local information by using shared weights; it utilizes the optimized context-aware weights to enhance the local features. The integration of the attention mechanism in the global and local branches enables the tracker to perceive both high-frequency local information and low-frequency global information simultaneously. Experimental comparisons show that the tracker achieves superior results and proves the generalization ability and effectiveness. Code will be available at <span><span>https://github.com/WangJun-CV/GLDTrack</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104725"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146079536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
All-in-focus image fusion using graph wavelet transform for multi-modal light field 基于图小波变换的多模态光场全焦图像融合
IF 3.1 4区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-03-01 Epub Date: 2026-01-15 DOI: 10.1016/j.jvcir.2026.104722
Jinjin Li , Baiyuan Qing , Kun Zhang , Xinyuan Yang , Xiangui Yin , Yichang Liu
The multi-modal nature of light field imaging produces a refocused image stack, but each image suffers from a limited depth-of-field. All-in-focus (AIF) fusion aims to create a single, sharp image from this stack, a task challenged by irregular depth boundaries and degraded spatial resolution. We propose a novel fusion framework based on the graph wavelet transform (GWT). Unlike traditional methods, our approach adaptively models pixel correlations to better handle irregular boundaries while preserving details. The method decomposes each image using a fast GWT. Low-frequency components are fused via a multi-layer strategy, while high-frequency components are merged using an integrated weighting scheme enhanced by guided filtering. Finally, the AIF image is reconstructed via an inverse GWT. Experimental results on light field datasets demonstrate superior performance over existing methods, achieving average EI, QY, and SSIM scores of 44.939, 0.9941, and 0.8719, respectively, showing its potential for practical applications.
光场成像的多模态特性产生了重新聚焦的图像堆栈,但是每个图像都受到景深的限制。全聚焦(AIF)融合旨在从这些叠加中创建一个单一的、清晰的图像,这是一项受到不规则深度边界和空间分辨率下降的挑战的任务。提出了一种基于图小波变换(GWT)的融合框架。与传统方法不同,我们的方法自适应建模像素相关性,在保留细节的同时更好地处理不规则边界。该方法使用快速GWT分解每个图像。低频分量通过多层融合策略融合,高频分量通过引导滤波增强的综合加权方案融合。最后,通过逆GWT重构AIF图像。在光场数据集上的实验结果表明,该方法优于现有方法,平均EI、QY和SSIM得分分别为44.939、0.9941和0.8719,显示了该方法的实际应用潜力。
{"title":"All-in-focus image fusion using graph wavelet transform for multi-modal light field","authors":"Jinjin Li ,&nbsp;Baiyuan Qing ,&nbsp;Kun Zhang ,&nbsp;Xinyuan Yang ,&nbsp;Xiangui Yin ,&nbsp;Yichang Liu","doi":"10.1016/j.jvcir.2026.104722","DOIUrl":"10.1016/j.jvcir.2026.104722","url":null,"abstract":"<div><div>The multi-modal nature of light field imaging produces a refocused image stack, but each image suffers from a limited depth-of-field. All-in-focus (AIF) fusion aims to create a single, sharp image from this stack, a task challenged by irregular depth boundaries and degraded spatial resolution. We propose a novel fusion framework based on the graph wavelet transform (GWT). Unlike traditional methods, our approach adaptively models pixel correlations to better handle irregular boundaries while preserving details. The method decomposes each image using a fast GWT. Low-frequency components are fused via a multi-layer strategy, while high-frequency components are merged using an integrated weighting scheme enhanced by guided filtering. Finally, the AIF image is reconstructed via an inverse GWT. Experimental results on light field datasets demonstrate superior performance over existing methods, achieving average EI, <span><math><msub><mrow><mi>Q</mi></mrow><mrow><mi>Y</mi></mrow></msub></math></span>, and SSIM scores of 44.939, 0.9941, and 0.8719, respectively, showing its potential for practical applications.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"116 ","pages":"Article 104722"},"PeriodicalIF":3.1,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Visual Communication and Image Representation
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1