首页 > 最新文献

Computer Vision and Image Understanding最新文献

英文 中文
Multi-domain awareness for compressed deepfake videos detection over social networks guided by common mechanisms between artifacts 以人工制品之间的共同机制为指导,利用多域感知技术检测社交网络上的压缩深度伪造视频
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-10 DOI: 10.1016/j.cviu.2024.104072

The viral spread of massive deepfake videos over social networks has caused serious security problems. Despite the remarkable advancements achieved by existing deepfake detection algorithms, deepfake videos over social networks are inevitably influenced by compression factors. This causes deepfake detection performance to be limited by the following challenging issues: (a) interfering with compression artifacts, (b) loss of feature information, and (c) aliasing of feature distributions. In this paper, we analyze the common mechanism between compression artifacts and deepfake artifacts, revealing the structural similarity between them and providing a reliable theoretical basis for enhancing the robustness of deepfake detection models against compression. Firstly, based on the common mechanism between artifacts, we design a frequency domain adaptive notch filter to eliminate the interference of compression artifacts on specific frequency bands. Secondly, to reduce the sensitivity of deepfake detection models to unknown noise, we propose a spatial residual denoising strategy. Thirdly, to exploit the intrinsic correlation between feature vectors in the frequency domain branch and the spatial domain branch, we enhance deepfake features using an attention-based feature fusion method. Finally, we adopt a multi-task decision approach to enhance the discriminative power of the latent space representation of deepfakes, achieving deepfake detection with robustness against compression. Extensive experiments show that compared with the baseline methods, the detection performance of the proposed algorithm on compressed deepfake videos has been significantly improved. In particular, our model is resistant to various types of noise disturbances and can be easily combined with baseline detection models to improve their robustness.

社交网络上大量深度伪造视频的病毒式传播引发了严重的安全问题。尽管现有的深度伪造检测算法取得了显著进步,但社交网络上的深度伪造视频不可避免地受到压缩因素的影响。这导致深度伪造检测性能受到以下挑战性问题的限制:(a)压缩伪影的干扰,(b)特征信息的丢失,以及(c)特征分布的混叠。本文分析了压缩伪影与深度伪影的共同机理,揭示了二者在结构上的相似性,为增强深度伪影检测模型对抗压缩的鲁棒性提供了可靠的理论依据。首先,基于伪影之间的共同机制,我们设计了一种频域自适应陷波滤波器来消除压缩伪影对特定频段的干扰。其次,为了降低深度伪造检测模型对未知噪声的敏感性,我们提出了一种空间残差去噪策略。第三,为了利用频域分支和空间域分支中特征向量之间的内在相关性,我们采用了一种基于注意力的特征融合方法来增强深度伪造特征。最后,我们采用了一种多任务决策方法来增强潜空间表征对深度伪造的判别能力,从而实现了具有抗压缩鲁棒性的深度伪造检测。大量实验表明,与基线方法相比,所提算法在压缩 deepfake 视频上的检测性能有了显著提高。特别是,我们的模型能抵御各种类型的噪声干扰,并能轻松地与基线检测模型相结合,以提高其鲁棒性。
{"title":"Multi-domain awareness for compressed deepfake videos detection over social networks guided by common mechanisms between artifacts","authors":"","doi":"10.1016/j.cviu.2024.104072","DOIUrl":"10.1016/j.cviu.2024.104072","url":null,"abstract":"<div><p>The viral spread of massive deepfake videos over social networks has caused serious security problems. Despite the remarkable advancements achieved by existing deepfake detection algorithms, deepfake videos over social networks are inevitably influenced by compression factors. This causes deepfake detection performance to be limited by the following challenging issues: (a) interfering with compression artifacts, (b) loss of feature information, and (c) aliasing of feature distributions. In this paper, we analyze the common mechanism between compression artifacts and deepfake artifacts, revealing the structural similarity between them and providing a reliable theoretical basis for enhancing the robustness of deepfake detection models against compression. Firstly, based on the common mechanism between artifacts, we design a frequency domain adaptive notch filter to eliminate the interference of compression artifacts on specific frequency bands. Secondly, to reduce the sensitivity of deepfake detection models to unknown noise, we propose a spatial residual denoising strategy. Thirdly, to exploit the intrinsic correlation between feature vectors in the frequency domain branch and the spatial domain branch, we enhance deepfake features using an attention-based feature fusion method. Finally, we adopt a multi-task decision approach to enhance the discriminative power of the latent space representation of deepfakes, achieving deepfake detection with robustness against compression. Extensive experiments show that compared with the baseline methods, the detection performance of the proposed algorithm on compressed deepfake videos has been significantly improved. In particular, our model is resistant to various types of noise disturbances and can be easily combined with baseline detection models to improve their robustness.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141715999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision and Structured-Language Pretraining for Cross-Modal Food Retrieval 跨模态食物检索的视觉和结构化语言预训练
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-09 DOI: 10.1016/j.cviu.2024.104071

Vision-Language Pretraining (VLP) and Foundation models have been the go-to recipe for achieving SoTA performance on general benchmarks. However, leveraging these powerful techniques for more complex vision-language tasks, such as cooking applications, with more structured input data, is still little investigated. In this work, we propose to leverage these techniques for structured-text based computational cuisine tasks. Our strategy, dubbed VLPCook, first transforms existing image-text pairs to image and structured-text pairs. This allows to pretrain our VLPCook model using VLP objectives adapted to the structured data of the resulting datasets, then finetuning it on downstream computational cooking tasks. During finetuning, we also enrich the visual encoder, leveraging pretrained foundation models (e.g. CLIP) to provide local and global textual context. VLPCook outperforms current SoTA by a significant margin (+3.3 Recall@1 absolute improvement) on the task of Cross-Modal Food Retrieval on the large Recipe1M dataset. We conduct further experiments on VLP to validate their importance, especially on the Recipe1M+ dataset. Finally, we validate the generalization of the approach to other tasks (i.e, Food Recognition) and domains with structured text such as the Medical domain on the ROCO dataset. The code will be made publicly available.

视觉语言预训练(VLP)和基础模型一直是在一般基准上实现 SoTA 性能的常用方法。然而,利用这些强大的技术来完成更复杂的视觉语言任务(如烹饪应用)以及结构化程度更高的输入数据的研究仍然很少。在这项工作中,我们建议将这些技术用于基于结构化文本的计算烹饪任务。我们的策略被称为 VLPCook,首先将现有的图像-文本对转换为图像和结构化文本对。这样,我们就可以使用 VLP 目标对 VLPCook 模型进行预训练,以适应由此产生的数据集的结构化数据,然后在下游计算烹饪任务中对其进行微调。在微调过程中,我们还丰富了视觉编码器,利用预训练的基础模型(如 CLIP)提供局部和全局文本上下文。在大型 Recipe1M 数据集的跨模态食物检索任务中,VLPCook 的表现明显优于当前的 SoTA(+3.3 Recall@1 absolute improvement)。我们对 VLP 进行了进一步实验,以验证其重要性,尤其是在 Recipe1M+ 数据集上。最后,我们在 ROCO 数据集上验证了该方法在其他任务(即食品识别)和具有结构化文本的领域(如医疗领域)中的通用性。代码将公开发布。
{"title":"Vision and Structured-Language Pretraining for Cross-Modal Food Retrieval","authors":"","doi":"10.1016/j.cviu.2024.104071","DOIUrl":"10.1016/j.cviu.2024.104071","url":null,"abstract":"<div><p>Vision-Language Pretraining (VLP) and Foundation models have been the go-to recipe for achieving SoTA performance on general benchmarks. However, leveraging these powerful techniques for more complex vision-language tasks, such as cooking applications, with more structured input data, is still little investigated. In this work, we propose to leverage these techniques for structured-text based computational cuisine tasks. Our strategy, dubbed VLPCook, first transforms existing image-text pairs to image and structured-text pairs. This allows to pretrain our VLPCook model using VLP objectives adapted to the structured data of the resulting datasets, then finetuning it on downstream computational cooking tasks. During finetuning, we also enrich the visual encoder, leveraging pretrained foundation models (<em>e.g.</em> CLIP) to provide local and global textual context. VLPCook outperforms current SoTA by a significant margin (+3.3 Recall@1 absolute improvement) on the task of Cross-Modal Food Retrieval on the large Recipe1M dataset. We conduct further experiments on VLP to validate their importance, especially on the Recipe1M+ dataset. Finally, we validate the generalization of the approach to other tasks (<em>i.e</em>, Food Recognition) and domains with structured text such as the Medical domain on the ROCO dataset. The code will be made publicly available.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141962288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modality adaptation via feature difference learning for depth human parsing 通过特征差异学习进行深度人类解析的模态适应
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-08 DOI: 10.1016/j.cviu.2024.104070

In the field of human parsing, depth data offers unique advantages over RGB data due to its illumination invariance and geometric detail, which motivates us to explore human parsing with only depth input. However, depth data is challenging to collect at scale due to the specialized equipment required. In contrast, RGB data is readily available in large quantities, presenting an opportunity to enhance depth-only parsing models with semantic knowledge learned from RGB data. However, fully finetuning the RGB-pretrained encoder leads to high training costs and inflexible domain generalization, while keeping the encoder frozen suffers from a large RGB-depth modality gap and restricts the parsing performance. To alleviate the limitations of these naive approaches, we introduce a Modality Adaptation pipeline via Feature Difference Learning (MAFDL) which leverages the RGB knowledge to facilitate depth human parsing. A Difference-Guided Depth Adapter (DGDA) is proposed within MAFDL to learn the feature differences between RGB and depth modalities, adapting depth features into RGB feature space to bridge the modality gap. Furthermore, we also design a Feature Alignment Constraint (FAC) to impose explicit alignment supervision at pixel and batch levels, making the modality adaptation more comprehensive. Extensive experiments on the NTURGBD-Parsing-4K dataset show that our method surpasses previous state-of-the-art approaches.

在人类解析领域,深度数据因其光照不变性和几何细节而比 RGB 数据具有独特的优势,这促使我们探索仅使用深度输入进行人类解析的方法。然而,由于需要专业设备,深度数据的大规模收集具有挑战性。相比之下,RGB 数据则很容易大量获得,这为利用从 RGB 数据中学到的语义知识来增强纯深度解析模型提供了机会。然而,对 RGB 预训练编码器进行完全微调会导致高昂的训练成本和不灵活的领域泛化,而保持编码器冻结则会造成巨大的 RGB 深度模态差距,并限制解析性能。为了缓解这些幼稚方法的局限性,我们引入了通过特征差分学习(MAFDL)进行模态适应的管道,利用 RGB 知识促进深度人类解析。我们在 MAFDL 中提出了差异引导深度适配器 (DGDA),用于学习 RGB 和深度模态之间的特征差异,将深度特征适配到 RGB 特征空间中,以弥合模态差距。此外,我们还设计了特征对齐约束 (FAC),在像素和批次级别实施明确的对齐监督,使模态适应更加全面。在 NTURGBD-Parsing-4K 数据集上进行的广泛实验表明,我们的方法超越了以前的先进方法。
{"title":"Modality adaptation via feature difference learning for depth human parsing","authors":"","doi":"10.1016/j.cviu.2024.104070","DOIUrl":"10.1016/j.cviu.2024.104070","url":null,"abstract":"<div><p>In the field of human parsing, depth data offers unique advantages over RGB data due to its illumination invariance and geometric detail, which motivates us to explore human parsing with only depth input. However, depth data is challenging to collect at scale due to the specialized equipment required. In contrast, RGB data is readily available in large quantities, presenting an opportunity to enhance depth-only parsing models with semantic knowledge learned from RGB data. However, fully finetuning the RGB-pretrained encoder leads to high training costs and inflexible domain generalization, while keeping the encoder frozen suffers from a large RGB-depth modality gap and restricts the parsing performance. To alleviate the limitations of these naive approaches, we introduce a Modality Adaptation pipeline via Feature Difference Learning (MAFDL) which leverages the RGB knowledge to facilitate depth human parsing. A Difference-Guided Depth Adapter (DGDA) is proposed within MAFDL to learn the feature differences between RGB and depth modalities, adapting depth features into RGB feature space to bridge the modality gap. Furthermore, we also design a Feature Alignment Constraint (FAC) to impose explicit alignment supervision at pixel and batch levels, making the modality adaptation more comprehensive. Extensive experiments on the NTURGBD-Parsing-4K dataset show that our method surpasses previous state-of-the-art approaches.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141637936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implicit and explicit commonsense for multi-sentence video captioning 多句式视频字幕的隐性和显性常识
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-05 DOI: 10.1016/j.cviu.2024.104064

Existing dense or paragraph video captioning approaches rely on holistic representations of videos, possibly coupled with learned object/action representations, to condition hierarchical language decoders. However, they fundamentally lack the commonsense knowledge of the world required to reason about progression of events, causality, and even the function of certain objects within a scene. To address this limitation we propose a novel video captioning Transformer-based model, that takes into account both implicit (visuo-lingual and purely linguistic) and explicit (knowledge-base) commonsense knowledge. We show that these forms of knowledge, in isolation and in combination, enhance the quality of produced captions. Further, inspired by imitation learning, we propose a new task of instruction generation, where the goal is to produce a set of linguistic instructions from a video demonstration of its performance. We formalize the task using the ALFRED dataset generated using an AI2-THOR environment. While instruction generation is conceptually similar to paragraph captioning, it differs in the fact that it exhibits stronger object persistence, as well as spatially-aware and causal sentence structure. We show that our commonsense knowledge enhanced approach produces significant improvements on this task (up to 57% in METEOR and 8.5% in CIDEr), as well as the state-of-the-art result on more traditional video captioning in the ActivityNet Captions dataset.

现有的密集或段落视频字幕方法依赖于视频的整体表征,并可能与学习到的物体/动作表征相结合,为分层语言解码器提供条件。然而,这些方法从根本上缺乏推理事件进展、因果关系,甚至场景中某些物体的功能所需的常识性知识。为了解决这一局限性,我们提出了一种基于转换器的新型视频字幕模型,该模型同时考虑了隐性(视觉语言和纯语言)和显性(知识库)常识知识。我们的研究表明,这些知识形式无论是单独使用还是结合使用,都能提高字幕的质量。此外,受模仿学习的启发,我们提出了一项新的指令生成任务,其目标是从视频演示中生成一组语言指令。我们利用在 AI2-THOR 环境中生成的 ALFRED 数据集正式确定了这一任务。虽然指令生成在概念上与段落标题相似,但其不同之处在于它表现出更强的对象持久性,以及空间感知和因果关系句子结构。我们的研究表明,我们的常识性知识增强方法在这项任务中取得了显著的改进(在 METEOR 中达到了 57%,在 CIDEr 中达到了 8.5%),在 ActivityNet Captions 数据集中的更传统的视频字幕上也取得了最先进的结果。
{"title":"Implicit and explicit commonsense for multi-sentence video captioning","authors":"","doi":"10.1016/j.cviu.2024.104064","DOIUrl":"10.1016/j.cviu.2024.104064","url":null,"abstract":"<div><p>Existing dense or paragraph video captioning approaches rely on holistic representations of videos, possibly coupled with learned object/action representations, to condition hierarchical language decoders. However, they fundamentally lack the commonsense knowledge of the world required to reason about progression of events, causality, and even the function of certain objects within a scene. To address this limitation we propose a novel video captioning Transformer-based model, that takes into account both implicit (visuo-lingual and purely linguistic) and explicit (knowledge-base) commonsense knowledge. We show that these forms of knowledge, in isolation and in combination, enhance the quality of produced captions. Further, inspired by imitation learning, we propose a new task of instruction generation, where the goal is to produce a set of linguistic instructions from a video demonstration of its performance. We formalize the task using the ALFRED dataset generated using an AI2-THOR environment. While instruction generation is conceptually similar to paragraph captioning, it differs in the fact that it exhibits stronger object persistence, as well as spatially-aware and causal sentence structure. We show that our commonsense knowledge enhanced approach produces significant improvements on this task (up to 57% in METEOR and 8.5% in CIDEr), as well as the state-of-the-art result on more traditional video captioning in the ActivityNet Captions dataset.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1077314224001450/pdfft?md5=10aaeba9fc35361fc930cce1f7a01fe8&pid=1-s2.0-S1077314224001450-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141732463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Classroom teacher action recognition based on spatio-temporal dual-branch feature fusion 基于时空双分支特征融合的课堂教师动作识别
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-04 DOI: 10.1016/j.cviu.2024.104068
Di Wu , Jun Wang , Wei Zou , Shaodong Zou , Juxiang Zhou , Jianhou Gan

The classroom teaching action recognition task refers to recognizing and understanding teacher action through video temporal and spatial information. Due to complex backgrounds and significant occlusions, recognizing teacher action in the classroom environment poses substantial challenges. In this study, we propose a classroom teacher action recognition approach based on a spatio-temporal dual-branch feature fusion architecture, where the core task involves utilizing continuous human keypoint heatmap information and single-frame image information. Specifically, we fuse features from two modalities to propose a method combining image spatial information with temporal human keypoint heatmap information for teacher action recognition. Our approach ensures recognition accuracy while reducing the model’s parameters and computational complexity. Additionally, we constructed a teacher action dataset (CTA) in a real classroom environment, comprising 12 action categories, 13k+ video segments, and a total duration exceeding 15 h. The experimental results on the CTA dataset validate the effectiveness of our proposed method. Our research explores action recognition tasks in real complex classroom environments, providing a technical framework for classroom teaching intelligent analysis.

课堂教学动作识别任务是指通过视频的时间和空间信息来识别和理解教师的动作。由于背景复杂、遮挡严重,识别教室环境中的教师动作面临巨大挑战。在本研究中,我们提出了一种基于时空双分支特征融合架构的课堂教师动作识别方法,其核心任务是利用连续的人体关键点热图信息和单帧图像信息。具体来说,我们融合了两种模式的特征,提出了一种将图像空间信息与时间人类关键点热图信息相结合的方法,用于教师动作识别。我们的方法在降低模型参数和计算复杂度的同时,确保了识别的准确性。此外,我们还在真实教室环境中构建了一个教师动作数据集(CTA),其中包括 12 个动作类别、1300 多个视频片段,总时长超过 15 小时。我们的研究探索了真实复杂课堂环境中的动作识别任务,为课堂教学智能分析提供了一个技术框架。
{"title":"Classroom teacher action recognition based on spatio-temporal dual-branch feature fusion","authors":"Di Wu ,&nbsp;Jun Wang ,&nbsp;Wei Zou ,&nbsp;Shaodong Zou ,&nbsp;Juxiang Zhou ,&nbsp;Jianhou Gan","doi":"10.1016/j.cviu.2024.104068","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104068","url":null,"abstract":"<div><p>The classroom teaching action recognition task refers to recognizing and understanding teacher action through video temporal and spatial information. Due to complex backgrounds and significant occlusions, recognizing teacher action in the classroom environment poses substantial challenges. In this study, we propose a classroom teacher action recognition approach based on a spatio-temporal dual-branch feature fusion architecture, where the core task involves utilizing continuous human keypoint heatmap information and single-frame image information. Specifically, we fuse features from two modalities to propose a method combining image spatial information with temporal human keypoint heatmap information for teacher action recognition. Our approach ensures recognition accuracy while reducing the model’s parameters and computational complexity. Additionally, we constructed a teacher action dataset (CTA) in a real classroom environment, comprising 12 action categories, 13k+ video segments, and a total duration exceeding 15 h. The experimental results on the CTA dataset validate the effectiveness of our proposed method. Our research explores action recognition tasks in real complex classroom environments, providing a technical framework for classroom teaching intelligent analysis.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141596268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced dual contrast representation learning with cell separation and merging for breast cancer diagnosis 利用细胞分离与合并增强双重对比表征学习,用于乳腺癌诊断
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-02 DOI: 10.1016/j.cviu.2024.104065

Breast cancer remains a prevalent malignancy impacting a substantial number of individuals globally. In recent times, there has been a growing trend of combining deep learning methods with breast cancer diagnosis. Nevertheless, this integration encounters challenges, including limited data availability, class imbalance, and the absence of fine-grained labels to safeguard patient privacy and accommodate experience-dependent detection. To address these issues, we propose an effective framework by a dual contrast representation learning with a cell separation and merging strategy. The proposed algorithm comprises three main components: the cell separation and merging part, the dual contrast representation learning part, and the multi-category classification part. The cell separation and merging part takes an unpaired set of histopathological images as input and produces two sets of separated image layers, through the exploration of latent semantic information using SAM. Subsequently, these separated image layers are utilized to generate two new unpaired histopathological images via a cell separation and merging approach based on the linear superimposition model, with an inpainting network being employed to refine image details. Thus the class imbalance problem is alleviated and the data size is enlarged for a sufficient CNN training. The second part introduces a dual contrast representation learning framework for these generated images, with one branch designed for the positive samples (tumor cells) and the other for the negative samples (normal cells). The contrast learning network effectively minimizes the distance between two generated positive samples while maximizing the similarity of intra-class images to enhance feature representation. Leveraging the facilitated feature representation acquired from the dual contrast representation learning part, a pre-trained classifier is further fine-tuned to predict breast cancer categories. Extensive quantitative and qualitative experimental results validates the superiority of our proposed method compared to other state-of-the-art methods on the BreaKHis dataset in terms of four measurement metrics.

乳腺癌仍然是一种流行的恶性肿瘤,影响着全球相当多的人。近来,将深度学习方法与乳腺癌诊断相结合的趋势日益明显。然而,这种结合也遇到了一些挑战,包括数据可用性有限、类不平衡、缺乏细粒度标签以保护患者隐私和适应依赖经验的检测。为了解决这些问题,我们提出了一种有效的框架,即通过细胞分离与合并策略进行双重对比表示学习。所提出的算法包括三个主要部分:细胞分离与合并部分、双重对比度表征学习部分和多类别分类部分。细胞分离与合并部分以一组未配对的组织病理学图像为输入,通过使用 SAM 挖掘潜在语义信息,生成两组分离的图像层。随后,利用这些分离的图像层,通过基于线性叠加模型的细胞分离与合并方法,生成两幅新的未配对组织病理学图像,并利用内绘网络完善图像细节。因此,类不平衡问题得到了缓解,数据量也得到了扩大,从而实现了充分的 CNN 训练。第二部分针对这些生成的图像引入了双重对比度表示学习框架,其中一个分支针对阳性样本(肿瘤细胞),另一个分支针对阴性样本(正常细胞)。对比度学习网络能有效地最小化两个生成的阳性样本之间的距离,同时最大化类内图像的相似性,从而增强特征表示。利用双对比度表征学习部分获得的便利特征表征,对预训练分类器进行进一步微调,以预测乳腺癌类别。广泛的定量和定性实验结果验证了我们提出的方法在 BreaKHis 数据集上与其他先进方法相比在四个测量指标上的优越性。
{"title":"Enhanced dual contrast representation learning with cell separation and merging for breast cancer diagnosis","authors":"","doi":"10.1016/j.cviu.2024.104065","DOIUrl":"10.1016/j.cviu.2024.104065","url":null,"abstract":"<div><p>Breast cancer remains a prevalent malignancy impacting a substantial number of individuals globally. In recent times, there has been a growing trend of combining deep learning methods with breast cancer diagnosis. Nevertheless, this integration encounters challenges, including limited data availability, class imbalance, and the absence of fine-grained labels to safeguard patient privacy and accommodate experience-dependent detection. To address these issues, we propose an effective framework by a dual contrast representation learning with a cell separation and merging strategy. The proposed algorithm comprises three main components: the cell separation and merging part, the dual contrast representation learning part, and the multi-category classification part. The cell separation and merging part takes an unpaired set of histopathological images as input and produces two sets of separated image layers, through the exploration of latent semantic information using SAM. Subsequently, these separated image layers are utilized to generate two new unpaired histopathological images via a cell separation and merging approach based on the linear superimposition model, with an inpainting network being employed to refine image details. Thus the class imbalance problem is alleviated and the data size is enlarged for a sufficient CNN training. The second part introduces a dual contrast representation learning framework for these generated images, with one branch designed for the positive samples (tumor cells) and the other for the negative samples (normal cells). The contrast learning network effectively minimizes the distance between two generated positive samples while maximizing the similarity of intra-class images to enhance feature representation. Leveraging the facilitated feature representation acquired from the dual contrast representation learning part, a pre-trained classifier is further fine-tuned to predict breast cancer categories. Extensive quantitative and qualitative experimental results validates the superiority of our proposed method compared to other state-of-the-art methods on the BreaKHis dataset in terms of four measurement metrics.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141700981","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-label image classification using adaptive graph convolutional networks: From a single domain to multiple domains 利用自适应图卷积网络进行多标签图像分类:从单域到多域
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-07-01 DOI: 10.1016/j.cviu.2024.104062
Inder Pal Singh , Enjie Ghorbel , Oyebade Oyedotun , Djamila Aouada

This paper proposes an adaptive graph-based approach for multi-label image classification. Graph-based methods have been largely exploited in the field of multi-label classification, given their ability to model label correlations. Specifically, their effectiveness has been proven not only when considering a single domain but also when taking into account multiple domains. However, the topology of the used graph is not optimal as it is pre-defined heuristically. In addition, consecutive Graph Convolutional Network (GCN) aggregations tend to destroy the feature similarity. To overcome these issues, an architecture for learning the graph connectivity in an end-to-end fashion is introduced. This is done by integrating an attention-based mechanism and a similarity-preserving strategy. The proposed framework is then extended to multiple domains using an adversarial training scheme. Numerous experiments are reported on well-known single-domain and multi-domain benchmarks. The results demonstrate that our approach achieves competitive results in terms of mean Average Precision (mAP) and model size as compared to the state-of-the-art. The code will be made publicly available.

本文提出了一种基于图的自适应多标签图像分类方法。基于图的方法具有标签相关性建模能力,因此在多标签分类领域得到了广泛应用。具体来说,这些方法不仅在考虑单个领域时有效,在考虑多个领域时也同样有效。然而,所使用的图的拓扑结构并不是最佳的,因为它是预先启发式定义的。此外,连续的图卷积网络(GCN)聚合往往会破坏特征的相似性。为了克服这些问题,我们引入了一种以端到端方式学习图连接性的架构。这是通过整合基于注意力的机制和保持相似性的策略来实现的。然后,利用对抗训练方案将所提出的框架扩展到多个领域。报告在著名的单域和多域基准上进行了大量实验。结果表明,与最先进的方法相比,我们的方法在平均精度(mAP)和模型大小方面都取得了有竞争力的结果。代码将公开发布。
{"title":"Multi-label image classification using adaptive graph convolutional networks: From a single domain to multiple domains","authors":"Inder Pal Singh ,&nbsp;Enjie Ghorbel ,&nbsp;Oyebade Oyedotun ,&nbsp;Djamila Aouada","doi":"10.1016/j.cviu.2024.104062","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104062","url":null,"abstract":"<div><p>This paper proposes an adaptive graph-based approach for multi-label image classification. Graph-based methods have been largely exploited in the field of multi-label classification, given their ability to model label correlations. Specifically, their effectiveness has been proven not only when considering a single domain but also when taking into account multiple domains. However, the topology of the used graph is not optimal as it is pre-defined heuristically. In addition, consecutive Graph Convolutional Network (GCN) aggregations tend to destroy the feature similarity. To overcome these issues, an architecture for learning the graph connectivity in an end-to-end fashion is introduced. This is done by integrating an attention-based mechanism and a similarity-preserving strategy. The proposed framework is then extended to multiple domains using an adversarial training scheme. Numerous experiments are reported on well-known single-domain and multi-domain benchmarks. The results demonstrate that our approach achieves competitive results in terms of mean Average Precision (mAP) and model size as compared to the state-of-the-art. The code will be made publicly available.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1077314224001437/pdfft?md5=0c261f58e8fe19e830f04f80492395f1&pid=1-s2.0-S1077314224001437-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141607611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pseudo initialization based Few-Shot Class Incremental Learning 基于伪初始化的 "几枪 "类增量学习
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-27 DOI: 10.1016/j.cviu.2024.104067
Mingwen Shao , Xinkai Zhuang , Lixu Zhang , Wangmeng Zuo

Few-Shot Class Incremental Learning (FSCIL) aims to recognize sequentially arriving new classes without catastrophic forgetting old classes. The incremental new classes only contain very few labeled examples for updating the model, which causes overfitting problem. Current popular reserving embedding space method Forward Compatible Training preserves feature space for novel classes. Base class is pushed away from the most similar virtual class, preparing for the incoming novel classes. However, this can lead to pushing the base class to other similar virtual classes. In this paper, we propose a novel FSCIL method in order to overcome the aforementioned problem. Specifically, our core idea is pushing base classes away from the most similar top-K virtual classes to reserve feature space and provide pseudo initialization for the incoming novel classes. To further encourage learning new classes without forgetting, an additional regularization is applied to limit the extent of model updating. Extensive experiments are conducted on CUB200, CIFAR100 and mini-ImageNet, illustrating the performance of our proposed method. The results show that our method outperforms the state-of-the-art method and achieves significant improvement.

少量类增量学习(FSCIL)旨在识别连续出现的新类,而不会灾难性地遗忘旧类。增量新类只包含极少量用于更新模型的标注示例,这会导致过拟合问题。目前流行的保留嵌入空间方法是前向兼容训练(Forward Compatible Training),为新类别保留特征空间。基类被推离最相似的虚拟类,为新进入的类做准备。然而,这可能导致将基类推向其他相似的虚拟类。本文提出了一种新颖的 FSCIL 方法,以克服上述问题。具体来说,我们的核心思想是将基类推离最相似的 Top-K 虚拟类,以保留特征空间,并为新进入的新类提供伪初始化。为了进一步鼓励在不遗忘的情况下学习新类,我们还采用了额外的正则化来限制模型更新的范围。我们在 CUB200、CIFAR100 和 mini-ImageNet 上进行了广泛的实验,以说明我们提出的方法的性能。结果表明,我们的方法优于最先进的方法,并取得了显著的改进。
{"title":"Pseudo initialization based Few-Shot Class Incremental Learning","authors":"Mingwen Shao ,&nbsp;Xinkai Zhuang ,&nbsp;Lixu Zhang ,&nbsp;Wangmeng Zuo","doi":"10.1016/j.cviu.2024.104067","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104067","url":null,"abstract":"<div><p>Few-Shot Class Incremental Learning (FSCIL) aims to recognize sequentially arriving new classes without catastrophic forgetting old classes. The incremental new classes only contain very few labeled examples for updating the model, which causes overfitting problem. Current popular reserving embedding space method Forward Compatible Training preserves feature space for novel classes. Base class is pushed away from the most similar virtual class, preparing for the incoming novel classes. However, this can lead to pushing the base class to other similar virtual classes. In this paper, we propose a novel FSCIL method in order to overcome the aforementioned problem. Specifically, our core idea is pushing base classes away from the most similar top-K virtual classes to reserve feature space and provide pseudo initialization for the incoming novel classes. To further encourage learning new classes without forgetting, an additional regularization is applied to limit the extent of model updating. Extensive experiments are conducted on CUB200, CIFAR100 and mini-ImageNet, illustrating the performance of our proposed method. The results show that our method outperforms the state-of-the-art method and achieves significant improvement.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141596270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Camouflaged object segmentation with prior via two-stage training 通过两阶段训练进行带先验的伪装物体分割
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-26 DOI: 10.1016/j.cviu.2024.104061
Rui Wang , Caijuan Shi , Changyu Duan , Weixiang Gao , Hongli Zhu , Yunchao Wei , Meiqin Liu

The camouflaged object segmentation (COS) task aims to segment objects visually embedded within the background. Existing models usually rely on prior information as an auxiliary means to identify camouflaged objects. However, low-quality priors and the singular guidance form hinder the effective utilization of prior information. To address these issues, we propose a novel approach for prior generation and guidance, named prior-guided transformer (PGT). For prior generation, we design a prior generation subnetwork consisting of a Transformer backbone and simple convolutions to obtain higher-quality priors at a lower cost. In addition, to fully exploit the backbone’s understanding capabilities of the camouflage characteristics, a novel two-stage training method is proposed to achieve the backbone’s deep supervision. For prior guidance, we design a prior guidance modules (PGM), with distinct space token mixers to respectively capture global dependencies of location priors and local details of boundary priors. Additionally, we introduce a cross-level prior in the form of features to facilitate inter-level communication of backbone features. Extensive experiments have been conducted and experimental results illustrate the effectiveness and superiority of our method. The code is available at https://github.com/Ray3417/PGT.

伪装物体分割(COS)任务旨在分割视觉上嵌入背景中的物体。现有模型通常依赖先验信息作为识别伪装物体的辅助手段。然而,低质量的先验信息和单一的引导形式阻碍了先验信息的有效利用。为了解决这些问题,我们提出了一种新的先验生成和引导方法,命名为先验引导变换器(PGT)。在先验生成方面,我们设计了一个由变换器主干和简单卷积组成的先验生成子网络,以较低的成本获得更高质量的先验。此外,为了充分利用骨干网对伪装特征的理解能力,我们提出了一种新颖的两阶段训练方法,以实现骨干网的深度监督。在先验引导方面,我们设计了一个先验引导模块(PGM),它具有不同的空间令牌混合器,分别捕捉位置先验的全局依赖性和边界先验的局部细节。此外,我们还以特征的形式引入了跨层级先验,以促进骨干特征的层级间交流。我们进行了广泛的实验,实验结果表明了我们方法的有效性和优越性。代码见 https://github.com/Ray3417/PGT。
{"title":"Camouflaged object segmentation with prior via two-stage training","authors":"Rui Wang ,&nbsp;Caijuan Shi ,&nbsp;Changyu Duan ,&nbsp;Weixiang Gao ,&nbsp;Hongli Zhu ,&nbsp;Yunchao Wei ,&nbsp;Meiqin Liu","doi":"10.1016/j.cviu.2024.104061","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104061","url":null,"abstract":"<div><p>The camouflaged object segmentation (COS) task aims to segment objects visually embedded within the background. Existing models usually rely on prior information as an auxiliary means to identify camouflaged objects. However, low-quality priors and the singular guidance form hinder the effective utilization of prior information. To address these issues, we propose a novel approach for prior generation and guidance, named prior-guided transformer (PGT). For prior generation, we design a prior generation subnetwork consisting of a Transformer backbone and simple convolutions to obtain higher-quality priors at a lower cost. In addition, to fully exploit the backbone’s understanding capabilities of the camouflage characteristics, a novel two-stage training method is proposed to achieve the backbone’s deep supervision. For prior guidance, we design a prior guidance modules (PGM), with distinct space token mixers to respectively capture global dependencies of location priors and local details of boundary priors. Additionally, we introduce a cross-level prior in the form of features to facilitate inter-level communication of backbone features. Extensive experiments have been conducted and experimental results illustrate the effectiveness and superiority of our method. The code is available at <span>https://github.com/Ray3417/PGT</span><svg><path></path></svg>.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141540520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A proxy-data-based hierarchical adversarial patch generation method 基于代理数据的分层对抗补丁生成方法
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-24 DOI: 10.1016/j.cviu.2024.104066
Jiawei Liu , Xun Gong , Tingting Wang , Yunfeng Hu , Hong Chen

Current training data-dependent physical attacks have limited applicability to privacy-critical situations when attackers lack access to neural networks’ training data. To address this issue, this paper presents a hierarchical adversarial patch generation framework considering data privacy, utilizing proxy datasets while assuming that the training data is blinded. In the upper layer, Average Patch Saliency (APS) is introduced as a quantitative metric to determine the best proxy dataset for patch generation from a set of publicly available datasets. In the lower layer, Expectation of Transformation Plus (EoT+) method is developed to generate patches while accounting for perturbing background simulation and sensitivity alleviation. Evaluation results obtained in digital settings show that the proposed proxy-data-based framework achieves comparable targeted attack results to the data-dependent benchmark method. Finally, the framework’s validity is comprehensively evaluated in the physical world, where the corresponding experimental videos and code can be found at here.

当攻击者无法访问神经网络的训练数据时,当前依赖训练数据的物理攻击对隐私关键型情况的适用性有限。为了解决这个问题,本文提出了一个考虑到数据隐私的分层对抗补丁生成框架,利用代理数据集,同时假设训练数据是盲目的。在上层,引入平均补丁显著性(APS)作为量化指标,从一组公开可用的数据集中确定生成补丁的最佳代理数据集。在下层,开发了期望变换加(EoT+)方法来生成补丁,同时考虑到扰动背景模拟和灵敏度降低。在数字环境中获得的评估结果表明,所提出的基于代理数据的框架可实现与依赖数据的基准方法相当的目标攻击结果。最后,在物理世界中对该框架的有效性进行了全面评估,相应的实验视频和代码可在此处找到。
{"title":"A proxy-data-based hierarchical adversarial patch generation method","authors":"Jiawei Liu ,&nbsp;Xun Gong ,&nbsp;Tingting Wang ,&nbsp;Yunfeng Hu ,&nbsp;Hong Chen","doi":"10.1016/j.cviu.2024.104066","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104066","url":null,"abstract":"<div><p>Current <em>training data-dependent</em> physical attacks have limited applicability to privacy-critical situations when attackers lack access to neural networks’ training data. To address this issue, this paper presents a hierarchical adversarial patch generation framework considering data privacy, utilizing <em>proxy datasets</em> while assuming that the training data is blinded. In the upper layer, <strong>Average Patch Saliency</strong> (<strong>APS</strong>) is introduced as a quantitative metric to determine the best proxy dataset for patch generation from a set of publicly available datasets. In the lower layer, <strong>Expectation of Transformation Plus</strong> (<strong>EoT+</strong>) method is developed to generate patches while accounting for perturbing background simulation and sensitivity alleviation. Evaluation results obtained in digital settings show that the proposed proxy-data-based framework achieves comparable targeted attack results to the data-dependent benchmark method. Finally, the framework’s validity is comprehensively evaluated in the physical world, where the corresponding experimental videos and code can be found at <span>here</span><svg><path></path></svg>.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141480368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer Vision and Image Understanding
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1