首页 > 最新文献

Computer Vision and Image Understanding最新文献

英文 中文
Self-supervised network for low-light traffic image enhancement based on deep noise and artifacts removal 基于深度去噪和去伪的低照度交通图像增强自监督网络
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-22 DOI: 10.1016/j.cviu.2024.104063
Houwang Zhang , Kai-Fu Yang , Yong-Jie Li , Leanne Lai-Hang Chan

In the intelligent transportation system (ITS), detecting vehicles and pedestrians in low-light conditions is challenging due to the low contrast between objects and the background. Recently, many works have enhanced low-light images using deep learning-based methods, but these methods require paired images during training, which are impractical to obtain in real-world traffic scenarios. Therefore, we propose a self-supervised network (SSN) for low-light traffic image enhancement that can be trained without paired images. To avoid amplifying noise and artifacts in the processed image during enhancement, we first proposed a denoising net to reduce the noise and artifacts in the input image. Then the processed image can be enhanced by the enhancement net. Considering the compression of the traffic image, we designed an artifacts removal net to improve the quality of the enhanced image. We proposed several effective and differential losses to make SSN trainable with low-light images only. To better integrate the extracted features from different levels in the network, we also proposed an attention module named the multi-head non-local block. In experiments, we evaluated SSN and other low-light image enhancement methods on two low-light traffic image sets: the Berkeley Deep Drive (BDD) dataset and the Hong Kong night-time multi-class vehicle (HK) dataset. The results indicated that SSN significantly improves upon other methods in visual comparison and some blind image quality metrics. We also conducted comparisons on classical ITS tasks like vehicle detection on the images enhanced by SSN and other methods, which further verified its effectiveness.

在智能交通系统(ITS)中,由于物体与背景之间的对比度较低,在弱光条件下检测车辆和行人是一项挑战。最近,许多研究利用基于深度学习的方法增强了弱光图像,但这些方法在训练过程中需要配对图像,而这在现实世界的交通场景中是不切实际的。因此,我们提出了一种用于弱光交通图像增强的自监督网络(SSN),无需配对图像即可进行训练。为了避免在增强过程中放大处理图像中的噪声和伪影,我们首先提出了一个去噪网络,以减少输入图像中的噪声和伪影。然后通过增强网对处理后的图像进行增强。考虑到交通图像的压缩问题,我们设计了一个去伪网络来提高增强图像的质量。我们提出了几种有效的差分损耗,使 SSN 仅能在低照度图像下进行训练。为了更好地整合网络中不同层次的提取特征,我们还提出了一个名为多头非本地块的注意力模块。在实验中,我们在两个弱光交通图像集上评估了 SSN 和其他弱光图像增强方法:伯克利深度驾驶(BDD)数据集和香港夜间多类车辆(HK)数据集。结果表明,在视觉对比和一些盲图像质量指标方面,SSN 明显优于其他方法。我们还在经 SSN 和其他方法增强的图像上进行了经典 ITS 任务(如车辆检测)的比较,进一步验证了 SSN 的有效性。
{"title":"Self-supervised network for low-light traffic image enhancement based on deep noise and artifacts removal","authors":"Houwang Zhang ,&nbsp;Kai-Fu Yang ,&nbsp;Yong-Jie Li ,&nbsp;Leanne Lai-Hang Chan","doi":"10.1016/j.cviu.2024.104063","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104063","url":null,"abstract":"<div><p>In the intelligent transportation system (ITS), detecting vehicles and pedestrians in low-light conditions is challenging due to the low contrast between objects and the background. Recently, many works have enhanced low-light images using deep learning-based methods, but these methods require paired images during training, which are impractical to obtain in real-world traffic scenarios. Therefore, we propose a self-supervised network (SSN) for low-light traffic image enhancement that can be trained without paired images. To avoid amplifying noise and artifacts in the processed image during enhancement, we first proposed a denoising net to reduce the noise and artifacts in the input image. Then the processed image can be enhanced by the enhancement net. Considering the compression of the traffic image, we designed an artifacts removal net to improve the quality of the enhanced image. We proposed several effective and differential losses to make SSN trainable with low-light images only. To better integrate the extracted features from different levels in the network, we also proposed an attention module named the multi-head non-local block. In experiments, we evaluated SSN and other low-light image enhancement methods on two low-light traffic image sets: the Berkeley Deep Drive (BDD) dataset and the Hong Kong night-time multi-class vehicle (HK) dataset. The results indicated that SSN significantly improves upon other methods in visual comparison and some blind image quality metrics. We also conducted comparisons on classical ITS tasks like vehicle detection on the images enhanced by SSN and other methods, which further verified its effectiveness.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1077314224001449/pdfft?md5=ede1ba1e9b1e2fb3d3bbb81d8d671711&pid=1-s2.0-S1077314224001449-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141480370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ModelNet-O: A large-scale synthetic dataset for occlusion-aware point cloud classification ModelNet-O:用于遮挡感知点云分类的大规模合成数据集
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-19 DOI: 10.1016/j.cviu.2024.104060
Zhongbin Fang , Xia Li , Xiangtai Li , Shen Zhao , Mengyuan Liu

Recently, 3D point cloud classification has made significant progress with the help of many datasets. However, these datasets do not reflect the incomplete nature of real-world point clouds caused by occlusion, which limits the practical application of current methods. To bridge this gap, we propose ModelNet-O, a large-scale synthetic dataset of 123,041 samples that emulates real-world point clouds with self-occlusion caused by scanning from monocular cameras. ModelNet-O is 10 times larger than existing datasets and offers more challenging cases to evaluate the robustness of existing methods. Our observation on ModelNet-O reveals that well-designed sparse structures can preserve structural information of point clouds under occlusion, motivating us to propose a robust point cloud processing method that leverages a critical point sampling (CPS) strategy in a multi-level manner. We term our method PointMLS. Through extensive experiments, we demonstrate that our PointMLS achieves state-of-the-art results on ModelNet-O and competitive results on regular datasets such as ModelNet40 and ScanObjectNN, and we also demonstrate its robustness and effectiveness. Code available: https://github.com/fanglaosi/ModelNet-O_PointMLS.

最近,在许多数据集的帮助下,三维点云分类取得了重大进展。然而,这些数据集并不能反映真实世界点云因遮挡而造成的不完整性,这限制了当前方法的实际应用。为了弥补这一缺陷,我们提出了 ModelNet-O,这是一个由 123,041 个样本组成的大规模合成数据集,它模拟了真实世界中由单目摄像头扫描引起的自闭塞点云。ModelNet-O 比现有数据集大 10 倍,为评估现有方法的鲁棒性提供了更具挑战性的案例。我们对 ModelNet-O 的观察发现,精心设计的稀疏结构可以在遮挡情况下保留点云的结构信息,这促使我们提出了一种稳健的点云处理方法,该方法以多层次的方式利用临界点采样(CPS)策略。我们将这种方法称为 PointMLS。通过大量实验,我们证明了我们的 PointMLS 在 ModelNet-O 上取得了最先进的结果,在 ModelNet40 和 ScanObjectNN 等常规数据集上也取得了有竞争力的结果,我们还证明了它的鲁棒性和有效性。可用代码:https://github.com/fanglaosi/ModelNet-O_PointMLS。
{"title":"ModelNet-O: A large-scale synthetic dataset for occlusion-aware point cloud classification","authors":"Zhongbin Fang ,&nbsp;Xia Li ,&nbsp;Xiangtai Li ,&nbsp;Shen Zhao ,&nbsp;Mengyuan Liu","doi":"10.1016/j.cviu.2024.104060","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104060","url":null,"abstract":"<div><p>Recently, 3D point cloud classification has made significant progress with the help of many datasets. However, these datasets do not reflect the incomplete nature of real-world point clouds caused by <strong><em>occlusion</em></strong>, which limits the practical application of current methods. To bridge this gap, we propose ModelNet-O, <strong>a large-scale synthetic dataset of 123,041 samples</strong> that emulates real-world point clouds with self-occlusion caused by scanning from monocular cameras. ModelNet-O is <strong><em>10 times</em></strong> larger than existing datasets and offers more challenging cases to evaluate the robustness of existing methods. Our observation on ModelNet-O reveals that <strong>well-designed sparse structures can preserve structural information of point clouds under occlusion</strong>, motivating us to propose a robust point cloud processing method that leverages a critical point sampling (CPS) strategy in a multi-level manner. We term our method PointMLS. Through extensive experiments, we demonstrate that our PointMLS achieves state-of-the-art results on ModelNet-O and competitive results on regular datasets such as ModelNet40 and ScanObjectNN, and we also demonstrate its robustness and effectiveness. Code available: <span>https://github.com/fanglaosi/ModelNet-O_PointMLS</span><svg><path></path></svg>.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141480369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Confidence sharing adaptation for out-of-domain human pose and shape estimation 用于域外人体姿态和形状估计的置信度共享自适应
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-06-14 DOI: 10.1016/j.cviu.2024.104051
Tianyi Yue, Keyan Ren, Yu Shi, Hu Zhao, Qingyun Bian

3D human pose and shape estimation is often impacted by distribution bias in real-world scenarios due to factors such as bone length, camera parameters, background, and occlusion. To address this issue, we propose the Confidence Sharing Adaptation (CSA) algorithm, which corrects model bias using unlabeled images from the test domain before testing. However, the lack of annotation constraints in the adaptive training process poses a significant challenge, making it susceptible to model collapse. CSA utilizes a decoupled dual-branch learning framework to provide pseudo-labels and remove noise samples based on the confidence scores of the inference results. By sharing the most confident prior knowledge between the dual-branch networks, CSA effectively mitigates distribution bias. CSA is also remarkably adaptable to severely occluded scenes, thanks to two auxiliary techniques: a self-attentive parametric regressor that ensures robustness to occlusion of local body parts and a rendered surface texture loss that regulates the relationship between occlusion of human joint positions. Evaluation results show that CSA successfully adapts to scenarios beyond the training domain and achieves state-of-the-art performance on both occlusion-specific and general benchmarks. Code and pre-trained models are available for research at https://github.com/bodymapper/csa.git

在现实世界中,由于骨骼长度、相机参数、背景和遮挡等因素的影响,三维人体姿态和形状估计经常会受到分布偏差的影响。为了解决这个问题,我们提出了信心共享自适应(CSA)算法,该算法在测试前使用来自测试域的未标注图像纠正模型偏差。然而,在自适应训练过程中缺乏注释约束是一个重大挑战,使其容易出现模型崩溃。CSA 利用解耦双分支学习框架提供伪标签,并根据推理结果的置信度分数去除噪声样本。通过在双分支网络之间共享最有置信度的先验知识,CSA 有效地减轻了分布偏差。CSA 还能很好地适应严重遮挡的场景,这要归功于两种辅助技术:一种是自注意参数回归器,可确保对局部身体部位遮挡的鲁棒性;另一种是渲染表面纹理损失,可调节人体关节位置遮挡之间的关系。评估结果表明,CSA 成功地适应了训练领域以外的场景,并在特定遮挡和一般基准测试中取得了最先进的性能。代码和预训练模型可通过 https://github.com/bodymapper/csa.git 进行研究。
{"title":"Confidence sharing adaptation for out-of-domain human pose and shape estimation","authors":"Tianyi Yue,&nbsp;Keyan Ren,&nbsp;Yu Shi,&nbsp;Hu Zhao,&nbsp;Qingyun Bian","doi":"10.1016/j.cviu.2024.104051","DOIUrl":"10.1016/j.cviu.2024.104051","url":null,"abstract":"<div><p>3D human pose and shape estimation is often impacted by distribution bias in real-world scenarios due to factors such as bone length, camera parameters, background, and occlusion. To address this issue, we propose the Confidence Sharing Adaptation (CSA) algorithm, which corrects model bias using unlabeled images from the test domain before testing. However, the lack of annotation constraints in the adaptive training process poses a significant challenge, making it susceptible to model collapse. CSA utilizes a decoupled dual-branch learning framework to provide pseudo-labels and remove noise samples based on the confidence scores of the inference results. By sharing the most confident prior knowledge between the dual-branch networks, CSA effectively mitigates distribution bias. CSA is also remarkably adaptable to severely occluded scenes, thanks to two auxiliary techniques: a self-attentive parametric regressor that ensures robustness to occlusion of local body parts and a rendered surface texture loss that regulates the relationship between occlusion of human joint positions. Evaluation results show that CSA successfully adapts to scenarios beyond the training domain and achieves state-of-the-art performance on both occlusion-specific and general benchmarks. Code and pre-trained models are available for research at <span>https://github.com/bodymapper/csa.git</span><svg><path></path></svg></p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.3,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141392614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lv-Adapter: Adapting Vision Transformers for Visual Classification with Linear-layers and Vectors Lv-Adapter:利用线性层和向量调整视觉变换器以进行视觉分类
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-06-07 DOI: 10.1016/j.cviu.2024.104049
Guangyi Xu, Junyong Ye, Xinyuan Liu, Xubin Wen, Youwei Li, Jingjing Wang

Large pre-trained models based on Vision Transformers (ViTs) contain nearly billions of parameters, demanding substantial computational resources and storage space. This restricts their transferability across different tasks. Recent approaches try to use adapter fine-tuning to address this drawback. However, there is still potential to improve the number of tunable parameters and the accuracy in these methods. To address this challenge, we propose an adapter fine-tuning module called Lv-Adapter, which consists of a linear layer and vector. This module enables targeted parameter fine-tuning of pretrained models by learning both the prior knowledge of pre-trained task and the information from downstream specific task, to adapt to various downstream tasks in image and video tasks while transfer learning. Compared to full fine-tuning methods, Lv-Adapter has several appealing advantages. Firstly, by adding only about 3% extra parameters to ViT, Lv-Adapter achieves comparable accuracy to full fine-tuning methods and even significantly surpasses them on action recognition benchmarks. Secondly, Lv-Adapter is a lightweight module that can be plug-and-play in different transformer models due to its simplicity. Finally, to validate these claims, extensive experiments were conducted on five image and video datasets in this study, providing evidence for the effectiveness of Lv-Adapter. When only 3.5% of the extra parameters are updated, it respectively achieves a relative boost of about 13% and 24% compared to the fully fine-tuned model on SSv2 and HMDB51.

基于视觉转换器(ViT)的大型预训练模型包含近数十亿个参数,需要大量的计算资源和存储空间。这限制了它们在不同任务中的可移植性。最近的方法尝试使用适配器微调来解决这一缺点。然而,这些方法在可调参数数量和精确度方面仍有改进的空间。为了应对这一挑战,我们提出了一种名为 Lv-Adapter 的适配器微调模块,它由线性层和向量组成。该模块通过学习预训练任务的先验知识和下游特定任务的信息,对预训练模型进行有针对性的参数微调,从而在迁移学习的同时适应图像和视频任务中的各种下游任务。与完全微调方法相比,Lv-Adapter 有几个吸引人的优点。首先,Lv-Adapter 只需在 ViT 中增加约 3% 的额外参数,就能达到与完全微调方法相当的准确率,甚至在动作识别基准测试中大大超过它们。其次,Lv-Adapter 是一个轻量级模块,由于其简单性,可以在不同型号的变压器中即插即用。最后,为了验证这些说法,本研究在五个图像和视频数据集上进行了大量实验,为 Lv-Adapter 的有效性提供了证据。当仅更新 3.5% 的额外参数时,与 SSv2 和 HMDB51 上的完全微调模型相比,它分别实现了约 13% 和 24% 的相对提升。
{"title":"Lv-Adapter: Adapting Vision Transformers for Visual Classification with Linear-layers and Vectors","authors":"Guangyi Xu,&nbsp;Junyong Ye,&nbsp;Xinyuan Liu,&nbsp;Xubin Wen,&nbsp;Youwei Li,&nbsp;Jingjing Wang","doi":"10.1016/j.cviu.2024.104049","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104049","url":null,"abstract":"<div><p>Large pre-trained models based on Vision Transformers (ViTs) contain nearly billions of parameters, demanding substantial computational resources and storage space. This restricts their transferability across different tasks. Recent approaches try to use adapter fine-tuning to address this drawback. However, there is still potential to improve the number of tunable parameters and the accuracy in these methods. To address this challenge, we propose an adapter fine-tuning module called Lv-Adapter, which consists of a linear layer and vector. This module enables targeted parameter fine-tuning of pretrained models by learning both the prior knowledge of pre-trained task and the information from downstream specific task, to adapt to various downstream tasks in image and video tasks while transfer learning. Compared to full fine-tuning methods, Lv-Adapter has several appealing advantages. Firstly, by adding only about 3% extra parameters to ViT, Lv-Adapter achieves comparable accuracy to full fine-tuning methods and even significantly surpasses them on action recognition benchmarks. Secondly, Lv-Adapter is a lightweight module that can be plug-and-play in different transformer models due to its simplicity. Finally, to validate these claims, extensive experiments were conducted on five image and video datasets in this study, providing evidence for the effectiveness of Lv-Adapter. When only 3.5% of the extra parameters are updated, it respectively achieves a relative boost of about 13% and 24% compared to the fully fine-tuned model on SSv2 and HMDB51.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141323899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Skeleton Cluster Tracking for robust multi-view multi-person 3D human pose estimation 用于多视角多人三维人体姿态稳健估算的骨架集群跟踪技术
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-06-07 DOI: 10.1016/j.cviu.2024.104059
Zehai Niu , Ke Lu , Jian Xue , Jinbao Wang

The multi-view 3D human pose estimation task relies on 2D human pose estimation for each view; however, severe occlusion, truncation, and human interaction lead to incorrect 2D human pose estimation for some views. The traditional “Matching-Lifting-Tracking” paradigm amplifies the incorrect 2D human pose into an incorrect 3D human pose, which significantly challenges the robustness of multi-view 3D human pose estimation. In this paper, we propose a novel method that tackles the inherent difficulties of the traditional paradigm. This method is rooted in the newly devised “Skeleton Pooling-Clustering-Tracking (SPCT)” paradigm. It initiates a 2D human pose estimation for each perspective. Then a symmetrical dilated network is created for skeleton pool estimation. Upon clustering the skeleton pool, we introduce and implement an innovative tracking method that is explicitly designed for the SPCT paradigm. The tracking method refines and filters the skeleton clusters, thereby enhancing the robustness of the multi-person 3D human pose estimation results. By coupling the skeleton pool with the tracking refinement process, our method obtains high-quality multi-person 3D human pose estimation results despite severe occlusions that produce erroneous 2D and 3D estimates. By employing the proposed SPCT paradigm and a computationally efficient network architecture, our method outperformed existing approaches regarding robustness on the Shelf, 4D Association, and CMU Panoptic datasets, and could be applied in practical scenarios such as markerless motion capture and animation production.

多视角三维人体姿态估计任务依赖于每个视角的二维人体姿态估计;然而,严重的遮挡、截断和人机交互会导致某些视角的二维人体姿态估计不正确。传统的 "匹配-提升-跟踪 "范式会将错误的二维人体姿态放大为错误的三维人体姿态,这对多视角三维人体姿态估计的鲁棒性提出了极大的挑战。在本文中,我们提出了一种新方法来解决传统模式的固有难题。这种方法植根于新设计的 "骨架池-聚类-跟踪(SPCT)"范式。它首先对每个视角进行二维人体姿态估计。然后创建一个对称的扩张网络,用于骨架池估算。在对骨架池进行聚类后,我们引入并实施了一种明确针对 SPCT 范例设计的创新跟踪方法。该跟踪方法对骨架集群进行细化和过滤,从而增强了多人三维人体姿态估计结果的鲁棒性。通过将骨架池与跟踪细化过程相结合,我们的方法可以获得高质量的多人三维人体姿态估计结果,尽管严重的遮挡会产生错误的二维和三维估计结果。通过采用建议的 SPCT 范式和计算效率高的网络架构,我们的方法在 Shelf、4D Association 和 CMU Panoptic 数据集上的鲁棒性优于现有方法,可应用于无标记动作捕捉和动画制作等实际场景。
{"title":"Skeleton Cluster Tracking for robust multi-view multi-person 3D human pose estimation","authors":"Zehai Niu ,&nbsp;Ke Lu ,&nbsp;Jian Xue ,&nbsp;Jinbao Wang","doi":"10.1016/j.cviu.2024.104059","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104059","url":null,"abstract":"<div><p>The multi-view 3D human pose estimation task relies on 2D human pose estimation for each view; however, severe occlusion, truncation, and human interaction lead to incorrect 2D human pose estimation for some views. The traditional “Matching-Lifting-Tracking” paradigm amplifies the incorrect 2D human pose into an incorrect 3D human pose, which significantly challenges the robustness of multi-view 3D human pose estimation. In this paper, we propose a novel method that tackles the inherent difficulties of the traditional paradigm. This method is rooted in the newly devised “Skeleton Pooling-Clustering-Tracking (SPCT)” paradigm. It initiates a 2D human pose estimation for each perspective. Then a symmetrical dilated network is created for skeleton pool estimation. Upon clustering the skeleton pool, we introduce and implement an innovative tracking method that is explicitly designed for the SPCT paradigm. The tracking method refines and filters the skeleton clusters, thereby enhancing the robustness of the multi-person 3D human pose estimation results. By coupling the skeleton pool with the tracking refinement process, our method obtains high-quality multi-person 3D human pose estimation results despite severe occlusions that produce erroneous 2D and 3D estimates. By employing the proposed SPCT paradigm and a computationally efficient network architecture, our method outperformed existing approaches regarding robustness on the Shelf, 4D Association, and CMU Panoptic datasets, and could be applied in practical scenarios such as markerless motion capture and animation production.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141324067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic-driven diffusion for sign language production with gloss-pose latent spaces alignment 语义驱动的手语生产扩散与词汇潜在空间对齐
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-06-07 DOI: 10.1016/j.cviu.2024.104050
Sheng Chen, Qingshan Wang, Qi Wang

Sign Language Production (SLP) aims to translate spoken language into visual sign language sequences. The most challenging process in SLP is the transformation of a sequence of sign glosses into corresponding sign poses (G2P). Existing approaches on G2P mainly focus on constructing mappings of sign language glosses to frame-level sign pose representations, while neglecting gloss is just a weak annotation of the sequence of sign poses. To address this problem, this paper proposes the semantic-driven diffusion model with gloss-pose latent spaces alignment (SDD-GPLA) for G2P. G2P is divided into two phases. In the first phase, we design the gloss-pose latent spaces alignment (GPLA) to model the sign pose latent representations with glosses dependency. In the second phase, we propose semantic-driven diffusion (SDD) with supervised pose reconstruction guidance as a mapping between the gloss and sign poses latent features. In addition, we propose the sign pose decoder (Decoderp) to progressively generate high-resolution sign poses from latent sign pose features and to guide the SDD training process. We evaluated SDD-GPLA on a self-collected dataset of Daily Chinese Sign Language (DCSL) and a public dataset called RWTH-Phoenix-Weather-2014T. Compared with the state-of-the-art G2P methods, we obtain at least 22.9% and 2.3% improvement in WER scores on the above two datasets, respectively.

手语制作(SLP)旨在将口语转化为视觉手语序列。手语制作中最具挑战性的过程是将手语词汇序列转换为相应的手语姿势(G2P)。现有的 G2P 方法主要侧重于构建手语词汇到帧级手势表示的映射,而忽略了词汇只是手势序列的一个弱注释。针对这一问题,本文提出了针对 G2P 的语义驱动扩散模型与词汇-姿势潜空间配准(SDD-GPLA)。G2P 分为两个阶段。在第一阶段,我们设计了词汇-姿势-潜在空间配准(GPLA)来模拟具有词汇依赖性的符号姿势潜在表征。在第二阶段,我们提出了语义驱动扩散(SDD),将监督姿势重构指导作为词汇和符号姿势潜特征之间的映射。此外,我们还提出了符号姿势解码器(Decoderp),以便从潜在符号姿势特征逐步生成高分辨率符号姿势,并指导 SDD 的训练过程。我们在《每日中国手语》(DCSL)自收集数据集和名为 RWTH-Phoenix-Weather-2014T 的公共数据集上对 SDD-GPLA 进行了评估。与最先进的 G2P 方法相比,我们在上述两个数据集上的 WER 分数分别提高了至少 22.9% 和 2.3%。
{"title":"Semantic-driven diffusion for sign language production with gloss-pose latent spaces alignment","authors":"Sheng Chen,&nbsp;Qingshan Wang,&nbsp;Qi Wang","doi":"10.1016/j.cviu.2024.104050","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104050","url":null,"abstract":"<div><p>Sign Language Production (SLP) aims to translate spoken language into visual sign language sequences. The most challenging process in SLP is the transformation of a sequence of sign glosses into corresponding sign poses (G2P). Existing approaches on G2P mainly focus on constructing mappings of sign language glosses to frame-level sign pose representations, while neglecting gloss is just a weak annotation of the sequence of sign poses. To address this problem, this paper proposes the semantic-driven diffusion model with gloss-pose latent spaces alignment (SDD-GPLA) for G2P. G2P is divided into two phases. In the first phase, we design the gloss-pose latent spaces alignment (GPLA) to model the sign pose latent representations with glosses dependency. In the second phase, we propose semantic-driven diffusion (SDD) with supervised pose reconstruction guidance as a mapping between the gloss and sign poses latent features. In addition, we propose the sign pose decoder (<span><math><msup><mrow><mtext>Decoder</mtext></mrow><mrow><mi>p</mi></mrow></msup></math></span>) to progressively generate high-resolution sign poses from latent sign pose features and to guide the SDD training process. We evaluated SDD-GPLA on a self-collected dataset of Daily Chinese Sign Language (DCSL) and a public dataset called RWTH-Phoenix-Weather-2014T. Compared with the state-of-the-art G2P methods, we obtain at least 22.9% and 2.3% improvement in WER scores on the above two datasets, respectively.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141328561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identity-preserving editing of multiple facial attributes by learning global edit directions and local adjustments 通过学习全局编辑方向和局部调整,对多个面部属性进行身份保护编辑
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-06-01 DOI: 10.1016/j.cviu.2024.104047
Najmeh Mohammadbagheri, Fardin Ayar, Ahmad Nickabadi, Reza Safabakhsh

Semantic facial attribute editing using pre-trained Generative Adversarial Networks (GANs) has attracted a great deal of attention and effort from researchers in recent years. Due to the high quality of face images generated by StyleGANs, much work has focused on the StyleGANs’ latent space and the proposed methods for facial image editing. Although these methods have achieved satisfying results for manipulating user-intended attributes, they have not fulfilled the goal of preserving the identity, which is an important challenge. We present ID-Style, a new architecture capable of addressing the problem of identity loss during attribute manipulation. The key components of ID-Style include a Learnable Global Direction (LGD) module, which finds a shared and semi-sparse direction for each attribute, and an Instance-Aware Intensity Predictor (IAIP) network, which finetunes the global direction according to the input instance. Furthermore, we introduce two losses during training to enforce the LGD and IAIP to find semi-sparse semantic directions that preserve the identity of the input instance. Despite reducing the size of the network by roughly 95% as compared to similar state-of-the-art works, ID-Style outperforms baselines by 10% and 7% in identity preserving metric (FRS) and average accuracy of manipulation (mACC), respectively.

近年来,使用预训练生成对抗网络(GANs)进行语义面部属性编辑吸引了研究人员的大量关注和努力。由于 StyleGANs 生成的人脸图像质量很高,许多工作都集中在 StyleGANs 的潜在空间和所提出的人脸图像编辑方法上。虽然这些方法在处理用户意图属性方面取得了令人满意的结果,但它们并没有实现保留身份的目标,而这正是一个重要的挑战。我们提出的 ID-Style 是一种能够解决属性操作过程中身份丢失问题的新架构。ID-Style 的关键组件包括一个可学习全局方向(LGD)模块和一个实例感知强度预测器(IAIP)网络,前者可为每个属性找到一个共享的半稀疏方向,后者可根据输入实例对全局方向进行微调。此外,我们还在训练过程中引入了两种损失,以强制 LGD 和 IAIP 找到保留输入实例特征的半稀疏语义方向。尽管与同类最先进的研究相比,ID-Style 网络的规模缩小了约 95%,但在身份保留度量(FRS)和平均操作准确率(mACC)方面,ID-Style 分别比基线高出 10% 和 7%。
{"title":"Identity-preserving editing of multiple facial attributes by learning global edit directions and local adjustments","authors":"Najmeh Mohammadbagheri,&nbsp;Fardin Ayar,&nbsp;Ahmad Nickabadi,&nbsp;Reza Safabakhsh","doi":"10.1016/j.cviu.2024.104047","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104047","url":null,"abstract":"<div><p>Semantic facial attribute editing using pre-trained Generative Adversarial Networks (GANs) has attracted a great deal of attention and effort from researchers in recent years. Due to the high quality of face images generated by StyleGANs, much work has focused on the StyleGANs’ latent space and the proposed methods for facial image editing. Although these methods have achieved satisfying results for manipulating user-intended attributes, they have not fulfilled the goal of preserving the identity, which is an important challenge. We present ID-Style, a new architecture capable of addressing the problem of identity loss during attribute manipulation. The key components of ID-Style include a Learnable Global Direction (LGD) module, which finds a shared and semi-sparse direction for each attribute, and an Instance-Aware Intensity Predictor (IAIP) network, which finetunes the global direction according to the input instance. Furthermore, we introduce two losses during training to enforce the LGD and IAIP to find semi-sparse semantic directions that preserve the identity of the input instance. Despite reducing the size of the network by roughly 95% as compared to similar state-of-the-art works, ID-Style outperforms baselines by 10% and 7% in identity preserving metric (FRS) and average accuracy of manipulation (mACC), respectively.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141303008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-supervised monocular depth estimation with self-distillation and dense skip connection 利用自颤动和密集跳接进行自我监督单目深度估计
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-06-01 DOI: 10.1016/j.cviu.2024.104048
Xuezhi Xiang , Wei Li , Yao Wang , Abdulmotaleb El Saddik

Monocular depth estimation (MDE) is crucial in a wide range of applications, including robotics, autonomous driving and virtual reality. Self-supervised monocular depth estimation has emerged as a promising MDE approach without requiring hard-to-obtain depth labels during training, and multi-scale photometric loss is widely used for self-supervised monocular depth estimation as the self-supervised signal. However, multi-photometric loss is a weak training signal and might disturb the good intermediate features representation. In this paper, we propose a successive depth map self-distillation(SDM-SD) loss, which combines with the single-scale photometric loss to replace the multi-scale photometric loss. Moreover, considering that multi-stage feature representations are essential for dense prediction tasks such as depth estimation, we also propose a dense skip connection, which can efficiently fuse the intermediate features of the encoder and fully utilize them in each stage of the decoder in our encoder–decoder architecture. By applying successive depth map self-distillation loss and dense skip connection, our proposed method can achieve state-of-the-art performance on the KITTI benchmark, and exhibit the best generalization ability on the challenging indoor dataset NYUv2 dataset.

单目深度估计(MDE)在机器人、自动驾驶和虚拟现实等广泛应用中至关重要。自监督单目深度估计已成为一种前景广阔的 MDE 方法,它无需在训练过程中使用难以获得的深度标签,多尺度光度损失作为自监督信号被广泛用于自监督单目深度估计。然而,多尺度光度损失是一种弱训练信号,可能会干扰良好的中间特征表示。本文提出了一种连续深度图自抖动(SDM-SD)损失,它与单尺度光度损失相结合,取代了多尺度光度损失。此外,考虑到多阶段特征表示对于深度估计等密集预测任务至关重要,我们还提出了密集跳转连接,它可以有效地融合编码器的中间特征,并在我们的编码器-解码器架构中的解码器的每个阶段充分利用这些特征。通过应用连续深度图自抖动损耗和密集跳转连接,我们提出的方法在 KITTI 基准测试中取得了最先进的性能,并在具有挑战性的室内数据集 NYUv2 数据集上表现出最佳的泛化能力。
{"title":"Self-supervised monocular depth estimation with self-distillation and dense skip connection","authors":"Xuezhi Xiang ,&nbsp;Wei Li ,&nbsp;Yao Wang ,&nbsp;Abdulmotaleb El Saddik","doi":"10.1016/j.cviu.2024.104048","DOIUrl":"10.1016/j.cviu.2024.104048","url":null,"abstract":"<div><p>Monocular depth estimation (MDE) is crucial in a wide range of applications, including robotics, autonomous driving and virtual reality. Self-supervised monocular depth estimation has emerged as a promising MDE approach without requiring hard-to-obtain depth labels during training, and multi-scale photometric loss is widely used for self-supervised monocular depth estimation as the self-supervised signal. However, multi-photometric loss is a weak training signal and might disturb the good intermediate features representation. In this paper, we propose a successive depth map self-distillation(SDM-SD) loss, which combines with the single-scale photometric loss to replace the multi-scale photometric loss. Moreover, considering that multi-stage feature representations are essential for dense prediction tasks such as depth estimation, we also propose a dense skip connection, which can efficiently fuse the intermediate features of the encoder and fully utilize them in each stage of the decoder in our encoder–decoder architecture. By applying successive depth map self-distillation loss and dense skip connection, our proposed method can achieve state-of-the-art performance on the KITTI benchmark, and exhibit the best generalization ability on the challenging indoor dataset NYUv2 dataset.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141282083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DHBSR: A deep hybrid representation-based network for blind image super resolution DHBSR:基于深度混合表示的盲图像超分辨率网络
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-05-28 DOI: 10.1016/j.cviu.2024.104034
Alireza Esmaeilzehi , Farshid Nooshi , Hossein Zaredar , M. Omair Ahmad

Image super resolution involves enhancing the spatial resolution of low-quality images and improving their visual quality. As in many real-life situations, the image degradation process is unknown, performing the task of image super resolution in a blind manner is of paramount importance. Deep neural networks provide high performances for the task of blind image super resolution, in view of their end-to-end learning capability between the low-resolution images and their ground truth versions. Generally speaking, deep blind image super resolution networks initially estimate the parameters of the image degradation process, such as blurring kernel, and then use them for super-resolving the low-resolution images. In this paper, we develop a novel deep learning-based scheme for the task of blind image super resolution, in which the idea of leveraging the hybrid representations is utilized. Specifically, we employ the deterministic and stochastic representations of the blurring kernel parameters to train a deep blind super resolution network in an effective manner. The results of extensive experiments prove the effectiveness of various ideas used in the development of the proposed deep blind image super resolution network.

图像超级分辨率涉及增强低质量图像的空间分辨率并改善其视觉质量。在现实生活中的许多情况下,图像降解过程是未知的,因此以盲法执行图像超分辨率任务至关重要。鉴于深度神经网络在低分辨率图像及其地面实况版本之间的端到端学习能力,它能为图像盲超分辨率任务提供高性能。一般来说,深度盲图像超分辨率网络最初会估计图像降解过程的参数,如模糊核,然后利用这些参数对低分辨率图像进行超分辨率处理。在本文中,我们针对图像盲超分辨率任务开发了一种基于深度学习的新方案,其中利用了混合表征的思想。具体来说,我们利用模糊核参数的确定性和随机性表示,以有效的方式训练深度盲超分辨率网络。大量的实验结果证明了在开发拟议的深度盲图像超级分辨率网络时所采用的各种理念的有效性。
{"title":"DHBSR: A deep hybrid representation-based network for blind image super resolution","authors":"Alireza Esmaeilzehi ,&nbsp;Farshid Nooshi ,&nbsp;Hossein Zaredar ,&nbsp;M. Omair Ahmad","doi":"10.1016/j.cviu.2024.104034","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104034","url":null,"abstract":"<div><p>Image super resolution involves enhancing the spatial resolution of low-quality images and improving their visual quality. As in many real-life situations, the image degradation process is unknown, performing the task of image super resolution in a blind manner is of paramount importance. Deep neural networks provide high performances for the task of blind image super resolution, in view of their end-to-end learning capability between the low-resolution images and their ground truth versions. Generally speaking, deep blind image super resolution networks initially estimate the parameters of the image degradation process, such as blurring kernel, and then use them for super-resolving the low-resolution images. In this paper, we develop a novel deep learning-based scheme for the task of blind image super resolution, in which the idea of leveraging the hybrid representations is utilized. Specifically, we employ the deterministic and stochastic representations of the blurring kernel parameters to train a deep blind super resolution network in an effective manner. The results of extensive experiments prove the effectiveness of various ideas used in the development of the proposed deep blind image super resolution network.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141308042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
POTLoc: Pseudo-label Oriented Transformer for point-supervised temporal Action Localization POTLoc:用于点监督时间动作定位的伪标签定向变换器
IF 4.5 3区 计算机科学 Q1 Computer Science Pub Date : 2024-05-28 DOI: 10.1016/j.cviu.2024.104044
Elahe Vahdani, Yingli Tian

This paper tackles the challenge of point-supervised temporal action detection, wherein only a single frame is annotated for each action instance in the training set. Most of the current methods, hindered by the sparse nature of annotated points, struggle to effectively represent the continuous structure of actions or the inherent temporal and semantic dependencies within action instances. Consequently, these methods frequently learn merely the most distinctive segments of actions, leading to the creation of incomplete action proposals. This paper proposes POTLoc, a Pseudo-label Oriented Transformer for weakly-supervised Action Localization utilizing only point-level annotation. POTLoc is designed to identify and track continuous action structures via a self-training strategy. The base model begins by generating action proposals solely with point-level supervision. These proposals undergo refinement and regression to enhance the precision of the estimated action boundaries, which subsequently results in the production of ‘pseudo-labels’ to serve as supplementary supervisory signals. The architecture of the model integrates a transformer with a temporal feature pyramid to capture video snippet dependencies and model actions of varying duration. The pseudo-labels, providing information about the coarse locations and boundaries of actions, assist in guiding the transformer for enhanced learning of action dynamics. POTLoc outperforms the state-of-the-art point-supervised methods on THUMOS’14 and ActivityNet-v1.2 datasets.

本文探讨了点监督时态动作检测的挑战,在这种检测中,训练集中的每个动作实例只注释了一个帧。目前的大多数方法都受到注释点稀疏性的阻碍,难以有效表示动作的连续结构或动作实例中固有的时间和语义依赖关系。因此,这些方法往往只能学习到动作中最独特的片段,从而产生不完整的动作建议。本文提出的 POTLoc 是一种面向伪标签的转换器,用于仅利用点级注释的弱监督动作定位。POTLoc 设计用于通过自我训练策略识别和跟踪连续动作结构。基础模型首先仅通过点级监督生成动作建议。这些建议经过完善和回归,以提高估计动作边界的精确度,随后产生 "伪标签",作为补充监督信号。该模型的架构将变压器与时间特征金字塔整合在一起,以捕捉视频片段的相关性并对不同持续时间的动作进行建模。伪标签提供了有关动作粗略位置和边界的信息,有助于引导变换器加强动作动态学习。在 THUMOS'14 和 ActivityNet-v1.2 数据集上,POTLoc 的表现优于最先进的点监督方法。
{"title":"POTLoc: Pseudo-label Oriented Transformer for point-supervised temporal Action Localization","authors":"Elahe Vahdani,&nbsp;Yingli Tian","doi":"10.1016/j.cviu.2024.104044","DOIUrl":"https://doi.org/10.1016/j.cviu.2024.104044","url":null,"abstract":"<div><p>This paper tackles the challenge of point-supervised temporal action detection, wherein only a single frame is annotated for each action instance in the training set. Most of the current methods, hindered by the sparse nature of annotated points, struggle to effectively represent the continuous structure of actions or the inherent temporal and semantic dependencies within action instances. Consequently, these methods frequently learn merely the most distinctive segments of actions, leading to the creation of incomplete action proposals. This paper proposes POTLoc, a <strong>P</strong>seudo-label <strong>O</strong>riented <strong>T</strong>ransformer for weakly-supervised Action <strong>Loc</strong>alization utilizing only point-level annotation. POTLoc is designed to identify and track continuous action structures via a self-training strategy. The base model begins by generating action proposals solely with point-level supervision. These proposals undergo refinement and regression to enhance the precision of the estimated action boundaries, which subsequently results in the production of ‘pseudo-labels’ to serve as supplementary supervisory signals. The architecture of the model integrates a transformer with a temporal feature pyramid to capture video snippet dependencies and model actions of varying duration. The pseudo-labels, providing information about the coarse locations and boundaries of actions, assist in guiding the transformer for enhanced learning of action dynamics. POTLoc outperforms the state-of-the-art point-supervised methods on THUMOS’14 and ActivityNet-v1.2 datasets.</p></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":null,"pages":null},"PeriodicalIF":4.5,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141264274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer Vision and Image Understanding
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1