首页 > 最新文献

Computer Vision and Image Understanding最新文献

英文 中文
Generative adversarial network for semi-supervised image captioning 半监督图像标题生成对抗网络
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-04 DOI: 10.1016/j.cviu.2024.104199
Xu Liang, Chen Li, Lihua Tian
Traditional supervised image captioning methods usually rely on a large number of images and paired captions for training. However, the creation of such datasets necessitates considerable temporal and human resources. Therefore, we propose a new semi-supervised image captioning algorithm to solve this problem. The proposed method uses a generative adversarial network to generate images that match captions, and uses these generated images and captions as new training data. This avoids the error accumulation problem when generating pseudo captions with autoregressive method and the network can directly perform backpropagation. At the same time, in order to ensure the correlation between the generated images and captions, we introduced the CLIP model for constraints. The CLIP model has been pre-trained on a large amount of image–text data, so it shows excellent performance in semantic alignment of images and text. To verify the effectiveness of our method, we validate on MSCOCO offline “Karpathy” test split. Experiment results show that our method can significantly improve the performance of the model when using 1% paired data, with the CIDEr score increasing from 69.5% to 77.7%. This shows that our method can effectively utilize unlabeled data for image caption tasks.
传统的有监督图像标题方法通常依赖大量图像和配对标题进行训练。然而,创建这样的数据集需要大量的时间和人力资源。因此,我们提出了一种新的半监督图像字幕算法来解决这个问题。建议的方法使用生成式对抗网络生成与标题匹配的图像,并将这些生成的图像和标题作为新的训练数据。这就避免了用自回归法生成伪标题时的误差积累问题,网络可以直接进行反向传播。同时,为了确保生成的图像和字幕之间的相关性,我们引入了 CLIP 模型进行约束。CLIP 模型已在大量图像-文本数据上进行了预训练,因此在图像和文本的语义配准方面表现出色。为了验证我们方法的有效性,我们在 MSCOCO 离线 "Karpathy "测试分片上进行了验证。实验结果表明,当使用 1%的配对数据时,我们的方法能显著提高模型的性能,CIDEr 分数从 69.5% 提高到 77.7%。这表明我们的方法可以有效地利用无标记数据来完成图像标题任务。
{"title":"Generative adversarial network for semi-supervised image captioning","authors":"Xu Liang,&nbsp;Chen Li,&nbsp;Lihua Tian","doi":"10.1016/j.cviu.2024.104199","DOIUrl":"10.1016/j.cviu.2024.104199","url":null,"abstract":"<div><div>Traditional supervised image captioning methods usually rely on a large number of images and paired captions for training. However, the creation of such datasets necessitates considerable temporal and human resources. Therefore, we propose a new semi-supervised image captioning algorithm to solve this problem. The proposed method uses a generative adversarial network to generate images that match captions, and uses these generated images and captions as new training data. This avoids the error accumulation problem when generating pseudo captions with autoregressive method and the network can directly perform backpropagation. At the same time, in order to ensure the correlation between the generated images and captions, we introduced the CLIP model for constraints. The CLIP model has been pre-trained on a large amount of image–text data, so it shows excellent performance in semantic alignment of images and text. To verify the effectiveness of our method, we validate on MSCOCO offline “Karpathy” test split. Experiment results show that our method can significantly improve the performance of the model when using 1% paired data, with the CIDEr score increasing from 69.5% to 77.7%. This shows that our method can effectively utilize unlabeled data for image caption tasks.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104199"},"PeriodicalIF":4.3,"publicationDate":"2024-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BundleMoCap++: Efficient, robust and smooth motion capture from sparse multiview videos BundleMoCap++:从稀疏的多视角视频中高效、稳健、流畅地捕捉动作
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-04 DOI: 10.1016/j.cviu.2024.104190
Georgios Albanis , Nikolaos Zioulis , Kostas Kolomvatsos
Producing smooth and accurate motions from sparse videos without requiring specialized equipment and markers is a long-standing problem in the research community. Most approaches typically involve complex processes such as temporal constraints, multiple stages combining data-driven regression and optimization techniques, and bundle solving over temporal windows. These increase the computational burden and introduce the challenge of hyperparameter tuning for the different objective terms. In contrast, BundleMoCap++ offers a simple yet effective approach to this problem. It solves the motion in a single stage, eliminating the need for temporal smoothness objectives while still delivering smooth motions without compromising accuracy. BundleMoCap++ outperforms the state-of-the-art without increasing complexity. Our approach is based on manifold interpolation between latent keyframes. By relying on a local manifold smoothness assumption and appropriate interpolation schemes, we efficiently solve a bundle of frames using two or more latent codes. Additionally, the method is implemented as a sliding window optimization and requires only the first frame to be properly initialized, reducing the overall computational burden. BundleMoCap++’s strength lies in achieving high-quality motion capture results with fewer computational resources. To do this efficiently, we propose a novel human pose prior that focuses on the geometric aspect of the latent space, modeling it as a hypersphere, allowing for the introduction of sophisticated interpolation techniques. We also propose an algorithm for optimizing the latent variables directly on the learned manifold, improving convergence and performance. Finally, we introduce high-order interpolation techniques adapted for the hypersphere, allowing us to increase the solving temporal window, enhancing performance and efficiency.
无需专业设备和标记,就能从稀疏视频中生成平滑准确的运动图像,是研究界长期存在的问题。大多数方法通常涉及复杂的过程,如时间约束、结合数据驱动回归和优化技术的多阶段,以及在时间窗口上的捆绑求解。这些都增加了计算负担,并带来了针对不同目标项调整超参数的挑战。相比之下,BundleMoCap++ 为这一问题提供了一种简单而有效的方法。它只需一个阶段就能解决运动问题,无需时间平滑目标,同时还能在不影响精度的情况下实现平滑运动。在不增加复杂性的情况下,BundleMoCap++ 超越了最先进的技术。我们的方法基于潜在关键帧之间的流形插值。通过依赖局部流形平滑性假设和适当的插值方案,我们使用两个或更多潜在代码高效地解决了帧束问题。此外,该方法是以滑动窗口优化的方式实现的,只需要对第一帧进行适当的初始化,从而减轻了整体的计算负担。BundleMoCap++ 的优势在于用较少的计算资源获得高质量的运动捕捉结果。为了高效地实现这一目标,我们提出了一种新颖的人体姿态先验,该先验侧重于潜空间的几何方面,将其建模为超球,从而可以引入复杂的插值技术。我们还提出了一种直接在所学流形上优化潜变量的算法,从而提高了收敛性和性能。最后,我们引入了适用于超球的高阶插值技术,允许我们增加求解时间窗口,从而提高性能和效率。
{"title":"BundleMoCap++: Efficient, robust and smooth motion capture from sparse multiview videos","authors":"Georgios Albanis ,&nbsp;Nikolaos Zioulis ,&nbsp;Kostas Kolomvatsos","doi":"10.1016/j.cviu.2024.104190","DOIUrl":"10.1016/j.cviu.2024.104190","url":null,"abstract":"<div><div>Producing smooth and accurate motions from sparse videos without requiring specialized equipment and markers is a long-standing problem in the research community. Most approaches typically involve complex processes such as temporal constraints, multiple stages combining data-driven regression and optimization techniques, and bundle solving over temporal windows. These increase the computational burden and introduce the challenge of hyperparameter tuning for the different objective terms. In contrast, BundleMoCap++ offers a simple yet effective approach to this problem. It solves the motion in a single stage, eliminating the need for temporal smoothness objectives while still delivering smooth motions without compromising accuracy. BundleMoCap++ outperforms the state-of-the-art without increasing complexity. Our approach is based on manifold interpolation between latent keyframes. By relying on a local manifold smoothness assumption and appropriate interpolation schemes, we efficiently solve a bundle of frames using two or more latent codes. Additionally, the method is implemented as a sliding window optimization and requires only the first frame to be properly initialized, reducing the overall computational burden. BundleMoCap++’s strength lies in achieving high-quality motion capture results with fewer computational resources. To do this efficiently, we propose a novel human pose prior that focuses on the geometric aspect of the latent space, modeling it as a hypersphere, allowing for the introduction of sophisticated interpolation techniques. We also propose an algorithm for optimizing the latent variables directly on the learned manifold, improving convergence and performance. Finally, we introduce high-order interpolation techniques adapted for the hypersphere, allowing us to increase the solving temporal window, enhancing performance and efficiency.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104190"},"PeriodicalIF":4.3,"publicationDate":"2024-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142442023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel image inpainting method based on a modified Lengyel–Epstein model 基于改进的 Lengyel-Epstein 模型的新型图像着色方法
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-03 DOI: 10.1016/j.cviu.2024.104195
Jian Wang , Mengyu Luo , Xinlei Chen , Heming Xu , Junseok Kim
With the increasing popularity of digital images, developing advanced algorithms that can accurately reconstruct damaged images while maintaining high visual quality is crucial. Traditional image restoration algorithms often struggle with complex structures and details, while recent deep learning methods, though effective, face significant challenges related to high data dependency and computational costs. To resolve these challenges, we propose a novel image inpainting model, which is based on a modified Lengyel–Epstein (LE) model. We discretize the modified LE model by using an explicit Euler algorithm. A series of restoration experiments are conducted on various image types, including binary images, grayscale images, index images, and color images. The experimental results demonstrate the effectiveness and robustness of the method, and even under complex conditions of noise interference and local damage, the proposed method can exhibit excellent repair performance. To quantify the fidelity of these restored images, we use the peak signal-to-noise ratio (PSNR), a widely accepted metric in image processing. The calculation results further demonstrate the applicability of our model across different image types. Moreover, by evaluating CPU time, our method can achieve ideal repair results within a remarkably brief duration. The proposed method validates significant potential for real-world applications in diverse domains of image restoration and enhancement.
随着数字图像的日益普及,开发既能准确重建受损图像又能保持高视觉质量的先进算法至关重要。传统的图像修复算法往往难以处理复杂的结构和细节,而最新的深度学习方法虽然有效,却面临着与高数据依赖性和计算成本相关的重大挑战。为了解决这些难题,我们提出了一种基于改进的伦盖尔-爱泼斯坦(Lengyel-Epstein,LE)模型的新型图像内绘模型。我们使用显式欧拉算法对修改后的 LE 模型进行离散化。我们在二值图像、灰度图像、索引图像和彩色图像等各种类型的图像上进行了一系列修复实验。实验结果证明了该方法的有效性和鲁棒性,即使在噪声干扰和局部损伤的复杂条件下,所提出的方法也能表现出优异的修复性能。为了量化这些修复图像的保真度,我们使用了峰值信噪比 (PSNR),这是一个在图像处理中被广泛接受的指标。计算结果进一步证明了我们的模型适用于不同类型的图像。此外,通过评估 CPU 时间,我们的方法可以在极短的时间内实现理想的修复效果。所提出的方法验证了在图像修复和增强等不同领域的实际应用中的巨大潜力。
{"title":"A novel image inpainting method based on a modified Lengyel–Epstein model","authors":"Jian Wang ,&nbsp;Mengyu Luo ,&nbsp;Xinlei Chen ,&nbsp;Heming Xu ,&nbsp;Junseok Kim","doi":"10.1016/j.cviu.2024.104195","DOIUrl":"10.1016/j.cviu.2024.104195","url":null,"abstract":"<div><div>With the increasing popularity of digital images, developing advanced algorithms that can accurately reconstruct damaged images while maintaining high visual quality is crucial. Traditional image restoration algorithms often struggle with complex structures and details, while recent deep learning methods, though effective, face significant challenges related to high data dependency and computational costs. To resolve these challenges, we propose a novel image inpainting model, which is based on a modified Lengyel–Epstein (LE) model. We discretize the modified LE model by using an explicit Euler algorithm. A series of restoration experiments are conducted on various image types, including binary images, grayscale images, index images, and color images. The experimental results demonstrate the effectiveness and robustness of the method, and even under complex conditions of noise interference and local damage, the proposed method can exhibit excellent repair performance. To quantify the fidelity of these restored images, we use the peak signal-to-noise ratio (PSNR), a widely accepted metric in image processing. The calculation results further demonstrate the applicability of our model across different image types. Moreover, by evaluating CPU time, our method can achieve ideal repair results within a remarkably brief duration. The proposed method validates significant potential for real-world applications in diverse domains of image restoration and enhancement.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104195"},"PeriodicalIF":4.3,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WGS-YOLO: A real-time object detector based on YOLO framework for autonomous driving WGS-YOLO:基于 YOLO 框架的自动驾驶实时物体检测器
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-03 DOI: 10.1016/j.cviu.2024.104200
Shiqin Yue , Ziyi Zhang , Ying Shi , Yonghua Cai
The safety and reliability of autonomous driving depends on the precision and efficiency of object detection systems. In this paper, a refined adaptation of the YOLO architecture (WGS-YOLO) is developed to improve the detection of pedestrians and vehicles. Specifically, its information fusion is enhanced by incorporating the Weighted Efficient Layer Aggregation Network (W-ELAN) module, an innovative dynamic weighted feature fusion module using channel shuffling. Meanwhile, the computational demands and parameters of the proposed WGS-YOLO are significantly reduced by employing the Space-to-Depth Convolution (SPD-Conv) and the Grouped Spatial Pyramid Pooling (GSPP) modules that have been strategically designed. The performance of our model is evaluated with the BDD100k and DAIR-V2X-V datasets. In terms of mean Average Precision (mAP0.5), the proposed model outperforms the baseline Yolov7 by 12%. Furthermore, extensive experiments are conducted to verify our analysis and the model’s robustness across diverse scenarios.
自动驾驶的安全性和可靠性取决于物体检测系统的精度和效率。本文对 YOLO 架构(WGS-YOLO)进行了改进,以提高行人和车辆的检测能力。具体来说,通过加入加权高效层聚合网络(Weighted Efficient Layer Aggregation Network,W-ELAN)模块(一种使用信道洗牌的创新动态加权特征融合模块),增强了其信息融合能力。同时,通过采用战略性设计的空深卷积(SPD-Conv)和分组空间金字塔池化(GSPP)模块,大大降低了拟议 WGS-YOLO 的计算需求和参数。我们使用 BDD100k 和 DAIR-V2X-V 数据集评估了模型的性能。就平均精度(mAP0.5)而言,所提出的模型比基准 Yolov7 高出 12%。此外,我们还进行了大量实验,以验证我们的分析和模型在不同场景下的鲁棒性。
{"title":"WGS-YOLO: A real-time object detector based on YOLO framework for autonomous driving","authors":"Shiqin Yue ,&nbsp;Ziyi Zhang ,&nbsp;Ying Shi ,&nbsp;Yonghua Cai","doi":"10.1016/j.cviu.2024.104200","DOIUrl":"10.1016/j.cviu.2024.104200","url":null,"abstract":"<div><div>The safety and reliability of autonomous driving depends on the precision and efficiency of object detection systems. In this paper, a refined adaptation of the YOLO architecture (WGS-YOLO) is developed to improve the detection of pedestrians and vehicles. Specifically, its information fusion is enhanced by incorporating the Weighted Efficient Layer Aggregation Network (W-ELAN) module, an innovative dynamic weighted feature fusion module using channel shuffling. Meanwhile, the computational demands and parameters of the proposed WGS-YOLO are significantly reduced by employing the Space-to-Depth Convolution (SPD-Conv) and the Grouped Spatial Pyramid Pooling (GSPP) modules that have been strategically designed. The performance of our model is evaluated with the BDD100k and DAIR-V2X-V datasets. In terms of mean Average Precision (<span><math><msub><mrow><mtext>mAP</mtext></mrow><mrow><mn>0</mn><mo>.</mo><mn>5</mn></mrow></msub></math></span>), the proposed model outperforms the baseline Yolov7 by 12%. Furthermore, extensive experiments are conducted to verify our analysis and the model’s robustness across diverse scenarios.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104200"},"PeriodicalIF":4.3,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Found missing semantics: Supplemental prototype network for few-shot semantic segmentation 发现丢失的语义:用于少量语义分割的补充原型网络
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-03 DOI: 10.1016/j.cviu.2024.104191
Chen Liang, Shuang Bai
Few-shot semantic segmentation alleviates the problem of massive data requirements and high costs in semantic segmentation tasks. By learning from support set, few-shot semantic segmentation can segment new classes. However, existing few-shot semantic segmentation methods suffer from information loss during the process of mask average pooling. To address this problem, we propose a supplemental prototype network (SPNet). The SPNet aggregates the lost information from global prototypes to create a supplemental prototype, which enhances the segmentation performance for the current class. In addition, we utilize mutual attention to enhance the similarity between the support and the query feature maps, allowing the model to better identify the target to be segmented. Finally, we introduce a Self-correcting auxiliary, which utilizes the data more effectively to improve segmentation accuracy. We conducted extensive experiments on PASCAL-5i and COCO-20i, which demonstrated the effectiveness of SPNet. And our method achieved state-of-the-art results in the 1-shot and 5-shot semantic segmentation settings.
少量语义分割缓解了语义分割任务中的海量数据需求和高成本问题。通过从支持集学习,少量语义分割可以分割出新的类别。然而,现有的几次语义分割方法在掩码平均池化过程中存在信息丢失问题。为了解决这个问题,我们提出了一种补充原型网络(SPNet)。SPNet 汇集了全局原型丢失的信息,创建了一个补充原型,从而提高了当前类别的分割性能。此外,我们还利用相互关注来增强支持特征图和查询特征图之间的相似性,使模型能够更好地识别待分割目标。最后,我们引入了自校正辅助工具,它能更有效地利用数据来提高分割准确性。我们在 PASCAL-5i 和 COCO-20i 上进行了大量实验,证明了 SPNet 的有效性。我们的方法在 1 次和 5 次语义分割设置中都取得了一流的结果。
{"title":"Found missing semantics: Supplemental prototype network for few-shot semantic segmentation","authors":"Chen Liang,&nbsp;Shuang Bai","doi":"10.1016/j.cviu.2024.104191","DOIUrl":"10.1016/j.cviu.2024.104191","url":null,"abstract":"<div><div>Few-shot semantic segmentation alleviates the problem of massive data requirements and high costs in semantic segmentation tasks. By learning from support set, few-shot semantic segmentation can segment new classes. However, existing few-shot semantic segmentation methods suffer from information loss during the process of mask average pooling. To address this problem, we propose a supplemental prototype network (SPNet). The SPNet aggregates the lost information from global prototypes to create a supplemental prototype, which enhances the segmentation performance for the current class. In addition, we utilize mutual attention to enhance the similarity between the support and the query feature maps, allowing the model to better identify the target to be segmented. Finally, we introduce a Self-correcting auxiliary, which utilizes the data more effectively to improve segmentation accuracy. We conducted extensive experiments on PASCAL-5i and COCO-20i, which demonstrated the effectiveness of SPNet. And our method achieved state-of-the-art results in the 1-shot and 5-shot semantic segmentation settings.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104191"},"PeriodicalIF":4.3,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MultiSubjects: A multi-subject video dataset for single-person basketball action recognition from basketball gym 多主体:用于篮球馆单人篮球动作识别的多主体视频数据集
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-03 DOI: 10.1016/j.cviu.2024.104193
Zhijie Han , Wansong Qin , Yalu Wang , Qixiang Wang , Yongbin Shi
Computer vision technology is becoming a research focus in the field of basketball. Despite the abundance of datasets centered on basketball games, there remains a significant gap in the availability of a large-scale, multi-subject, and fine-grained dataset for the recognition of basketball actions in real-world sports scenarios, particularly for amateur players. Such datasets are crucial for advancing the application of computer vision tasks in the real world. To address this gap, we deployed multi-view cameras in a civilian basketball gym, constructed a real basketball data acquisition platform, and acquired a challenging multi-subject video dataset, named MultiSubjects. The MultiSubjects v1.0 dataset features a variety of ages, body types, attire, genders, and basketball actions, providing researchers with a high-quality and diverse resource of basketball action data. We collected a total of 1,000 distinct subjects from video data between September and December 2023, classified and labeled three basic basketball actions, and assigned a unique identity ID to each subject, provided a total of 6,144 video clips, 436,460 frames, and labeled 6,144 instances of actions with clear temporal boundaries using 436,460 human body bounding boxes. Additionally, complete frame-wise skeleton keypoint coordinates for the entire action are provided. We used some representative video action recognition algorithms as well as skeleton-based action recognition algorithms on the MultiSubjects v1.0 dataset and analyzed the results. The results confirm that the quality of our dataset surpasses that of popular video action recognition datasets, it also presents that skeleton-based action recognition remains a challenging task. The link to our dataset is: https://huggingface.co/datasets/Henu-Software/Henu-MultiSubjects.
计算机视觉技术正成为篮球领域的研究重点。尽管有大量以篮球比赛为中心的数据集,但大规模、多主体、细粒度的数据集在现实世界运动场景中的篮球动作识别(尤其是业余球员的动作识别)方面仍存在巨大差距。此类数据集对于推动计算机视觉任务在现实世界中的应用至关重要。为了填补这一空白,我们在民用篮球馆部署了多视角摄像机,构建了一个真实的篮球数据采集平台,并获取了一个具有挑战性的多主体视频数据集,命名为 "MultiSubjects"。MultiSubjects v1.0 数据集包含各种年龄、体型、服装、性别和篮球动作,为研究人员提供了高质量、多样化的篮球动作数据资源。我们从 2023 年 9 月至 12 月期间的视频数据中收集了总共 1000 个不同的受试者,对三个基本篮球动作进行了分类和标记,并为每个受试者分配了唯一的身份 ID,共提供了 6,144 个视频片段、436,460 个帧,并使用 436,460 个人体边界框标记了 6,144 个具有明确时间界限的动作实例。此外,我们还提供了整个动作的完整帧骨架关键点坐标。我们在多主体 v1.0 数据集上使用了一些具有代表性的视频动作识别算法以及基于骨架的动作识别算法,并对结果进行了分析。结果证实,我们数据集的质量超过了流行的视频动作识别数据集,同时也表明基于骨架的动作识别仍然是一项具有挑战性的任务。我们数据集的链接是:https://huggingface.co/datasets/Henu-Software/Henu-MultiSubjects。
{"title":"MultiSubjects: A multi-subject video dataset for single-person basketball action recognition from basketball gym","authors":"Zhijie Han ,&nbsp;Wansong Qin ,&nbsp;Yalu Wang ,&nbsp;Qixiang Wang ,&nbsp;Yongbin Shi","doi":"10.1016/j.cviu.2024.104193","DOIUrl":"10.1016/j.cviu.2024.104193","url":null,"abstract":"<div><div>Computer vision technology is becoming a research focus in the field of basketball. Despite the abundance of datasets centered on basketball games, there remains a significant gap in the availability of a large-scale, multi-subject, and fine-grained dataset for the recognition of basketball actions in real-world sports scenarios, particularly for amateur players. Such datasets are crucial for advancing the application of computer vision tasks in the real world. To address this gap, we deployed multi-view cameras in a civilian basketball gym, constructed a real basketball data acquisition platform, and acquired a challenging multi-subject video dataset, named MultiSubjects. The MultiSubjects v1.0 dataset features a variety of ages, body types, attire, genders, and basketball actions, providing researchers with a high-quality and diverse resource of basketball action data. We collected a total of 1,000 distinct subjects from video data between September and December 2023, classified and labeled three basic basketball actions, and assigned a unique identity ID to each subject, provided a total of 6,144 video clips, 436,460 frames, and labeled 6,144 instances of actions with clear temporal boundaries using 436,460 human body bounding boxes. Additionally, complete frame-wise skeleton keypoint coordinates for the entire action are provided. We used some representative video action recognition algorithms as well as skeleton-based action recognition algorithms on the MultiSubjects v1.0 dataset and analyzed the results. The results confirm that the quality of our dataset surpasses that of popular video action recognition datasets, it also presents that skeleton-based action recognition remains a challenging task. The link to our dataset is: <span><span>https://huggingface.co/datasets/Henu-Software/Henu-MultiSubjects</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104193"},"PeriodicalIF":4.3,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distance-based loss function for deep feature space learning of convolutional neural networks 基于距离的卷积神经网络深度特征空间学习损失函数
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-28 DOI: 10.1016/j.cviu.2024.104184
Eduardo S. Ribeiro , Lourenço R.G. Araújo , Gabriel T.L. Chaves , Antônio P. Braga
Convolutional Neural Networks (CNNs) have been on the forefront of neural network research in recent years. Their breakthrough performance in fields such as image classification has gathered efforts in the development of new CNN-based architectures, but recently more attention has been directed to the study of new loss functions. Softmax loss remains the most popular loss function due mainly to its efficiency in class separation, but the function is unsatisfactory in terms of intra-class compactness. While some studies have addressed this problem, most solutions attempt to refine softmax loss or combine it with other approaches. We present a novel loss function based on distance matrices (LDMAT), softmax independent, that maximizes interclass distance and minimizes intraclass distance. The loss function operates directly on deep features, allowing their use on arbitrary classifiers. LDMAT minimizes the distance between two distance matrices, one constructed with the model’s deep features and the other calculated from the labels. The use of a distance matrix in the loss function allows a two-dimensional representation of features and imposes a fixed distance between classes, while improving intra-class compactness. A regularization method applied to the distance matrix of labels is also presented, that allows a degree of relaxation of the solution and leads to a better spreading of features in the separation space. Efficient feature extraction was observed on datasets such as MNIST, CIFAR10 and CIFAR100.
卷积神经网络(CNN)近年来一直处于神经网络研究的前沿。它们在图像分类等领域的突破性表现为开发基于 CNN 的新架构集聚了力量,但最近更多的注意力被引导到新损失函数的研究上。Softmax 损失函数仍然是最受欢迎的损失函数,这主要是因为它在类分离方面的高效性,但该函数在类内紧凑性方面并不令人满意。虽然一些研究已经解决了这一问题,但大多数解决方案都试图改进 softmax 损失函数或将其与其他方法相结合。我们提出了一种基于距离矩阵(LDMAT)、独立于 softmax 的新型损失函数,它能最大化类间距离,最小化类内距离。该损失函数直接作用于深度特征,可用于任意分类器。LDMAT 将两个距离矩阵之间的距离最小化,其中一个距离矩阵由模型的深度特征构建,另一个距离矩阵由标签计算得出。在损失函数中使用距离矩阵可实现特征的二维表示,并在改善类内紧凑性的同时,强加类之间的固定距离。此外,还介绍了一种应用于标签距离矩阵的正则化方法,这种方法可以在一定程度上放松解决方案,并使特征在分离空间中得到更好的分布。在 MNIST、CIFAR10 和 CIFAR100 等数据集上观察到了高效的特征提取。
{"title":"Distance-based loss function for deep feature space learning of convolutional neural networks","authors":"Eduardo S. Ribeiro ,&nbsp;Lourenço R.G. Araújo ,&nbsp;Gabriel T.L. Chaves ,&nbsp;Antônio P. Braga","doi":"10.1016/j.cviu.2024.104184","DOIUrl":"10.1016/j.cviu.2024.104184","url":null,"abstract":"<div><div>Convolutional Neural Networks (CNNs) have been on the forefront of neural network research in recent years. Their breakthrough performance in fields such as image classification has gathered efforts in the development of new CNN-based architectures, but recently more attention has been directed to the study of new loss functions. Softmax loss remains the most popular loss function due mainly to its efficiency in class separation, but the function is unsatisfactory in terms of intra-class compactness. While some studies have addressed this problem, most solutions attempt to refine softmax loss or combine it with other approaches. We present a novel loss function based on distance matrices (LDMAT), softmax independent, that maximizes interclass distance and minimizes intraclass distance. The loss function operates directly on deep features, allowing their use on arbitrary classifiers. LDMAT minimizes the distance between two distance matrices, one constructed with the model’s deep features and the other calculated from the labels. The use of a distance matrix in the loss function allows a two-dimensional representation of features and imposes a fixed distance between classes, while improving intra-class compactness. A regularization method applied to the distance matrix of labels is also presented, that allows a degree of relaxation of the solution and leads to a better spreading of features in the separation space. Efficient feature extraction was observed on datasets such as MNIST, CIFAR10 and CIFAR100.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104184"},"PeriodicalIF":4.3,"publicationDate":"2024-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient degradation representation learning network for remote sensing image super-resolution 用于遥感图像超分辨率的高效降级表示学习网络
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-28 DOI: 10.1016/j.cviu.2024.104182
Xuan Wang , Lijun Sun , Jinglei Yi , Yongchao Song , Qiang Zheng , Abdellah Chehri
The advancements in convolutional neural networks have led to significant progress in image super-resolution (SR) techniques. Nevertheless, it is crucial to acknowledge that current SR methods operate under the assumption of bicubic downsampling as a degradation factor in low-resolution (LR) images and train models accordingly. However, this approach does not account for the unknown degradation patterns present in real-world scenes. To address this problem, we propose an efficient degradation representation learning network (EDRLN). Specifically, we adopt a contrast learning approach, which enables the model to distinguish and learn various degradation representations in realistic images to obtain critical degradation information. We also introduce streamlined and efficient pixel attention to strengthen the feature extraction capability of the model. In addition, we optimize our model with mutual affine convolution layers instead of ordinary convolution layers to make it more lightweight while minimizing performance loss. Experimental results on remote sensing and benchmark datasets show that our proposed EDRLN exhibits good performance for different degradation scenarios, while the lightweight version minimizes the performance loss as much as possible. The Code will be available at: https://github.com/Leilei11111/EDRLN.
卷积神经网络的进步使图像超分辨率(SR)技术取得了重大进展。然而,必须承认的是,当前的超分辨率方法是在假设低分辨率(LR)图像的降解因素为双三次降采样的情况下运行的,并据此训练模型。然而,这种方法并没有考虑到真实世界场景中存在的未知降解模式。为了解决这个问题,我们提出了一种高效降解表示学习网络(EDRLN)。具体来说,我们采用了一种对比学习方法,使模型能够区分和学习现实图像中的各种退化表征,从而获得关键的退化信息。我们还引入了精简高效的像素关注,以加强模型的特征提取能力。此外,我们用互仿卷积层代替普通卷积层对模型进行了优化,使其更加轻便,同时将性能损失降到最低。在遥感和基准数据集上的实验结果表明,我们提出的 EDRLN 在不同的退化场景下都表现出了良好的性能,而轻量级版本则尽可能减少了性能损失。代码见:https://github.com/Leilei11111/EDRLN。
{"title":"Efficient degradation representation learning network for remote sensing image super-resolution","authors":"Xuan Wang ,&nbsp;Lijun Sun ,&nbsp;Jinglei Yi ,&nbsp;Yongchao Song ,&nbsp;Qiang Zheng ,&nbsp;Abdellah Chehri","doi":"10.1016/j.cviu.2024.104182","DOIUrl":"10.1016/j.cviu.2024.104182","url":null,"abstract":"<div><div>The advancements in convolutional neural networks have led to significant progress in image super-resolution (SR) techniques. Nevertheless, it is crucial to acknowledge that current SR methods operate under the assumption of bicubic downsampling as a degradation factor in low-resolution (LR) images and train models accordingly. However, this approach does not account for the unknown degradation patterns present in real-world scenes. To address this problem, we propose an efficient degradation representation learning network (EDRLN). Specifically, we adopt a contrast learning approach, which enables the model to distinguish and learn various degradation representations in realistic images to obtain critical degradation information. We also introduce streamlined and efficient pixel attention to strengthen the feature extraction capability of the model. In addition, we optimize our model with mutual affine convolution layers instead of ordinary convolution layers to make it more lightweight while minimizing performance loss. Experimental results on remote sensing and benchmark datasets show that our proposed EDRLN exhibits good performance for different degradation scenarios, while the lightweight version minimizes the performance loss as much as possible. The Code will be available at: <span><span>https://github.com/Leilei11111/EDRLN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104182"},"PeriodicalIF":4.3,"publicationDate":"2024-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An efficient feature reuse distillation network for lightweight image super-resolution 用于轻量级图像超分辨率的高效特征重用蒸馏网络
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-27 DOI: 10.1016/j.cviu.2024.104178
Chunying Liu , Guangwei Gao , Fei Wu , Zhenhua Guo , Yi Yu
In recent research, single-image super-resolution (SISR) using deep Convolutional Neural Networks (CNN) has seen significant advancements. While previous methods excelled at learning complex mappings between low-resolution (LR) and high-resolution (HR) images, they often required substantial computational and memory resources. We propose the Efficient Feature Reuse Distillation Network (EFRDN) to alleviate these challenges. EFRDN primarily comprises Asymmetric Convolutional Distillation Modules (ACDM), incorporating the Multiple Self-Calibrating Convolution (MSCC) units for spatial and channel feature extraction. It includes an Asymmetric Convolution Residual Block (ACRB) to enhance the skeleton information of the square convolution kernel and a Feature Fusion Lattice Block (FFLB) to convert low-order input signals into higher-order representations. Introducing a Transformer module for global features, we enhance feature reuse and gradient flow, improving model performance and efficiency. Extensive experimental results demonstrate that EFRDN outperforms existing methods in performance while conserving computing and memory resources.
近年来,使用深度卷积神经网络(CNN)的单图像超分辨率(SISR)研究取得了重大进展。虽然以前的方法擅长学习低分辨率(LR)和高分辨率(HR)图像之间的复杂映射,但它们往往需要大量的计算和内存资源。我们提出了高效特征重用蒸馏网络(EFRDN)来缓解这些挑战。EFRDN 主要由非对称卷积蒸馏模块(ACDM)组成,其中包含用于空间和信道特征提取的多重自校准卷积(MSCC)单元。它包括一个非对称卷积残差块(ACRB),用于增强方形卷积核的骨架信息,以及一个特征融合网格块(FFLB),用于将低阶输入信号转换为高阶表示。通过引入全局特征变换器模块,我们增强了特征重用和梯度流,从而提高了模型的性能和效率。广泛的实验结果表明,EFRDN 的性能优于现有方法,同时节约了计算和内存资源。
{"title":"An efficient feature reuse distillation network for lightweight image super-resolution","authors":"Chunying Liu ,&nbsp;Guangwei Gao ,&nbsp;Fei Wu ,&nbsp;Zhenhua Guo ,&nbsp;Yi Yu","doi":"10.1016/j.cviu.2024.104178","DOIUrl":"10.1016/j.cviu.2024.104178","url":null,"abstract":"<div><div>In recent research, single-image super-resolution (SISR) using deep Convolutional Neural Networks (CNN) has seen significant advancements. While previous methods excelled at learning complex mappings between low-resolution (LR) and high-resolution (HR) images, they often required substantial computational and memory resources. We propose the Efficient Feature Reuse Distillation Network (EFRDN) to alleviate these challenges. EFRDN primarily comprises Asymmetric Convolutional Distillation Modules (ACDM), incorporating the Multiple Self-Calibrating Convolution (MSCC) units for spatial and channel feature extraction. It includes an Asymmetric Convolution Residual Block (ACRB) to enhance the skeleton information of the square convolution kernel and a Feature Fusion Lattice Block (FFLB) to convert low-order input signals into higher-order representations. Introducing a Transformer module for global features, we enhance feature reuse and gradient flow, improving model performance and efficiency. Extensive experimental results demonstrate that EFRDN outperforms existing methods in performance while conserving computing and memory resources.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104178"},"PeriodicalIF":4.3,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty guided test-time training for face forgery detection 不确定性指导下的人脸伪造检测测试时间训练
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-27 DOI: 10.1016/j.cviu.2024.104185
Pengxiang Xu, Yang He, Jian Yang, Shanshan Zhang
The rapid development of generative image modeling poses security risks of spreading unreal visual information, even though those techniques make a lot of applications possible in positive aspects. To provide alerts and maintain a secure social environment, forgery detection has been an urgent and crucial solution to deal with this situation and try to avoid any negative effects, especially for human faces, owing to potential severe results when malicious creators spread disinformation widely. In spite of the success of recent works w.r.t. model design and feature engineering, detecting face forgery from novel image creation methods or data distributions remains unresolved, because well-trained models are typically not robust to the distribution shift during test-time. In this work, we aim to alleviate the sensitivity of an existing face forgery detector to new domains, and then boost real-world detection under unknown test situations. In specific, we leverage test examples, selected by uncertainty values, to fine-tune the model before making a final prediction. Therefore, it leads to a test-time training based approach for face forgery detection, that our framework incorporates an uncertainty-driven test sample selection with self-training to adapt a classifier onto target domains. To demonstrate the effectiveness of our framework and compare with previous methods, we conduct extensive experiments on public datasets, including FaceForensics++, Celeb-DF-v2, ForgeryNet and DFDC. Our results clearly show that the proposed framework successfully improves many state-of-the-art methods in terms of better overall performance as well as stronger robustness to novel data distributions.
生成图像建模技术的快速发展带来了传播虚假视觉信息的安全风险,尽管这些技术在积极方面使许多应用成为可能。为了提供警示并维护安全的社会环境,伪造检测已成为应对这种情况的一个紧迫而关键的解决方案,并努力避免任何负面影响,尤其是对人脸的影响,因为当恶意制造者广泛传播虚假信息时,可能会造成严重后果。尽管最近的工作在模型设计和特征工程方面取得了成功,但从新的图像创建方法或数据分布中检测人脸伪造仍然是一个悬而未决的问题,因为训练有素的模型通常对测试期间的分布变化不具有鲁棒性。在这项工作中,我们的目标是减轻现有人脸伪造检测器对新领域的敏感性,然后提高未知测试环境下的实际检测能力。具体来说,我们利用根据不确定值选择的测试实例,在做出最终预测之前对模型进行微调。因此,我们的框架将不确定性驱动的测试样本选择与自我训练相结合,使分类器适应目标领域,从而形成了一种基于测试时间训练的人脸伪造检测方法。为了证明我们的框架的有效性并与之前的方法进行比较,我们在公共数据集上进行了广泛的实验,包括 FaceForensics++、Celeb-DF-v2、ForgeryNet 和 DFDC。我们的结果清楚地表明,所提出的框架成功地改进了许多最先进的方法,不仅整体性能更好,而且对新的数据分布具有更强的鲁棒性。
{"title":"Uncertainty guided test-time training for face forgery detection","authors":"Pengxiang Xu,&nbsp;Yang He,&nbsp;Jian Yang,&nbsp;Shanshan Zhang","doi":"10.1016/j.cviu.2024.104185","DOIUrl":"10.1016/j.cviu.2024.104185","url":null,"abstract":"<div><div>The rapid development of generative image modeling poses security risks of spreading unreal visual information, even though those techniques make a lot of applications possible in positive aspects. To provide alerts and maintain a secure social environment, forgery detection has been an urgent and crucial solution to deal with this situation and try to avoid any negative effects, especially for human faces, owing to potential severe results when malicious creators spread disinformation widely. In spite of the success of recent works w.r.t. model design and feature engineering, detecting face forgery from novel image creation methods or data distributions remains unresolved, because well-trained models are typically not robust to the distribution shift during test-time. In this work, we aim to alleviate the sensitivity of an existing face forgery detector to new domains, and then boost real-world detection under unknown test situations. In specific, we leverage test examples, selected by uncertainty values, to fine-tune the model before making a final prediction. Therefore, it leads to a test-time training based approach for face forgery detection, that our framework incorporates an uncertainty-driven test sample selection with self-training to adapt a classifier onto target domains. To demonstrate the effectiveness of our framework and compare with previous methods, we conduct extensive experiments on public datasets, including FaceForensics++, Celeb-DF-v2, ForgeryNet and DFDC. Our results clearly show that the proposed framework successfully improves many state-of-the-art methods in terms of better overall performance as well as stronger robustness to novel data distributions.</div></div>","PeriodicalId":50633,"journal":{"name":"Computer Vision and Image Understanding","volume":"249 ","pages":"Article 104185"},"PeriodicalIF":4.3,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142422250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer Vision and Image Understanding
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1