首页 > 最新文献

IET Computer Vision最新文献

英文 中文
Guest Editorial: Learning from limited annotations for computer vision tasks 客座编辑:从计算机视觉任务的有限注释中学习
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-16 DOI: 10.1049/cvi2.12229
Yazhou Yao, Wenguan Wang, Qiang Wu, Dongfang Liu, Jin Zheng

The past decade has witnessed remarkable achievements in computer vision, owing to the fast development of deep learning. With the advancement of computing power and deep learning algorithms, we can process and apply millions or even hundreds of millions of large-scale data to train robust and advanced deep learning models. In spite of the impressive success, current deep learning methods tend to rely on massive annotated training data and lack the capability of learning from limited exemplars.

However, constructing a million-scale annotated dataset like ImageNet is time-consuming, labour-intensive and even infeasible in many applications. In certain fields, very limited annotated examples can be gathered due to various reasons such as privacy or ethical issues. Consequently, one of the pressing challenges in computer vision is to develop approaches that are capable of learning from limited annotated data. The purpose of this Special Issue is to collect high-quality articles on learning from limited annotations for computer vision tasks (e.g. image classification, object detection, semantic segmentation, instance segmentation and many others), publish new ideas, theories, solutions and insights on this topic and showcase their applications.

In this Special Issue we received 29 papers, all of which underwent peer review. Of the 29 originally submitted papers, 9 have been accepted.

The nine accepted papers can be clustered into two main categories: theoretical and applications. The papers that fall into the first category are by Liu et al., Li et al. and He et al. The second category of papers offers a direct solution to various computer vision tasks. These papers are by Ma et al., Wu et al., Rao et al., Sun et al., Hou et al. and Gong et al. A brief presentation of each of the papers in this Special Issue follows.

All of the papers selected for this Special Issue show that the field of learning from limited annotations for computer vision tasks is steadily moving forward. The possibility of a weakly supervised learning paradigm will remain a source of inspiration for new techniques in the years to come.

由于深度学习的快速发展,过去十年在计算机视觉方面取得了显著成就。随着计算能力和深度学习算法的进步,我们可以处理和应用数百万甚至数亿的大规模数据,以训练健壮、先进的深度学习模型。尽管取得了令人印象深刻的成功,但当前的深度学习方法往往依赖于大量注释的训练数据,缺乏从有限的样本中学习的能力。然而,构建像ImageNet这样的百万规模注释数据集是耗时、劳动密集型的,在许多应用中甚至是不可行的。在某些领域,由于隐私或道德问题等各种原因,可以收集到非常有限的注释示例。因此,计算机视觉的一个紧迫挑战是开发能够从有限的注释数据中学习的方法。本期特刊的目的是收集关于从计算机视觉任务(如图像分类、对象检测、语义分割、实例分割等)的有限注释中学习的高质量文章,发表有关该主题的新思想、理论、解决方案和见解,并展示其应用。在本期特刊中,我们收到了29篇论文,所有论文都经过了同行评审。在最初提交的29篇论文中,有9篇已被接受。九篇被接受的论文可以分为两大类:理论和应用。属于第一类的论文是刘等人。,李等。和He等人。第二类论文提供了各种计算机视觉任务的直接解决方案。这些论文由Ma等人。,吴等。,Rao等人。,Sun等人。,Hou等人。Gong等人。以下是本期特刊中每一篇论文的简要介绍。本期特刊所选的所有论文都表明,从计算机视觉任务的有限注释中学习的领域正在稳步发展。弱监督学习范式的可能性在未来几年仍将是新技术的灵感来源。
{"title":"Guest Editorial: Learning from limited annotations for computer vision tasks","authors":"Yazhou Yao,&nbsp;Wenguan Wang,&nbsp;Qiang Wu,&nbsp;Dongfang Liu,&nbsp;Jin Zheng","doi":"10.1049/cvi2.12229","DOIUrl":"https://doi.org/10.1049/cvi2.12229","url":null,"abstract":"<p>The past decade has witnessed remarkable achievements in computer vision, owing to the fast development of deep learning. With the advancement of computing power and deep learning algorithms, we can process and apply millions or even hundreds of millions of large-scale data to train robust and advanced deep learning models. In spite of the impressive success, current deep learning methods tend to rely on massive annotated training data and lack the capability of learning from limited exemplars.</p><p>However, constructing a million-scale annotated dataset like ImageNet is time-consuming, labour-intensive and even infeasible in many applications. In certain fields, very limited annotated examples can be gathered due to various reasons such as privacy or ethical issues. Consequently, one of the pressing challenges in computer vision is to develop approaches that are capable of learning from limited annotated data. The purpose of this Special Issue is to collect high-quality articles on learning from limited annotations for computer vision tasks (e.g. image classification, object detection, semantic segmentation, instance segmentation and many others), publish new ideas, theories, solutions and insights on this topic and showcase their applications.</p><p>In this Special Issue we received 29 papers, all of which underwent peer review. Of the 29 originally submitted papers, 9 have been accepted.</p><p>The nine accepted papers can be clustered into two main categories: theoretical and applications. The papers that fall into the first category are by Liu et al., Li et al. and He et al. The second category of papers offers a direct solution to various computer vision tasks. These papers are by Ma et al., Wu et al., Rao et al., Sun et al., Hou et al. and Gong et al. A brief presentation of each of the papers in this Special Issue follows.</p><p>All of the papers selected for this Special Issue show that the field of learning from limited annotations for computer vision tasks is steadily moving forward. The possibility of a weakly supervised learning paradigm will remain a source of inspiration for new techniques in the years to come.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"17 5","pages":"509-512"},"PeriodicalIF":1.7,"publicationDate":"2023-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12229","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50151226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Point completion by a Stack-Style Folding Network with multi-scaled graphical features 具有多尺度图形特征的堆叠式折叠网络的点补全
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-11 DOI: 10.1049/cvi2.12196
Yunbo Rao, Ping Xu, Shaoning Zeng, Jianping Gou
Point cloud completion is prevalent due to the insufficient results from current point cloud acquisition equipments, where a large number of point data failed to represent a relatively complete shape. Existing point cloud completion algorithms, mostly encoder‐decoder structures with grids transform (also presented as folding operation), can hardly obtain a persuasive representation of input clouds due to the issue that their bottleneck‐shape result cannot tell a precise relationship between the global and local structures. For this reason, this article proposes a novel point cloud completion model based on a Stack‐Style Folding Network (SSFN). Firstly, to enhance the deep latent feature extraction, SSFN enhances the exploitation of shape feature extractor by integrating both low‐level point feature and high‐level graphical feature. Next, a precise presentation is obtained from a high dimensional semantic space to improve the reconstruction ability. Finally, a refining module is designed to make a more evenly distributed result. Experimental results shows that our SSFN produces the most promising results of multiple representative metrics with a smaller scale parameters than current models.
由于当前点云采集设备的结果不充分,大量的点数据不能代表一个相对完整的形状,所以点云补全很普遍。现有的点云补全算法,主要是具有网格变换的编码器-解码器结构(也称为折叠操作),由于其瓶颈形状的结果无法告诉全局和局部结构之间的精确关系,因此很难获得有说服力的输入云表示。为此,本文提出了一种基于堆栈式折叠网络(SSFN)的点云补全模型。首先,为了增强深度潜在特征提取,SSFN通过融合低水平点特征和高水平图形特征来增强形状特征提取器的开发;其次,从高维语义空间获得精确的表示,提高重构能力;最后,设计了细化模块,使结果分布更加均匀。实验结果表明,我们的SSFN产生了比现有模型更小尺度参数的多个代表性指标的最有希望的结果。
{"title":"Point completion by a Stack-Style Folding Network with multi-scaled graphical features","authors":"Yunbo Rao, Ping Xu, Shaoning Zeng, Jianping Gou","doi":"10.1049/cvi2.12196","DOIUrl":"https://doi.org/10.1049/cvi2.12196","url":null,"abstract":"Point cloud completion is prevalent due to the insufficient results from current point cloud acquisition equipments, where a large number of point data failed to represent a relatively complete shape. Existing point cloud completion algorithms, mostly encoder‐decoder structures with grids transform (also presented as folding operation), can hardly obtain a persuasive representation of input clouds due to the issue that their bottleneck‐shape result cannot tell a precise relationship between the global and local structures. For this reason, this article proposes a novel point cloud completion model based on a Stack‐Style Folding Network (SSFN). Firstly, to enhance the deep latent feature extraction, SSFN enhances the exploitation of shape feature extractor by integrating both low‐level point feature and high‐level graphical feature. Next, a precise presentation is obtained from a high dimensional semantic space to improve the reconstruction ability. Finally, a refining module is designed to make a more evenly distributed result. Experimental results shows that our SSFN produces the most promising results of multiple representative metrics with a smaller scale parameters than current models.","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"17 1","pages":"576-585"},"PeriodicalIF":1.7,"publicationDate":"2023-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"57700209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Point completion by a Stack-Style Folding Network with multi-scaled graphical features 具有多比例图形特征的堆叠式折叠网络的点完成
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-11 DOI: 10.1049/cvi2.12196
Yunbo Rao, Ping Xu, Shaoning Zeng, Jianping Gou

Point cloud completion is prevalent due to the insufficient results from current point cloud acquisition equipments, where a large number of point data failed to represent a relatively complete shape. Existing point cloud completion algorithms, mostly encoder-decoder structures with grids transform (also presented as folding operation), can hardly obtain a persuasive representation of input clouds due to the issue that their bottleneck-shape result cannot tell a precise relationship between the global and local structures. For this reason, this article proposes a novel point cloud completion model based on a Stack-Style Folding Network (SSFN). Firstly, to enhance the deep latent feature extraction, SSFN enhances the exploitation of shape feature extractor by integrating both low-level point feature and high-level graphical feature. Next, a precise presentation is obtained from a high dimensional semantic space to improve the reconstruction ability. Finally, a refining module is designed to make a more evenly distributed result. Experimental results shows that our SSFN produces the most promising results of multiple representative metrics with a smaller scale parameters than current models.

由于目前点云采集设备的结果不充分,大量的点数据无法代表相对完整的形状,点云完成普遍存在。现有的点云完成算法,主要是具有网格变换的编码器-解码器结构(也称为折叠操作),由于其瓶颈形状结果不能说明全局结构和局部结构之间的精确关系,因此很难获得输入云的有说服力的表示。为此,本文提出了一种新的基于堆叠式折叠网络(SSFN)的点云完成模型。首先,为了增强深层潜在特征提取,SSFN通过集成低级点特征和高级图形特征来增强形状特征提取器的开发。接下来,从高维语义空间中获得精确的表示,以提高重建能力。最后,设计了一个细化模块,使结果分布更加均匀。实验结果表明,我们的SSFN以比当前模型更小的尺度参数产生了多个代表性度量的最有希望的结果。
{"title":"Point completion by a Stack-Style Folding Network with multi-scaled graphical features","authors":"Yunbo Rao,&nbsp;Ping Xu,&nbsp;Shaoning Zeng,&nbsp;Jianping Gou","doi":"10.1049/cvi2.12196","DOIUrl":"https://doi.org/10.1049/cvi2.12196","url":null,"abstract":"<p>Point cloud completion is prevalent due to the insufficient results from current point cloud acquisition equipments, where a large number of point data failed to represent a relatively complete shape. Existing point cloud completion algorithms, mostly encoder-decoder structures with grids transform (also presented as folding operation), can hardly obtain a persuasive representation of input clouds due to the issue that their bottleneck-shape result cannot tell a precise relationship between the global and local structures. For this reason, this article proposes a novel point cloud completion model based on a Stack-Style Folding Network (SSFN). Firstly, to enhance the deep latent feature extraction, SSFN enhances the exploitation of shape feature extractor by integrating both low-level point feature and high-level graphical feature. Next, a precise presentation is obtained from a high dimensional semantic space to improve the reconstruction ability. Finally, a refining module is designed to make a more evenly distributed result. Experimental results shows that our SSFN produces the most promising results of multiple representative metrics with a smaller scale parameters than current models.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"17 5","pages":"576-585"},"PeriodicalIF":1.7,"publicationDate":"2023-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12196","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50128438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Low-rank preserving embedding regression for robust image feature extraction 用于稳健图像特征提取的低秩保持嵌入回归
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-08 DOI: 10.1049/cvi2.12228
Tao Zhang, Chen-Feng Long, Yang-Jun Deng, Wei-Ye Wang, Si-Qiao Tan, Heng-Chao Li

Although low-rank representation (LRR)-based subspace learning has been widely applied for feature extraction in computer vision, how to enhance the discriminability of the low-dimensional features extracted by LRR based subspace learning methods is still a problem that needs to be further investigated. Therefore, this paper proposes a novel low-rank preserving embedding regression (LRPER) method by integrating LRR, linear regression, and projection learning into a unified framework. In LRPER, LRR can reveal the underlying structure information to strengthen the robustness of projection learning. The robust metric L2,1-norm is employed to measure the low-rank reconstruction error and regression loss for moulding the noise and occlusions. An embedding regression is proposed to make full use of the prior information for improving the discriminability of the learned projection. In addition, an alternative iteration algorithm is designed to optimise the proposed model, and the computational complexity of the optimisation algorithm is briefly analysed. The convergence of the optimisation algorithm is theoretically and numerically studied. At last, extensive experiments on four types of image datasets are carried out to demonstrate the effectiveness of LRPER, and the experimental results demonstrate that LRPER performs better than some state-of-the-art feature extraction methods.

尽管基于低秩表示(LRR)的子空间学习已被广泛应用于计算机视觉中的特征提取,但如何提高基于LRR的子空间方法提取的低维特征的可分辨性仍然是一个需要进一步研究的问题。因此,本文提出了一种新的低秩保持嵌入回归(LRPER)方法,将LRR、线性回归和投影学习集成到一个统一的框架中。在LRPER中,LRR可以揭示底层结构信息,以增强投影学习的鲁棒性。鲁棒度量L2,1范数用于测量低阶重建误差和回归损失,用于建模噪声和遮挡。提出了一种嵌入回归,以充分利用先验信息来提高学习投影的可分辨性。此外,设计了一种替代迭代算法来优化所提出的模型,并简要分析了优化算法的计算复杂性。对优化算法的收敛性进行了理论和数值研究。最后,在四种类型的图像数据集上进行了广泛的实验,以证明LRPER的有效性,实验结果表明LRPER比一些最先进的特征提取方法表现更好。
{"title":"Low-rank preserving embedding regression for robust image feature extraction","authors":"Tao Zhang,&nbsp;Chen-Feng Long,&nbsp;Yang-Jun Deng,&nbsp;Wei-Ye Wang,&nbsp;Si-Qiao Tan,&nbsp;Heng-Chao Li","doi":"10.1049/cvi2.12228","DOIUrl":"10.1049/cvi2.12228","url":null,"abstract":"<p>Although low-rank representation (LRR)-based subspace learning has been widely applied for feature extraction in computer vision, how to enhance the discriminability of the low-dimensional features extracted by LRR based subspace learning methods is still a problem that needs to be further investigated. Therefore, this paper proposes a novel low-rank preserving embedding regression (LRPER) method by integrating LRR, linear regression, and projection learning into a unified framework. In LRPER, LRR can reveal the underlying structure information to strengthen the robustness of projection learning. The robust metric <i>L</i><sub>2,1</sub>-norm is employed to measure the low-rank reconstruction error and regression loss for moulding the noise and occlusions. An embedding regression is proposed to make full use of the prior information for improving the discriminability of the learned projection. In addition, an alternative iteration algorithm is designed to optimise the proposed model, and the computational complexity of the optimisation algorithm is briefly analysed. The convergence of the optimisation algorithm is theoretically and numerically studied. At last, extensive experiments on four types of image datasets are carried out to demonstrate the effectiveness of LRPER, and the experimental results demonstrate that LRPER performs better than some state-of-the-art feature extraction methods.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"18 1","pages":"124-140"},"PeriodicalIF":1.7,"publicationDate":"2023-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12228","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46800153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Visual privacy behaviour recognition for social robots based on an improved generative adversarial network 基于改进生成对抗网络的社交机器人视觉隐私行为识别
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-04 DOI: 10.1049/cvi2.12231
Guanci Yang, Jiacheng Lin, Zhidong Su, Yang Li

Although social robots equipped with visual devices may leak user information, countermeasures for ensuring privacy are not readily available, making visual privacy protection problematic. In this article, a semi-supervised learning algorithm is proposed for visual privacy behaviour recognition based on an improved generative adversarial network for social robots; it is called PBR-GAN. A 9-layer residual generator network enhances the data quality, and a 10-layer discriminator network strengthens the feature extraction. A tailored objective function, loss function, and strategy are proposed to dynamically adjust the learning rate to guarantee high performance. A social robot platform and architecture for visual privacy recognition and protection are implemented. The recognition accuracy of the proposed PBR-GAN is compared with Inception_v3, SS-GAN, and SF-GAN. The average recognition accuracy of the proposed PBR-GAN is 85.91%, which is improved by 3.93%, 9.91%, and 1.73% compared with the performance of Inception_v3, SS-GAN, and SF-GAN respectively. Through a case study, seven situations are considered related to privacy at home, and develop training and test datasets with 8,720 and 1,280 images, respectively, are developed. The proposed PBR-GAN recognises the designed visual privacy information with an average accuracy of 89.91%.

尽管配备了视觉设备的社交机器人可能会泄露用户信息,但确保隐私的对策并不容易获得,这使得视觉隐私保护成为问题。本文提出了一种基于改进的社交机器人生成对抗性网络的视觉隐私行为识别半监督学习算法;它被称为PBR-GAN。9层残差生成器网络增强了数据质量,10层鉴别器网络增强了特征提取。提出了一种定制的目标函数、损失函数和策略来动态调整学习率,以确保高性能。实现了一个用于视觉隐私识别和保护的社交机器人平台和架构。将所提出的PBR‐GAN的识别精度与Inception_v3、SS‐GAN和SF‐GAN进行了比较。所提出的PBR-GAN的平均识别准确率为85.91%,与Inception_v3、SS‐GAN和SF‐GAN的性能相比,分别提高了3.93%、9.91%和1.73%。通过案例研究,考虑了七种与家庭隐私有关的情况,并分别开发了8720张和1280张图像的训练和测试数据集。所提出的PBR-GAN识别设计的视觉隐私信息的平均准确率为89.91%。
{"title":"Visual privacy behaviour recognition for social robots based on an improved generative adversarial network","authors":"Guanci Yang,&nbsp;Jiacheng Lin,&nbsp;Zhidong Su,&nbsp;Yang Li","doi":"10.1049/cvi2.12231","DOIUrl":"10.1049/cvi2.12231","url":null,"abstract":"<p>Although social robots equipped with visual devices may leak user information, countermeasures for ensuring privacy are not readily available, making visual privacy protection problematic. In this article, a semi-supervised learning algorithm is proposed for visual privacy behaviour recognition based on an improved generative adversarial network for social robots; it is called PBR-GAN. A 9-layer residual generator network enhances the data quality, and a 10-layer discriminator network strengthens the feature extraction. A tailored objective function, loss function, and strategy are proposed to dynamically adjust the learning rate to guarantee high performance. A social robot platform and architecture for visual privacy recognition and protection are implemented. The recognition accuracy of the proposed PBR-GAN is compared with Inception_v3, SS-GAN, and SF-GAN. The average recognition accuracy of the proposed PBR-GAN is 85.91%, which is improved by 3.93%, 9.91%, and 1.73% compared with the performance of Inception_v3, SS-GAN, and SF-GAN respectively. Through a case study, seven situations are considered related to privacy at home, and develop training and test datasets with 8,720 and 1,280 images, respectively, are developed. The proposed PBR-GAN recognises the designed visual privacy information with an average accuracy of 89.91%.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"18 1","pages":"110-123"},"PeriodicalIF":1.7,"publicationDate":"2023-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12231","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47731526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Determining the proper number of proposals for individual images 为单个图像确定适当数量的建议
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-03 DOI: 10.1049/cvi2.12230
Zihang He, Yong Li

The region proposal network is indispensable to two-stage object detection methods. It generates a fixed number of proposals that are to be classified and regressed by detection heads to produce detection boxes. However, the fixed number of proposals may be too large when an image contains only a few objects but too small when it contains much more objects. Considering this, the authors explored determining a proper number of proposals according to the number of objects in an image to reduce the computational cost while improving the detection accuracy. Since the number of ground truth objects is unknown at the inference stage, the authors designed a simple but effective module to predict the number of foreground regions, which will be substituted for the number of objects for determining the proposal number. Experimental results of various two-stage detection methods on different datasets, including MS-COCO, PASCAL VOC, and CrowdHuman showed that equipping the designed module increased the detection accuracy while decreasing the FLOPs of the detection head. For example, experimental results on the PASCAL VOC dataset showed that applying the designed module to Libra R-CNN and Grid R-CNN increased over 1.5 AP50 while decreasing the FLOPs of detection heads from 28.6 G to nearly 9.0 G.

区域建议网络对于两阶段目标检测方法是必不可少的。它生成固定数量的建议,这些建议将由检测头进行分类和回归,以产生检测盒。然而,当图像仅包含少数对象时,固定数量的建议可能太大,而当图像包含更多对象时,建议可能太小。考虑到这一点,作者探索根据图像中对象的数量确定适当数量的建议,以降低计算成本,同时提高检测精度。由于在推理阶段,地面实况对象的数量是未知的,作者设计了一个简单但有效的模块来预测前景区域的数量,该模块将取代对象的数量来确定提案数量。各种两阶段检测方法在不同数据集(包括MS‐COCO、PASCAL VOC和CrowdHuman)上的实验结果表明,配备所设计的模块提高了检测精度,同时降低了检测头的FLOP。例如,PASCAL VOC数据集的实验结果表明,将设计的模块应用于Libra R‐CNN和Grid R‐CNN时,AP50增加了1.5以上,同时检测头的FLOP从28.6 G降低到近9.0 G。
{"title":"Determining the proper number of proposals for individual images","authors":"Zihang He,&nbsp;Yong Li","doi":"10.1049/cvi2.12230","DOIUrl":"10.1049/cvi2.12230","url":null,"abstract":"<p>The region proposal network is indispensable to two-stage object detection methods. It generates a fixed number of proposals that are to be classified and regressed by detection heads to produce detection boxes. However, the fixed number of proposals may be too large when an image contains only a few objects but too small when it contains much more objects. Considering this, the authors explored determining a proper number of proposals according to the number of objects in an image to reduce the computational cost while improving the detection accuracy. Since the number of ground truth objects is unknown at the inference stage, the authors designed a simple but effective module to predict the number of foreground regions, which will be substituted for the number of objects for determining the proposal number. Experimental results of various two-stage detection methods on different datasets, including MS-COCO, PASCAL VOC, and CrowdHuman showed that equipping the designed module increased the detection accuracy while decreasing the FLOPs of the detection head. For example, experimental results on the PASCAL VOC dataset showed that applying the designed module to Libra R-CNN and Grid R-CNN increased over 1.5 AP<sub>50</sub> while decreasing the FLOPs of detection heads from 28.6 G to nearly 9.0 G.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"18 1","pages":"141-149"},"PeriodicalIF":1.7,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12230","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43408162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Zero-shot temporal event localisation: Label-free, training-free, domain-free 零时间事件定位:无标签、无训练、无领域
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-03 DOI: 10.1049/cvi2.12224
Li Sun, Ping Wang, Liuan Wang, Jun Sun, Takayuki Okatani
Temporal event localisation (TEL) has recently attracted increasing attention due to the rapid development of video platforms. Existing methods are based on either fully/weakly supervised or unsupervised learning, and thus they rely on expensive data annotation and time‐consuming training. Moreover, these models, which are trained on specific domain data, limit the model generalisation to data distribution shifts. To cope with these difficulties, the authors propose a zero‐shot TEL method that can operate without training data or annotations. Leveraging large‐scale vision and language pre‐trained models, for example, CLIP, we solve the two key problems: (1) how to find the relevant region where the event is likely to occur; (2) how to determine event duration after we find the relevant region. Query guided optimisation for local frame relevance relying on the query‐to‐frame relationship is proposed to find the most relevant frame region where the event is most likely to occur. Proposal generation method relying on the frame‐to‐frame relationship is proposed to determine the event duration. The authors also propose a greedy event sampling strategy to predict multiple durations with high reliability for the given event. The authors’ methodology is unique, offering a label‐free, training‐free, and domain‐free approach. It enables the application of TEL purely at the testing stage. The practical results show it achieves competitive performance on the standard Charades‐STA and ActivityCaptions datasets.
随着视频平台的快速发展,时间事件定位(TEL)越来越受到人们的关注。现有的方法要么基于完全/弱监督学习,要么基于无监督学习,因此它们依赖于昂贵的数据注释和耗时的训练。此外,这些模型是在特定领域数据上训练的,限制了模型的泛化到数据分布的变化。为了应对这些困难,作者提出了一种零射击TEL方法,该方法可以在没有训练数据或注释的情况下运行。利用大规模的视觉和语言预训练模型,例如CLIP,我们解决了两个关键问题:(1)如何找到事件可能发生的相关区域;(2)找到相关区域后,如何确定事件持续时间。提出了基于查询-帧关系的局部帧相关性的查询导向优化,以找到事件最有可能发生的最相关的帧区域。提出了一种基于帧间关系的提案生成方法来确定事件持续时间。作者还提出了一种贪婪事件采样策略,以高可靠性预测给定事件的多个持续时间。作者的方法是独一无二的,提供了一个标签自由,培训自由,和领域自由的方法。它使TEL的应用纯粹在测试阶段。实际结果表明,该方法在标准Charades - STA和ActivityCaptions数据集上取得了具有竞争力的性能。
{"title":"Zero-shot temporal event localisation: Label-free, training-free, domain-free","authors":"Li Sun, Ping Wang, Liuan Wang, Jun Sun, Takayuki Okatani","doi":"10.1049/cvi2.12224","DOIUrl":"https://doi.org/10.1049/cvi2.12224","url":null,"abstract":"Temporal event localisation (TEL) has recently attracted increasing attention due to the rapid development of video platforms. Existing methods are based on either fully/weakly supervised or unsupervised learning, and thus they rely on expensive data annotation and time‐consuming training. Moreover, these models, which are trained on specific domain data, limit the model generalisation to data distribution shifts. To cope with these difficulties, the authors propose a zero‐shot TEL method that can operate without training data or annotations. Leveraging large‐scale vision and language pre‐trained models, for example, CLIP, we solve the two key problems: (1) how to find the relevant region where the event is likely to occur; (2) how to determine event duration after we find the relevant region. Query guided optimisation for local frame relevance relying on the query‐to‐frame relationship is proposed to find the most relevant frame region where the event is most likely to occur. Proposal generation method relying on the frame‐to‐frame relationship is proposed to determine the event duration. The authors also propose a greedy event sampling strategy to predict multiple durations with high reliability for the given event. The authors’ methodology is unique, offering a label‐free, training‐free, and domain‐free approach. It enables the application of TEL purely at the testing stage. The practical results show it achieves competitive performance on the standard Charades‐STA and ActivityCaptions datasets.","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"17 1","pages":"599-613"},"PeriodicalIF":1.7,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"57700647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Zero-shot temporal event localisation: Label-free, training-free, domain-free 零样本时间事件本地化:无标签、无训练、无域
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-08-03 DOI: 10.1049/cvi2.12224
Li Sun, Ping Wang, Liuan Wang, Jun Sun, Takayuki Okatani

Temporal event localisation (TEL) has recently attracted increasing attention due to the rapid development of video platforms. Existing methods are based on either fully/weakly supervised or unsupervised learning, and thus they rely on expensive data annotation and time-consuming training. Moreover, these models, which are trained on specific domain data, limit the model generalisation to data distribution shifts. To cope with these difficulties, the authors propose a zero-shot TEL method that can operate without training data or annotations. Leveraging large-scale vision and language pre-trained models, for example, CLIP, we solve the two key problems: (1) how to find the relevant region where the event is likely to occur; (2) how to determine event duration after we find the relevant region. Query guided optimisation for local frame relevance relying on the query-to-frame relationship is proposed to find the most relevant frame region where the event is most likely to occur. Proposal generation method relying on the frame-to-frame relationship is proposed to determine the event duration. The authors also propose a greedy event sampling strategy to predict multiple durations with high reliability for the given event. The authors’ methodology is unique, offering a label-free, training-free, and domain-free approach. It enables the application of TEL purely at the testing stage. The practical results show it achieves competitive performance on the standard Charades-STA and ActivityCaptions datasets.

由于视频平台的快速发展,时间事件本地化(TEL)最近引起了越来越多的关注。现有的方法基于完全/弱监督或无监督的学习,因此它们依赖于昂贵的数据注释和耗时的训练。此外,这些基于特定领域数据训练的模型将模型泛化限制在数据分布变化上。为了应对这些困难,作者提出了一种零样本TEL方法,该方法可以在没有训练数据或注释的情况下操作。利用大规模的视觉和语言预训练模型,例如CLIP,我们解决了两个关键问题:(1)如何找到事件可能发生的相关区域;(2) 如何在找到相关区域后确定事件持续时间。提出了基于查询-帧关系的局部帧相关性的查询导向优化,以找到事件最有可能发生的最相关的帧区域。提出了一种基于帧间关系的建议生成方法来确定事件持续时间。作者还提出了一种贪婪事件采样策略,以预测给定事件的多个高可靠性持续时间。作者的方法是独特的,提供了一种无标签、无训练和无领域的方法。它使TEL的应用完全处于测试阶段。实际结果表明,它在标准Charades STA和ActivityCaptions数据集上实现了具有竞争力的性能。
{"title":"Zero-shot temporal event localisation: Label-free, training-free, domain-free","authors":"Li Sun,&nbsp;Ping Wang,&nbsp;Liuan Wang,&nbsp;Jun Sun,&nbsp;Takayuki Okatani","doi":"10.1049/cvi2.12224","DOIUrl":"https://doi.org/10.1049/cvi2.12224","url":null,"abstract":"<p>Temporal event localisation (TEL) has recently attracted increasing attention due to the rapid development of video platforms. Existing methods are based on either fully/weakly supervised or unsupervised learning, and thus they rely on expensive data annotation and time-consuming training. Moreover, these models, which are trained on specific domain data, limit the model generalisation to data distribution shifts. To cope with these difficulties, the authors propose a zero-shot TEL method that can operate without training data or annotations. Leveraging large-scale vision and language pre-trained models, for example, CLIP, we solve the two key problems: (1) how to find the relevant region where the event is likely to occur; (2) how to determine event duration after we find the relevant region. Query guided optimisation for local frame relevance relying on the query-to-frame relationship is proposed to find the most relevant frame region where the event is most likely to occur. Proposal generation method relying on the frame-to-frame relationship is proposed to determine the event duration. The authors also propose a greedy event sampling strategy to predict multiple durations with high reliability for the given event. The authors’ methodology is unique, offering a label-free, training-free, and domain-free approach. It enables the application of TEL purely at the testing stage. The practical results show it achieves competitive performance on the standard Charades-STA and ActivityCaptions datasets.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"17 5","pages":"599-613"},"PeriodicalIF":1.7,"publicationDate":"2023-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12224","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50123617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving object detection by enhancing the effect of localisation quality evaluation on detection confidence 通过增强定位质量评价对检测置信度的影响来改进目标检测
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-31 DOI: 10.1049/cvi2.12227
Zuyi Wang, Wei Zhao, Li Xu

The one-stage object detector has been widely applied in many computer vision applications due to its high detection efficiency and simple framework. However, one-stage detectors heavily rely on Non-maximum Suppression to remove the duplicated predictions for the same objects, and the detectors produce detection confidence to measure the quality of those predictions. The localisation quality is an important factor to evaluate the predicted bounding boxes, but its role has not been fully utilised in previous works. To alleviate the problem, the Quality Prediction Block (QPB), a lightweight sub-network, is designed by the authors, which strengthens the effect of localisation quality evaluation on detection confidence by leveraging the features of predicted bounding boxes. The QPB is simple in structure and applies to different forms of detection confidence. Extensive experiments are conducted on the public benchmarks, MS COCO, PASCAL VOC and Berkeley DeepDrive. The results demonstrate the effectiveness of our method in the detectors with various forms of detection confidence. The proposed approach also achieves better performance in the stronger one-stage detectors.

单级物体检测器因其检测效率高、框架简单而被广泛应用于许多计算机视觉应用中。然而,单级检测器在很大程度上依赖于 "非最大抑制"(Non-maximum Suppression)来去除对同一物体的重复预测,并且检测器会产生检测置信度来衡量这些预测的质量。定位质量是评估预测边界框的一个重要因素,但在以前的工作中并未充分发挥其作用。为了缓解这一问题,作者设计了一个轻量级子网络--质量预测块(QPB),通过利用预测边界框的特征来加强定位质量评估对检测可信度的影响。QPB 结构简单,适用于不同形式的检测可信度。我们在 MS COCO、PASCAL VOC 和 Berkeley DeepDrive 等公共基准上进行了广泛的实验。结果表明,我们的方法在具有不同检测置信度的检测器中都很有效。所提出的方法在更强的单级检测器中也取得了更好的性能。
{"title":"Improving object detection by enhancing the effect of localisation quality evaluation on detection confidence","authors":"Zuyi Wang,&nbsp;Wei Zhao,&nbsp;Li Xu","doi":"10.1049/cvi2.12227","DOIUrl":"10.1049/cvi2.12227","url":null,"abstract":"<p>The one-stage object detector has been widely applied in many computer vision applications due to its high detection efficiency and simple framework. However, one-stage detectors heavily rely on Non-maximum Suppression to remove the duplicated predictions for the same objects, and the detectors produce detection confidence to measure the quality of those predictions. The localisation quality is an important factor to evaluate the predicted bounding boxes, but its role has not been fully utilised in previous works. To alleviate the problem, the Quality Prediction Block (QPB), a lightweight sub-network, is designed by the authors, which strengthens the effect of localisation quality evaluation on detection confidence by leveraging the features of predicted bounding boxes. The QPB is simple in structure and applies to different forms of detection confidence. Extensive experiments are conducted on the public benchmarks, MS COCO, PASCAL VOC and Berkeley DeepDrive. The results demonstrate the effectiveness of our method in the detectors with various forms of detection confidence. The proposed approach also achieves better performance in the stronger one-stage detectors.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"18 1","pages":"97-109"},"PeriodicalIF":1.7,"publicationDate":"2023-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12227","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44989905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lite-weight semantic segmentation with AG self-attention 具有AG自注意的Lite权重语义分割
IF 1.7 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-07-28 DOI: 10.1049/cvi2.12225
Bing Liu, Yansheng Gao, Hai Li, Zhaohao Zhong, Hongwei Zhao

Due to the large computational and GPUs memory cost of semantic segmentation, some works focus on designing a lite weight model to achieve a good trade-off between computational cost and accuracy. A common method is to combined CNN and vision transformer. However, these methods ignore the contextual information of multi receptive fields. And existing methods often fail to inject detailed information losses in the downsampling of multi-scale feature. To fix these issues, we propose AG Self-Attention, which is Enhanced Atrous Self-Attention (EASA), and Gate Attention. AG Self-Attention adds the contextual information of multi receptive fields into the global semantic feature. Specifically, the Enhanced Atrous Self-Attention uses weight shared atrous convolution with different atrous rates to get the contextual information under the specific different receptive fields. Gate Attention introduces gating mechanism to inject detailed information into the global semantic feature and filter detailed information by producing “fusion” gate and “update” gate. In order to prove our insight. We conduct numerous experiments in common semantic segmentation datasets, consisting of ADE20 K, COCO-stuff, PASCAL Context, Cityscapes, to show that our method achieves state-of-the-art performance and achieve a good trade-off between computational cost and accuracy.

由于语义分割的计算成本和 GPU 内存成本都很高,因此一些研究集中于设计一个轻量级模型,以便在计算成本和准确性之间实现良好的权衡。一种常见的方法是将 CNN 与视觉转换器相结合。然而,这些方法忽略了多感受野的上下文信息。而且现有的方法往往无法在多尺度特征的下采样中注入细节信息损失。为了解决这些问题,我们提出了 AG 自我注意(AG Self-Attention),即增强型自注意(Enhanced Atrous Self-Attention,EASA)和门注意(Gate Attention)。AG 自我注意将多感受野的上下文信息添加到全局语义特征中。具体来说,增强型无齿自注意使用不同无齿率的权重共享无齿卷积来获取特定不同感受野下的上下文信息。门注意(Gate Attention)引入了门机制,通过产生 "融合 "门和 "更新 "门,向全局语义特征注入详细信息并过滤详细信息。为了证明我们的见解。我们在常见的语义分割数据集(包括 ADE20 K、COCO-stuff、PASCAL Context 和 Cityscapes)中进行了大量实验,结果表明我们的方法达到了最先进的性能,并在计算成本和准确性之间实现了良好的权衡。
{"title":"Lite-weight semantic segmentation with AG self-attention","authors":"Bing Liu,&nbsp;Yansheng Gao,&nbsp;Hai Li,&nbsp;Zhaohao Zhong,&nbsp;Hongwei Zhao","doi":"10.1049/cvi2.12225","DOIUrl":"10.1049/cvi2.12225","url":null,"abstract":"<p>Due to the large computational and GPUs memory cost of semantic segmentation, some works focus on designing a lite weight model to achieve a good trade-off between computational cost and accuracy. A common method is to combined CNN and vision transformer. However, these methods ignore the contextual information of multi receptive fields. And existing methods often fail to inject detailed information losses in the downsampling of multi-scale feature. To fix these issues, we propose AG Self-Attention, which is Enhanced Atrous Self-Attention (EASA), and Gate Attention. AG Self-Attention adds the contextual information of multi receptive fields into the global semantic feature. Specifically, the Enhanced Atrous Self-Attention uses weight shared atrous convolution with different atrous rates to get the contextual information under the specific different receptive fields. Gate Attention introduces gating mechanism to inject detailed information into the global semantic feature and filter detailed information by producing “fusion” gate and “update” gate. In order to prove our insight. We conduct numerous experiments in common semantic segmentation datasets, consisting of ADE20 K, COCO-stuff, PASCAL Context, Cityscapes, to show that our method achieves state-of-the-art performance and achieve a good trade-off between computational cost and accuracy.</p>","PeriodicalId":56304,"journal":{"name":"IET Computer Vision","volume":"18 1","pages":"72-83"},"PeriodicalIF":1.7,"publicationDate":"2023-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cvi2.12225","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45119461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IET Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1