首页 > 最新文献

Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence最新文献

英文 中文
Lightweight Object Detection Method for Mobile Robot Platform 移动机器人平台轻量化目标检测方法
Yuncheng Sang, Han Huang, Shuangqing Ma, Shouwen Cai, Zhen Shi
∗We present some experienced improvements to YOLOv5s for mobile robot platforms that occupy many system resources and are difficult to meet the requirements of the actual application. Firstly, the FPN + PAN structure is redesigned to replace this complex structure into a dilated residual module with fewer parameters and calculations. The dilated residual module is composed of dilated residual blocks with different dilation rates stacked together. Secondly, we switch some convolution modules to improved Ghost modules in the backbone. The improved Ghost modules concatenate feature maps obtained by convolution with ones generated by a linear transformation. Then, the two parts of feature maps are shuffled to boost information fusion. The model is trained on the COCO dataset. In this paper, mAP_0.5 is 56.1%, mAP_0.5:0.95 is 35.7%, and the speed is 6.1% faster than YOLOv5s. The experimental results show that the method can further increase the inference speed to ensure detection accuracy. It can well solve the task of object detection on the mobile robot platform.
对于占用大量系统资源且难以满足实际应用要求的移动机器人平台,我们提出了对YOLOv5s的一些经验改进。首先,对FPN + PAN结构进行了重新设计,将该复杂结构替换为一个参数和计算量更少的扩展剩余模块。膨胀残差模块由不同膨胀率的膨胀残差块堆叠而成。其次,我们将一些卷积模块转换为改进的Ghost模块。改进的Ghost模块将卷积得到的特征映射与线性变换生成的特征映射连接起来。然后,对特征图的两个部分进行洗牌,增强信息融合。该模型是在COCO数据集上训练的。在本文中,mAP_0.5为56.1%,mAP_0.5:0.95为35.7%,速度比YOLOv5s快6.1%。实验结果表明,该方法可以进一步提高推理速度,保证检测精度。它可以很好地解决移动机器人平台上的目标检测任务。
{"title":"Lightweight Object Detection Method for Mobile Robot Platform","authors":"Yuncheng Sang, Han Huang, Shuangqing Ma, Shouwen Cai, Zhen Shi","doi":"10.1145/3507548.3507550","DOIUrl":"https://doi.org/10.1145/3507548.3507550","url":null,"abstract":"∗We present some experienced improvements to YOLOv5s for mobile robot platforms that occupy many system resources and are difficult to meet the requirements of the actual application. Firstly, the FPN + PAN structure is redesigned to replace this complex structure into a dilated residual module with fewer parameters and calculations. The dilated residual module is composed of dilated residual blocks with different dilation rates stacked together. Secondly, we switch some convolution modules to improved Ghost modules in the backbone. The improved Ghost modules concatenate feature maps obtained by convolution with ones generated by a linear transformation. Then, the two parts of feature maps are shuffled to boost information fusion. The model is trained on the COCO dataset. In this paper, mAP_0.5 is 56.1%, mAP_0.5:0.95 is 35.7%, and the speed is 6.1% faster than YOLOv5s. The experimental results show that the method can further increase the inference speed to ensure detection accuracy. It can well solve the task of object detection on the mobile robot platform.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127593062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Texture Dataset Construction and Texture Image Retrieval based on Deep Learning 基于深度学习的纹理数据集构建与纹理图像检索
Zhisheng Zhang, Huaijing Qu, Hengbin Wang, Jia Xu, Jiwei Wang, Yanan Wei
In the deep texture image retrieval, to address the problem that the retrieval performance is affected by the lack of sufficiently large texture image dataset used for the effective training of deep neural network, a deep learning based texture dataset construction and texture image retrieval method is proposed in this paper. First, a large-scale texture image dataset containing rich texture information is constructed based on the DTD texture image dataset, and used as the source dataset for pre-training deep neural networks. To effectively characterize the information of the source texture dataset, a revised version of the VGG16 model, called ReV-VGG16, is adaptively designed. Then, the pre-trained ReV-VGG16 model is combined with the target texture image datasets for the transfer learning, and the probability values of the output from the classification layer of the model are used for the computation of the similarity measurement to achieve the retrieval of the target texture image dataset. Finally, the retrieval experiments are conducted on four typical texture image datasets, namely, VisTex, Brodatz, STex and ALOT. The experimental results show that our method outperforms the existing state-of-the-art texture image retrieval approaches in terms of the retrieval performance.
在深度纹理图像检索中,针对缺乏足够大的纹理图像数据集用于深度神经网络的有效训练而影响检索性能的问题,提出了一种基于深度学习的纹理数据集构建和纹理图像检索方法。首先,基于DTD纹理图像数据集构建了包含丰富纹理信息的大规模纹理图像数据集,并将其作为深度神经网络预训练的源数据集;为了有效表征源纹理数据集的信息,自适应设计了VGG16模型的修正版本ReV-VGG16。然后,将预训练好的ReV-VGG16模型与目标纹理图像数据集结合进行迁移学习,利用模型分类层输出的概率值进行相似性度量计算,实现目标纹理图像数据集的检索。最后,在VisTex、Brodatz、STex和ALOT四种典型纹理图像数据集上进行检索实验。实验结果表明,该方法在检索性能方面优于现有的最先进的纹理图像检索方法。
{"title":"Texture Dataset Construction and Texture Image Retrieval based on Deep Learning","authors":"Zhisheng Zhang, Huaijing Qu, Hengbin Wang, Jia Xu, Jiwei Wang, Yanan Wei","doi":"10.1145/3507548.3507564","DOIUrl":"https://doi.org/10.1145/3507548.3507564","url":null,"abstract":"In the deep texture image retrieval, to address the problem that the retrieval performance is affected by the lack of sufficiently large texture image dataset used for the effective training of deep neural network, a deep learning based texture dataset construction and texture image retrieval method is proposed in this paper. First, a large-scale texture image dataset containing rich texture information is constructed based on the DTD texture image dataset, and used as the source dataset for pre-training deep neural networks. To effectively characterize the information of the source texture dataset, a revised version of the VGG16 model, called ReV-VGG16, is adaptively designed. Then, the pre-trained ReV-VGG16 model is combined with the target texture image datasets for the transfer learning, and the probability values of the output from the classification layer of the model are used for the computation of the similarity measurement to achieve the retrieval of the target texture image dataset. Finally, the retrieval experiments are conducted on four typical texture image datasets, namely, VisTex, Brodatz, STex and ALOT. The experimental results show that our method outperforms the existing state-of-the-art texture image retrieval approaches in terms of the retrieval performance.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130991618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Regression Algorithm Based on Self-Distillation and Ensemble Learning 基于自蒸馏和集成学习的回归算法
Yaqi Li, Qiwen Dong, Gang Liu
Low-dimensional feature regression is a common problem in various disciplines, such as chemistry, kinetics, and medicine, etc. Most common solutions are based on machine learning, but as deep learning evolves, there is room for performance improvements. A few researchers have proposed deep learning-based solutions such as ResidualNet, GrowNet and EnsembleNet. The latter two methods are both boost methods, which are more suitable for shallow network, and the model performance is basically determined by the first model, with limited effect of subsequent boosting steps. We propose a method based on self-distillation and bagging, which selects the well-performing base model and distills several student models by appropriate regression distillation algorithm. Finally, the output of these student models is averaged as the final result. This integration method can be applied to any form of network. The method achieves good results in the CASP dataset, and the R2(Coefficient of Determination) of the model is improved from (0.65) to (0.70) in comparison with the best base model ResidualNet.
低维特征回归是化学、动力学、医学等学科中常见的问题。大多数常见的解决方案都是基于机器学习的,但随着深度学习的发展,性能还有改进的空间。一些研究人员提出了基于深度学习的解决方案,如ResidualNet、GrowNet和EnsembleNet。后两种方法都是boost方法,更适合于浅层网络,模型性能基本由第一个模型决定,后续boost步骤的影响有限。我们提出了一种基于自蒸馏和装袋的方法,该方法选择性能较好的基本模型,并通过适当的回归蒸馏算法对多个学生模型进行蒸馏。最后,将这些学生模型的输出平均作为最终结果。这种集成方法适用于任何形式的网络。该方法在CASP数据集上取得了较好的效果,与最佳基础模型residalnet相比,模型的R2(决定系数)由(0.65)提高到(0.70)。
{"title":"Regression Algorithm Based on Self-Distillation and Ensemble Learning","authors":"Yaqi Li, Qiwen Dong, Gang Liu","doi":"10.1145/3507548.3507580","DOIUrl":"https://doi.org/10.1145/3507548.3507580","url":null,"abstract":"Low-dimensional feature regression is a common problem in various disciplines, such as chemistry, kinetics, and medicine, etc. Most common solutions are based on machine learning, but as deep learning evolves, there is room for performance improvements. A few researchers have proposed deep learning-based solutions such as ResidualNet, GrowNet and EnsembleNet. The latter two methods are both boost methods, which are more suitable for shallow network, and the model performance is basically determined by the first model, with limited effect of subsequent boosting steps. We propose a method based on self-distillation and bagging, which selects the well-performing base model and distills several student models by appropriate regression distillation algorithm. Finally, the output of these student models is averaged as the final result. This integration method can be applied to any form of network. The method achieves good results in the CASP dataset, and the R2(Coefficient of Determination) of the model is improved from (0.65) to (0.70) in comparison with the best base model ResidualNet.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128657555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Research of Predicting Student's Academic Performance Based on Educational Data 基于教育数据的学生学习成绩预测研究
Yubo Zhang, Yanfang Liu
In recent years, with the continuous improvement of teaching informatization, online teaching or online and offline hybrid teaching has become the new normal of teaching in some schools. However, the biggest problem in online teaching is that it is difficult to predict students' academic performance. Therefore, it is necessary to design an effective method to predict students' academic performance more accurately. In this paper, a student academic level prediction method based on stacking model fusion is proposed. Logistic regression, random forest, XGBoost, and Naive Bayes are selected as base learners according to optimal fusion criteria and model properties. Furthermore, the structure and distribution of the features of the dataset are optimized by data preprocessing, feature coding, and feature selection, and the upper limit of the model expression is effectively raised. On this basis, according to the features of dataset and model performance, we select the appropriate model for model fusion, and further improve the prediction effect. Experiments are conducted on OULAD and xAPI datasets, and the results show that the prediction accuracy of the proposed method is better than that of traditional prediction methods. Finally, we analyze the factors that affect academic performance and give some specific suggestions.
近年来,随着教学信息化程度的不断提高,在线教学或线上线下混合教学已成为一些学校教学的新常态。然而,在线教学最大的问题是很难预测学生的学习成绩。因此,有必要设计一种有效的方法来更准确地预测学生的学习成绩。提出了一种基于叠加模型融合的学生学业水平预测方法。根据最优融合准则和模型特性,选择逻辑回归、随机森林、XGBoost和朴素贝叶斯作为基础学习器。通过数据预处理、特征编码、特征选择等方法对数据集特征的结构和分布进行优化,有效提高了模型表达的上限。在此基础上,根据数据集和模型性能的特点,选择合适的模型进行模型融合,进一步提高预测效果。在OULAD和xAPI数据集上进行了实验,结果表明,该方法的预测精度优于传统的预测方法。最后,分析了影响学生学习成绩的因素,并提出了具体的建议。
{"title":"The Research of Predicting Student's Academic Performance Based on Educational Data","authors":"Yubo Zhang, Yanfang Liu","doi":"10.1145/3507548.3507578","DOIUrl":"https://doi.org/10.1145/3507548.3507578","url":null,"abstract":"In recent years, with the continuous improvement of teaching informatization, online teaching or online and offline hybrid teaching has become the new normal of teaching in some schools. However, the biggest problem in online teaching is that it is difficult to predict students' academic performance. Therefore, it is necessary to design an effective method to predict students' academic performance more accurately. In this paper, a student academic level prediction method based on stacking model fusion is proposed. Logistic regression, random forest, XGBoost, and Naive Bayes are selected as base learners according to optimal fusion criteria and model properties. Furthermore, the structure and distribution of the features of the dataset are optimized by data preprocessing, feature coding, and feature selection, and the upper limit of the model expression is effectively raised. On this basis, according to the features of dataset and model performance, we select the appropriate model for model fusion, and further improve the prediction effect. Experiments are conducted on OULAD and xAPI datasets, and the results show that the prediction accuracy of the proposed method is better than that of traditional prediction methods. Finally, we analyze the factors that affect academic performance and give some specific suggestions.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"220 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128727897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human Fall Detection Model with Lightweight Network and Tracking in Video 基于轻量级网络和视频跟踪的人体跌倒检测模型
Xiaoli Ren, Yunjie Zhang, Yanrong Yang
In order to real time and accurately detect the action of human falling, combined with lightweight detection network, Kalman filter tracking, posture estimation network and spatiotemporal graph convolutional network, a joint algorithm for human fall detection in video is proposed. Firstly, the lightweight YOLOv3-Tiny algorithm is used to locate the frame of human in video, which can quickly detect the human-frame; among them, for the situation that the human body is likely to be missed in video, the Kalman filter tracking algorithm is integrated into the stage of target-detection and the accuracy of detecting is improved. Secondly, the human-frame detected or tracked in video is sent to the AlphaPose network to estimate the posture graph about human body. Finally, the spatiotemporal graph convolutional network is exploited to extract the spatiotemporal features of the human body, and eventually the result for classification is output. Experimental results show that the algorithm proposed in this paper, which is more appealing and successful than the other algorithm.
为了实时准确地检测人体跌倒动作,结合轻量级检测网络、卡尔曼滤波跟踪、姿态估计网络和时空图卷积网络,提出了一种视频中人体跌倒检测的联合算法。首先,采用轻量级的YOLOv3-Tiny算法对视频中的人体帧进行定位,能够快速检测出人体帧;其中,针对视频中人体容易被遗漏的情况,将卡尔曼滤波跟踪算法集成到目标检测阶段,提高了检测精度。其次,将视频中检测到或跟踪到的人体帧发送到AlphaPose网络,估计人体姿态图;最后,利用时空图卷积网络提取人体的时空特征,并输出分类结果。实验结果表明,本文提出的算法比其他算法更具吸引力和有效性。
{"title":"Human Fall Detection Model with Lightweight Network and Tracking in Video","authors":"Xiaoli Ren, Yunjie Zhang, Yanrong Yang","doi":"10.1145/3507548.3507549","DOIUrl":"https://doi.org/10.1145/3507548.3507549","url":null,"abstract":"In order to real time and accurately detect the action of human falling, combined with lightweight detection network, Kalman filter tracking, posture estimation network and spatiotemporal graph convolutional network, a joint algorithm for human fall detection in video is proposed. Firstly, the lightweight YOLOv3-Tiny algorithm is used to locate the frame of human in video, which can quickly detect the human-frame; among them, for the situation that the human body is likely to be missed in video, the Kalman filter tracking algorithm is integrated into the stage of target-detection and the accuracy of detecting is improved. Secondly, the human-frame detected or tracked in video is sent to the AlphaPose network to estimate the posture graph about human body. Finally, the spatiotemporal graph convolutional network is exploited to extract the spatiotemporal features of the human body, and eventually the result for classification is output. Experimental results show that the algorithm proposed in this paper, which is more appealing and successful than the other algorithm.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121085950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Enhanced Efficient YOLOv3-tiny for Object Detection 增强高效YOLOv3-tiny的目标检测
Huanqia Cai, Lele Xu, Lili Guo
Lightweight object detection models have great application prospects in resource-restricted scenarios such as mobile and embedded devices, and have been a hot topic in the computer vision community. However, most existing lightweight object detection methods show poor detection accuracy. In this study, we put forward a lightweight objection detection model named Enhanced-YOLOv3-tiny to improve the detection accuracy and reduce the model complexity at the same time. In Enhanced-YOLOv3-tiny, we propose a new backbone named GhostDarkNet based on DarkNet53 and Ghost Module for decreasing the model parameters, which helps to get more representative features compared with YOLOv3-tiny. Furthermore, we put forward a new Multiscale Head, which adds three more heads and includes Ghost Module in each head to fuse multi-scale features. Experiments on the Priority Research Application dataset from the real scenes in driving show that the proposed Enhanced-YOLOv3-tiny outperforms the state-of-the-art YOLOv3-tiny by 8.4% improvement in AP metric and decreases parameters from 8.8M to 3.9M, demonstrating the application potentials of our proposed method in resource-constrained scenarios.
轻量级目标检测模型在移动和嵌入式设备等资源受限场景中具有很大的应用前景,一直是计算机视觉界的研究热点。然而,现有的大多数轻量化目标检测方法检测精度较差。在本研究中,我们提出了一种轻量级的目标检测模型Enhanced-YOLOv3-tiny,在提高检测精度的同时降低模型复杂度。在Enhanced-YOLOv3-tiny中,我们提出了一种基于DarkNet53和Ghost Module的新主干GhostDarkNet,以减少模型参数,从而获得比YOLOv3-tiny更具代表性的特征。在此基础上,我们提出了一种新的多尺度磁头,该磁头增加了3个磁头,并在每个磁头中加入Ghost Module以融合多尺度特征。在Priority Research Application真实驾驶场景数据集上的实验表明,本文提出的Enhanced-YOLOv3-tiny在AP度量上比最先进的YOLOv3-tiny提高了8.4%,并将参数从8.8M降至3.9M,证明了本文提出的方法在资源受限场景下的应用潜力。
{"title":"Enhanced Efficient YOLOv3-tiny for Object Detection","authors":"Huanqia Cai, Lele Xu, Lili Guo","doi":"10.1145/3507548.3507551","DOIUrl":"https://doi.org/10.1145/3507548.3507551","url":null,"abstract":"Lightweight object detection models have great application prospects in resource-restricted scenarios such as mobile and embedded devices, and have been a hot topic in the computer vision community. However, most existing lightweight object detection methods show poor detection accuracy. In this study, we put forward a lightweight objection detection model named Enhanced-YOLOv3-tiny to improve the detection accuracy and reduce the model complexity at the same time. In Enhanced-YOLOv3-tiny, we propose a new backbone named GhostDarkNet based on DarkNet53 and Ghost Module for decreasing the model parameters, which helps to get more representative features compared with YOLOv3-tiny. Furthermore, we put forward a new Multiscale Head, which adds three more heads and includes Ghost Module in each head to fuse multi-scale features. Experiments on the Priority Research Application dataset from the real scenes in driving show that the proposed Enhanced-YOLOv3-tiny outperforms the state-of-the-art YOLOv3-tiny by 8.4% improvement in AP metric and decreases parameters from 8.8M to 3.9M, demonstrating the application potentials of our proposed method in resource-constrained scenarios.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"149 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115587522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LCR-GAN: Learning Crucial Representation for Anomaly Detection LCR-GAN:学习异常检测的关键表征
Shuo Liu, Li-wen Xu, Jin-Rong Wang, Yan Sun, Zeran Qin
Anomaly detection is pivotal and challenging in artificial intelligence, which aims to determine whether a query sample comes from the same class, given a set of normal samples from a particular class. There are a plethora of anomaly detection methods based on generative models; however, these methods aim to make the reconstruction error of the training samples smaller or extract more information from the training samples. We believe that it is more important for anomaly detection to extract crucial representation from normal samples rather than more information, so we propose a semi-supervised method named LCR-GAN. We conducted extensive experiments on four image datasets and 15 tabular datasets to demonstrate the effectiveness of the proposed method. Meanwhile, we also carried out an anti-noise study to demonstrate the robustness of the proposed method.
异常检测在人工智能中是关键和具有挑战性的,其目的是确定查询样本是否来自同一类,给定一组来自特定类的正常样本。有很多基于生成模型的异常检测方法;然而,这些方法的目的是使训练样本的重构误差更小或从训练样本中提取更多的信息。我们认为,对于异常检测来说,从正常样本中提取关键的表征比提取更多的信息更重要,因此我们提出了一种半监督方法LCR-GAN。我们在4个图像数据集和15个表格数据集上进行了广泛的实验,以证明所提出方法的有效性。同时,我们还进行了抗噪声研究,以证明该方法的鲁棒性。
{"title":"LCR-GAN: Learning Crucial Representation for Anomaly Detection","authors":"Shuo Liu, Li-wen Xu, Jin-Rong Wang, Yan Sun, Zeran Qin","doi":"10.1145/3507548.3508229","DOIUrl":"https://doi.org/10.1145/3507548.3508229","url":null,"abstract":"Anomaly detection is pivotal and challenging in artificial intelligence, which aims to determine whether a query sample comes from the same class, given a set of normal samples from a particular class. There are a plethora of anomaly detection methods based on generative models; however, these methods aim to make the reconstruction error of the training samples smaller or extract more information from the training samples. We believe that it is more important for anomaly detection to extract crucial representation from normal samples rather than more information, so we propose a semi-supervised method named LCR-GAN. We conducted extensive experiments on four image datasets and 15 tabular datasets to demonstrate the effectiveness of the proposed method. Meanwhile, we also carried out an anti-noise study to demonstrate the robustness of the proposed method.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127504005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sampling May Not Always Increase Detector Performance: A Study on Collecting Training Examples 抽样不一定能提高检测器的性能:一种收集训练样本的研究
Jun Liu, Shuang Lai
In recent years, the research of computer vision is popular. However, the image data that can be used for computer vision training is very limited, so it is necessary to find an effective method to expand the datasets based on the existing image data. In this paper, we study methods to collect more training data from existing datasets and compare detectors’ performance trained with datasets generated by different methods. One method is to perform sampling-based on statistical properties of feature descriptors. For every feature, the underlying assumption is that a probability density function (PDF) exists, such PDF is approximated with existing training examples and new training examples are sampled from the approximated PDF. The other method is simply to expand the existing datasets by flipping each training example along its symmetric axis. Locally Adaptive Regression Kernel (LARK) feature is used in this paper because it is robust against illumination changes and noise. Our experimental results demonstrate that an expanded training dataset is not always preferable, even if the expanded dataset includes all original training data.
近年来,计算机视觉的研究非常热门。然而,可用于计算机视觉训练的图像数据非常有限,因此有必要在现有图像数据的基础上寻找一种有效的方法来扩展数据集。在本文中,我们研究了从现有数据集中收集更多训练数据的方法,并比较了不同方法生成的数据集训练后的检测器性能。一种方法是基于特征描述符的统计性质进行抽样。对于每一个特征,其基本假设是存在一个概率密度函数(PDF),用已有的训练样例对概率密度函数进行近似,并从近似的概率密度函数中采样新的训练样例。另一种方法是通过沿其对称轴翻转每个训练示例来扩展现有数据集。局部自适应回归核(LARK)特征对光照变化和噪声具有较强的鲁棒性。我们的实验结果表明,扩展的训练数据集并不总是更好的,即使扩展的数据集包括所有原始训练数据。
{"title":"Sampling May Not Always Increase Detector Performance: A Study on Collecting Training Examples","authors":"Jun Liu, Shuang Lai","doi":"10.1145/3507548.3507568","DOIUrl":"https://doi.org/10.1145/3507548.3507568","url":null,"abstract":"In recent years, the research of computer vision is popular. However, the image data that can be used for computer vision training is very limited, so it is necessary to find an effective method to expand the datasets based on the existing image data. In this paper, we study methods to collect more training data from existing datasets and compare detectors’ performance trained with datasets generated by different methods. One method is to perform sampling-based on statistical properties of feature descriptors. For every feature, the underlying assumption is that a probability density function (PDF) exists, such PDF is approximated with existing training examples and new training examples are sampled from the approximated PDF. The other method is simply to expand the existing datasets by flipping each training example along its symmetric axis. Locally Adaptive Regression Kernel (LARK) feature is used in this paper because it is robust against illumination changes and noise. Our experimental results demonstrate that an expanded training dataset is not always preferable, even if the expanded dataset includes all original training data.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"45 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126277872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on Digital Exhibition Design of Former Residence Memorial Hall based on IPOP Theory 基于IPOP理论的故居纪念馆数字化展示设计研究
Xia Wang, Zhengqing Jiang
This paper takes visitor experience as the center to study the specific strategies to enhance the digital display effect of the former residence Memorial Hall, in order to deal with the plight of its digital display development lag. Based on the theory of IPOP, this paper makes an empirical study of Tsou Jung Memorial Hall through questionnaire survey and field observation, this paper discusses the differences of four kinds of audience on the digital display of the former residence Memorial Hall. There are significant differences and correlations among the four kinds of experiences dimensions among the audiences with different preference types. Through the analysis of their internal relations, this paper explores the application possibility of IPOP theory in the digital display experience of museums, it provides a general plan for the evaluation of demonstration effect and the standard of technology application.
本文以游客体验为中心,研究提升故居纪念馆数字展示效果的具体策略,以应对其数字展示发展滞后的困境。本文以IPOP理论为基础,通过问卷调查和实地观察,对祖正纪念馆进行了实证研究,探讨了四种观众对故居纪念馆数字展示的差异。不同偏好类型的受众在四种体验维度上存在显著的差异和相关性。通过对二者内在关系的分析,探索IPOP理论在博物馆数字展示体验中的应用可能性,为示范效果评价和技术应用标准提供总体规划。
{"title":"Research on Digital Exhibition Design of Former Residence Memorial Hall based on IPOP Theory","authors":"Xia Wang, Zhengqing Jiang","doi":"10.1145/3507548.3507606","DOIUrl":"https://doi.org/10.1145/3507548.3507606","url":null,"abstract":"This paper takes visitor experience as the center to study the specific strategies to enhance the digital display effect of the former residence Memorial Hall, in order to deal with the plight of its digital display development lag. Based on the theory of IPOP, this paper makes an empirical study of Tsou Jung Memorial Hall through questionnaire survey and field observation, this paper discusses the differences of four kinds of audience on the digital display of the former residence Memorial Hall. There are significant differences and correlations among the four kinds of experiences dimensions among the audiences with different preference types. Through the analysis of their internal relations, this paper explores the application possibility of IPOP theory in the digital display experience of museums, it provides a general plan for the evaluation of demonstration effect and the standard of technology application.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130280681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GCN-Seq2Seq: A Spatio-Temporal feature-fused model for surface water quality prediction GCN-Seq2Seq:地表水水质时空特征融合预测模型
Ying Chen, Ping Yang, Chengxu Ye, Zhikun Miao
Aiming at the complex dependence of water quality data in space and time, we propose a GCN-Seq2Seq model for surface water quality prediction. The model uses Graph Convolutional Network (GCN) to capture the spatial feature of water quality monitoring sites, uses the sequence to sequence (Seq2Seq) model constructed by GRU to extract the temporal feature of the water quality data sequence, and predicts multi-step water quality time series. Experiments were carried out with data from 6 water quality monitoring stations in the Huangshui River and surrounding areas in Xining City, Qinghai Province from November 2020 to June 2021, and compared with the baseline model. experimental results show that the proposed model can effectively improve the accuracy of multi-step prediction of surface water quality.
针对地表水水质数据在空间和时间上的复杂依赖性,提出了一种用于地表水水质预测的GCN-Seq2Seq模型。该模型利用图卷积网络(Graph Convolutional Network, GCN)捕捉水质监测点的空间特征,利用GRU构建的序列到序列(sequence to sequence, Seq2Seq)模型提取水质数据序列的时间特征,并对多步水质时间序列进行预测。利用青海省西宁市湟水河及周边地区6个水质监测站2020年11月至2021年6月的数据进行实验,并与基线模型进行对比。实验结果表明,该模型能有效提高地表水水质多步预测的精度。
{"title":"GCN-Seq2Seq: A Spatio-Temporal feature-fused model for surface water quality prediction","authors":"Ying Chen, Ping Yang, Chengxu Ye, Zhikun Miao","doi":"10.1145/3507548.3507597","DOIUrl":"https://doi.org/10.1145/3507548.3507597","url":null,"abstract":"Aiming at the complex dependence of water quality data in space and time, we propose a GCN-Seq2Seq model for surface water quality prediction. The model uses Graph Convolutional Network (GCN) to capture the spatial feature of water quality monitoring sites, uses the sequence to sequence (Seq2Seq) model constructed by GRU to extract the temporal feature of the water quality data sequence, and predicts multi-step water quality time series. Experiments were carried out with data from 6 water quality monitoring stations in the Huangshui River and surrounding areas in Xining City, Qinghai Province from November 2020 to June 2021, and compared with the baseline model. experimental results show that the proposed model can effectively improve the accuracy of multi-step prediction of surface water quality.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130873026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1