首页 > 最新文献

Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence最新文献

英文 中文
Multi-objective evolutionary algorithm based on decomposition with integration strategy 基于分解与集成策略的多目标进化算法
Xinwen Fang, Yuan xia Shen, Xue Feng Zhang
To improve the precision in the later stage of population evolution for multi-objective evolutionary algorithm based on decomposition (MOEA/D), a MOEA/D with integration strategy (MOEA/D-IS) is proposed. The proposed algorithm adopts multiple updating strategies, including a novel first-order differential learning strategy, the individual learning strategy, and the binary and polynomial crossover mutation strategy. The penalty-based boundary intersection approach and Chebyshev approach are used to alternately evaluate individuals. The proposed algorithm and five improved MOEA algorithms are tested on 21 functions. Simulation results show that MOEA/D-IS has good performance in diversity and convergence accuracy.
为了提高基于分解的多目标进化算法(MOEA/D)在种群进化后期的精度,提出了一种基于分解的多目标进化算法(MOEA/D- is)。该算法采用了多种更新策略,包括一阶差分学习策略、个体学习策略以及二叉多项式交叉突变策略。采用基于惩罚的边界交叉法和Chebyshev法交替评价个体。在21个函数上对该算法和5种改进的MOEA算法进行了测试。仿真结果表明,MOEA/D-IS具有良好的分集性能和收敛精度。
{"title":"Multi-objective evolutionary algorithm based on decomposition with integration strategy","authors":"Xinwen Fang, Yuan xia Shen, Xue Feng Zhang","doi":"10.1145/3507548.3507581","DOIUrl":"https://doi.org/10.1145/3507548.3507581","url":null,"abstract":"To improve the precision in the later stage of population evolution for multi-objective evolutionary algorithm based on decomposition (MOEA/D), a MOEA/D with integration strategy (MOEA/D-IS) is proposed. The proposed algorithm adopts multiple updating strategies, including a novel first-order differential learning strategy, the individual learning strategy, and the binary and polynomial crossover mutation strategy. The penalty-based boundary intersection approach and Chebyshev approach are used to alternately evaluate individuals. The proposed algorithm and five improved MOEA algorithms are tested on 21 functions. Simulation results show that MOEA/D-IS has good performance in diversity and convergence accuracy.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128095342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Vessel Pattern Recognition Using Trajectory Shape Feature 基于轨迹形状特征的船舶模式识别
Jia Li, Haiyan Liu, Xiaohui Chen, Jing Li, Junhong Xiang
In the era of big data, analyzing vessels patterns using massive trajectory data has become the main method of mining activity pattern. Trajectory shape feature, as one of the important features of vessel trajectory data, can be used to identify the vessel activity patterns. But most of research only focused on the features such as standard deviation of latitude and longitude, navigation heading to the analysis of vessels trajectories. Therefore, considering the spatial-temporal feature of vessels data, we propose a method based on Sevcik fractal dimension to extract shape feature for identifying vessels activity types. Firstly, we segment the vessel trajectories to form the sub-trajectory according to the speed and temporal threshold. Secondly, we construct the feature vector of trajectory shape using the improved Sevcik fractal dimension algorithm. Then, we select the standard deviation of latitude and longitude and shape features extracted by Sevcik fractal dimension as the comparison features, and observe the performance in K-means and GMM algorithms respectively to verify the effectiveness of shape feature vectors we proposed. Finally, we select the simulation data and two real data sets for experimental analysis. The results show that the shape feature extraction algorithm can extract the shape features of trajectories, and the performance in classification algorithm is better than the standard deviation and Sevcik fractal dimension. So the method we proposed can realize the pattern recognition of vessel and abnormal trajectory analysist.
在大数据时代,利用海量轨迹数据分析船舶模式已成为挖掘活动模式的主要方法。轨迹形状特征是舰船轨迹数据的重要特征之一,可以用来识别舰船的活动模式。但大多数研究只关注经纬度标准差、航行航向等特征,对船舶轨迹进行分析。因此,考虑到血管数据的时空特征,我们提出了一种基于Sevcik分形维数提取血管活动类型形状特征的方法。首先,我们根据速度和时间阈值对血管轨迹进行分割,形成子轨迹;其次,利用改进的Sevcik分形维数算法构造轨迹形状特征向量;然后选取经纬度标准差和Sevcik分形维数提取的形状特征作为比较特征,分别观察K-means和GMM算法的性能,验证所提形状特征向量的有效性。最后,选取仿真数据和两个真实数据集进行实验分析。结果表明,形状特征提取算法能够提取轨迹的形状特征,分类算法的性能优于标准差和Sevcik分形维数。因此,我们提出的方法可以实现船舶的模式识别和异常轨迹分析。
{"title":"Vessel Pattern Recognition Using Trajectory Shape Feature","authors":"Jia Li, Haiyan Liu, Xiaohui Chen, Jing Li, Junhong Xiang","doi":"10.1145/3507548.3507561","DOIUrl":"https://doi.org/10.1145/3507548.3507561","url":null,"abstract":"In the era of big data, analyzing vessels patterns using massive trajectory data has become the main method of mining activity pattern. Trajectory shape feature, as one of the important features of vessel trajectory data, can be used to identify the vessel activity patterns. But most of research only focused on the features such as standard deviation of latitude and longitude, navigation heading to the analysis of vessels trajectories. Therefore, considering the spatial-temporal feature of vessels data, we propose a method based on Sevcik fractal dimension to extract shape feature for identifying vessels activity types. Firstly, we segment the vessel trajectories to form the sub-trajectory according to the speed and temporal threshold. Secondly, we construct the feature vector of trajectory shape using the improved Sevcik fractal dimension algorithm. Then, we select the standard deviation of latitude and longitude and shape features extracted by Sevcik fractal dimension as the comparison features, and observe the performance in K-means and GMM algorithms respectively to verify the effectiveness of shape feature vectors we proposed. Finally, we select the simulation data and two real data sets for experimental analysis. The results show that the shape feature extraction algorithm can extract the shape features of trajectories, and the performance in classification algorithm is better than the standard deviation and Sevcik fractal dimension. So the method we proposed can realize the pattern recognition of vessel and abnormal trajectory analysist.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132670301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Research on Traffic Sign Recognition based on Convolutional Neural Network 基于卷积神经网络的交通标志识别研究
Wanjun Liu, Jiaxin Li, Haicheng Qu
Traffic sign recognition has a wide application prospect in the field of automatic driving. External factors such as illumination, Angle and occlusion will affect the recognition effect of small traffic signs. In order to solve these problems, this paper designs a multi-scale fusion convolutional neural network model (SQ-RCNN) based on feature extraction network Faster RCNN. Firstly, the multi-scale Atrous Spatial Pyramid Pooling (SASPP) module is added to the basic feature extraction network. After multi-scale cavity convolution sampling, the amount of information under each feature is not changed. In this way, the loss of resolution can be reduced and the context information of the same image can be captured. Secondly, the combination structure of two convolution layers and one pooling layer in the original VGG16 model was improved, and the concat operation was adopted to enrich the number of features by merging the number of channels, so as to realize the fusion of features at different scales and improve the accuracy of identifying small targets. In addition, a dropout layer is added to prevent overfitting. The experimental results show that: In this paper, a new network structure SQ-RCNN was used to extract features from CCTSDB data set, the mean average accuracy of traffic sign identification reached 86.96%, at the same time, effectively shorten the training time.
交通标志识别在自动驾驶领域有着广阔的应用前景。照明、角度、遮挡等外部因素会影响小型交通标志的识别效果。为了解决这些问题,本文设计了一种基于特征提取网络Faster RCNN的多尺度融合卷积神经网络模型(SQ-RCNN)。首先,在基本特征提取网络中加入多尺度空间金字塔池(SASPP)模块;多尺度空腔卷积采样后,各特征下的信息量不变。这样可以减少分辨率的损失,并且可以捕获同一图像的上下文信息。其次,对原有VGG16模型中两个卷积层和一个池化层的组合结构进行改进,采用concat操作,通过合并通道数来丰富特征数量,从而实现不同尺度特征的融合,提高小目标识别的精度。此外,还添加了一个dropout层来防止过拟合。实验结果表明:本文采用一种新的网络结构SQ-RCNN对CCTSDB数据集进行特征提取,交通标志识别的平均准确率达到86.96%,同时有效缩短了训练时间。
{"title":"Research on Traffic Sign Recognition based on Convolutional Neural Network","authors":"Wanjun Liu, Jiaxin Li, Haicheng Qu","doi":"10.1145/3507548.3507559","DOIUrl":"https://doi.org/10.1145/3507548.3507559","url":null,"abstract":"Traffic sign recognition has a wide application prospect in the field of automatic driving. External factors such as illumination, Angle and occlusion will affect the recognition effect of small traffic signs. In order to solve these problems, this paper designs a multi-scale fusion convolutional neural network model (SQ-RCNN) based on feature extraction network Faster RCNN. Firstly, the multi-scale Atrous Spatial Pyramid Pooling (SASPP) module is added to the basic feature extraction network. After multi-scale cavity convolution sampling, the amount of information under each feature is not changed. In this way, the loss of resolution can be reduced and the context information of the same image can be captured. Secondly, the combination structure of two convolution layers and one pooling layer in the original VGG16 model was improved, and the concat operation was adopted to enrich the number of features by merging the number of channels, so as to realize the fusion of features at different scales and improve the accuracy of identifying small targets. In addition, a dropout layer is added to prevent overfitting. The experimental results show that: In this paper, a new network structure SQ-RCNN was used to extract features from CCTSDB data set, the mean average accuracy of traffic sign identification reached 86.96%, at the same time, effectively shorten the training time.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"405 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133048654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictive Screening of Accident Black Spots based on Deep Neural Models of Road Networks and Facilities: A Case Study based on a District in Hong Kong 基于道路网络和设施深度神经模型的事故黑点预测筛选:以香港某地区为例
Andrew Kwok-Fai Lui, Y. Chan, K. Lo, Wang-To Cheng, Hang-Tak Cheung
The screening of road accident black spots is to predict accident prone locations in the road network, with the aim of preventing further accidents with remedial measures. As black spots are linked to a location, certain features of the location and its nearby branches of the network should be capable of explaining the black spots. Several open data sources now provide feature-rich road network and facilities datasets. This paper proposes a data-driven machine learning solution for black spot screening using features of road network and facilities. The accident neighborhood is a concept introduced in the paper that represents the nearby locations associated with the happening of accidents. The concept has been realized as graph embeddings of road network, which, together with a deep neural network classifier, are the two major components of the solution. An evaluation of the solution using data from a Hong Kong district indicates that recognition of both the surrounding road network structure and the local features near accident sites can yield accurate models for black spot prediction.
筛选道路交通意外黑点的目的,是预测道路网中容易发生意外的地点,以采取补救措施,防止进一步发生意外。当黑点与一个地点相关联时,该地点及其附近网络分支的某些特征应该能够解释黑点。现在有几个开放的数据源提供功能丰富的道路网络和设施数据集。本文提出了一种利用路网和设施特征进行黑点筛选的数据驱动机器学习解决方案。事故街区是本文引入的一个概念,表示与事故发生有关的附近地点。该概念已被实现为道路网络的图嵌入,它与深度神经网络分类器一起是解决方案的两个主要组成部分。使用香港地区的数据对该解决方案进行的评估表明,识别周围的道路网络结构和事故现场附近的局部特征可以产生准确的黑点预测模型。
{"title":"Predictive Screening of Accident Black Spots based on Deep Neural Models of Road Networks and Facilities: A Case Study based on a District in Hong Kong","authors":"Andrew Kwok-Fai Lui, Y. Chan, K. Lo, Wang-To Cheng, Hang-Tak Cheung","doi":"10.1145/3507548.3507613","DOIUrl":"https://doi.org/10.1145/3507548.3507613","url":null,"abstract":"The screening of road accident black spots is to predict accident prone locations in the road network, with the aim of preventing further accidents with remedial measures. As black spots are linked to a location, certain features of the location and its nearby branches of the network should be capable of explaining the black spots. Several open data sources now provide feature-rich road network and facilities datasets. This paper proposes a data-driven machine learning solution for black spot screening using features of road network and facilities. The accident neighborhood is a concept introduced in the paper that represents the nearby locations associated with the happening of accidents. The concept has been realized as graph embeddings of road network, which, together with a deep neural network classifier, are the two major components of the solution. An evaluation of the solution using data from a Hong Kong district indicates that recognition of both the surrounding road network structure and the local features near accident sites can yield accurate models for black spot prediction.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116272457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AZY-GCN: Multi-scale feature suppression attentional diagram convolutional network for human pose prediction 基于多尺度特征抑制注意图卷积网络的人体姿态预测
Yang Zhang, Fan Xiao Shan, Gang He
Due to the randomness and non-periodic nature of the future posture of the human body, the prediction of the posture of the human body has always been a very challenging task. In the latest research, graph convolution is proved to be an effective method to capture the dynamic relationship between the human body posture joints, which is helpful for the human body posture prediction. Moreover, graph convolution can abstract the pose of the human body to obtain a multi-scale pose set. As the level of abstraction increases, the posture movement will become more stable. Although the average prediction accuracy has improved significantly in recent years, there is still much room for exploration in the application of graph convolution in pose prediction. In this work, we propose a new multi-scale feature suppression attention map convolutional network (AZY-GCN) for end-to-end human pose prediction tasks. We use GCN to extract features from the fine-grained scale to the coarse-grained scale and then from the coarse-grained scale to the fine-grained scale. Then we combine and decode the extracted features at each scale to obtain the residual between the input and the target pose. We also performed intermediate supervision on all predicted poses so that the network can learn more representative features. In addition, we also propose a new feature suppression attention module (FISA-block), which can effectively extract relevant information from neighboring nodes while suppressing poor GCN learning noise. Our proposed method was evaluated on the public data sets of Human3.6M and CMU Mocap. After a large number of experiments, it is shown that our method has achieved relatively advanced performance.
由于人体未来姿态的随机性和非周期性,对人体姿态的预测一直是一项非常具有挑战性的任务。在最新的研究中,图卷积被证明是一种捕获人体姿态关节之间动态关系的有效方法,有助于人体姿态的预测。此外,图卷积可以对人体的姿态进行抽象,得到多尺度的姿态集。随着抽象水平的提高,姿势的运动将变得更加稳定。虽然近年来平均预测精度有了明显提高,但图卷积在位姿预测中的应用仍有很大的探索空间。在这项工作中,我们提出了一种新的多尺度特征抑制注意图卷积网络(AZY-GCN),用于端到端人体姿势预测任务。我们使用GCN将特征从细粒度尺度提取到粗粒度尺度,再从粗粒度尺度提取到细粒度尺度。然后在每个尺度上对提取的特征进行组合和解码,得到输入和目标姿态之间的残差。我们还对所有预测的姿势进行了中间监督,以便网络可以学习到更多具有代表性的特征。此外,我们还提出了一种新的特征抑制注意模块(FISA-block),该模块可以有效地从相邻节点提取相关信息,同时抑制GCN学习噪声。在Human3.6M和CMU Mocap的公共数据集上对我们提出的方法进行了评估。经过大量的实验表明,我们的方法取得了比较先进的性能。
{"title":"AZY-GCN: Multi-scale feature suppression attentional diagram convolutional network for human pose prediction","authors":"Yang Zhang, Fan Xiao Shan, Gang He","doi":"10.1145/3507548.3507565","DOIUrl":"https://doi.org/10.1145/3507548.3507565","url":null,"abstract":"Due to the randomness and non-periodic nature of the future posture of the human body, the prediction of the posture of the human body has always been a very challenging task. In the latest research, graph convolution is proved to be an effective method to capture the dynamic relationship between the human body posture joints, which is helpful for the human body posture prediction. Moreover, graph convolution can abstract the pose of the human body to obtain a multi-scale pose set. As the level of abstraction increases, the posture movement will become more stable. Although the average prediction accuracy has improved significantly in recent years, there is still much room for exploration in the application of graph convolution in pose prediction. In this work, we propose a new multi-scale feature suppression attention map convolutional network (AZY-GCN) for end-to-end human pose prediction tasks. We use GCN to extract features from the fine-grained scale to the coarse-grained scale and then from the coarse-grained scale to the fine-grained scale. Then we combine and decode the extracted features at each scale to obtain the residual between the input and the target pose. We also performed intermediate supervision on all predicted poses so that the network can learn more representative features. In addition, we also propose a new feature suppression attention module (FISA-block), which can effectively extract relevant information from neighboring nodes while suppressing poor GCN learning noise. Our proposed method was evaluated on the public data sets of Human3.6M and CMU Mocap. After a large number of experiments, it is shown that our method has achieved relatively advanced performance.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"158 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123613339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Real-time Activity Recognition System based on Dynamic Adaptive Windows using WiFi Signals 基于动态自适应窗口的WiFi信号实时活动识别系统
Shiming Chen, Chunjing Xiao, Yanhui Han, Xianghe Du
WiFi Chanel State Information (CSI)-based activity recognition has attracted much attention in recent years. And it is extremely vital to recognize activities in time, especially for dangerous activities such as fall. In this paper, we present a real-time activity recognition system. In this system, we design a dynamic threshold-based activity segmentation method, which can address the problems of the fixed threshold and single window, and accurately detect start and end points of activities. The experiments demonstrate that our system acquires expected recognition performance.
基于WiFi香奈儿状态信息(CSI)的活动识别近年来备受关注。及时识别活动是非常重要的,特别是对于摔倒等危险活动。在本文中,我们提出了一个实时活动识别系统。在该系统中,我们设计了一种基于动态阈值的活动分割方法,解决了固定阈值和单一窗口的问题,能够准确地检测活动的起始点和结束点。实验表明,该系统取得了预期的识别性能。
{"title":"A Real-time Activity Recognition System based on Dynamic Adaptive Windows using WiFi Signals","authors":"Shiming Chen, Chunjing Xiao, Yanhui Han, Xianghe Du","doi":"10.1145/3507548.3507566","DOIUrl":"https://doi.org/10.1145/3507548.3507566","url":null,"abstract":"WiFi Chanel State Information (CSI)-based activity recognition has attracted much attention in recent years. And it is extremely vital to recognize activities in time, especially for dangerous activities such as fall. In this paper, we present a real-time activity recognition system. In this system, we design a dynamic threshold-based activity segmentation method, which can address the problems of the fixed threshold and single window, and accurately detect start and end points of activities. The experiments demonstrate that our system acquires expected recognition performance.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122923689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lightweight Object Detection Method for Mobile Robot Platform 移动机器人平台轻量化目标检测方法
Yuncheng Sang, Han Huang, Shuangqing Ma, Shouwen Cai, Zhen Shi
∗We present some experienced improvements to YOLOv5s for mobile robot platforms that occupy many system resources and are difficult to meet the requirements of the actual application. Firstly, the FPN + PAN structure is redesigned to replace this complex structure into a dilated residual module with fewer parameters and calculations. The dilated residual module is composed of dilated residual blocks with different dilation rates stacked together. Secondly, we switch some convolution modules to improved Ghost modules in the backbone. The improved Ghost modules concatenate feature maps obtained by convolution with ones generated by a linear transformation. Then, the two parts of feature maps are shuffled to boost information fusion. The model is trained on the COCO dataset. In this paper, mAP_0.5 is 56.1%, mAP_0.5:0.95 is 35.7%, and the speed is 6.1% faster than YOLOv5s. The experimental results show that the method can further increase the inference speed to ensure detection accuracy. It can well solve the task of object detection on the mobile robot platform.
对于占用大量系统资源且难以满足实际应用要求的移动机器人平台,我们提出了对YOLOv5s的一些经验改进。首先,对FPN + PAN结构进行了重新设计,将该复杂结构替换为一个参数和计算量更少的扩展剩余模块。膨胀残差模块由不同膨胀率的膨胀残差块堆叠而成。其次,我们将一些卷积模块转换为改进的Ghost模块。改进的Ghost模块将卷积得到的特征映射与线性变换生成的特征映射连接起来。然后,对特征图的两个部分进行洗牌,增强信息融合。该模型是在COCO数据集上训练的。在本文中,mAP_0.5为56.1%,mAP_0.5:0.95为35.7%,速度比YOLOv5s快6.1%。实验结果表明,该方法可以进一步提高推理速度,保证检测精度。它可以很好地解决移动机器人平台上的目标检测任务。
{"title":"Lightweight Object Detection Method for Mobile Robot Platform","authors":"Yuncheng Sang, Han Huang, Shuangqing Ma, Shouwen Cai, Zhen Shi","doi":"10.1145/3507548.3507550","DOIUrl":"https://doi.org/10.1145/3507548.3507550","url":null,"abstract":"∗We present some experienced improvements to YOLOv5s for mobile robot platforms that occupy many system resources and are difficult to meet the requirements of the actual application. Firstly, the FPN + PAN structure is redesigned to replace this complex structure into a dilated residual module with fewer parameters and calculations. The dilated residual module is composed of dilated residual blocks with different dilation rates stacked together. Secondly, we switch some convolution modules to improved Ghost modules in the backbone. The improved Ghost modules concatenate feature maps obtained by convolution with ones generated by a linear transformation. Then, the two parts of feature maps are shuffled to boost information fusion. The model is trained on the COCO dataset. In this paper, mAP_0.5 is 56.1%, mAP_0.5:0.95 is 35.7%, and the speed is 6.1% faster than YOLOv5s. The experimental results show that the method can further increase the inference speed to ensure detection accuracy. It can well solve the task of object detection on the mobile robot platform.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127593062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Single Image Super-Resolution via Residual Dictionary Learning 基于残差字典学习的单幅图像超分辨率
Yanrong Yang, Yunjie Zhang, Xiaoli Ren
Aiming at the shortcomings of traditional learning-based super-resolution (SR) reconstruction algorithms, single image super-resolution via residual dictionary learning is proposed. This method adds residual image learning to the super-resolution algorithm of beta process joint dictionary learning for coupled feature spaces. The residual dictionary pairs are learned by combining the high-resolution (HR) and low-resolution (LR) images in the external training set, which can improve the reconstruction quality and speed up the dictionary training. According to the experimental results, compared with these traditional algorithms, the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) of the proposed algorithm are significantly improved, and the visual effect is also improved.
针对传统基于学习的超分辨率重建算法的不足,提出了基于残差字典学习的单幅图像超分辨率重建算法。该方法将残差图像学习加入到耦合特征空间的β过程联合字典学习超分辨率算法中。将外部训练集中的高分辨率(HR)和低分辨率(LR)图像结合起来学习残差字典对,提高了重建质量,加快了字典训练速度。实验结果表明,与这些传统算法相比,本文算法的峰值信噪比(PSNR)和结构相似度指标(SSIM)均有显著提高,视觉效果也有所改善。
{"title":"Single Image Super-Resolution via Residual Dictionary Learning","authors":"Yanrong Yang, Yunjie Zhang, Xiaoli Ren","doi":"10.1145/3507548.3507563","DOIUrl":"https://doi.org/10.1145/3507548.3507563","url":null,"abstract":"Aiming at the shortcomings of traditional learning-based super-resolution (SR) reconstruction algorithms, single image super-resolution via residual dictionary learning is proposed. This method adds residual image learning to the super-resolution algorithm of beta process joint dictionary learning for coupled feature spaces. The residual dictionary pairs are learned by combining the high-resolution (HR) and low-resolution (LR) images in the external training set, which can improve the reconstruction quality and speed up the dictionary training. According to the experimental results, compared with these traditional algorithms, the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) of the proposed algorithm are significantly improved, and the visual effect is also improved.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132871046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A DNN-Based Method for Sea Clutter Doppler Parameters Prediction 一种基于dnn的海杂波多普勒参数预测方法
Xiaoyu Li, Yushi Zhang, Jinpeng Zhang
With the dramatic development of information technology and rapid growth of computation performances, artificial intelligent techniques have been gradually applied in all aspects of industrial research, especially in radar signal processing. However, deep learning methods utilized in radar sea clutter are just beginning, and related researches on Doppler characteristics of sea clutter remain sparse. In this paper, artificial intelligent research on sea clutter Doppler parameters prediction is developed based on real data. Firstly, classical signal processing methods for sea clutter spectral parameters extraction are introduced. Secondly, a deep neural network model is built to predict sea clutter Doppler parameters. Finally, the raised DNN model is compared to three other classical machine learning models which are widely used in regression prediction. After comprehensive comparisons with other models in different metrics, it can be concluded that DNN model built in this paper achieves better prediction results.
随着信息技术的飞速发展和计算性能的飞速提高,人工智能技术已逐渐应用于工业研究的各个方面,尤其是雷达信号处理领域。然而,深度学习方法在雷达海杂波中的应用才刚刚起步,对海杂波多普勒特性的相关研究还比较少。本文以实际数据为基础,开展了海杂波多普勒参数预测的人工智能研究。首先,介绍了海杂波频谱参数提取的经典信号处理方法。其次,建立深度神经网络模型,预测海杂波多普勒参数;最后,将提出的深度神经网络模型与其他三种广泛用于回归预测的经典机器学习模型进行了比较。通过与其他模型在不同指标上的综合比较,可以得出本文构建的DNN模型具有更好的预测效果。
{"title":"A DNN-Based Method for Sea Clutter Doppler Parameters Prediction","authors":"Xiaoyu Li, Yushi Zhang, Jinpeng Zhang","doi":"10.1145/3507548.3507595","DOIUrl":"https://doi.org/10.1145/3507548.3507595","url":null,"abstract":"With the dramatic development of information technology and rapid growth of computation performances, artificial intelligent techniques have been gradually applied in all aspects of industrial research, especially in radar signal processing. However, deep learning methods utilized in radar sea clutter are just beginning, and related researches on Doppler characteristics of sea clutter remain sparse. In this paper, artificial intelligent research on sea clutter Doppler parameters prediction is developed based on real data. Firstly, classical signal processing methods for sea clutter spectral parameters extraction are introduced. Secondly, a deep neural network model is built to predict sea clutter Doppler parameters. Finally, the raised DNN model is compared to three other classical machine learning models which are widely used in regression prediction. After comprehensive comparisons with other models in different metrics, it can be concluded that DNN model built in this paper achieves better prediction results.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"415 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133382712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Texture Dataset Construction and Texture Image Retrieval based on Deep Learning 基于深度学习的纹理数据集构建与纹理图像检索
Zhisheng Zhang, Huaijing Qu, Hengbin Wang, Jia Xu, Jiwei Wang, Yanan Wei
In the deep texture image retrieval, to address the problem that the retrieval performance is affected by the lack of sufficiently large texture image dataset used for the effective training of deep neural network, a deep learning based texture dataset construction and texture image retrieval method is proposed in this paper. First, a large-scale texture image dataset containing rich texture information is constructed based on the DTD texture image dataset, and used as the source dataset for pre-training deep neural networks. To effectively characterize the information of the source texture dataset, a revised version of the VGG16 model, called ReV-VGG16, is adaptively designed. Then, the pre-trained ReV-VGG16 model is combined with the target texture image datasets for the transfer learning, and the probability values of the output from the classification layer of the model are used for the computation of the similarity measurement to achieve the retrieval of the target texture image dataset. Finally, the retrieval experiments are conducted on four typical texture image datasets, namely, VisTex, Brodatz, STex and ALOT. The experimental results show that our method outperforms the existing state-of-the-art texture image retrieval approaches in terms of the retrieval performance.
在深度纹理图像检索中,针对缺乏足够大的纹理图像数据集用于深度神经网络的有效训练而影响检索性能的问题,提出了一种基于深度学习的纹理数据集构建和纹理图像检索方法。首先,基于DTD纹理图像数据集构建了包含丰富纹理信息的大规模纹理图像数据集,并将其作为深度神经网络预训练的源数据集;为了有效表征源纹理数据集的信息,自适应设计了VGG16模型的修正版本ReV-VGG16。然后,将预训练好的ReV-VGG16模型与目标纹理图像数据集结合进行迁移学习,利用模型分类层输出的概率值进行相似性度量计算,实现目标纹理图像数据集的检索。最后,在VisTex、Brodatz、STex和ALOT四种典型纹理图像数据集上进行检索实验。实验结果表明,该方法在检索性能方面优于现有的最先进的纹理图像检索方法。
{"title":"Texture Dataset Construction and Texture Image Retrieval based on Deep Learning","authors":"Zhisheng Zhang, Huaijing Qu, Hengbin Wang, Jia Xu, Jiwei Wang, Yanan Wei","doi":"10.1145/3507548.3507564","DOIUrl":"https://doi.org/10.1145/3507548.3507564","url":null,"abstract":"In the deep texture image retrieval, to address the problem that the retrieval performance is affected by the lack of sufficiently large texture image dataset used for the effective training of deep neural network, a deep learning based texture dataset construction and texture image retrieval method is proposed in this paper. First, a large-scale texture image dataset containing rich texture information is constructed based on the DTD texture image dataset, and used as the source dataset for pre-training deep neural networks. To effectively characterize the information of the source texture dataset, a revised version of the VGG16 model, called ReV-VGG16, is adaptively designed. Then, the pre-trained ReV-VGG16 model is combined with the target texture image datasets for the transfer learning, and the probability values of the output from the classification layer of the model are used for the computation of the similarity measurement to achieve the retrieval of the target texture image dataset. Finally, the retrieval experiments are conducted on four typical texture image datasets, namely, VisTex, Brodatz, STex and ALOT. The experimental results show that our method outperforms the existing state-of-the-art texture image retrieval approaches in terms of the retrieval performance.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130991618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1