首页 > 最新文献

International Conference on Image Processing and Intelligent Control最新文献

英文 中文
A neural network model for adversarial defense based on deep learning 基于深度学习的对抗防御神经网络模型
Pub Date : 2023-08-09 DOI: 10.1117/12.3000789
Zhiying Wang, Yong Wang
Deep learning has achieved great success in many fields, such as image classification and target detection. Adding small disturbance which is hard to be detected by the human eyes to original images can make the neural network output error results with high confidence. An image after adding small disturbance is an adversarial example. The existence of adversarial examples brings a huge security problem to deep learning. In order to effectively defend against adversarial examples attacks, an adversarial example defense method based on image reconstruction is proposed by analyzing the existing adversarial examples attack methods and defense methods. Our data set is based on ImageNet 1k data set, and some filtering and expansion are carried out. Four attack modes, FGSM, BIM, DeepFool and C&W are selected to test the defense method. Based on the EDSR network, multi-scale feature fusion module and subspace attention module are added. By capturing the global correlation information of the image, the disturbance can be removed, while the image texture details can be better preserved, and the defense performance can be improved. The experimental results show that the proposed method has good defense effect.
深度学习在图像分类、目标检测等领域取得了巨大的成功。在原始图像中加入人眼难以察觉的微小干扰,可以使神经网络输出的误差结果具有较高的置信度。加入小扰动后的图像是一个对抗性的例子。对抗性示例的存在给深度学习带来了巨大的安全问题。为了有效防御对抗性样例攻击,在分析现有对抗性样例攻击方法和防御方法的基础上,提出了一种基于图像重构的对抗性样例防御方法。我们的数据集基于ImageNet 1k数据集,并进行了一些过滤和扩展。选择FGSM、BIM、DeepFool和C&W四种攻击模式进行防御方法测试。在EDSR网络的基础上,增加了多尺度特征融合模块和子空间关注模块。通过捕获图像的全局相关信息,可以去除干扰,同时更好地保留图像纹理细节,提高防御性能。实验结果表明,该方法具有良好的防御效果。
{"title":"A neural network model for adversarial defense based on deep learning","authors":"Zhiying Wang, Yong Wang","doi":"10.1117/12.3000789","DOIUrl":"https://doi.org/10.1117/12.3000789","url":null,"abstract":"Deep learning has achieved great success in many fields, such as image classification and target detection. Adding small disturbance which is hard to be detected by the human eyes to original images can make the neural network output error results with high confidence. An image after adding small disturbance is an adversarial example. The existence of adversarial examples brings a huge security problem to deep learning. In order to effectively defend against adversarial examples attacks, an adversarial example defense method based on image reconstruction is proposed by analyzing the existing adversarial examples attack methods and defense methods. Our data set is based on ImageNet 1k data set, and some filtering and expansion are carried out. Four attack modes, FGSM, BIM, DeepFool and C&W are selected to test the defense method. Based on the EDSR network, multi-scale feature fusion module and subspace attention module are added. By capturing the global correlation information of the image, the disturbance can be removed, while the image texture details can be better preserved, and the defense performance can be improved. The experimental results show that the proposed method has good defense effect.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126726056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on optical detection technology for underwater archaeology 水下考古光学探测技术研究
Pub Date : 2023-08-09 DOI: 10.1117/12.3002208
Wei Mu, Ruohan Zheng, Wenrui Zhang
In response to the problem that the current image processing technology and underwater target recognition algorithms are not yet mature enough in the field of underwater archaeology, this article innovatively applies object detection and underwater image clarity technology to the field of underwater archaeology. We propose a method for detecting and recognizing underwater cultural heritage based on optical devices. The method includes ocean image preprocessing and underwater cultural heritage object recognition based on YOLO V4. The results of experiments demonstrate that the proposed method can effectively and accurately detect and recognize targets in the underwater cultural heritage scene, and the clear image of the underwater relics after image preprocessing can better assist archaeologists in observing the species and distribution of samples in the real scene.
针对目前水下考古领域的图像处理技术和水下目标识别算法还不够成熟的问题,本文创新性地将目标检测和水下图像清晰度技术应用于水下考古领域。提出了一种基于光学装置的水下文化遗产探测与识别方法。该方法包括海洋图像预处理和基于YOLO V4的水下文物目标识别。实验结果表明,该方法能够有效、准确地检测和识别水下文物场景中的目标,图像预处理后的水下文物清晰图像能够更好地辅助考古学家观察真实场景中样本的种类和分布。
{"title":"Research on optical detection technology for underwater archaeology","authors":"Wei Mu, Ruohan Zheng, Wenrui Zhang","doi":"10.1117/12.3002208","DOIUrl":"https://doi.org/10.1117/12.3002208","url":null,"abstract":"In response to the problem that the current image processing technology and underwater target recognition algorithms are not yet mature enough in the field of underwater archaeology, this article innovatively applies object detection and underwater image clarity technology to the field of underwater archaeology. We propose a method for detecting and recognizing underwater cultural heritage based on optical devices. The method includes ocean image preprocessing and underwater cultural heritage object recognition based on YOLO V4. The results of experiments demonstrate that the proposed method can effectively and accurately detect and recognize targets in the underwater cultural heritage scene, and the clear image of the underwater relics after image preprocessing can better assist archaeologists in observing the species and distribution of samples in the real scene.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115529067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Video description method with fusion of instance-aware temporal features 融合实例感知时间特征的视频描述方法
Pub Date : 2023-08-09 DOI: 10.1117/12.3000765
Ju Huang, He Yan, Lingkun Liu, Yuhan Liu
There are still challenges in the field of video understanding today, especially how to use natural language to describe the visual content in videos. Existing video encoder-decoder models struggle to extract deep semantic information and effectively understand the complex contextual semantics in a video sequence. Furthermore, different visual elements in the video contribute differently to the generation of video text descriptions. In this paper, we propose a video description method that fuses instance-aware temporal features. We extract local features of instances on the temporal sequence to enhance perception of temporal instances. We also employ spatial attention to perform weighted fusion of temporal features. Finally, we use bidirectional long short-term memory networks to encode the contextual semantic information of the video sequence, thereby helping to generate higher quality descriptive text. Experimental results on two public datasets demonstrate that our method achieves good performance on various evaluation metrics.
目前,视频理解领域仍存在诸多挑战,特别是如何使用自然语言来描述视频中的视觉内容。现有的视频编码器-解码器模型难以提取深度语义信息并有效理解视频序列中复杂的上下文语义。此外,视频中不同的视觉元素对视频文本描述的生成也有不同的贡献。本文提出了一种融合实例感知时间特征的视频描述方法。我们在时间序列上提取实例的局部特征,以增强对时间实例的感知。我们还利用空间注意力对时间特征进行加权融合。最后,我们使用双向长短期记忆网络对视频序列的上下文语义信息进行编码,从而有助于生成更高质量的描述性文本。在两个公共数据集上的实验结果表明,我们的方法在各种评估指标上都取得了良好的性能。
{"title":"Video description method with fusion of instance-aware temporal features","authors":"Ju Huang, He Yan, Lingkun Liu, Yuhan Liu","doi":"10.1117/12.3000765","DOIUrl":"https://doi.org/10.1117/12.3000765","url":null,"abstract":"There are still challenges in the field of video understanding today, especially how to use natural language to describe the visual content in videos. Existing video encoder-decoder models struggle to extract deep semantic information and effectively understand the complex contextual semantics in a video sequence. Furthermore, different visual elements in the video contribute differently to the generation of video text descriptions. In this paper, we propose a video description method that fuses instance-aware temporal features. We extract local features of instances on the temporal sequence to enhance perception of temporal instances. We also employ spatial attention to perform weighted fusion of temporal features. Finally, we use bidirectional long short-term memory networks to encode the contextual semantic information of the video sequence, thereby helping to generate higher quality descriptive text. Experimental results on two public datasets demonstrate that our method achieves good performance on various evaluation metrics.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114131405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D target detection based on dynamic occlusion processing 基于动态遮挡处理的三维目标检测
Pub Date : 2023-08-09 DOI: 10.1117/12.3000786
Jishen Peng, Jun Ma, Li Li
In order to solve the multi-vehicle mutual occlusion problem encountered in 3D target detection by self-driving vehicles, this paper proposes a monocular 3D detection method that includes dynamic occlusion determination. The method adds a dynamic occlusion processing module to the CenterNet3D network framework to improve the accuracy of 3D target detection of occluded vehicles in the road. Specifically, the occlusion determination module of the method uses the 2D detection results extracted from target detection as the occlusion relationship determination condition, wherein the method of changing the occlusion determination threshold with the depth value is introduced. Then the occlusion compensation module is used to compensate and adjust the 3D detection results of the occurring occluded vehicles, and finally the 3D target detection results are output. The experimental results show that the method improves the accuracy of both vehicle center point detection and 3D dimensional detection results in the case of long-distance continuous vehicle occlusion. And compared with other existing methods, the accuracy of 3D detection results and bird's-eye view detection results are improved by 1%-2.64% in the case of intersection over union of 0.5. The method can compensate for the occluded vehicles in 3D target detection and improve the accuracy
为了解决自动驾驶车辆在三维目标检测中遇到的多车相互遮挡问题,本文提出了一种包含动态遮挡确定的单目三维检测方法。该方法在CenterNet3D网络框架中增加了动态遮挡处理模块,提高了道路中遮挡车辆的三维目标检测精度。具体而言,该方法的遮挡确定模块以目标检测提取的二维检测结果作为遮挡关系确定条件,其中引入了用深度值改变遮挡确定阈值的方法。然后利用遮挡补偿模块对发生遮挡的车辆进行三维检测结果的补偿和调整,最后输出三维目标检测结果。实验结果表明,该方法提高了长距离连续遮挡情况下车辆中心点检测和三维尺寸检测结果的精度。与其他现有方法相比,在交点比并度为0.5的情况下,三维检测结果和鸟瞰检测结果的精度提高了1% ~ 2.64%。该方法可以补偿被遮挡车辆在三维目标检测中的影响,提高检测精度
{"title":"3D target detection based on dynamic occlusion processing","authors":"Jishen Peng, Jun Ma, Li Li","doi":"10.1117/12.3000786","DOIUrl":"https://doi.org/10.1117/12.3000786","url":null,"abstract":"In order to solve the multi-vehicle mutual occlusion problem encountered in 3D target detection by self-driving vehicles, this paper proposes a monocular 3D detection method that includes dynamic occlusion determination. The method adds a dynamic occlusion processing module to the CenterNet3D network framework to improve the accuracy of 3D target detection of occluded vehicles in the road. Specifically, the occlusion determination module of the method uses the 2D detection results extracted from target detection as the occlusion relationship determination condition, wherein the method of changing the occlusion determination threshold with the depth value is introduced. Then the occlusion compensation module is used to compensate and adjust the 3D detection results of the occurring occluded vehicles, and finally the 3D target detection results are output. The experimental results show that the method improves the accuracy of both vehicle center point detection and 3D dimensional detection results in the case of long-distance continuous vehicle occlusion. And compared with other existing methods, the accuracy of 3D detection results and bird's-eye view detection results are improved by 1%-2.64% in the case of intersection over union of 0.5. The method can compensate for the occluded vehicles in 3D target detection and improve the accuracy","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114918538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research and application of 3D simulation of truck formation based on Unreal Engine 基于虚幻引擎的卡车编队三维仿真研究与应用
Pub Date : 2023-08-09 DOI: 10.1117/12.3001392
Zhenzhou Wang, Fang Wu, Jiangnan Zhang, Jianguang Wu
In order to show the transport conditions of goods on different roads and provide more real and three-dimensional transport information for situation inference users, this paper proposes a simple and PID controlled three-dimensional simulation method for truck formation based on Unreal Engine. Firstly, based on the basic theory of automatic control [1] , the longitudinal lollipop controller and the transverse PID controller are designed respectively based on the lollipop control and PID control ideas, and the perception-decision framework is combined to realize the automatic driving of the truck along the spline line on the road. On this basis, a truck controller is designed to realize the truck formation driving with high recovery degree based on the leader-follower strategy. The results show that the truck based on PID control can accurately drive along the road line. With the cooperation of truck formation controller, the whole process of formation, maintenance and driving of truck formation can be basically restored.
为了展示货物在不同道路上的运输状况,为情景推理用户提供更加真实、立体的运输信息,本文提出了一种基于虚幻引擎的简单、PID控制的货车编队三维仿真方法。首先,在自动控制基本理论的基础上[1],基于棒棒糖控制和PID控制思想,分别设计纵向棒棒糖控制器和横向PID控制器,并结合感知-决策框架,实现货车沿样条线在道路上的自动驾驶。在此基础上,设计了基于leader-follower策略的货车编队高回收度驾驶控制器。结果表明,基于PID控制的卡车能够准确地沿道路行驶。在卡车编队控制器的配合下,可以基本恢复卡车编队的形成、维护和行驶的整个过程。
{"title":"Research and application of 3D simulation of truck formation based on Unreal Engine","authors":"Zhenzhou Wang, Fang Wu, Jiangnan Zhang, Jianguang Wu","doi":"10.1117/12.3001392","DOIUrl":"https://doi.org/10.1117/12.3001392","url":null,"abstract":"In order to show the transport conditions of goods on different roads and provide more real and three-dimensional transport information for situation inference users, this paper proposes a simple and PID controlled three-dimensional simulation method for truck formation based on Unreal Engine. Firstly, based on the basic theory of automatic control [1] , the longitudinal lollipop controller and the transverse PID controller are designed respectively based on the lollipop control and PID control ideas, and the perception-decision framework is combined to realize the automatic driving of the truck along the spline line on the road. On this basis, a truck controller is designed to realize the truck formation driving with high recovery degree based on the leader-follower strategy. The results show that the truck based on PID control can accurately drive along the road line. With the cooperation of truck formation controller, the whole process of formation, maintenance and driving of truck formation can be basically restored.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128967080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of design factors of an interactive interface of intangible cultural heritage APP based on user experience 基于用户体验的非物质文化遗产APP交互界面设计因素评价
Pub Date : 2023-08-09 DOI: 10.1117/12.3000771
Chengjun Zhou, Ruowei Li
In this paper, the non-cultural material heritage mobile terminal APP interface is the carrier, according to the user experience of the interactive interface design. By using user interview, observation, qualitative research and quantitative research, and based on the theoretical model of user experience, the author conducted data collection and analysis using user interview and questionnaire survey to obtain four evaluation indexes and eight sub-criteria for users' interaction interface of intangible cultural heritage apps. The analytic hierarchy process was introduced into weight calculation. The weight of each evaluation factor is obtained through investigation and calculation, and the evaluation level of each element is determined by referring to the Likert scale. The evaluation data of the design scheme is obtained through the questionnaire method, the fuzzy analysis is carried out on the results of the questionnaire, and the final evaluation results are obtained according to the principle of full membership to provide implementable improvement suggestions for the interactive interface design to improve the user experience. The research results have theoretical guiding significance for the interactive interface design of intangible cultural heritage apps.
本文以非文化物质遗产移动端APP界面为载体,根据用户体验进行交互界面设计。采用用户访谈、观察、定性研究、定量研究等方法,以用户体验理论模型为基础,采用用户访谈、问卷调查等方式进行数据收集和分析,得出非物质文化遗产类app用户交互界面的4项评价指标和8个子标准。将层次分析法引入权重计算。通过调查计算得出各评价因子的权重,参照李克特量表确定各要素的评价等级。通过问卷法获得设计方案的评价数据,对问卷结果进行模糊分析,根据全隶属原则得出最终的评价结果,为交互界面设计提供可实施的改进建议,提高用户体验。研究成果对非物质文化遗产app交互界面设计具有理论指导意义。
{"title":"Evaluation of design factors of an interactive interface of intangible cultural heritage APP based on user experience","authors":"Chengjun Zhou, Ruowei Li","doi":"10.1117/12.3000771","DOIUrl":"https://doi.org/10.1117/12.3000771","url":null,"abstract":"In this paper, the non-cultural material heritage mobile terminal APP interface is the carrier, according to the user experience of the interactive interface design. By using user interview, observation, qualitative research and quantitative research, and based on the theoretical model of user experience, the author conducted data collection and analysis using user interview and questionnaire survey to obtain four evaluation indexes and eight sub-criteria for users' interaction interface of intangible cultural heritage apps. The analytic hierarchy process was introduced into weight calculation. The weight of each evaluation factor is obtained through investigation and calculation, and the evaluation level of each element is determined by referring to the Likert scale. The evaluation data of the design scheme is obtained through the questionnaire method, the fuzzy analysis is carried out on the results of the questionnaire, and the final evaluation results are obtained according to the principle of full membership to provide implementable improvement suggestions for the interactive interface design to improve the user experience. The research results have theoretical guiding significance for the interactive interface design of intangible cultural heritage apps.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114128327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of Videolog visualization technology in workover operation 视频可视化技术在修井作业中的应用
Pub Date : 2023-08-09 DOI: 10.1117/12.3001436
Ying Zhang, Jiatian Zhang, Wenhao Jin
The actual underground situation is of great significance for workover operation. Videolog visualization technology can clearly and accurately obtain the underground color video information, and provide effective guidance for workover operation. This paper introduces the system composition, working principle and functional parameters of Videolog equipment, and gives an example of its practical application in workover operation, which shows that Videolog visualization technology is more efficient, safe and intuitive than traditional downhole video technology, and has a good application prospect in workover operation field.
井下实际情况对修井作业具有重要意义。视频可视化技术能够清晰准确地获取井下彩色视频信息,为修井作业提供有效指导。介绍了Videolog设备的系统组成、工作原理和功能参数,并给出了其在修井作业中的实际应用实例,表明Videolog可视化技术比传统井下视频技术更高效、安全、直观,在修井作业领域具有良好的应用前景。
{"title":"Application of Videolog visualization technology in workover operation","authors":"Ying Zhang, Jiatian Zhang, Wenhao Jin","doi":"10.1117/12.3001436","DOIUrl":"https://doi.org/10.1117/12.3001436","url":null,"abstract":"The actual underground situation is of great significance for workover operation. Videolog visualization technology can clearly and accurately obtain the underground color video information, and provide effective guidance for workover operation. This paper introduces the system composition, working principle and functional parameters of Videolog equipment, and gives an example of its practical application in workover operation, which shows that Videolog visualization technology is more efficient, safe and intuitive than traditional downhole video technology, and has a good application prospect in workover operation field.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116248229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ORB feature extraction and feature matching based on geometric constraints 基于几何约束的ORB特征提取与特征匹配
Pub Date : 2023-08-09 DOI: 10.1117/12.3000969
Zhenyu Wu, Xueqian Wu
This paper studies feature extraction and feature matching in visual odometry. Aiming at the problems that ORB feature extraction does not have illumination invariance and feature distribution is uneven, an adaptive threshold algorithm for feature extraction is added, and a quadtree is used to manage feature points. Aiming at the problem of high time cost of the feature matching algorithm, an outlier removal algorithm based on geometric constraints is proposed, and the constraint set is constructed by using the slope, distance, and descriptor distance between the matching feature point pairs. Tested on the TUM dataset, the feature extraction algorithm can adapt to scenes with different brightness, and the robustness is improved. The time taken by outlier removal algorithm based on geometric constraints is about 10% of RANSAC. After that, combined with RANSAC, the running time of RANSAC can be reduced by 60%. Our algorithm can improve the estimation accuracy and robustness of the system.
本文研究了视觉里程计中的特征提取和特征匹配。针对ORB特征提取不具有光照不变性和特征分布不均匀的问题,加入自适应阈值算法进行特征提取,并采用四叉树对特征点进行管理。针对特征匹配算法耗时高的问题,提出了一种基于几何约束的离群值去除算法,并利用匹配特征点对之间的斜率、距离和描述符距离构造约束集。在TUM数据集上测试,特征提取算法能够适应不同亮度的场景,鲁棒性得到了提高。基于几何约束的离群点去除算法所需的时间约为RANSAC算法的10%。之后,与RANSAC结合使用,RANSAC的运行时间可缩短60%。该算法可以提高系统的估计精度和鲁棒性。
{"title":"ORB feature extraction and feature matching based on geometric constraints","authors":"Zhenyu Wu, Xueqian Wu","doi":"10.1117/12.3000969","DOIUrl":"https://doi.org/10.1117/12.3000969","url":null,"abstract":"This paper studies feature extraction and feature matching in visual odometry. Aiming at the problems that ORB feature extraction does not have illumination invariance and feature distribution is uneven, an adaptive threshold algorithm for feature extraction is added, and a quadtree is used to manage feature points. Aiming at the problem of high time cost of the feature matching algorithm, an outlier removal algorithm based on geometric constraints is proposed, and the constraint set is constructed by using the slope, distance, and descriptor distance between the matching feature point pairs. Tested on the TUM dataset, the feature extraction algorithm can adapt to scenes with different brightness, and the robustness is improved. The time taken by outlier removal algorithm based on geometric constraints is about 10% of RANSAC. After that, combined with RANSAC, the running time of RANSAC can be reduced by 60%. Our algorithm can improve the estimation accuracy and robustness of the system.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121496173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on auxiliary decision-making for sea striking of naval aviation based on deep reinforcement learning 基于深度强化学习的海军航空兵海上打击辅助决策研究
Pub Date : 2023-08-09 DOI: 10.1117/12.3000933
Minjie Wu, D. Yin
The situation of the future naval battlefield will become more and more complex, and it will become a trend to develop various military auxiliary decision-making systems based on artificial intelligence and big data technology. This paper sorts out the key technologies of the auxiliary decision-making system based on deep reinforcement learning. On this basis, it proposes the construction method of the naval aviation sea-striking agent model, and completes the construction of the training framework with the combat deduction system as the environment. Finally, it summarizes and prospects some of future work.
未来海军战场形势将越来越复杂,发展基于人工智能和大数据技术的各种军事辅助决策系统将成为一种趋势。对基于深度强化学习的辅助决策系统的关键技术进行了梳理。在此基础上,提出了海军航空兵海上打击主体模型的构建方法,并以作战演绎系统为环境,完成了训练框架的构建。最后,对今后的工作进行了总结和展望。
{"title":"Research on auxiliary decision-making for sea striking of naval aviation based on deep reinforcement learning","authors":"Minjie Wu, D. Yin","doi":"10.1117/12.3000933","DOIUrl":"https://doi.org/10.1117/12.3000933","url":null,"abstract":"The situation of the future naval battlefield will become more and more complex, and it will become a trend to develop various military auxiliary decision-making systems based on artificial intelligence and big data technology. This paper sorts out the key technologies of the auxiliary decision-making system based on deep reinforcement learning. On this basis, it proposes the construction method of the naval aviation sea-striking agent model, and completes the construction of the training framework with the combat deduction system as the environment. Finally, it summarizes and prospects some of future work.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133063780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Infrared small target recognition in waterways based on YOLOv5 algorithm 基于YOLOv5算法的水路红外小目标识别
Pub Date : 2023-08-09 DOI: 10.1117/12.3002081
Yikai Fan, Yingjun Zhang
YOLOv5 is one of the target detection algorithms with fast detection speed and high accuracy, but it has the problems of insufficient sensory field and low accuracy of small target detection. In order to solve above problems, an improved YOLOv5 network model, i.e., an improved YOLOv5-TI model based on the attention mechanism, is proposed. The attention module is added to the backbone network when extracting features to improve the target detection accuracy, and the input features are shifted windowed for self-attention calculation to effectively utilize the features and improve the small target detection accuracy; the proposed model YOLOv5-TI is experimented on the self-built inland infrared dataset, and the mAP value reaches 95.5%, and the results show that YOLOv5-TI can effectively improve the target detection accuracy. The inland vessels equipped with visual intelligent perception system can effectively identify the targets on water, and they have wide applications in the fields of surface exploration and autonomous search and rescue.
YOLOv5是一种检测速度快、精度高的目标检测算法,但存在感觉场不足、小目标检测精度低等问题。为了解决上述问题,本文提出了一种改进的YOLOv5网络模型,即基于注意机制的改进YOLOv5- ti模型。在提取特征时在骨干网络中加入注意模块,提高目标检测精度,对输入特征进行移窗自注意计算,有效利用特征,提高小目标检测精度;本文提出的模型YOLOv5-TI在自建的内陆红外数据集上进行了实验,mAP值达到95.5%,结果表明YOLOv5-TI能有效提高目标检测精度。配备视觉智能感知系统的内河船舶能够有效识别水面目标,在水面探测和自主搜救等领域有着广泛的应用。
{"title":"Infrared small target recognition in waterways based on YOLOv5 algorithm","authors":"Yikai Fan, Yingjun Zhang","doi":"10.1117/12.3002081","DOIUrl":"https://doi.org/10.1117/12.3002081","url":null,"abstract":"YOLOv5 is one of the target detection algorithms with fast detection speed and high accuracy, but it has the problems of insufficient sensory field and low accuracy of small target detection. In order to solve above problems, an improved YOLOv5 network model, i.e., an improved YOLOv5-TI model based on the attention mechanism, is proposed. The attention module is added to the backbone network when extracting features to improve the target detection accuracy, and the input features are shifted windowed for self-attention calculation to effectively utilize the features and improve the small target detection accuracy; the proposed model YOLOv5-TI is experimented on the self-built inland infrared dataset, and the mAP value reaches 95.5%, and the results show that YOLOv5-TI can effectively improve the target detection accuracy. The inland vessels equipped with visual intelligent perception system can effectively identify the targets on water, and they have wide applications in the fields of surface exploration and autonomous search and rescue.","PeriodicalId":210802,"journal":{"name":"International Conference on Image Processing and Intelligent Control","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130373002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Conference on Image Processing and Intelligent Control
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1