首页 > 最新文献

2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)最新文献

英文 中文
Research on Product Style Design Based on Genetic Algorithm 基于遗传算法的产品风格设计研究
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177467
Cheng Yang, Lei Kong
To satisfy the differentiated requirements of different user groups and even individuals needs abundant design solutions, leading to a significant increase in design time and labor costs. Through the interactive evolutionary algorithm, a new product intelligent design method is proposed to solve the problem of intelligent generation of batched appearance design schemes, and the design scheme “population” that matches the target style image is directly obtained. First, model the stylized parameters of the product. Then, a neural network is used to establish a mapping model of the style image space and the appearance parameters of the product. Finally, through the collaborative evolution mechanism, the intelligent generation of product design solutions is realized. The results show that this method greatly eases the evaluation fatigue problem in evolutionary calculations while obtaining the target style product design scheme.
为了满足不同用户群体甚至个人的差异化需求,需要丰富的设计方案,导致设计时间和人工成本显著增加。通过交互进化算法,提出了一种新的产品智能设计方法,解决了批量外观设计方案的智能生成问题,直接获得了与目标风格图像匹配的设计方案“种群”。首先,对产品的风格化参数进行建模。然后,利用神经网络建立了样式图像空间与产品外观参数的映射模型。最后,通过协同演化机制,实现产品设计方案的智能生成。结果表明,该方法在获得目标样式产品设计方案时,极大地缓解了进化计算中的评估疲劳问题。
{"title":"Research on Product Style Design Based on Genetic Algorithm","authors":"Cheng Yang, Lei Kong","doi":"10.1109/ICIVC50857.2020.9177467","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177467","url":null,"abstract":"To satisfy the differentiated requirements of different user groups and even individuals needs abundant design solutions, leading to a significant increase in design time and labor costs. Through the interactive evolutionary algorithm, a new product intelligent design method is proposed to solve the problem of intelligent generation of batched appearance design schemes, and the design scheme “population” that matches the target style image is directly obtained. First, model the stylized parameters of the product. Then, a neural network is used to establish a mapping model of the style image space and the appearance parameters of the product. Finally, through the collaborative evolution mechanism, the intelligent generation of product design solutions is realized. The results show that this method greatly eases the evaluation fatigue problem in evolutionary calculations while obtaining the target style product design scheme.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"6 1","pages":"317-321"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73029350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Application of CEM Algorithm in the Field of Tunnel Crack Identification CEM算法在隧道裂缝识别中的应用
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177491
Bingqing Niu, Hongtao Wu, Ying Meng
Cracks are one of the most common and serious diseases of tunnel lining, which seriously threatens the safety of vehicles and requires regular inspection and measurement. In view of the problems of underexposure, uneven illumination and serious noise of the collected images in the tunnel, after the image is evenly processed, a denoising method combined with median filtering and bilateral filtering is constructed, which can filter out a lot of noise on the basis of protecting the details of the crack edge. Due to the large number of mechanical scratches and disturbing textures in the tunnel lining, EMAP is used to enhance features after Gabor filtering, and the improved CEM segmentation algorithm is used to effectively overcome the inaccurate segmentation of traditional algorithms and obtain binary images of cracks. The experimental results show that the proposed algorithm can identify the accuracy of tunnel lining cracks by more than 92%, which verifies the effectiveness of the proposed algorithm.
裂缝是隧道衬砌最常见、最严重的病害之一,严重威胁着车辆的安全,需要定期检查和测量。针对隧道内采集图像存在曝光不足、光照不均匀、噪声严重等问题,在对图像进行均匀处理后,构建了一种中值滤波和双边滤波相结合的去噪方法,在保护裂缝边缘细节的基础上滤除大量噪声。针对隧道衬砌中存在大量的机械划痕和扰动纹理,采用EMAP对Gabor滤波后的特征进行增强,并采用改进的CEM分割算法有效克服传统算法分割不准确的问题,获得裂纹的二值图像。实验结果表明,该算法对隧道衬砌裂缝的识别准确率达到92%以上,验证了该算法的有效性。
{"title":"Application of CEM Algorithm in the Field of Tunnel Crack Identification","authors":"Bingqing Niu, Hongtao Wu, Ying Meng","doi":"10.1109/ICIVC50857.2020.9177491","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177491","url":null,"abstract":"Cracks are one of the most common and serious diseases of tunnel lining, which seriously threatens the safety of vehicles and requires regular inspection and measurement. In view of the problems of underexposure, uneven illumination and serious noise of the collected images in the tunnel, after the image is evenly processed, a denoising method combined with median filtering and bilateral filtering is constructed, which can filter out a lot of noise on the basis of protecting the details of the crack edge. Due to the large number of mechanical scratches and disturbing textures in the tunnel lining, EMAP is used to enhance features after Gabor filtering, and the improved CEM segmentation algorithm is used to effectively overcome the inaccurate segmentation of traditional algorithms and obtain binary images of cracks. The experimental results show that the proposed algorithm can identify the accuracy of tunnel lining cracks by more than 92%, which verifies the effectiveness of the proposed algorithm.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"1 1","pages":"232-236"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85011439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Real-Time Measurement of Thread Number of Rail Fastener 钢轨紧固件螺纹数的实时测量
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177448
X. Luo, Ke-bin Jia, Pengyu Liu, Daoquan Xiong, Xiuchen Tian
A real-time measurement method for the thread number of rail fastener is proposed. The measurement is achieved by the processing of thread images. After analysis, the fastener thread has the elliptical feature, so two parameters‐‐‐elliptical similarity and elliptical integrity are proposed. Based on the two parameters, the thread region is located. Combining with the light intensity distribution of thread region and thread positions, the thread number is measured. 200 samples are chosen to verify the validity of this method. Experimental data show that the limitation error of this method can reach 0.1985 threads.
提出了一种钢轨扣件螺纹数的实时测量方法。测量是通过对螺纹图像的处理来实现的。经分析,紧固件螺纹具有椭圆特征,提出了椭圆相似度和椭圆完整性两个参数。根据这两个参数定位线程区域。结合螺纹区域的光强分布和螺纹位置,测量螺纹数。选取200个样本验证了该方法的有效性。实验数据表明,该方法的限制误差可达0.1985个线程。
{"title":"Real-Time Measurement of Thread Number of Rail Fastener","authors":"X. Luo, Ke-bin Jia, Pengyu Liu, Daoquan Xiong, Xiuchen Tian","doi":"10.1109/ICIVC50857.2020.9177448","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177448","url":null,"abstract":"A real-time measurement method for the thread number of rail fastener is proposed. The measurement is achieved by the processing of thread images. After analysis, the fastener thread has the elliptical feature, so two parameters‐‐‐elliptical similarity and elliptical integrity are proposed. Based on the two parameters, the thread region is located. Combining with the light intensity distribution of thread region and thread positions, the thread number is measured. 200 samples are chosen to verify the validity of this method. Experimental data show that the limitation error of this method can reach 0.1985 threads.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"9 1","pages":"312-316"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85559113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D Object Recognition Method Based on Improved Canny Edge Detection Algorithm in Augmented Reality 增强现实中基于改进Canny边缘检测算法的三维目标识别方法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177488
Tianhang Gao, Zhenhao Yang
Augmented reality (AR) superimposes computer-generated virtual objects on real scenes to gain immersive experience. Effective recognition of 3D objects in real scenes is the fundamental requirement in AR. The traditional Canny edge detection algorithm ignores the important boundary information about the object, thus decreasing the recognition accuracy. In this paper, we improve Canny to propose a novel 3D object recognition method, where median filtering is adopted in order to extract the contour of the object instead of Gaussian fuzzy. An operator based on wedge template is designed to improve the boundary detection effect of the corner. Local feature descriptors are then introduced to describe the local feature points of the object. Finally, SLAM technology is conducted to ensure that the virtual model is stably superimposed above the 3D object. The experimental results show that the proposed method is able to retain the edge information of the object well and can be combined with local feature descriptors to accurately recognize 3D objects.
增强现实(AR)将计算机生成的虚拟物体叠加在真实场景上,以获得身临其境的体验。有效识别真实场景中的三维物体是增强现实的基本要求,传统的Canny边缘检测算法忽略了物体的重要边界信息,从而降低了识别精度。本文对Canny进行改进,提出了一种新的三维物体识别方法,该方法采用中值滤波来提取物体的轮廓,而不是高斯模糊。为了提高边角的边界检测效果,设计了一种基于楔形模板的算子。然后引入局部特征描述符来描述目标的局部特征点。最后进行SLAM技术,保证虚拟模型稳定叠加在三维物体之上。实验结果表明,该方法能够很好地保留物体的边缘信息,并能与局部特征描述子相结合,实现对三维物体的准确识别。
{"title":"3D Object Recognition Method Based on Improved Canny Edge Detection Algorithm in Augmented Reality","authors":"Tianhang Gao, Zhenhao Yang","doi":"10.1109/ICIVC50857.2020.9177488","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177488","url":null,"abstract":"Augmented reality (AR) superimposes computer-generated virtual objects on real scenes to gain immersive experience. Effective recognition of 3D objects in real scenes is the fundamental requirement in AR. The traditional Canny edge detection algorithm ignores the important boundary information about the object, thus decreasing the recognition accuracy. In this paper, we improve Canny to propose a novel 3D object recognition method, where median filtering is adopted in order to extract the contour of the object instead of Gaussian fuzzy. An operator based on wedge template is designed to improve the boundary detection effect of the corner. Local feature descriptors are then introduced to describe the local feature points of the object. Finally, SLAM technology is conducted to ensure that the virtual model is stably superimposed above the 3D object. The experimental results show that the proposed method is able to retain the edge information of the object well and can be combined with local feature descriptors to accurately recognize 3D objects.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"22 1","pages":"19-23"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87898239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Deep Learning Approach in Gregg Shorthand Word to English-Word Conversion Gregg速记词到英语单词转换的深度学习方法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177452
Dionis A. Padilla, Nicole Kim U. Vitug, Julius Benito S. Marquez
Shorthand or Stenography has been used in a variety of fields of practice, particularly by court stenographers. To record every detail of the hearing, a stenographer must write fast and accurate In the Philippines, the stenographers still used the conventional way of writing shorthand, which is by hand. Transcribing shorthand writing is time-consuming and sometimes confusing because of a lot of characters or words to be transcribed. Another problem is that only a stenographer can understand and translate shorthand writing. What if there is no stenographer available to decipher a document? A deep learning approach was used to implement and developed an automated Gregg shorthand word to English-word conversion. The Convolutional Neural Network (CNN) model used was the Inception-v3 in TensorFlow platform, an open-source algorithm used for object classification. The training datasets consist of 135 Legal Terminologies with 120 images per word with a total of 16,200 datasets. The trained model achieved a validation accuracy of 91%. For testing, 10 trials per legal terminology were executed with a total of 1,350 handwritten Gregg Shorthand words tested. The system correctly translated a total of 739 words resulting in 54.74% accuracy.
速记或速记术已被用于各种领域的实践,特别是法庭速记员。为了记录听证会的每一个细节,速记员必须写得又快又准确。在菲律宾,速记员仍然使用传统的手写速记方式。抄写速记是费时的,有时由于要抄写很多字符或单词而令人困惑。另一个问题是,只有速记员才能理解和翻译速记。如果没有速记员可以破译一份文件怎么办?采用深度学习方法实现并开发了Gregg速记词到英语单词的自动转换。使用的卷积神经网络(CNN)模型是TensorFlow平台中的Inception-v3,这是一种用于对象分类的开源算法。训练数据集包括135个法律术语,每个词120个图像,总共16,200个数据集。训练后的模型验证准确率达到91%。为了进行测试,每个法律术语执行了10次测试,总共测试了1,350个手写的Gregg速记单词。该系统正确翻译了739个单词,准确率为54.74%。
{"title":"Deep Learning Approach in Gregg Shorthand Word to English-Word Conversion","authors":"Dionis A. Padilla, Nicole Kim U. Vitug, Julius Benito S. Marquez","doi":"10.1109/ICIVC50857.2020.9177452","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177452","url":null,"abstract":"Shorthand or Stenography has been used in a variety of fields of practice, particularly by court stenographers. To record every detail of the hearing, a stenographer must write fast and accurate In the Philippines, the stenographers still used the conventional way of writing shorthand, which is by hand. Transcribing shorthand writing is time-consuming and sometimes confusing because of a lot of characters or words to be transcribed. Another problem is that only a stenographer can understand and translate shorthand writing. What if there is no stenographer available to decipher a document? A deep learning approach was used to implement and developed an automated Gregg shorthand word to English-word conversion. The Convolutional Neural Network (CNN) model used was the Inception-v3 in TensorFlow platform, an open-source algorithm used for object classification. The training datasets consist of 135 Legal Terminologies with 120 images per word with a total of 16,200 datasets. The trained model achieved a validation accuracy of 91%. For testing, 10 trials per legal terminology were executed with a total of 1,350 handwritten Gregg Shorthand words tested. The system correctly translated a total of 739 words resulting in 54.74% accuracy.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"35 1","pages":"204-210"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82796768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Remote Sensing Scene Classification with Dual Attention-Aware Network 基于双注意感知网络的遥感场景分类
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177460
Yue Gao, Jun Shi, Jun Li, Ruoyu Wang
Remote sensing scene classification is of great importance to remote sensing image analysis. Most existing methods based on Convolutional Neural Network (CNN) fail to discriminate the crucial information from the complex scene content due to the intraclass diversity. In this paper, we propose a dual attention-aware network for remote sensing scene classification. Specifically, we use two kinds of attention modules (i.e. channel and spatial attentions) to explore the contextual dependencies from the channel and spatial dimensions respectively. The channel attention module intends to capture the channel-wise feature dependencies and further exploit the significant semantic attention. On the other hand, the spatial attention module aims to concentrate the attentive spatial locations and thus discover the discriminative parts inside the scene. The outputs of two attention modules are finally integrated as the attention-aware feature representation for improving classification performance. Experimental results on RSSCN7 and AID benchmark datasets show the effectiveness and superiority of the proposed methods for scene classification in remote sensing imagery.
遥感场景分类是遥感图像分析的重要内容。现有的基于卷积神经网络(CNN)的方法由于类内多样性,无法从复杂的场景内容中识别出关键信息。本文提出了一种用于遥感场景分类的双注意感知网络。具体而言,我们使用两种注意模块(即通道注意和空间注意)分别从通道和空间维度探索语境依赖关系。通道注意模块旨在捕获通道特征依赖,并进一步利用重要的语义注意。另一方面,空间注意模块旨在集中注意的空间位置,从而发现场景内部的判别部分。最后将两个注意模块的输出集成为注意感知特征表示,以提高分类性能。在RSSCN7和AID基准数据集上的实验结果表明了所提方法在遥感影像场景分类中的有效性和优越性。
{"title":"Remote Sensing Scene Classification with Dual Attention-Aware Network","authors":"Yue Gao, Jun Shi, Jun Li, Ruoyu Wang","doi":"10.1109/ICIVC50857.2020.9177460","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177460","url":null,"abstract":"Remote sensing scene classification is of great importance to remote sensing image analysis. Most existing methods based on Convolutional Neural Network (CNN) fail to discriminate the crucial information from the complex scene content due to the intraclass diversity. In this paper, we propose a dual attention-aware network for remote sensing scene classification. Specifically, we use two kinds of attention modules (i.e. channel and spatial attentions) to explore the contextual dependencies from the channel and spatial dimensions respectively. The channel attention module intends to capture the channel-wise feature dependencies and further exploit the significant semantic attention. On the other hand, the spatial attention module aims to concentrate the attentive spatial locations and thus discover the discriminative parts inside the scene. The outputs of two attention modules are finally integrated as the attention-aware feature representation for improving classification performance. Experimental results on RSSCN7 and AID benchmark datasets show the effectiveness and superiority of the proposed methods for scene classification in remote sensing imagery.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"109 1","pages":"171-175"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88648111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Air Quality Inference with Deep Convolutional Conditional Random Field 基于深度卷积条件随机场的空气质量推断
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177449
Zhe Luo, You Yu, Daijun Zhang, Shijie Feng, H. Yu, Yongxin Chang, Wei Shen
Conditional Random Field is a discriminative model for time series data. In this paper, we propose an improved CRF and apply it to the task of air quality inference. Different from the classical CRF, our linear chain CRF is a supervised learning based on the deep convolution neural network, which has a strong learning ability and fast processing speed for the engineering big data. Specifically, we model the state feature function and the state transition feature function with deep convolutional neural network. The parameter space can store more feature expressions learned from a large number of data. For the state transition feature function of linear conditional random field, we add the influence of input sequence on this function. Through the modelling and learning both vertex features and edge features from data, we obtain a more powerful and more efficient CRF. Experiments on natural language and air quality data show our CRF can achieve higher accuracy.
条件随机场是时间序列数据的判别模型。在本文中,我们提出了一种改进的CRF,并将其应用于空气质量推断任务。与经典的CRF不同,我们的线性链CRF是基于深度卷积神经网络的监督学习,对于工程大数据具有很强的学习能力和快速的处理速度。具体来说,我们利用深度卷积神经网络对状态特征函数和状态转移特征函数进行建模。参数空间可以存储更多从大量数据中学习到的特征表达式。对于线性条件随机场的状态转移特征函数,我们加入了输入序列对该函数的影响。通过建模并从数据中学习顶点特征和边缘特征,我们得到了一个更强大、更高效的CRF。在自然语言和空气质量数据上的实验表明,该算法可以达到较高的准确率。
{"title":"Air Quality Inference with Deep Convolutional Conditional Random Field","authors":"Zhe Luo, You Yu, Daijun Zhang, Shijie Feng, H. Yu, Yongxin Chang, Wei Shen","doi":"10.1109/ICIVC50857.2020.9177449","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177449","url":null,"abstract":"Conditional Random Field is a discriminative model for time series data. In this paper, we propose an improved CRF and apply it to the task of air quality inference. Different from the classical CRF, our linear chain CRF is a supervised learning based on the deep convolution neural network, which has a strong learning ability and fast processing speed for the engineering big data. Specifically, we model the state feature function and the state transition feature function with deep convolutional neural network. The parameter space can store more feature expressions learned from a large number of data. For the state transition feature function of linear conditional random field, we add the influence of input sequence on this function. Through the modelling and learning both vertex features and edge features from data, we obtain a more powerful and more efficient CRF. Experiments on natural language and air quality data show our CRF can achieve higher accuracy.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"9 1","pages":"296-302"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74354346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Method to Improve the Lining Images Quality in Complex Tunnel Scenes 一种提高复杂隧道场景中衬砌图像质量的方法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177462
Ying Meng, Hongtao Wu, Bingqing Niu
The lining image collected by the tunnel detection equipment will be degraded by the uneven gray distribution of the collected image due to the restriction of the site environment and hardware resources of the tunnel. In serious cases, the whole image is dim and fuzzy, and the disease feature information cannot be identified from the image background. In order to solve these problems, this paper proposes an image adaptive smoothing and image high frequency edge preserving optimization algorithm for tunnel lining environment. Compared with the traditional image preprocessing and image denoising algorithm, this algorithm improves the problem of the disease gray feature information jumping and information loss in the tunnel lining image due to the imbalance of gray level and the noise interference, and ensures the effectiveness of the original image interested in the disease target area information. Compared with a large number of experimental data, the improved algorithm has a great improvement in convergence speed and image quality.
由于隧道现场环境和硬件资源的限制,隧道检测设备采集的衬砌图像灰度分布不均匀,会导致图像质量下降。严重的情况下,整个图像暗淡模糊,无法从图像背景中识别疾病特征信息。为了解决这些问题,本文提出了一种适用于隧道衬砌环境的图像自适应平滑和图像高频保边优化算法。与传统的图像预处理和图像去噪算法相比,该算法改善了隧道衬砌图像中由于灰度不平衡和噪声干扰导致的疾病灰度特征信息跳跃和信息丢失的问题,保证了原始图像对疾病目标区域信息感兴趣的有效性。与大量实验数据相比,改进后的算法在收敛速度和图像质量方面都有很大提高。
{"title":"A Method to Improve the Lining Images Quality in Complex Tunnel Scenes","authors":"Ying Meng, Hongtao Wu, Bingqing Niu","doi":"10.1109/ICIVC50857.2020.9177462","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177462","url":null,"abstract":"The lining image collected by the tunnel detection equipment will be degraded by the uneven gray distribution of the collected image due to the restriction of the site environment and hardware resources of the tunnel. In serious cases, the whole image is dim and fuzzy, and the disease feature information cannot be identified from the image background. In order to solve these problems, this paper proposes an image adaptive smoothing and image high frequency edge preserving optimization algorithm for tunnel lining environment. Compared with the traditional image preprocessing and image denoising algorithm, this algorithm improves the problem of the disease gray feature information jumping and information loss in the tunnel lining image due to the imbalance of gray level and the noise interference, and ensures the effectiveness of the original image interested in the disease target area information. Compared with a large number of experimental data, the improved algorithm has a great improvement in convergence speed and image quality.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"57 5 1","pages":"199-203"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79830096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale Fast Detection of Objects in High Resolution Remote Sensing Images 高分辨率遥感图像中目标的多尺度快速检测
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177484
Longwei Li, Jiangbo Xi, Wandong Jiang, Ming Cong, Ling Han, Yun Yang
Objects detection in high resolution (HR) remote sensing images plays an important role in modern military, national defense, and commercial field. Because of a variety of object types and different sizes, it is difficulty to realize the rapid detection of multi-scale high resolution remote sensing objects, and provides support for succeeding decision making responses. This paper proposes a multi-scale fast detection method of remote sensing image objects with deep learning model, named YOLOv3. The COCO data model is used to establish the high resolution remote sensing image set based on the NWPU data. The proposed model can realize automatic learning of object features, which has good properties on generalization and robustness. It can also overcome the deficiency of traditional object detection method needing manual feature design for different objects. The experimental results show that the average detection accuracy of objects with different sizes in high resolution remote sensing images can reach 93.50%, which demonstrates that the proposed method can achieve rapid detection of different types of multi-scale objects.
高分辨率遥感图像目标检测在现代军事、国防和商业领域发挥着重要作用。由于目标类型多样、大小不一,难以实现多尺度高分辨率遥感目标的快速检测,为后续的决策响应提供支持。本文提出了一种基于深度学习模型的遥感影像目标多尺度快速检测方法YOLOv3。利用COCO数据模型,建立了基于NWPU数据的高分辨率遥感影像集。该模型能够实现对象特征的自动学习,具有良好的泛化和鲁棒性。它还可以克服传统目标检测方法需要针对不同目标进行人工特征设计的不足。实验结果表明,高分辨率遥感图像中不同尺寸目标的平均检测精度可达93.50%,表明该方法可以实现不同类型多尺度目标的快速检测。
{"title":"Multi-scale Fast Detection of Objects in High Resolution Remote Sensing Images","authors":"Longwei Li, Jiangbo Xi, Wandong Jiang, Ming Cong, Ling Han, Yun Yang","doi":"10.1109/ICIVC50857.2020.9177484","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177484","url":null,"abstract":"Objects detection in high resolution (HR) remote sensing images plays an important role in modern military, national defense, and commercial field. Because of a variety of object types and different sizes, it is difficulty to realize the rapid detection of multi-scale high resolution remote sensing objects, and provides support for succeeding decision making responses. This paper proposes a multi-scale fast detection method of remote sensing image objects with deep learning model, named YOLOv3. The COCO data model is used to establish the high resolution remote sensing image set based on the NWPU data. The proposed model can realize automatic learning of object features, which has good properties on generalization and robustness. It can also overcome the deficiency of traditional object detection method needing manual feature design for different objects. The experimental results show that the average detection accuracy of objects with different sizes in high resolution remote sensing images can reach 93.50%, which demonstrates that the proposed method can achieve rapid detection of different types of multi-scale objects.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"18 1","pages":"5-10"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84101658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dictionary Learning for Visual Tracking with Dimensionality Reduction 基于降维的视觉跟踪字典学习
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177445
Jun Wang, Yuanyun Wang, Shaoquan Zhang, Chenguang Xu, Chengzhi Deng
Recently, visual tracking has seen much progress in either accuracy or speed. However, due to drastic illumination variation, partial occlusion, scale variation and out-of-plane rotation, visual tracking remains a challenging task. Dealing with complicated appearance variations is an open issue in visual tracking. Existing trackers represent target candidates by a combination of target templates or previous tracking results under some constraints. When a drastic appearance variation occurs or some appearance variations occur simultaneously, such target representations are not robust. In this paper, we present a discriminative dictionary learning based target representation. A target candidate is represented via a linear combination of atoms in a learnt dictionary. The online dictionary learning can learn the appearance variations in tracking processing. So, the learnt dictionary can cover all of kinds of appearance variations. Based on this kind of target representation, a novel tracking algorithm is proposed. Extensive experiments on challenging sequences in popular tracking benchmark demonstrate competing tracking performances against some state-of-the-art trackers.
最近,视觉追踪在准确性和速度上都取得了很大的进步。然而,由于强烈的光照变化、部分遮挡、尺度变化和面外旋转,视觉跟踪仍然是一项具有挑战性的任务。处理复杂的外观变化是视觉跟踪中的一个开放性问题。现有的跟踪器在一定的约束条件下,通过目标模板或先前跟踪结果的组合来表示目标候选者。当发生剧烈的外观变化或同时发生一些外观变化时,这种目标表征不具有鲁棒性。本文提出了一种基于判别字典学习的目标表示方法。目标候选对象通过学习字典中的原子的线性组合来表示。在线词典学习可以学习跟踪处理过程中的外观变化。所以,学到的字典可以涵盖所有的外观变化。基于这种目标表示,提出了一种新的目标跟踪算法。在流行的跟踪基准中对具有挑战性的序列进行了大量实验,证明了与一些最先进的跟踪器的跟踪性能存在竞争。
{"title":"Dictionary Learning for Visual Tracking with Dimensionality Reduction","authors":"Jun Wang, Yuanyun Wang, Shaoquan Zhang, Chenguang Xu, Chengzhi Deng","doi":"10.1109/ICIVC50857.2020.9177445","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177445","url":null,"abstract":"Recently, visual tracking has seen much progress in either accuracy or speed. However, due to drastic illumination variation, partial occlusion, scale variation and out-of-plane rotation, visual tracking remains a challenging task. Dealing with complicated appearance variations is an open issue in visual tracking. Existing trackers represent target candidates by a combination of target templates or previous tracking results under some constraints. When a drastic appearance variation occurs or some appearance variations occur simultaneously, such target representations are not robust. In this paper, we present a discriminative dictionary learning based target representation. A target candidate is represented via a linear combination of atoms in a learnt dictionary. The online dictionary learning can learn the appearance variations in tracking processing. So, the learnt dictionary can cover all of kinds of appearance variations. Based on this kind of target representation, a novel tracking algorithm is proposed. Extensive experiments on challenging sequences in popular tracking benchmark demonstrate competing tracking performances against some state-of-the-art trackers.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"7 1","pages":"251-255"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85812513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1