首页 > 最新文献

2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)最新文献

英文 中文
An Improved FAsT_Match Algorithm for Micro Parts Detection 微零件检测中一种改进的FAsT_Match算法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177440
Jia-yi Zhang, Y. Liu, Zhi-qiang Liu
Through analysis the characteristics of micro parts such as multi-categories, high detection frequency and similar shape in the classification process, based on the FAsT_Match algorithm, an improved algorithm of template matching recognition is proposed which is a Grid Region Optimized FAsT_Match (GRO FAsT_Match for short). Firstly, the method of gray level adjustment, global threshold image segmentation, boundary tracking and denoising is used to extract the smallest rectangle of the target part image as ROI area. Secondly, by calculating the scale relationship between ROI region and template image, the step sizes and limits of grid parameters for translation and scaling transformation are optimized. In order to improve the discrimination of normalized SAD distance for similar parts, uniform sampling of template image is adopted. The experimental data show that this algorithm features fast, precise, clear distinguish of similar parts, and meets the requirements of micro parts classification and detection. It has practical significance to improve the assembly efficiency of micro parts.
通过分析微零件分类过程中类别多、检测频率高、形状相似等特点,在FAsT_Match算法的基础上,提出了一种改进的模板匹配识别算法——网格区域优化FAsT_Match (GRO FAsT_Match)。首先,采用灰度调整、全局阈值图像分割、边界跟踪和去噪的方法提取目标部分图像的最小矩形作为感兴趣区域;其次,通过计算ROI区域与模板图像之间的尺度关系,优化平移和缩放变换网格参数的步长和限制;为了提高对相似部件归一化SAD距离的判别能力,对模板图像进行均匀采样。实验数据表明,该算法对相似零件的识别速度快、精度高、清晰,满足微细零件分类检测的要求。对提高微细零件的装配效率具有现实意义。
{"title":"An Improved FAsT_Match Algorithm for Micro Parts Detection","authors":"Jia-yi Zhang, Y. Liu, Zhi-qiang Liu","doi":"10.1109/ICIVC50857.2020.9177440","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177440","url":null,"abstract":"Through analysis the characteristics of micro parts such as multi-categories, high detection frequency and similar shape in the classification process, based on the FAsT_Match algorithm, an improved algorithm of template matching recognition is proposed which is a Grid Region Optimized FAsT_Match (GRO FAsT_Match for short). Firstly, the method of gray level adjustment, global threshold image segmentation, boundary tracking and denoising is used to extract the smallest rectangle of the target part image as ROI area. Secondly, by calculating the scale relationship between ROI region and template image, the step sizes and limits of grid parameters for translation and scaling transformation are optimized. In order to improve the discrimination of normalized SAD distance for similar parts, uniform sampling of template image is adopted. The experimental data show that this algorithm features fast, precise, clear distinguish of similar parts, and meets the requirements of micro parts classification and detection. It has practical significance to improve the assembly efficiency of micro parts.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"13 1","pages":"24-28"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90464670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pose Refinement of Occluded 3D Objects Based on Visible Surface Extraction 基于可见表面提取的遮挡三维物体姿态优化
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177481
Xunwei Tong, Ruifeng Li, Lianzheng Ge, Lijun Zhao, Ke Wang
In this paper, we propose a pose refinement method based on the visible surface extraction of 3D object. Given a rough estimation of object pose, the algorithm of iterative closet point (ICP) is often used to refine the pose by aligning the object model with test scene. To avoid the interference of invisible points on the ICP process, we only use the visible surface for pose refinement. It is especially necessary when occlusion occurs in the scene. Combining the technologies of image rendering and depth consistency verification, the visible surface can be effectively extracted. During the process of pose refinement, hypothesis verification methods are also used to eliminate unreasonable hypothetical poses as early as possible. The proposed method is evaluated on the public Tejani dataset. The experimental results show that our method improved the average F1-score by 0.2062, which proves that our method can obtain pose estimation results of high accuracy, even in the occluded scene.
本文提出了一种基于三维物体可见表面提取的姿态精细方法。在对目标姿态进行粗略估计的情况下,通常采用迭代封闭点(ICP)算法,通过将目标模型与测试场景对齐来改进姿态。为了避免不可见点对ICP过程的干扰,我们只使用可见表面进行姿态细化。当场景中出现遮挡时,这是特别必要的。结合图像绘制技术和深度一致性验证技术,可以有效提取可见表面。在姿态优化过程中,还采用假设验证方法,尽早消除不合理的假设姿态。所提出的方法在公共Tejani数据集上进行了评估。实验结果表明,我们的方法将平均f1分数提高了0.2062,证明了我们的方法即使在被遮挡的场景中也能获得高精度的姿态估计结果。
{"title":"Pose Refinement of Occluded 3D Objects Based on Visible Surface Extraction","authors":"Xunwei Tong, Ruifeng Li, Lianzheng Ge, Lijun Zhao, Ke Wang","doi":"10.1109/ICIVC50857.2020.9177481","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177481","url":null,"abstract":"In this paper, we propose a pose refinement method based on the visible surface extraction of 3D object. Given a rough estimation of object pose, the algorithm of iterative closet point (ICP) is often used to refine the pose by aligning the object model with test scene. To avoid the interference of invisible points on the ICP process, we only use the visible surface for pose refinement. It is especially necessary when occlusion occurs in the scene. Combining the technologies of image rendering and depth consistency verification, the visible surface can be effectively extracted. During the process of pose refinement, hypothesis verification methods are also used to eliminate unreasonable hypothetical poses as early as possible. The proposed method is evaluated on the public Tejani dataset. The experimental results show that our method improved the average F1-score by 0.2062, which proves that our method can obtain pose estimation results of high accuracy, even in the occluded scene.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"34 1","pages":"176-181"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76164413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
3D U-Net Brain Tumor Segmentation Using VAE Skip Connection 使用VAE跳跃连接的三维U-Net脑肿瘤分割
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177441
Ke Li, L. Kong, Yifeng Zhang
In clinical practice, the determination of the location, shape, and size of brain tumor can greatly assist the diagnosis, monitoring, and treatment of brain tumor. Therefore, accurate and reliable automatic brain tumor segmentation algorithm is of great significance for clinical diagnosis and treatment. With the rapid development of deep learning technology, more and more efficient image segmentation algorithms have also been applied in this field. It has been proven that U-Net model combined with variational auto-encoder can help to effectively regularize the shared encoder and thereby improve the performance. Based on the VAE-U-Net model, this paper proposes a structure called VAE skip connection. By fusing the position information contained in VAE branch into U-Net decoding stage, the network can retain more high-resolution detail information. In addition, we integrate ShakeDrop regularization into the networks to further alleviate the overfitting problem. The experimental results show that the networks after adding VAE skip connection and ShakeDrop can achieve competitive results on the BraTS 2018 dataset.
在临床实践中,确定脑肿瘤的位置、形状和大小对脑肿瘤的诊断、监测和治疗有很大的帮助。因此,准确可靠的脑肿瘤自动分割算法对临床诊断和治疗具有重要意义。随着深度学习技术的快速发展,越来越多高效的图像分割算法也被应用于该领域。研究表明,U-Net模型结合变分自编码器可以有效地对共享编码器进行正则化,从而提高编码器的性能。基于VAE- u - net模型,提出了一种VAE跳接结构。通过将VAE支路中包含的位置信息融合到U-Net解码阶段,可以保留更多高分辨率的细节信息。此外,我们将ShakeDrop正则化集成到网络中,以进一步缓解过拟合问题。实验结果表明,加入VAE跳跃连接和ShakeDrop后的网络在BraTS 2018数据集上可以取得有竞争力的结果。
{"title":"3D U-Net Brain Tumor Segmentation Using VAE Skip Connection","authors":"Ke Li, L. Kong, Yifeng Zhang","doi":"10.1109/ICIVC50857.2020.9177441","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177441","url":null,"abstract":"In clinical practice, the determination of the location, shape, and size of brain tumor can greatly assist the diagnosis, monitoring, and treatment of brain tumor. Therefore, accurate and reliable automatic brain tumor segmentation algorithm is of great significance for clinical diagnosis and treatment. With the rapid development of deep learning technology, more and more efficient image segmentation algorithms have also been applied in this field. It has been proven that U-Net model combined with variational auto-encoder can help to effectively regularize the shared encoder and thereby improve the performance. Based on the VAE-U-Net model, this paper proposes a structure called VAE skip connection. By fusing the position information contained in VAE branch into U-Net decoding stage, the network can retain more high-resolution detail information. In addition, we integrate ShakeDrop regularization into the networks to further alleviate the overfitting problem. The experimental results show that the networks after adding VAE skip connection and ShakeDrop can achieve competitive results on the BraTS 2018 dataset.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"3 1","pages":"97-101"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80228093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Design and Development of Spinning Bike Game Based on VR Technology 基于VR技术的动感单车游戏设计与开发
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177487
Xiong Chuan, Zhu Dandan, Peng Ziming
Virtual reality technology can provide a concurrent and multi-sensory interaction between humans and computers, enhancing the immersion, realism and fun in the process of human-computer interaction. The introduction of virtual reality technology into the design and development of fitness games can change the deficiencies of traditional fitness games such as low dimensionality, boringness, and low substitution. It makes the fitness process interesting and efficient. This article takes a VR spinning bike game designed and developed based on the Unity3D platform as an example to explore the application of VR technology in the fitness game industry.
虚拟现实技术可以提供人与计算机并行的多感官交互,增强人机交互过程中的沉浸感、真实感和趣味性。将虚拟现实技术引入健身游戏的设计与开发中,可以改变传统健身游戏低维度、无聊、低替代等不足。它使健身过程变得有趣和高效。本文以基于Unity3D平台设计开发的VR动感单车游戏为例,探索VR技术在健身游戏行业中的应用。
{"title":"Design and Development of Spinning Bike Game Based on VR Technology","authors":"Xiong Chuan, Zhu Dandan, Peng Ziming","doi":"10.1109/ICIVC50857.2020.9177487","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177487","url":null,"abstract":"Virtual reality technology can provide a concurrent and multi-sensory interaction between humans and computers, enhancing the immersion, realism and fun in the process of human-computer interaction. The introduction of virtual reality technology into the design and development of fitness games can change the deficiencies of traditional fitness games such as low dimensionality, boringness, and low substitution. It makes the fitness process interesting and efficient. This article takes a VR spinning bike game designed and developed based on the Unity3D platform as an example to explore the application of VR technology in the fitness game industry.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"43 1","pages":"227-231"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77409415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Biologically Inspired Channel Allocation Method for Image Acquisition in Cognitive Radio Sensor Networks 认知无线电传感器网络中基于生物学的图像采集信道分配方法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177490
Mengying Xu, Jie Zhou, Rui Yang
In recent years, cognitive radio sensor networks (CRSNs) have been commonly applied in environmental monitoring and image acquisition. However, recent advances in channel allocation have led to lower network reward, lifetime, and energy utilization rate. As a basic and fundamental problem to obtain image data in CRSNs, it governs the performance of CRSNs. To further improve the reward and throughput of obtaining image, this paper proposes an improved immune hybrid bat algorithm (IIHBA) based on bat algorithm. Furthermore, we develop a simulation environment and compared the performance of IIHBA with particle swarm optimization (PSO) and genetic algorithm (GA). Last but not the least, computational experiments showed that the reward is improved 11.36%, 27.20% respectively compared with GA and PSO when the number of users is 20 and the number of channels is 5. Based on the above findings, the proposed scheme can improve the reward of system, especially in terms of higher-throughput.
近年来,认知无线电传感器网络(CRSNs)在环境监测和图像采集中得到了广泛的应用。然而,最近在信道分配方面的进展导致了较低的网络回报、寿命和能源利用率。作为crsn中获取图像数据的基本问题,它决定着crsn的性能。为了进一步提高图像获取的奖励和吞吐量,本文提出了一种基于蝙蝠算法的改进免疫混合蝙蝠算法(IIHBA)。此外,我们开发了一个仿真环境,并将IIHBA的性能与粒子群优化(PSO)和遗传算法(GA)进行了比较。最后,计算实验表明,当用户数为20、通道数为5时,与遗传算法和粒子群算法相比,奖励分别提高了11.36%、27.20%。基于以上发现,提出的方案可以提高系统的奖励,特别是在更高的吞吐量方面。
{"title":"A Biologically Inspired Channel Allocation Method for Image Acquisition in Cognitive Radio Sensor Networks","authors":"Mengying Xu, Jie Zhou, Rui Yang","doi":"10.1109/ICIVC50857.2020.9177490","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177490","url":null,"abstract":"In recent years, cognitive radio sensor networks (CRSNs) have been commonly applied in environmental monitoring and image acquisition. However, recent advances in channel allocation have led to lower network reward, lifetime, and energy utilization rate. As a basic and fundamental problem to obtain image data in CRSNs, it governs the performance of CRSNs. To further improve the reward and throughput of obtaining image, this paper proposes an improved immune hybrid bat algorithm (IIHBA) based on bat algorithm. Furthermore, we develop a simulation environment and compared the performance of IIHBA with particle swarm optimization (PSO) and genetic algorithm (GA). Last but not the least, computational experiments showed that the reward is improved 11.36%, 27.20% respectively compared with GA and PSO when the number of users is 20 and the number of channels is 5. Based on the above findings, the proposed scheme can improve the reward of system, especially in terms of higher-throughput.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"12 1","pages":"267-271"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81950067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anisotropic Mesh Representation for Color Images 彩色图像的各向异性网格表示
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177477
Xianping Li
Triangular meshes have become popular in image representation and have been widely used in image processing to improve computational efficiency and accuracy. This paper introduces anisotropic mesh adaptation (AMA) to represent color images with fewer points while keeping good quality. Finite element interpolation is used in the image reconstruction from the triangular mesh-based representation. A few methods are proposed to deal with the different color channels for representation purpose. Experimental results for various images show that single triangular mesh can represent the color image as good as three-mesh representation and the differences are not significant.
三角形网格在图像表示中越来越流行,并被广泛应用于图像处理,以提高计算效率和精度。本文引入各向异性网格自适应(AMA)方法,在保证图像质量的前提下,用较少的点表示彩色图像。在基于三角网格表示的图像重建中,采用有限元插值。提出了几种方法来处理不同的颜色通道以达到表示的目的。对各种图像的实验结果表明,单三角网格对彩色图像的表示效果与三网格表示效果相当,且差异不显著。
{"title":"Anisotropic Mesh Representation for Color Images","authors":"Xianping Li","doi":"10.1109/ICIVC50857.2020.9177477","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177477","url":null,"abstract":"Triangular meshes have become popular in image representation and have been widely used in image processing to improve computational efficiency and accuracy. This paper introduces anisotropic mesh adaptation (AMA) to represent color images with fewer points while keeping good quality. Finite element interpolation is used in the image reconstruction from the triangular mesh-based representation. A few methods are proposed to deal with the different color channels for representation purpose. Experimental results for various images show that single triangular mesh can represent the color image as good as three-mesh representation and the differences are not significant.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"43 1","pages":"139-143"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79368535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Review of Remote Sensing Image Object Detection Algorithms Based on Deep Learning 基于深度学习的遥感图像目标检测算法综述
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177453
Zhe Zheng, Lin Lei, Hao Sun, Gangyao Kuang
Object detection is an important part of remote sensing image analysis. With the development of the earth observation technology and convolutional neural network, remote sensing image object detection technology based on deep learning has received more and more attention and research. At present, many excellent object detection algorithms have been proposed and applied in the field of remote sensing. In this paper, the object detection algorithms of remote sensing image is systematically summarized, the main contents include the traditional remote sensing image object detection method and the method based on deep learning, emphasis on summarize the remote sensing image object detection algorithm based on deep learning and its development course, then we introduced the rule of performance evaluation of object detection and datasets that commonly used. Finally, the future development trend is analyzed and prospected. It is hoped that this summary and analysis can provide some reference for future research on object detection technology in remote sensing field.
目标检测是遥感图像分析的重要组成部分。随着对地观测技术和卷积神经网络的发展,基于深度学习的遥感图像目标检测技术得到了越来越多的关注和研究。目前,已有许多优秀的目标检测算法被提出并应用于遥感领域。本文对遥感图像的目标检测算法进行了系统的总结,主要内容包括传统的遥感图像目标检测方法和基于深度学习的遥感图像目标检测方法,重点总结了基于深度学习的遥感图像目标检测算法及其发展历程,然后介绍了目标检测性能评价规则和常用的数据集。最后,对未来的发展趋势进行了分析和展望。希望本文的总结和分析能够为今后遥感领域目标检测技术的研究提供一些参考。
{"title":"A Review of Remote Sensing Image Object Detection Algorithms Based on Deep Learning","authors":"Zhe Zheng, Lin Lei, Hao Sun, Gangyao Kuang","doi":"10.1109/ICIVC50857.2020.9177453","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177453","url":null,"abstract":"Object detection is an important part of remote sensing image analysis. With the development of the earth observation technology and convolutional neural network, remote sensing image object detection technology based on deep learning has received more and more attention and research. At present, many excellent object detection algorithms have been proposed and applied in the field of remote sensing. In this paper, the object detection algorithms of remote sensing image is systematically summarized, the main contents include the traditional remote sensing image object detection method and the method based on deep learning, emphasis on summarize the remote sensing image object detection algorithm based on deep learning and its development course, then we introduced the rule of performance evaluation of object detection and datasets that commonly used. Finally, the future development trend is analyzed and prospected. It is hoped that this summary and analysis can provide some reference for future research on object detection technology in remote sensing field.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"4 1","pages":"34-43"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83083681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Feature Point Extraction and Matching Method Based on Akaze in Illumination Invariant Color Space 光照不变颜色空间中基于Akaze的特征点提取与匹配方法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177459
Yongyuan Xue, Tianhang Gao
Visual SLAM is the technology that complete self-localization and build environment map synchronously. The feature point extraction and matching of the input image is very important for visual SLAM to achieve pose calculation and map building. For most of the literature feature point extraction and matching algorithms, the change of illumination may have a great impact on the final matching results. To address the issue, this paper proposes a novel feature point extraction and matching method based on Akaze algorithm (IICS-Akaze). Histogram equalization and dark channel prior theory are combined to construct a color space with constant illumination. Akaze algorithm is adopted for fast multi-scale feature extraction to generate feature point descriptors. The feature points are then quickly matched through the FLANN, and RANSC is introduced to eliminate the mismatches. In addition, the experiments on open data set are conducted in terms of feature extraction quantity, matching accuracy, and illumination robustness among the related methods. The experimental results show that the proposed method is able to accurately extract and match image feature points when the illumination changes dramatically.
可视化SLAM是一种同步完成自定位和构建环境地图的技术。输入图像的特征点提取与匹配是视觉SLAM实现姿态计算和地图构建的关键。对于大多数文献中的特征点提取和匹配算法,光照的变化可能会对最终的匹配结果产生很大的影响。针对这一问题,本文提出了一种新的基于Akaze算法的特征点提取与匹配方法(IICS-Akaze)。将直方图均衡化和暗通道先验理论相结合,构造出具有恒定照度的色彩空间。采用Akaze算法进行快速多尺度特征提取,生成特征点描述子。然后通过FLANN快速匹配特征点,并引入RANSC来消除不匹配。此外,在开放数据集上对相关方法进行了特征提取量、匹配精度、光照鲁棒性等方面的实验。实验结果表明,该方法能够在光照剧烈变化的情况下准确提取和匹配图像特征点。
{"title":"Feature Point Extraction and Matching Method Based on Akaze in Illumination Invariant Color Space","authors":"Yongyuan Xue, Tianhang Gao","doi":"10.1109/ICIVC50857.2020.9177459","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177459","url":null,"abstract":"Visual SLAM is the technology that complete self-localization and build environment map synchronously. The feature point extraction and matching of the input image is very important for visual SLAM to achieve pose calculation and map building. For most of the literature feature point extraction and matching algorithms, the change of illumination may have a great impact on the final matching results. To address the issue, this paper proposes a novel feature point extraction and matching method based on Akaze algorithm (IICS-Akaze). Histogram equalization and dark channel prior theory are combined to construct a color space with constant illumination. Akaze algorithm is adopted for fast multi-scale feature extraction to generate feature point descriptors. The feature points are then quickly matched through the FLANN, and RANSC is introduced to eliminate the mismatches. In addition, the experiments on open data set are conducted in terms of feature extraction quantity, matching accuracy, and illumination robustness among the related methods. The experimental results show that the proposed method is able to accurately extract and match image feature points when the illumination changes dramatically.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"22 1","pages":"160-165"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75046774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Td-VOS: Tracking-Driven Single-Object Video Object Segmentation Td-VOS:跟踪驱动的单目标视频对象分割
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177471
Shaopan Xiong, Shengyang Li, Longxuan Kou, Weilong Guo, Zhuang Zhou, Zifei Zhao
This paper presents an approach to single-object video object segmentation, only using the first-frame bounding box (without mask) to initialize. The proposed method is a tracking-driven single-object video object segmentation, which combines an effective Box2Segmentation module with a general object tracking module. Just initialize the first frame box, the Box2Segmentation module can obtain the segmentation results based on the predicted tracking bounding box. Evaluations on the single-object video object segmentation dataset DAVIS2016 show that the proposed method achieves a competitive performance with a Region Similarity score of 75.4% and a Contour Accuracy score of 73.1%, only under the settings of first-frame bounding box initialization. The proposed method outperforms SiamMask which is the most competitive method for video object segmentation under the same settings, with Region Similarity score by 5.2% and Contour Accuracy score by 7.8%. Compared with the semi-supervised VOS methods without online fine-tuning initialized by a first frame mask, the proposed method also achieves comparable results.
本文提出了一种单目标视频对象分割方法,仅使用第一帧边界框(无掩码)进行初始化。该方法是一种跟踪驱动的单目标视频目标分割方法,它将有效的box2分割模块与通用的目标跟踪模块相结合。只需初始化第一帧框,Box2Segmentation模块就可以根据预测的跟踪边界框得到分割结果。对单目标视频目标分割数据集DAVIS2016的评估表明,仅在第一帧边界框初始化设置下,该方法的区域相似度得分为75.4%,轮廓精度得分为73.1%,具有较强的竞争力。在相同设置下,该方法优于最具竞争力的视频目标分割方法SiamMask,区域相似度得分提高5.2%,轮廓精度得分提高7.8%。与未使用第一帧掩码初始化在线微调的半监督VOS方法相比,该方法也取得了相当的效果。
{"title":"Td-VOS: Tracking-Driven Single-Object Video Object Segmentation","authors":"Shaopan Xiong, Shengyang Li, Longxuan Kou, Weilong Guo, Zhuang Zhou, Zifei Zhao","doi":"10.1109/ICIVC50857.2020.9177471","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177471","url":null,"abstract":"This paper presents an approach to single-object video object segmentation, only using the first-frame bounding box (without mask) to initialize. The proposed method is a tracking-driven single-object video object segmentation, which combines an effective Box2Segmentation module with a general object tracking module. Just initialize the first frame box, the Box2Segmentation module can obtain the segmentation results based on the predicted tracking bounding box. Evaluations on the single-object video object segmentation dataset DAVIS2016 show that the proposed method achieves a competitive performance with a Region Similarity score of 75.4% and a Contour Accuracy score of 73.1%, only under the settings of first-frame bounding box initialization. The proposed method outperforms SiamMask which is the most competitive method for video object segmentation under the same settings, with Region Similarity score by 5.2% and Contour Accuracy score by 7.8%. Compared with the semi-supervised VOS methods without online fine-tuning initialized by a first frame mask, the proposed method also achieves comparable results.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"51 1","pages":"102-107"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79154877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Point-Based Registration for Multi-stained Histology Images 基于点的多染色组织学图像配准
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177486
Jiehua Zhang, Zhang Li, Qifeng Yu
Image registration is a basic task in biological image processing. Different stained histology images contain different clinical information, which could assist pathologists to diagnose a certain disease. It is necessary to improve the accuracy of image registration. In this paper, we present a robust registration method that consists of three steps: 1) extracting match points; 2) a pre-alignment consisting of a rigid transformation and an affine transformation on the coarse level; 3) an accurate non-rigid registration optimized by the extracted points. The existing methods use the features of the image pair to initial alignment. We proposed a new metric for the non-rigid transformation which adding the part of optimizing extracting points into the original metric. We evaluate our method on the dataset from the ANHIR Registration Challenge and use MrTRE (median relative target registration error) to measure the performance on the training data. The test result illustrates that the presented method is accurate and robust.
图像配准是生物图像处理中的一项基本任务。不同的染色组织学图像包含不同的临床信息,这些信息可以帮助病理学家诊断某种疾病。提高图像配准的精度是很有必要的。本文提出了一种鲁棒配准方法,该方法分为三个步骤:1)提取匹配点;2)由粗层上的刚性变换和仿射变换组成的预对准;3)提取点优化的精确非刚性配准。现有的方法是利用图像对的特征进行初始对齐。提出了一种新的非刚性变换度量,将优化提取点的部分加入到原度量中。我们在来自ANHIR注册挑战的数据集上评估我们的方法,并使用MrTRE(中位数相对目标注册误差)来衡量训练数据上的性能。实验结果表明,该方法具有较好的鲁棒性和准确性。
{"title":"Point-Based Registration for Multi-stained Histology Images","authors":"Jiehua Zhang, Zhang Li, Qifeng Yu","doi":"10.1109/ICIVC50857.2020.9177486","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177486","url":null,"abstract":"Image registration is a basic task in biological image processing. Different stained histology images contain different clinical information, which could assist pathologists to diagnose a certain disease. It is necessary to improve the accuracy of image registration. In this paper, we present a robust registration method that consists of three steps: 1) extracting match points; 2) a pre-alignment consisting of a rigid transformation and an affine transformation on the coarse level; 3) an accurate non-rigid registration optimized by the extracted points. The existing methods use the features of the image pair to initial alignment. We proposed a new metric for the non-rigid transformation which adding the part of optimizing extracting points into the original metric. We evaluate our method on the dataset from the ANHIR Registration Challenge and use MrTRE (median relative target registration error) to measure the performance on the training data. The test result illustrates that the presented method is accurate and robust.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"41 1","pages":"92-96"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81318867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1