首页 > 最新文献

2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)最新文献

英文 中文
Point-Based Registration for Multi-stained Histology Images 基于点的多染色组织学图像配准
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177486
Jiehua Zhang, Zhang Li, Qifeng Yu
Image registration is a basic task in biological image processing. Different stained histology images contain different clinical information, which could assist pathologists to diagnose a certain disease. It is necessary to improve the accuracy of image registration. In this paper, we present a robust registration method that consists of three steps: 1) extracting match points; 2) a pre-alignment consisting of a rigid transformation and an affine transformation on the coarse level; 3) an accurate non-rigid registration optimized by the extracted points. The existing methods use the features of the image pair to initial alignment. We proposed a new metric for the non-rigid transformation which adding the part of optimizing extracting points into the original metric. We evaluate our method on the dataset from the ANHIR Registration Challenge and use MrTRE (median relative target registration error) to measure the performance on the training data. The test result illustrates that the presented method is accurate and robust.
图像配准是生物图像处理中的一项基本任务。不同的染色组织学图像包含不同的临床信息,这些信息可以帮助病理学家诊断某种疾病。提高图像配准的精度是很有必要的。本文提出了一种鲁棒配准方法,该方法分为三个步骤:1)提取匹配点;2)由粗层上的刚性变换和仿射变换组成的预对准;3)提取点优化的精确非刚性配准。现有的方法是利用图像对的特征进行初始对齐。提出了一种新的非刚性变换度量,将优化提取点的部分加入到原度量中。我们在来自ANHIR注册挑战的数据集上评估我们的方法,并使用MrTRE(中位数相对目标注册误差)来衡量训练数据上的性能。实验结果表明,该方法具有较好的鲁棒性和准确性。
{"title":"Point-Based Registration for Multi-stained Histology Images","authors":"Jiehua Zhang, Zhang Li, Qifeng Yu","doi":"10.1109/ICIVC50857.2020.9177486","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177486","url":null,"abstract":"Image registration is a basic task in biological image processing. Different stained histology images contain different clinical information, which could assist pathologists to diagnose a certain disease. It is necessary to improve the accuracy of image registration. In this paper, we present a robust registration method that consists of three steps: 1) extracting match points; 2) a pre-alignment consisting of a rigid transformation and an affine transformation on the coarse level; 3) an accurate non-rigid registration optimized by the extracted points. The existing methods use the features of the image pair to initial alignment. We proposed a new metric for the non-rigid transformation which adding the part of optimizing extracting points into the original metric. We evaluate our method on the dataset from the ANHIR Registration Challenge and use MrTRE (median relative target registration error) to measure the performance on the training data. The test result illustrates that the presented method is accurate and robust.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"41 1","pages":"92-96"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81318867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Td-VOS: Tracking-Driven Single-Object Video Object Segmentation Td-VOS:跟踪驱动的单目标视频对象分割
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177471
Shaopan Xiong, Shengyang Li, Longxuan Kou, Weilong Guo, Zhuang Zhou, Zifei Zhao
This paper presents an approach to single-object video object segmentation, only using the first-frame bounding box (without mask) to initialize. The proposed method is a tracking-driven single-object video object segmentation, which combines an effective Box2Segmentation module with a general object tracking module. Just initialize the first frame box, the Box2Segmentation module can obtain the segmentation results based on the predicted tracking bounding box. Evaluations on the single-object video object segmentation dataset DAVIS2016 show that the proposed method achieves a competitive performance with a Region Similarity score of 75.4% and a Contour Accuracy score of 73.1%, only under the settings of first-frame bounding box initialization. The proposed method outperforms SiamMask which is the most competitive method for video object segmentation under the same settings, with Region Similarity score by 5.2% and Contour Accuracy score by 7.8%. Compared with the semi-supervised VOS methods without online fine-tuning initialized by a first frame mask, the proposed method also achieves comparable results.
本文提出了一种单目标视频对象分割方法,仅使用第一帧边界框(无掩码)进行初始化。该方法是一种跟踪驱动的单目标视频目标分割方法,它将有效的box2分割模块与通用的目标跟踪模块相结合。只需初始化第一帧框,Box2Segmentation模块就可以根据预测的跟踪边界框得到分割结果。对单目标视频目标分割数据集DAVIS2016的评估表明,仅在第一帧边界框初始化设置下,该方法的区域相似度得分为75.4%,轮廓精度得分为73.1%,具有较强的竞争力。在相同设置下,该方法优于最具竞争力的视频目标分割方法SiamMask,区域相似度得分提高5.2%,轮廓精度得分提高7.8%。与未使用第一帧掩码初始化在线微调的半监督VOS方法相比,该方法也取得了相当的效果。
{"title":"Td-VOS: Tracking-Driven Single-Object Video Object Segmentation","authors":"Shaopan Xiong, Shengyang Li, Longxuan Kou, Weilong Guo, Zhuang Zhou, Zifei Zhao","doi":"10.1109/ICIVC50857.2020.9177471","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177471","url":null,"abstract":"This paper presents an approach to single-object video object segmentation, only using the first-frame bounding box (without mask) to initialize. The proposed method is a tracking-driven single-object video object segmentation, which combines an effective Box2Segmentation module with a general object tracking module. Just initialize the first frame box, the Box2Segmentation module can obtain the segmentation results based on the predicted tracking bounding box. Evaluations on the single-object video object segmentation dataset DAVIS2016 show that the proposed method achieves a competitive performance with a Region Similarity score of 75.4% and a Contour Accuracy score of 73.1%, only under the settings of first-frame bounding box initialization. The proposed method outperforms SiamMask which is the most competitive method for video object segmentation under the same settings, with Region Similarity score by 5.2% and Contour Accuracy score by 7.8%. Compared with the semi-supervised VOS methods without online fine-tuning initialized by a first frame mask, the proposed method also achieves comparable results.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"51 1","pages":"102-107"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79154877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Color Image Filtering in Bessel-Fourier Moments Domain 贝塞尔-傅里叶矩域彩色图像滤波
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177478
Tianpeng Xia, S. Liao
In this research, we have conducted a study on color image filtering in Bessel-Fourier moments domain. Bessel-Fourier moments of the two testing color images are computed independently from the three color channels (RGB), then lowpass and highpass filters are applied to the data in Bessel-Fourier moments domain for our investigation. For comparison, filters are applied in Fourier Frequency domain as well. The experimental results suggest that Bessel-Fourier moments of the lower orders contain mainly information of smooth varying components of images, while those of the higher orders are more related to details such as sharp transitions in intensity. It is also found that the Gaussian filters would reduce the ringing effect in Bessel-Fourier moments domain as they do in the Fourier Frequency domain.
在本研究中,我们对彩色图像的贝塞尔-傅里叶矩域滤波进行了研究。在三个颜色通道(RGB)中独立计算两幅测试彩色图像的贝塞尔-傅里叶矩,然后在贝塞尔-傅里叶矩域对数据应用低通和高通滤波器进行研究。为了比较,在傅里叶频域也应用了滤波器。实验结果表明,低阶贝塞尔-傅里叶矩主要包含图像平滑变化分量的信息,而高阶贝塞尔-傅里叶矩则更多地与图像强度的急剧变化等细节有关。研究还发现,高斯滤波器在贝塞尔-傅立叶矩域中和在傅立叶频域中一样,都能减小振铃效应。
{"title":"Color Image Filtering in Bessel-Fourier Moments Domain","authors":"Tianpeng Xia, S. Liao","doi":"10.1109/ICIVC50857.2020.9177478","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177478","url":null,"abstract":"In this research, we have conducted a study on color image filtering in Bessel-Fourier moments domain. Bessel-Fourier moments of the two testing color images are computed independently from the three color channels (RGB), then lowpass and highpass filters are applied to the data in Bessel-Fourier moments domain for our investigation. For comparison, filters are applied in Fourier Frequency domain as well. The experimental results suggest that Bessel-Fourier moments of the lower orders contain mainly information of smooth varying components of images, while those of the higher orders are more related to details such as sharp transitions in intensity. It is also found that the Gaussian filters would reduce the ringing effect in Bessel-Fourier moments domain as they do in the Fourier Frequency domain.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"112 1","pages":"75-81"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88776646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High Temperature Deformation Field Measurement Using 3D Digital Image Correlation Method 三维数字图像相关法测量高温变形场
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177479
Hongtao Deng, D. Jiang, Kai Wang, Q. Fei
A full-field, three-dimensional and non-contact deformation field measurement method under high temperature environment based on 3D digital image correlation (3D-DIC) is introduced. In order to reduce the impact of high temperature radiation on the image quality, a band-pass filter is placed in front of the camera lens. The two cameras simultaneously take pictures of the object before and after deformation, and use 3D-DIC to measure the three-dimensional deformation field of the object surface. The high temperature deformation field measurement test shows that 3D-DIC can accurately and conveniently measure the deformation field of an object under high temperature environment.
介绍了一种基于三维数字图像相关(3D- dic)的高温环境下全场、三维、非接触变形场测量方法。为了减少高温辐射对图像质量的影响,在相机镜头前放置了带通滤波器。两台相机同时拍摄物体变形前后的照片,并使用3D-DIC测量物体表面的三维变形场。高温变形场测量试验表明,3D-DIC能够准确、方便地测量物体在高温环境下的变形场。
{"title":"High Temperature Deformation Field Measurement Using 3D Digital Image Correlation Method","authors":"Hongtao Deng, D. Jiang, Kai Wang, Q. Fei","doi":"10.1109/ICIVC50857.2020.9177479","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177479","url":null,"abstract":"A full-field, three-dimensional and non-contact deformation field measurement method under high temperature environment based on 3D digital image correlation (3D-DIC) is introduced. In order to reduce the impact of high temperature radiation on the image quality, a band-pass filter is placed in front of the camera lens. The two cameras simultaneously take pictures of the object before and after deformation, and use 3D-DIC to measure the three-dimensional deformation field of the object surface. The high temperature deformation field measurement test shows that 3D-DIC can accurately and conveniently measure the deformation field of an object under high temperature environment.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"96 1","pages":"188-192"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88969096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Clone Chaotic Niche Evolutionary Algorithm for Duty Cycle Control Optimization in Wireless Multimedia Sensor Networks 无线多媒体传感器网络占空比控制优化的克隆混沌小生境进化算法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177435
Jie Zhou, Mengying Xu, Rui Yang
One of the most interesting issue regarding to wireless multimedia sensor networks (WMSNs) is to maximizing the network lifetime. Because sensor nodes are constrained in energy, it is very important and necessary to exploit novel duty cycle design algorithms. Such a problem is important in improving network lifetime in WMSNs. The new contribution of our paper is that we propose a clone chaotic niche evolutionary algorithm (CCNEA) for duty cycle design problem in WMSNs. Novel clone operator and chaotic operator have been designed to develop solutions randomly. The strategy merges the merits of clone selection, chaotic generation, and niche operator. CCNEA is a style of swarm algorithm, which has strong global exploit ability. CCNEA utilizes chaotic generation approach which targets to avoid local optima. Then, simulations are performed to verify the robust and efficacy performance of CCNEA compared to methods according to particle swarm optimization (PSO) and quantum genetic algorithm (QGA) under an WMSNs conditions. Simulation experiments denote that the presented CCNEA outperforms PSO and QGA under different conditions, especially for WMSNs that has large number of sensors.
无线多媒体传感器网络(wmsn)中最令人感兴趣的问题之一是如何使网络的生存时间最大化。由于传感器节点受到能量的限制,开发新的占空比设计算法是非常重要和必要的。该问题对于提高wmsn的网络生存时间具有重要意义。本文的新贡献是我们提出了一种克隆混沌生态位进化算法(CCNEA)来解决wmsn的占空比设计问题。设计了新颖的克隆算子和混沌算子来随机求解。该策略融合了克隆选择、混沌生成和小生境算子的优点。CCNEA是一种群算法,具有较强的全局攻击能力。CCNEA采用混沌生成方法,以避免局部最优为目标。在WMSNs条件下,对比粒子群优化(PSO)和量子遗传算法(QGA),仿真验证了CCNEA算法的鲁棒性和有效性。仿真实验表明,本文提出的CCNEA算法在不同条件下都优于粒子群算法和QGA算法,特别是对于具有大量传感器的wmsn。
{"title":"Clone Chaotic Niche Evolutionary Algorithm for Duty Cycle Control Optimization in Wireless Multimedia Sensor Networks","authors":"Jie Zhou, Mengying Xu, Rui Yang","doi":"10.1109/ICIVC50857.2020.9177435","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177435","url":null,"abstract":"One of the most interesting issue regarding to wireless multimedia sensor networks (WMSNs) is to maximizing the network lifetime. Because sensor nodes are constrained in energy, it is very important and necessary to exploit novel duty cycle design algorithms. Such a problem is important in improving network lifetime in WMSNs. The new contribution of our paper is that we propose a clone chaotic niche evolutionary algorithm (CCNEA) for duty cycle design problem in WMSNs. Novel clone operator and chaotic operator have been designed to develop solutions randomly. The strategy merges the merits of clone selection, chaotic generation, and niche operator. CCNEA is a style of swarm algorithm, which has strong global exploit ability. CCNEA utilizes chaotic generation approach which targets to avoid local optima. Then, simulations are performed to verify the robust and efficacy performance of CCNEA compared to methods according to particle swarm optimization (PSO) and quantum genetic algorithm (QGA) under an WMSNs conditions. Simulation experiments denote that the presented CCNEA outperforms PSO and QGA under different conditions, especially for WMSNs that has large number of sensors.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"15 1","pages":"278-282"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90767734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Facial Expression Recognition Based on Graph Neural Network 基于图神经网络的面部表情识别
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177430
Xuchou Xu, Zhou Ruan, Lei Yang
Facial expressions are one of the most powerful, natural and immediate means for human being to present their emotions and intensions. In this paper, we present a novel method for fully automatic facial expression recognition. The facial landmarks are detected for characterizing facial expressions. A graph convolutional neural network is proposed for feature extraction and facial expression recognition classification. The experiments were performed on the three facial expression databases. The result shows that the proposed FER method can achieve good recognition accuracy up to 95.85% using the proposed method.
面部表情是人类表达情感和意图的最有力、最自然、最直接的手段之一。本文提出了一种全自动面部表情识别的新方法。通过检测面部标志来表征面部表情。提出了一种用于特征提取和面部表情识别分类的图卷积神经网络。实验在三个面部表情数据库上进行。结果表明,该方法的识别准确率可达95.85%。
{"title":"Facial Expression Recognition Based on Graph Neural Network","authors":"Xuchou Xu, Zhou Ruan, Lei Yang","doi":"10.1109/ICIVC50857.2020.9177430","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177430","url":null,"abstract":"Facial expressions are one of the most powerful, natural and immediate means for human being to present their emotions and intensions. In this paper, we present a novel method for fully automatic facial expression recognition. The facial landmarks are detected for characterizing facial expressions. A graph convolutional neural network is proposed for feature extraction and facial expression recognition classification. The experiments were performed on the three facial expression databases. The result shows that the proposed FER method can achieve good recognition accuracy up to 95.85% using the proposed method.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"80 1","pages":"211-214"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88967613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Dual Stream Segmentation Network for Real-Time Semantic Segmentation 实时语义分割的双流分割网络
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177439
Changyuan Zhong, Zelin Hu, Miao Li, Hualong Li, Xuanjiang Yang, Fei Liu
Modern real-time segmentation methods employ two-branch framework to achieve good speed and accuracy trade-off. However, we observe that low-level features coming from the shallow layers go through less processing, producing a potential semantic gap between different levels of features. Meanwhile, a rigid fusion is less effective due to the absence of consideration for two-branch framework characteristics. In this paper, we propose two novel modules: Unified Interplay Module and Separate Pyramid Pooling Module to address those two issues respectively. Based on our proposed modules, we present a novel Dual Stream Segmentation Network (DSSNet), a two-branch framework for real-time semantic segmentation. Compared with BiSeNet, our DSSNet based on ResNet18 achieves better performance 76.45% mIoU on the Cityscapes test dataset while sharing similar computation costs with BiSeNet. Furthermore, our DSSNet with ResNet34 backbone outperforms previous real-time models, achieving 78.5% mIoU on the Cityscapes test dataset with speed of 39 FPS on GTX1080Ti.
现代实时分割方法采用双分支框架,以达到较好的速度和精度平衡。然而,我们观察到来自浅层的低级特征经过较少的处理,从而在不同级别的特征之间产生潜在的语义差距。同时,由于没有考虑两分支框架的特征,刚性融合的效果较差。本文提出了统一交互模块和分离金字塔池模块来解决这两个问题。基于我们提出的模块,我们提出了一个新的双流分割网络(DSSNet),一个实时语义分割的双分支框架。与BiSeNet相比,我们基于ResNet18的DSSNet在cityscape测试数据集上的性能达到76.45% mIoU,计算成本与BiSeNet相近。此外,我们采用ResNet34骨干网的DSSNet优于以前的实时模型,在GTX1080Ti上以39 FPS的速度在cityscape测试数据集上实现了78.5%的mIoU。
{"title":"Dual Stream Segmentation Network for Real-Time Semantic Segmentation","authors":"Changyuan Zhong, Zelin Hu, Miao Li, Hualong Li, Xuanjiang Yang, Fei Liu","doi":"10.1109/ICIVC50857.2020.9177439","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177439","url":null,"abstract":"Modern real-time segmentation methods employ two-branch framework to achieve good speed and accuracy trade-off. However, we observe that low-level features coming from the shallow layers go through less processing, producing a potential semantic gap between different levels of features. Meanwhile, a rigid fusion is less effective due to the absence of consideration for two-branch framework characteristics. In this paper, we propose two novel modules: Unified Interplay Module and Separate Pyramid Pooling Module to address those two issues respectively. Based on our proposed modules, we present a novel Dual Stream Segmentation Network (DSSNet), a two-branch framework for real-time semantic segmentation. Compared with BiSeNet, our DSSNet based on ResNet18 achieves better performance 76.45% mIoU on the Cityscapes test dataset while sharing similar computation costs with BiSeNet. Furthermore, our DSSNet with ResNet34 backbone outperforms previous real-time models, achieving 78.5% mIoU on the Cityscapes test dataset with speed of 39 FPS on GTX1080Ti.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"46 1","pages":"144-149"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91101276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Design of Face Recognition Attendance 人脸识别考勤系统的设计
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177492
Hai-Wu Lee, Wen-Tan Gu, Yuan-yuan Wang
In recent years, face recognition technology has developed rapidly, and its application range has become more and more extensive. It is one of the most important application fields in computer vision technology. However, there are still many technical factors that restrict the application and promotion of face recognition technology. For example: shadows, occlusions, light and dark areas, dark light, highlights and other factors will make the face recognition rate drop sharply. Therefore, face recognition has extremely high research and application value. We use the Local Binary Patterns (LBP) algorithms with histogram equalization to obtain high-resolution images and improve the recognition rate in different scenarios, and try to apply face recognition to attendance.
近年来,人脸识别技术发展迅速,其应用范围也越来越广泛。它是计算机视觉技术最重要的应用领域之一。然而,仍然有许多技术因素制约着人脸识别技术的应用和推广。例如:阴影、遮挡、明暗区域、暗光、高光等因素都会使人脸识别率急剧下降。因此,人脸识别具有极高的研究和应用价值。利用直方图均衡化的局部二值模式(LBP)算法获得高分辨率图像,提高不同场景下的识别率,并尝试将人脸识别应用于考勤。
{"title":"Design of Face Recognition Attendance","authors":"Hai-Wu Lee, Wen-Tan Gu, Yuan-yuan Wang","doi":"10.1109/ICIVC50857.2020.9177492","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177492","url":null,"abstract":"In recent years, face recognition technology has developed rapidly, and its application range has become more and more extensive. It is one of the most important application fields in computer vision technology. However, there are still many technical factors that restrict the application and promotion of face recognition technology. For example: shadows, occlusions, light and dark areas, dark light, highlights and other factors will make the face recognition rate drop sharply. Therefore, face recognition has extremely high research and application value. We use the Local Binary Patterns (LBP) algorithms with histogram equalization to obtain high-resolution images and improve the recognition rate in different scenarios, and try to apply face recognition to attendance.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"84 1","pages":"222-226"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83810027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
AI Illustrator: Art Illustration Generation Based on Generative Adversarial Network AI Illustrator:基于生成对抗网络的艺术插图生成
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177494
Zihan Chen, Lianghong Chen, Zhiyuan Zhao, Yue Wang
In recent years, people's pursuit of art has been on the rise. People want computers to be able to create artistic paintings based on descriptions. In this paper, we proposed a novel project, Painting Creator, which uses deep learning technology to enable the computer to generate artistic illustrations from a short piece of text. Our scheme includes two models, image generation model and style transfer model. In the real image generation model, inspired by the application of stack generative adversarial networks in text to image generation, we proposed an improved model, IStackGAN, to solve the problem of image generation. We added a classifier based on the original model and added image structure loss and feature extraction loss to improve the performance of the generator. The generator network can get additional hidden information from the classification information to produce better pictures. The loss of image structure can force the generator to restore the real image, and the loss of feature extraction can verify whether the generator network has extracted the features of the real image set. For the style transfer model, we improved the generator based on the original cycle generative adversarial networks and used the residual block to improve the stability and performance of the u-net generator. To improve the performance of the generator, we also added the cycle consistent loss with MS-SSIM. The experimental results show that our model is improved significantly based on the original paper, and the generated pictures are more vivid in detail, and pictures after the style transfer are more artistic to watch.
近年来,人们对艺术的追求一直在上升。人们希望计算机能够根据描述来创作艺术绘画。在本文中,我们提出了一个新颖的项目,Painting Creator,它使用深度学习技术使计算机能够从一小段文本中生成艺术插图。我们的方案包括两个模型,图像生成模型和风格迁移模型。在真实图像生成模型中,受文本中的堆栈生成对抗网络应用于图像生成的启发,我们提出了一种改进的模型IStackGAN来解决图像生成问题。我们在原有模型的基础上增加了分类器,并增加了图像结构损失和特征提取损失,提高了生成器的性能。生成器网络可以从分类信息中获取额外的隐藏信息,从而生成更好的图像。图像结构的丢失可以迫使生成器还原真实图像,特征提取的丢失可以验证生成器网络是否提取了真实图像集的特征。对于风格迁移模型,我们在原始循环生成对抗网络的基础上改进了生成器,并使用残差块来提高u-net生成器的稳定性和性能。为了提高发生器的性能,我们还增加了MS-SSIM的周期一致损耗。实验结果表明,我们的模型在原论文的基础上有了明显的改进,生成的图片细节更加生动,风格转换后的图片更具观赏性。
{"title":"AI Illustrator: Art Illustration Generation Based on Generative Adversarial Network","authors":"Zihan Chen, Lianghong Chen, Zhiyuan Zhao, Yue Wang","doi":"10.1109/ICIVC50857.2020.9177494","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177494","url":null,"abstract":"In recent years, people's pursuit of art has been on the rise. People want computers to be able to create artistic paintings based on descriptions. In this paper, we proposed a novel project, Painting Creator, which uses deep learning technology to enable the computer to generate artistic illustrations from a short piece of text. Our scheme includes two models, image generation model and style transfer model. In the real image generation model, inspired by the application of stack generative adversarial networks in text to image generation, we proposed an improved model, IStackGAN, to solve the problem of image generation. We added a classifier based on the original model and added image structure loss and feature extraction loss to improve the performance of the generator. The generator network can get additional hidden information from the classification information to produce better pictures. The loss of image structure can force the generator to restore the real image, and the loss of feature extraction can verify whether the generator network has extracted the features of the real image set. For the style transfer model, we improved the generator based on the original cycle generative adversarial networks and used the residual block to improve the stability and performance of the u-net generator. To improve the performance of the generator, we also added the cycle consistent loss with MS-SSIM. The experimental results show that our model is improved significantly based on the original paper, and the generated pictures are more vivid in detail, and pictures after the style transfer are more artistic to watch.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"69 1","pages":"155-159"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84354177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Analyzing Gully Planform Changes in GIS Based on Multi-level Topological Relations 基于多层次拓扑关系的GIS沟壑平台变化分析
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177461
Feng Guoqiang, Leng Liang, Ye Yinghui, Han Dong-liang
Topological relations can be used to describe qualitative geometric position relations between spatial objects in geospatial world, which plays important roles in spatial query, spatial analysis and spatial reasoning. People can apply topological relations to describe the morphological changes of real objects, such as changes of cadastral parcels, rivers, water systems, etc. Gully planform changes (GPCs) reflect the state of surface soil erosion, so it is important and valuable to describe GPCs in detail. In this paper, based on a hierarchical topological relation description method and combined with the features of GPCs in GIS, we propose a simple hierarchical topological relationship description method to describe GPCs. This method can be used to completely describe GPCs, and is more concise and efficient than the former hierarchical topological relation description method in describing GPCs.
拓扑关系可以定性地描述地理空间世界中空间对象之间的几何位置关系,在空间查询、空间分析和空间推理中起着重要作用。人们可以运用拓扑关系来描述真实物体的形态变化,如地籍地块、河流、水系等的变化。沟壑区台地变化反映了地表土壤侵蚀状况,对其进行详细描述具有重要意义和价值。本文在层次化拓扑关系描述方法的基础上,结合GIS中gpc的特点,提出了一种简单的层次化拓扑关系描述方法来描述gpc。该方法可以完整地描述gpc,并且在描述gpc时比以前的分层拓扑关系描述方法更简洁、高效。
{"title":"Analyzing Gully Planform Changes in GIS Based on Multi-level Topological Relations","authors":"Feng Guoqiang, Leng Liang, Ye Yinghui, Han Dong-liang","doi":"10.1109/ICIVC50857.2020.9177461","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177461","url":null,"abstract":"Topological relations can be used to describe qualitative geometric position relations between spatial objects in geospatial world, which plays important roles in spatial query, spatial analysis and spatial reasoning. People can apply topological relations to describe the morphological changes of real objects, such as changes of cadastral parcels, rivers, water systems, etc. Gully planform changes (GPCs) reflect the state of surface soil erosion, so it is important and valuable to describe GPCs in detail. In this paper, based on a hierarchical topological relation description method and combined with the features of GPCs in GIS, we propose a simple hierarchical topological relationship description method to describe GPCs. This method can be used to completely describe GPCs, and is more concise and efficient than the former hierarchical topological relation description method in describing GPCs.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"29 1","pages":"292-295"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82513028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1