首页 > 最新文献

2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)最新文献

英文 中文
Facial Expression Recognition Based on Graph Neural Network 基于图神经网络的面部表情识别
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177430
Xuchou Xu, Zhou Ruan, Lei Yang
Facial expressions are one of the most powerful, natural and immediate means for human being to present their emotions and intensions. In this paper, we present a novel method for fully automatic facial expression recognition. The facial landmarks are detected for characterizing facial expressions. A graph convolutional neural network is proposed for feature extraction and facial expression recognition classification. The experiments were performed on the three facial expression databases. The result shows that the proposed FER method can achieve good recognition accuracy up to 95.85% using the proposed method.
面部表情是人类表达情感和意图的最有力、最自然、最直接的手段之一。本文提出了一种全自动面部表情识别的新方法。通过检测面部标志来表征面部表情。提出了一种用于特征提取和面部表情识别分类的图卷积神经网络。实验在三个面部表情数据库上进行。结果表明,该方法的识别准确率可达95.85%。
{"title":"Facial Expression Recognition Based on Graph Neural Network","authors":"Xuchou Xu, Zhou Ruan, Lei Yang","doi":"10.1109/ICIVC50857.2020.9177430","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177430","url":null,"abstract":"Facial expressions are one of the most powerful, natural and immediate means for human being to present their emotions and intensions. In this paper, we present a novel method for fully automatic facial expression recognition. The facial landmarks are detected for characterizing facial expressions. A graph convolutional neural network is proposed for feature extraction and facial expression recognition classification. The experiments were performed on the three facial expression databases. The result shows that the proposed FER method can achieve good recognition accuracy up to 95.85% using the proposed method.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"80 1","pages":"211-214"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88967613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
High Temperature Deformation Field Measurement Using 3D Digital Image Correlation Method 三维数字图像相关法测量高温变形场
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177479
Hongtao Deng, D. Jiang, Kai Wang, Q. Fei
A full-field, three-dimensional and non-contact deformation field measurement method under high temperature environment based on 3D digital image correlation (3D-DIC) is introduced. In order to reduce the impact of high temperature radiation on the image quality, a band-pass filter is placed in front of the camera lens. The two cameras simultaneously take pictures of the object before and after deformation, and use 3D-DIC to measure the three-dimensional deformation field of the object surface. The high temperature deformation field measurement test shows that 3D-DIC can accurately and conveniently measure the deformation field of an object under high temperature environment.
介绍了一种基于三维数字图像相关(3D- dic)的高温环境下全场、三维、非接触变形场测量方法。为了减少高温辐射对图像质量的影响,在相机镜头前放置了带通滤波器。两台相机同时拍摄物体变形前后的照片,并使用3D-DIC测量物体表面的三维变形场。高温变形场测量试验表明,3D-DIC能够准确、方便地测量物体在高温环境下的变形场。
{"title":"High Temperature Deformation Field Measurement Using 3D Digital Image Correlation Method","authors":"Hongtao Deng, D. Jiang, Kai Wang, Q. Fei","doi":"10.1109/ICIVC50857.2020.9177479","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177479","url":null,"abstract":"A full-field, three-dimensional and non-contact deformation field measurement method under high temperature environment based on 3D digital image correlation (3D-DIC) is introduced. In order to reduce the impact of high temperature radiation on the image quality, a band-pass filter is placed in front of the camera lens. The two cameras simultaneously take pictures of the object before and after deformation, and use 3D-DIC to measure the three-dimensional deformation field of the object surface. The high temperature deformation field measurement test shows that 3D-DIC can accurately and conveniently measure the deformation field of an object under high temperature environment.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"96 1","pages":"188-192"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88969096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Color Image Filtering in Bessel-Fourier Moments Domain 贝塞尔-傅里叶矩域彩色图像滤波
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177478
Tianpeng Xia, S. Liao
In this research, we have conducted a study on color image filtering in Bessel-Fourier moments domain. Bessel-Fourier moments of the two testing color images are computed independently from the three color channels (RGB), then lowpass and highpass filters are applied to the data in Bessel-Fourier moments domain for our investigation. For comparison, filters are applied in Fourier Frequency domain as well. The experimental results suggest that Bessel-Fourier moments of the lower orders contain mainly information of smooth varying components of images, while those of the higher orders are more related to details such as sharp transitions in intensity. It is also found that the Gaussian filters would reduce the ringing effect in Bessel-Fourier moments domain as they do in the Fourier Frequency domain.
在本研究中,我们对彩色图像的贝塞尔-傅里叶矩域滤波进行了研究。在三个颜色通道(RGB)中独立计算两幅测试彩色图像的贝塞尔-傅里叶矩,然后在贝塞尔-傅里叶矩域对数据应用低通和高通滤波器进行研究。为了比较,在傅里叶频域也应用了滤波器。实验结果表明,低阶贝塞尔-傅里叶矩主要包含图像平滑变化分量的信息,而高阶贝塞尔-傅里叶矩则更多地与图像强度的急剧变化等细节有关。研究还发现,高斯滤波器在贝塞尔-傅立叶矩域中和在傅立叶频域中一样,都能减小振铃效应。
{"title":"Color Image Filtering in Bessel-Fourier Moments Domain","authors":"Tianpeng Xia, S. Liao","doi":"10.1109/ICIVC50857.2020.9177478","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177478","url":null,"abstract":"In this research, we have conducted a study on color image filtering in Bessel-Fourier moments domain. Bessel-Fourier moments of the two testing color images are computed independently from the three color channels (RGB), then lowpass and highpass filters are applied to the data in Bessel-Fourier moments domain for our investigation. For comparison, filters are applied in Fourier Frequency domain as well. The experimental results suggest that Bessel-Fourier moments of the lower orders contain mainly information of smooth varying components of images, while those of the higher orders are more related to details such as sharp transitions in intensity. It is also found that the Gaussian filters would reduce the ringing effect in Bessel-Fourier moments domain as they do in the Fourier Frequency domain.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"112 1","pages":"75-81"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88776646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Clone Chaotic Niche Evolutionary Algorithm for Duty Cycle Control Optimization in Wireless Multimedia Sensor Networks 无线多媒体传感器网络占空比控制优化的克隆混沌小生境进化算法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177435
Jie Zhou, Mengying Xu, Rui Yang
One of the most interesting issue regarding to wireless multimedia sensor networks (WMSNs) is to maximizing the network lifetime. Because sensor nodes are constrained in energy, it is very important and necessary to exploit novel duty cycle design algorithms. Such a problem is important in improving network lifetime in WMSNs. The new contribution of our paper is that we propose a clone chaotic niche evolutionary algorithm (CCNEA) for duty cycle design problem in WMSNs. Novel clone operator and chaotic operator have been designed to develop solutions randomly. The strategy merges the merits of clone selection, chaotic generation, and niche operator. CCNEA is a style of swarm algorithm, which has strong global exploit ability. CCNEA utilizes chaotic generation approach which targets to avoid local optima. Then, simulations are performed to verify the robust and efficacy performance of CCNEA compared to methods according to particle swarm optimization (PSO) and quantum genetic algorithm (QGA) under an WMSNs conditions. Simulation experiments denote that the presented CCNEA outperforms PSO and QGA under different conditions, especially for WMSNs that has large number of sensors.
无线多媒体传感器网络(wmsn)中最令人感兴趣的问题之一是如何使网络的生存时间最大化。由于传感器节点受到能量的限制,开发新的占空比设计算法是非常重要和必要的。该问题对于提高wmsn的网络生存时间具有重要意义。本文的新贡献是我们提出了一种克隆混沌生态位进化算法(CCNEA)来解决wmsn的占空比设计问题。设计了新颖的克隆算子和混沌算子来随机求解。该策略融合了克隆选择、混沌生成和小生境算子的优点。CCNEA是一种群算法,具有较强的全局攻击能力。CCNEA采用混沌生成方法,以避免局部最优为目标。在WMSNs条件下,对比粒子群优化(PSO)和量子遗传算法(QGA),仿真验证了CCNEA算法的鲁棒性和有效性。仿真实验表明,本文提出的CCNEA算法在不同条件下都优于粒子群算法和QGA算法,特别是对于具有大量传感器的wmsn。
{"title":"Clone Chaotic Niche Evolutionary Algorithm for Duty Cycle Control Optimization in Wireless Multimedia Sensor Networks","authors":"Jie Zhou, Mengying Xu, Rui Yang","doi":"10.1109/ICIVC50857.2020.9177435","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177435","url":null,"abstract":"One of the most interesting issue regarding to wireless multimedia sensor networks (WMSNs) is to maximizing the network lifetime. Because sensor nodes are constrained in energy, it is very important and necessary to exploit novel duty cycle design algorithms. Such a problem is important in improving network lifetime in WMSNs. The new contribution of our paper is that we propose a clone chaotic niche evolutionary algorithm (CCNEA) for duty cycle design problem in WMSNs. Novel clone operator and chaotic operator have been designed to develop solutions randomly. The strategy merges the merits of clone selection, chaotic generation, and niche operator. CCNEA is a style of swarm algorithm, which has strong global exploit ability. CCNEA utilizes chaotic generation approach which targets to avoid local optima. Then, simulations are performed to verify the robust and efficacy performance of CCNEA compared to methods according to particle swarm optimization (PSO) and quantum genetic algorithm (QGA) under an WMSNs conditions. Simulation experiments denote that the presented CCNEA outperforms PSO and QGA under different conditions, especially for WMSNs that has large number of sensors.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"15 1","pages":"278-282"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90767734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Dual Stream Segmentation Network for Real-Time Semantic Segmentation 实时语义分割的双流分割网络
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177439
Changyuan Zhong, Zelin Hu, Miao Li, Hualong Li, Xuanjiang Yang, Fei Liu
Modern real-time segmentation methods employ two-branch framework to achieve good speed and accuracy trade-off. However, we observe that low-level features coming from the shallow layers go through less processing, producing a potential semantic gap between different levels of features. Meanwhile, a rigid fusion is less effective due to the absence of consideration for two-branch framework characteristics. In this paper, we propose two novel modules: Unified Interplay Module and Separate Pyramid Pooling Module to address those two issues respectively. Based on our proposed modules, we present a novel Dual Stream Segmentation Network (DSSNet), a two-branch framework for real-time semantic segmentation. Compared with BiSeNet, our DSSNet based on ResNet18 achieves better performance 76.45% mIoU on the Cityscapes test dataset while sharing similar computation costs with BiSeNet. Furthermore, our DSSNet with ResNet34 backbone outperforms previous real-time models, achieving 78.5% mIoU on the Cityscapes test dataset with speed of 39 FPS on GTX1080Ti.
现代实时分割方法采用双分支框架,以达到较好的速度和精度平衡。然而,我们观察到来自浅层的低级特征经过较少的处理,从而在不同级别的特征之间产生潜在的语义差距。同时,由于没有考虑两分支框架的特征,刚性融合的效果较差。本文提出了统一交互模块和分离金字塔池模块来解决这两个问题。基于我们提出的模块,我们提出了一个新的双流分割网络(DSSNet),一个实时语义分割的双分支框架。与BiSeNet相比,我们基于ResNet18的DSSNet在cityscape测试数据集上的性能达到76.45% mIoU,计算成本与BiSeNet相近。此外,我们采用ResNet34骨干网的DSSNet优于以前的实时模型,在GTX1080Ti上以39 FPS的速度在cityscape测试数据集上实现了78.5%的mIoU。
{"title":"Dual Stream Segmentation Network for Real-Time Semantic Segmentation","authors":"Changyuan Zhong, Zelin Hu, Miao Li, Hualong Li, Xuanjiang Yang, Fei Liu","doi":"10.1109/ICIVC50857.2020.9177439","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177439","url":null,"abstract":"Modern real-time segmentation methods employ two-branch framework to achieve good speed and accuracy trade-off. However, we observe that low-level features coming from the shallow layers go through less processing, producing a potential semantic gap between different levels of features. Meanwhile, a rigid fusion is less effective due to the absence of consideration for two-branch framework characteristics. In this paper, we propose two novel modules: Unified Interplay Module and Separate Pyramid Pooling Module to address those two issues respectively. Based on our proposed modules, we present a novel Dual Stream Segmentation Network (DSSNet), a two-branch framework for real-time semantic segmentation. Compared with BiSeNet, our DSSNet based on ResNet18 achieves better performance 76.45% mIoU on the Cityscapes test dataset while sharing similar computation costs with BiSeNet. Furthermore, our DSSNet with ResNet34 backbone outperforms previous real-time models, achieving 78.5% mIoU on the Cityscapes test dataset with speed of 39 FPS on GTX1080Ti.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"46 1","pages":"144-149"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91101276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Foreground Mask Network for Cell Counting 一种用于细胞计数的前景掩码网络
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177433
Ni Jiang, Fei-hong Yu
Cell counting is important in medical image analysis for its meaningful information. In this paper, we propose a cell counting network to predict the number of cells in an image with the distribution of cells. The proposed network learns to predict the density map which has a direct relationship with the number of cells. A foreground mask is designed to filter the low-level feature maps and the favorable information is fed to the decoder to recover the spatial information better. The foreground mask is a probability map indicating the pixels are more likely to belong to cells. Experiments on three public datasets show that the proposed model can achieve promising performances. Especially the ablation study on the Adipocyte Cells demonstrates the necessity of the foreground mask.
细胞计数在医学图像分析中具有重要的意义。本文提出了一种细胞计数网络,利用细胞的分布来预测图像中细胞的数量。该网络学习预测与细胞数量有直接关系的密度图。设计前景掩模对底层特征图进行过滤,将有利信息送入解码器,更好地恢复空间信息。前景蒙版是一个概率图,表示像素更有可能属于细胞。在三个公共数据集上的实验表明,该模型可以取得良好的性能。特别是对脂肪细胞的消融研究,证明了前景掩膜的必要性。
{"title":"A Foreground Mask Network for Cell Counting","authors":"Ni Jiang, Fei-hong Yu","doi":"10.1109/ICIVC50857.2020.9177433","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177433","url":null,"abstract":"Cell counting is important in medical image analysis for its meaningful information. In this paper, we propose a cell counting network to predict the number of cells in an image with the distribution of cells. The proposed network learns to predict the density map which has a direct relationship with the number of cells. A foreground mask is designed to filter the low-level feature maps and the favorable information is fed to the decoder to recover the spatial information better. The foreground mask is a probability map indicating the pixels are more likely to belong to cells. Experiments on three public datasets show that the proposed model can achieve promising performances. Especially the ablation study on the Adipocyte Cells demonstrates the necessity of the foreground mask.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"4 1","pages":"128-132"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80141776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Craik-O'Brien Effect Based Lightness Modification Method Considering Color Distance for Dichromats 基于Craik-O'Brien效应的二色图颜色距离明度修正方法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177463
Yujun Liu, Shi Bao, Chuanying Yang, Shaoying Ma
Dichromacy, also called as visual impairment, is an inherited disease of defective or abnormal color vision characterized by an inability to recognize certain colors. In this paper, a new lightness modification method based on Craik-O'Brien (C-O) effect was proposed in order to improve the color recognition ability of dichromats. The main idea is to modify the lightness values of the contour parts of the regions which are easy to be confused for dichromats by establishing the objective function of considering color distance, and find the optimal lightness modification value by using the steepest descent method. The modified images will generate C-O effect, which will make the observers produce a visual lightness difference, thus improving the color recognition of the images. The proposed method can retain the information details and overall naturalness of the original color images, making it easier to obtain information and perceive color variations and overall characteristics in the original color images for dichromats. The effectiveness and feasibility of the proposed method is shown in the experimental part by means of the comparison and analysis among the test image, the dichromacy simulation images and the result images obtained by this method.
二色性,也被称为视觉障碍,是一种遗传性疾病,以无法识别某些颜色为特征的色觉缺陷或异常。为了提高二色体的颜色识别能力,提出了一种基于Craik-O'Brien (C-O)效应的亮度修正方法。其主要思想是通过建立考虑颜色距离的目标函数,对易混淆为二色者的区域轮廓部分的亮度值进行修改,并采用最陡下降法找到最优亮度修改值。修改后的图像会产生C-O效应,使观察者产生视觉亮度差异,从而提高图像的颜色识别能力。该方法可以保留原始彩色图像的信息细节和整体自然度,使二色者更容易获取信息,感知原始彩色图像中的颜色变化和整体特征。实验部分通过对测试图像、二色模拟图像和该方法得到的结果图像进行对比分析,证明了该方法的有效性和可行性。
{"title":"A Craik-O'Brien Effect Based Lightness Modification Method Considering Color Distance for Dichromats","authors":"Yujun Liu, Shi Bao, Chuanying Yang, Shaoying Ma","doi":"10.1109/ICIVC50857.2020.9177463","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177463","url":null,"abstract":"Dichromacy, also called as visual impairment, is an inherited disease of defective or abnormal color vision characterized by an inability to recognize certain colors. In this paper, a new lightness modification method based on Craik-O'Brien (C-O) effect was proposed in order to improve the color recognition ability of dichromats. The main idea is to modify the lightness values of the contour parts of the regions which are easy to be confused for dichromats by establishing the objective function of considering color distance, and find the optimal lightness modification value by using the steepest descent method. The modified images will generate C-O effect, which will make the observers produce a visual lightness difference, thus improving the color recognition of the images. The proposed method can retain the information details and overall naturalness of the original color images, making it easier to obtain information and perceive color variations and overall characteristics in the original color images for dichromats. The effectiveness and feasibility of the proposed method is shown in the experimental part by means of the comparison and analysis among the test image, the dichromacy simulation images and the result images obtained by this method.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"1 1","pages":"86-91"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84663772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Design of Face Recognition Attendance 人脸识别考勤系统的设计
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177492
Hai-Wu Lee, Wen-Tan Gu, Yuan-yuan Wang
In recent years, face recognition technology has developed rapidly, and its application range has become more and more extensive. It is one of the most important application fields in computer vision technology. However, there are still many technical factors that restrict the application and promotion of face recognition technology. For example: shadows, occlusions, light and dark areas, dark light, highlights and other factors will make the face recognition rate drop sharply. Therefore, face recognition has extremely high research and application value. We use the Local Binary Patterns (LBP) algorithms with histogram equalization to obtain high-resolution images and improve the recognition rate in different scenarios, and try to apply face recognition to attendance.
近年来,人脸识别技术发展迅速,其应用范围也越来越广泛。它是计算机视觉技术最重要的应用领域之一。然而,仍然有许多技术因素制约着人脸识别技术的应用和推广。例如:阴影、遮挡、明暗区域、暗光、高光等因素都会使人脸识别率急剧下降。因此,人脸识别具有极高的研究和应用价值。利用直方图均衡化的局部二值模式(LBP)算法获得高分辨率图像,提高不同场景下的识别率,并尝试将人脸识别应用于考勤。
{"title":"Design of Face Recognition Attendance","authors":"Hai-Wu Lee, Wen-Tan Gu, Yuan-yuan Wang","doi":"10.1109/ICIVC50857.2020.9177492","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177492","url":null,"abstract":"In recent years, face recognition technology has developed rapidly, and its application range has become more and more extensive. It is one of the most important application fields in computer vision technology. However, there are still many technical factors that restrict the application and promotion of face recognition technology. For example: shadows, occlusions, light and dark areas, dark light, highlights and other factors will make the face recognition rate drop sharply. Therefore, face recognition has extremely high research and application value. We use the Local Binary Patterns (LBP) algorithms with histogram equalization to obtain high-resolution images and improve the recognition rate in different scenarios, and try to apply face recognition to attendance.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"84 1","pages":"222-226"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83810027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Gain-Pixel Visualization Algorithm Designed for Computational Color Constancy Scheme 基于计算色彩常数方案的增益-像素可视化算法
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177464
S. Teng
Color constancy (CC) is an essential part of machine vision. Previously reported CC algorithms lacked consistent and clear-cut evaluation diagrams. This paper instead presents a gain-pixel visualization CC algorithm which uses optimization numerical analysis and 2D-3D graphical displays. This graph-based CC algorithm differs from others in that it gives a clear overall perspective on finding the appropriate amount of RGB gain adjustment to achieve image CC. The ground truth (GT) image, which is critical for data accuracy, has been used as a benchmark or a target in image CC. However, GT images in CC are often inconsistently determined or manually checked. This paper will illustrate that an accurate and specific GT image can be obtained or checked using an optimization scheme, namely the grayscale pixel maximization (GPM). Using previously published image CC results for evaluation and comparison, this paper demonstrates the usefulness, accuracy, and especially the forensic capability of this CC algorithm.
色彩恒常性(CC)是机器视觉的重要组成部分。先前报道的CC算法缺乏一致和明确的评估图。本文提出了一种利用优化数值分析和2D-3D图形显示的增益-像素可视化CC算法。这种基于图的CC算法与其他算法的不同之处在于,它给出了一个清晰的整体视角,如何找到合适的RGB增益调整来实现图像CC,对数据精度至关重要的ground truth (GT)图像被用作图像CC的基准或目标,但是CC中的GT图像往往是不一致的确定或人工检查。本文将说明使用优化方案,即灰度像素最大化(GPM),可以获得或检查精确和特定的GT图像。本文使用先前发表的图像CC结果进行评估和比较,证明了该CC算法的有用性,准确性,特别是取证能力。
{"title":"Gain-Pixel Visualization Algorithm Designed for Computational Color Constancy Scheme","authors":"S. Teng","doi":"10.1109/ICIVC50857.2020.9177464","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177464","url":null,"abstract":"Color constancy (CC) is an essential part of machine vision. Previously reported CC algorithms lacked consistent and clear-cut evaluation diagrams. This paper instead presents a gain-pixel visualization CC algorithm which uses optimization numerical analysis and 2D-3D graphical displays. This graph-based CC algorithm differs from others in that it gives a clear overall perspective on finding the appropriate amount of RGB gain adjustment to achieve image CC. The ground truth (GT) image, which is critical for data accuracy, has been used as a benchmark or a target in image CC. However, GT images in CC are often inconsistently determined or manually checked. This paper will illustrate that an accurate and specific GT image can be obtained or checked using an optimization scheme, namely the grayscale pixel maximization (GPM). Using previously published image CC results for evaluation and comparison, this paper demonstrates the usefulness, accuracy, and especially the forensic capability of this CC algorithm.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"20 1","pages":"237-246"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88969213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI Illustrator: Art Illustration Generation Based on Generative Adversarial Network AI Illustrator:基于生成对抗网络的艺术插图生成
Pub Date : 2020-07-01 DOI: 10.1109/ICIVC50857.2020.9177494
Zihan Chen, Lianghong Chen, Zhiyuan Zhao, Yue Wang
In recent years, people's pursuit of art has been on the rise. People want computers to be able to create artistic paintings based on descriptions. In this paper, we proposed a novel project, Painting Creator, which uses deep learning technology to enable the computer to generate artistic illustrations from a short piece of text. Our scheme includes two models, image generation model and style transfer model. In the real image generation model, inspired by the application of stack generative adversarial networks in text to image generation, we proposed an improved model, IStackGAN, to solve the problem of image generation. We added a classifier based on the original model and added image structure loss and feature extraction loss to improve the performance of the generator. The generator network can get additional hidden information from the classification information to produce better pictures. The loss of image structure can force the generator to restore the real image, and the loss of feature extraction can verify whether the generator network has extracted the features of the real image set. For the style transfer model, we improved the generator based on the original cycle generative adversarial networks and used the residual block to improve the stability and performance of the u-net generator. To improve the performance of the generator, we also added the cycle consistent loss with MS-SSIM. The experimental results show that our model is improved significantly based on the original paper, and the generated pictures are more vivid in detail, and pictures after the style transfer are more artistic to watch.
近年来,人们对艺术的追求一直在上升。人们希望计算机能够根据描述来创作艺术绘画。在本文中,我们提出了一个新颖的项目,Painting Creator,它使用深度学习技术使计算机能够从一小段文本中生成艺术插图。我们的方案包括两个模型,图像生成模型和风格迁移模型。在真实图像生成模型中,受文本中的堆栈生成对抗网络应用于图像生成的启发,我们提出了一种改进的模型IStackGAN来解决图像生成问题。我们在原有模型的基础上增加了分类器,并增加了图像结构损失和特征提取损失,提高了生成器的性能。生成器网络可以从分类信息中获取额外的隐藏信息,从而生成更好的图像。图像结构的丢失可以迫使生成器还原真实图像,特征提取的丢失可以验证生成器网络是否提取了真实图像集的特征。对于风格迁移模型,我们在原始循环生成对抗网络的基础上改进了生成器,并使用残差块来提高u-net生成器的稳定性和性能。为了提高发生器的性能,我们还增加了MS-SSIM的周期一致损耗。实验结果表明,我们的模型在原论文的基础上有了明显的改进,生成的图片细节更加生动,风格转换后的图片更具观赏性。
{"title":"AI Illustrator: Art Illustration Generation Based on Generative Adversarial Network","authors":"Zihan Chen, Lianghong Chen, Zhiyuan Zhao, Yue Wang","doi":"10.1109/ICIVC50857.2020.9177494","DOIUrl":"https://doi.org/10.1109/ICIVC50857.2020.9177494","url":null,"abstract":"In recent years, people's pursuit of art has been on the rise. People want computers to be able to create artistic paintings based on descriptions. In this paper, we proposed a novel project, Painting Creator, which uses deep learning technology to enable the computer to generate artistic illustrations from a short piece of text. Our scheme includes two models, image generation model and style transfer model. In the real image generation model, inspired by the application of stack generative adversarial networks in text to image generation, we proposed an improved model, IStackGAN, to solve the problem of image generation. We added a classifier based on the original model and added image structure loss and feature extraction loss to improve the performance of the generator. The generator network can get additional hidden information from the classification information to produce better pictures. The loss of image structure can force the generator to restore the real image, and the loss of feature extraction can verify whether the generator network has extracted the features of the real image set. For the style transfer model, we improved the generator based on the original cycle generative adversarial networks and used the residual block to improve the stability and performance of the u-net generator. To improve the performance of the generator, we also added the cycle consistent loss with MS-SSIM. The experimental results show that our model is improved significantly based on the original paper, and the generated pictures are more vivid in detail, and pictures after the style transfer are more artistic to watch.","PeriodicalId":6806,"journal":{"name":"2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)","volume":"69 1","pages":"155-159"},"PeriodicalIF":0.0,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84354177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1