首页 > 最新文献

2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)最新文献

英文 中文
Image classification based on log-Euclidean Fisher Vectors for covariance matrix descriptors 基于对数欧氏费雪向量协方差矩阵描述符的图像分类
Sara Akodad, L. Bombrun, C. Yaacoub, Y. Berthoumieu, C. Germain
This paper introduces an image classification method based on the encoding of a set of covariance matrices. This encoding relies on Fisher vectors adapted to the log-Euclidean metric: the log-Euclidean Fisher vectors (LE FV). This approach is next extended to full local Gaussian descriptors composed by a set of local mean vectors and local covariance matrices. For that, the local Gaussian model is transformed to a zero-mean Gaussian model with an augmented covariance matrix. All these approaches are used to encode handcrafted or deep learning features. Finally, they are applied in a remote sensing application on the UC Merced dataset which consists in classifying land cover images. A sensitivity analysis is carried out to evaluate the potential of the proposed LE FV.
介绍了一种基于协方差矩阵编码的图像分类方法。这种编码依赖于适合对数欧几里得度量的Fisher向量:对数欧几里得Fisher向量(LE FV)。然后将该方法扩展到由一组局部均值向量和局部协方差矩阵组成的全局部高斯描述子。为此,将局部高斯模型转化为具有增广协方差矩阵的零均值高斯模型。所有这些方法都用于对手工或深度学习特征进行编码。最后,将它们应用于UC Merced数据集的遥感应用,该数据集包括对土地覆盖图像进行分类。进行了敏感性分析,以评估拟议的LE FV的潜力。
{"title":"Image classification based on log-Euclidean Fisher Vectors for covariance matrix descriptors","authors":"Sara Akodad, L. Bombrun, C. Yaacoub, Y. Berthoumieu, C. Germain","doi":"10.1109/IPTA.2018.8608154","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608154","url":null,"abstract":"This paper introduces an image classification method based on the encoding of a set of covariance matrices. This encoding relies on Fisher vectors adapted to the log-Euclidean metric: the log-Euclidean Fisher vectors (LE FV). This approach is next extended to full local Gaussian descriptors composed by a set of local mean vectors and local covariance matrices. For that, the local Gaussian model is transformed to a zero-mean Gaussian model with an augmented covariance matrix. All these approaches are used to encode handcrafted or deep learning features. Finally, they are applied in a remote sensing application on the UC Merced dataset which consists in classifying land cover images. A sensitivity analysis is carried out to evaluate the potential of the proposed LE FV.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129340370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hyperspetral Anomaly Detection Incorporating Spatial Information 基于空间信息的高光谱异常检测
H. Ju, Zhigang Liu, Yang Wang
Most anomaly detection methods for hyperspectral image (HSI) have focused on the spectral information while ignoring the spatial information. In this paper, a novel anomaly detection method has been proposed in which the spatial information has been incorporated. Firstly, the dual windows are established to estimate the background of pixel under test (PUT). Secondly, the spectral distance is calculated between PUT and its background to measure its spectral anomaly degree. Then, the principal component analysis is performed on HSI and the spatial anomaly degree of PUT is measured on the first component by comparing the spatial structure similarity between PUT and its background. Lastly, combining the spectral anomaly degree and the spatial anomaly degree, the anomaly degree of PUT is obtained. Experimental results on two hyperspectral datasets confirm the proposed method is superior to three commonly used state-of-the-art anomaly detection methods in suppressing the background and detecting anomalies and is also quite robust to noise.
大多数高光谱图像异常检测方法都只关注光谱信息,而忽略了空间信息。本文提出了一种结合空间信息的异常检测方法。首先,建立双窗口估计被测像素的背景(PUT);其次,计算PUT与背景的光谱距离,测量其光谱异常程度;然后,对HSI进行主成分分析,通过对比PUT与其背景的空间结构相似性,在第一分量上测量PUT的空间异常程度。最后,结合光谱异常度和空间异常度,得到PUT异常度。在两个高光谱数据集上的实验结果表明,该方法在抑制背景和检测异常方面优于三种常用的最新异常检测方法,并且对噪声具有较强的鲁棒性。
{"title":"Hyperspetral Anomaly Detection Incorporating Spatial Information","authors":"H. Ju, Zhigang Liu, Yang Wang","doi":"10.1109/IPTA.2018.8608161","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608161","url":null,"abstract":"Most anomaly detection methods for hyperspectral image (HSI) have focused on the spectral information while ignoring the spatial information. In this paper, a novel anomaly detection method has been proposed in which the spatial information has been incorporated. Firstly, the dual windows are established to estimate the background of pixel under test (PUT). Secondly, the spectral distance is calculated between PUT and its background to measure its spectral anomaly degree. Then, the principal component analysis is performed on HSI and the spatial anomaly degree of PUT is measured on the first component by comparing the spatial structure similarity between PUT and its background. Lastly, combining the spectral anomaly degree and the spatial anomaly degree, the anomaly degree of PUT is obtained. Experimental results on two hyperspectral datasets confirm the proposed method is superior to three commonly used state-of-the-art anomaly detection methods in suppressing the background and detecting anomalies and is also quite robust to noise.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129667426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Measurement Method for Vehicle Queue Length of Intersection Based on Image Processing 基于图像处理的交叉口车辆队列长度测量方法
Zhan Qi, Maojun Li, Chongpei Liu, M. Zhao, Manyi Long
We propose a real-time detection method for vehicle queue length detection at intersections based on image processing technology. Firstly, we acquire the image of vehicle queue at the intersection, and apply the automatic brightness adjustment algorithm and lane line detection algorithm to reduce the affect of different light intensity and camera shake on the image respectively. And then, the preprocessed image is subtracted from the background image to obtain the foreground image of queue vehicle. Finally, we detect the vehicle queue length by the middle line, and the actual vehicle queue length is measured by the camera calibration method. The experimental results show that the proposed method has high accuracy rate and is fast enough for practical application.
提出了一种基于图像处理技术的交叉口车辆队列长度实时检测方法。首先,获取路口车辆队列图像,采用自动亮度调节算法和车道线检测算法,分别降低不同光强和相机抖动对图像的影响;然后将预处理后的图像与背景图像相减,得到队列车辆的前景图像。最后通过中线检测车辆队列长度,通过摄像机标定方法测量实际车辆队列长度。实验结果表明,该方法具有较高的准确率和速度,适合实际应用。
{"title":"A Measurement Method for Vehicle Queue Length of Intersection Based on Image Processing","authors":"Zhan Qi, Maojun Li, Chongpei Liu, M. Zhao, Manyi Long","doi":"10.1109/IPTA.2018.8608140","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608140","url":null,"abstract":"We propose a real-time detection method for vehicle queue length detection at intersections based on image processing technology. Firstly, we acquire the image of vehicle queue at the intersection, and apply the automatic brightness adjustment algorithm and lane line detection algorithm to reduce the affect of different light intensity and camera shake on the image respectively. And then, the preprocessed image is subtracted from the background image to obtain the foreground image of queue vehicle. Finally, we detect the vehicle queue length by the middle line, and the actual vehicle queue length is measured by the camera calibration method. The experimental results show that the proposed method has high accuracy rate and is fast enough for practical application.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129134462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
InNet: Learning to Detect Shadows with Injection Network InNet:学习用注入网络检测阴影
Xiaoyue Jiang, Zhongyun Hu, Yue Ni
Shadows bring great challenges but also play essential roles in image understanding. Most recent shadow detection methods are based on patches, then further reasoning method is required for the obtaining of a completed shadow detection result for an image. In this paper, an injection network is proposed to detect shadow regions for the whole image directly. In order to maintain as many as details, the skip structure is applied to directly inject the details from convolutional layers to de-convolutional layers. Meanwhile, a weighted loss function is proposed for the network training. With this adapted loss function, the network becomes more sensitive to errors of shadow regions. Thus the proposed network can focus on the learning of robust shadow features. Furthermore, a shadow refinement method is proposed to optimize the boundary region of shadows. In the experiments, the proposed methods are extensively evaluated on two popular datasets and shown better performance on shadow detection compared with current methods.
阴影给图像理解带来了巨大的挑战,但也发挥了重要作用。目前的阴影检测方法大多是基于小块的,为了得到完整的图像阴影检测结果,需要进一步的推理方法。本文提出了一种直接检测整个图像阴影区域的注入网络。为了保持尽可能多的细节,采用跳跃结构将细节从卷积层直接注入到去卷积层。同时,提出了一种用于网络训练的加权损失函数。利用这种自适应损失函数,网络对阴影区域的误差更加敏感。因此,该网络可以专注于鲁棒阴影特征的学习。在此基础上,提出了一种阴影细化方法来优化阴影边界区域。在实验中,本文提出的方法在两个流行的数据集上进行了广泛的评估,与现有方法相比,在阴影检测方面表现出更好的性能。
{"title":"InNet: Learning to Detect Shadows with Injection Network","authors":"Xiaoyue Jiang, Zhongyun Hu, Yue Ni","doi":"10.1109/IPTA.2018.8608155","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608155","url":null,"abstract":"Shadows bring great challenges but also play essential roles in image understanding. Most recent shadow detection methods are based on patches, then further reasoning method is required for the obtaining of a completed shadow detection result for an image. In this paper, an injection network is proposed to detect shadow regions for the whole image directly. In order to maintain as many as details, the skip structure is applied to directly inject the details from convolutional layers to de-convolutional layers. Meanwhile, a weighted loss function is proposed for the network training. With this adapted loss function, the network becomes more sensitive to errors of shadow regions. Thus the proposed network can focus on the learning of robust shadow features. Furthermore, a shadow refinement method is proposed to optimize the boundary region of shadows. In the experiments, the proposed methods are extensively evaluated on two popular datasets and shown better performance on shadow detection compared with current methods.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114247552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A New Database for Evaluating Underwater Image Processing Methods 评价水下图像处理方法的新数据库
Yupeng Ma, Xiaoyi Feng, Lujing Chao, Dong Huang, Zhaoqiang Xia, Xiaoyue Jiang
In this paper, we present a new, large-scale database on underwater image, which is called the NWPU underwater image database. This database contains 6240 underwater images of 40 objects. Each object is captured with 6 different levels of turbidity water, 4 lighting conditions and 6 different distances. Among them, we use the underwater images with turbidity value of 0 as Ground-truth. In addition, we captured the shadowless image of the object in the air and clear water. Different from other underwater databases, we capture underwater images with real high turbidity lake water instead of simulating the turbidity of water. This method ensures that the underwater images we captured are as close as possible to the real environment. We have given the database baseline which contains multi-scale Retinex with color restore (MSRCR) algorithms for enhancing images and four commonly used image quality evaluation criteria, including two full-references and two no-references methods. The four image quality evaluation methods include two no-reference and two full reference.
本文提出了一种新型的大型水下图像数据库,即NWPU水下图像数据库。该数据库包含40个物体的6240幅水下图像。每个物体都有6种不同的浑浊度,4种照明条件和6种不同的距离。其中,我们使用浊度值为0的水下图像作为Ground-truth。此外,我们在空气和清澈的水中捕捉到了物体的无影图像。与其他水下数据库不同的是,我们捕捉的是真实的高浊度湖水的水下图像,而不是模拟水体的浊度。这种方法确保我们捕获的水下图像尽可能接近真实环境。我们给出了包含多尺度Retinex带颜色恢复(MSRCR)算法的数据库基线和四种常用的图像质量评价标准,包括两种全参考和两种无参考方法。四种图像质量评价方法包括两种无参考和两种完全参考。
{"title":"A New Database for Evaluating Underwater Image Processing Methods","authors":"Yupeng Ma, Xiaoyi Feng, Lujing Chao, Dong Huang, Zhaoqiang Xia, Xiaoyue Jiang","doi":"10.1109/IPTA.2018.8608131","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608131","url":null,"abstract":"In this paper, we present a new, large-scale database on underwater image, which is called the NWPU underwater image database. This database contains 6240 underwater images of 40 objects. Each object is captured with 6 different levels of turbidity water, 4 lighting conditions and 6 different distances. Among them, we use the underwater images with turbidity value of 0 as Ground-truth. In addition, we captured the shadowless image of the object in the air and clear water. Different from other underwater databases, we capture underwater images with real high turbidity lake water instead of simulating the turbidity of water. This method ensures that the underwater images we captured are as close as possible to the real environment. We have given the database baseline which contains multi-scale Retinex with color restore (MSRCR) algorithms for enhancing images and four commonly used image quality evaluation criteria, including two full-references and two no-references methods. The four image quality evaluation methods include two no-reference and two full reference.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132745118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Similar Trademark Image Retrieval Based on Convolutional Neural Network and Constraint Theory 基于卷积神经网络和约束理论的相似商标图像检索
Tian Lan, Xiaoyi Feng, Lei Li, Zhaoqiang Xia
Trademarks are intellectual and industrial properties developed under the commodity economy, representing reputation, quality and reliability of firms. Therefore, in order to prevent the registration of new trademarks from having a high-degree similarity with registered ones, we propose a new trademark retrieval method. Based on the fact that the shape and color of a trademark are varied, our proposed method combines a metric convolutional neural network (CNN) and conventional hand-crafted features to describe the trademark images. More specifically, we first train the CNN based on Siamese and Triplet structures, and then extract the hand-crafted features from convolutional feature maps. For this research, we utilize a challenging trademark dataset that contains 7139 various color or gray images. Besides, extensive experiments on our dataset and the METU public dataset demonstrate the effectiveness of our method in trademark retrieval and achieve the state-of-the-art performance compared to traditional countermeasures.
商标是在商品经济下发展起来的知识产权和工业产权,代表着企业的信誉、质量和可靠性。因此,为了防止新注册商标与已注册商标高度相似,我们提出了一种新的商标检索方法。基于商标形状和颜色的多样性,我们提出了一种结合度量卷积神经网络(CNN)和传统手工特征来描述商标图像的方法。更具体地说,我们首先基于Siamese和Triplet结构训练CNN,然后从卷积特征映射中提取手工制作的特征。在这项研究中,我们使用了一个具有挑战性的商标数据集,该数据集包含7139个各种颜色或灰色图像。此外,在我们的数据集和METU公共数据集上进行的大量实验表明,我们的方法在商标检索中是有效的,与传统的对策相比,达到了最先进的性能。
{"title":"Similar Trademark Image Retrieval Based on Convolutional Neural Network and Constraint Theory","authors":"Tian Lan, Xiaoyi Feng, Lei Li, Zhaoqiang Xia","doi":"10.1109/IPTA.2018.8608162","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608162","url":null,"abstract":"Trademarks are intellectual and industrial properties developed under the commodity economy, representing reputation, quality and reliability of firms. Therefore, in order to prevent the registration of new trademarks from having a high-degree similarity with registered ones, we propose a new trademark retrieval method. Based on the fact that the shape and color of a trademark are varied, our proposed method combines a metric convolutional neural network (CNN) and conventional hand-crafted features to describe the trademark images. More specifically, we first train the CNN based on Siamese and Triplet structures, and then extract the hand-crafted features from convolutional feature maps. For this research, we utilize a challenging trademark dataset that contains 7139 various color or gray images. Besides, extensive experiments on our dataset and the METU public dataset demonstrate the effectiveness of our method in trademark retrieval and achieve the state-of-the-art performance compared to traditional countermeasures.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115834218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Pedestrian Detection Using Regional Proposal Network with Feature Fusion 基于特征融合的区域建议网络行人检测
Xiaogang Lv, Xiaotao Zhang, Yinghua Jiang, Jianxin Zhang
Pedestrian detection, which has broad application prospects in video security, robotics and self-driving vehicles etc., is one of the most important research fields in computer vision. Recently, deep learning methods, e.g., Region Proposal Network (RPN), have achieved major performance improvements in pedestrian detection. In order to further utilize the deep pedestrian features of RPN, this paper proposes a novel regional proposal network model based on feature fusion (RPN_FeaFus) for pedestrian detection. RPN_FeaFus adopts an asymmetric dual-path deep model, constructed by VGGNet and ZFNet, to extract pedestrian features in different levels, which are further combined through PCA dimension reduction and feature stacking to provide more discriminant representation. Then, the low-dimensional fusion features are adopted to detect the region proposals and train the classifier. Experimental results on three widely used pedestrian detection databases, i.e, Caltech database, Daimler database and TUD database, illuminate that RPN_FeaFus gains obvious performance improvements over its baseline RPN_BF, which is also competitive with the state-of-the-art methods.
行人检测是计算机视觉中最重要的研究领域之一,在视频安防、机器人、自动驾驶汽车等领域有着广阔的应用前景。最近,深度学习方法,如区域建议网络(RPN),在行人检测方面取得了重大的性能改进。为了进一步利用RPN的深度行人特征,本文提出了一种基于特征融合的区域建议网络模型(RPN_FeaFus)用于行人检测。RPN_FeaFus采用由VGGNet和ZFNet构建的非对称双路径深度模型提取不同层次的行人特征,再通过PCA降维和特征叠加进行组合,提供更具判别性的表示。然后,利用低维融合特征检测区域建议并训练分类器。在Caltech数据库、Daimler数据库和TUD数据库这三个被广泛使用的行人检测数据库上的实验结果表明,RPN_FeaFus比其基准RPN_BF有了明显的性能提升,与最先进的方法相比也具有竞争力。
{"title":"Pedestrian Detection Using Regional Proposal Network with Feature Fusion","authors":"Xiaogang Lv, Xiaotao Zhang, Yinghua Jiang, Jianxin Zhang","doi":"10.1109/IPTA.2018.8608159","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608159","url":null,"abstract":"Pedestrian detection, which has broad application prospects in video security, robotics and self-driving vehicles etc., is one of the most important research fields in computer vision. Recently, deep learning methods, e.g., Region Proposal Network (RPN), have achieved major performance improvements in pedestrian detection. In order to further utilize the deep pedestrian features of RPN, this paper proposes a novel regional proposal network model based on feature fusion (RPN_FeaFus) for pedestrian detection. RPN_FeaFus adopts an asymmetric dual-path deep model, constructed by VGGNet and ZFNet, to extract pedestrian features in different levels, which are further combined through PCA dimension reduction and feature stacking to provide more discriminant representation. Then, the low-dimensional fusion features are adopted to detect the region proposals and train the classifier. Experimental results on three widely used pedestrian detection databases, i.e, Caltech database, Daimler database and TUD database, illuminate that RPN_FeaFus gains obvious performance improvements over its baseline RPN_BF, which is also competitive with the state-of-the-art methods.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124491746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Spontaneous Facial Micro-expression Recognition via Deep Convolutional Network 基于深度卷积网络的自发面部微表情识别
Zhaoqiang Xia, Xiaoyi Feng, Xiaopeng Hong, Guoying Zhao
The automatic recognition of spontaneous facial micro-expressions becomes prevalent as it reveals the actual emotion of humans. However, handcrafted features employed for recognizing micro-expressions are designed for general applications and thus cannot well capture the subtle facial deformations of micro-expressions. To address this problem, we propose an end-to-end deep learning framework to suit the particular needs of micro-expression recognition (MER). In the deep model, re- current convolutional networks are utilized to learn the representation of subtle changes from image sequences. To guarantee the learning of deep model, we present a temporal jittering procedure to greatly enrich the training samples. Through performing the experiments on three spontaneous micro-expression datasets, i.e., SMIC, CASME, and CASME2, we verify the effectiveness of our proposed MER approach.
自发的面部微表情的自动识别因其反映了人类的真实情绪而变得普遍。然而,用于识别微表情的手工特征是为一般应用而设计的,因此不能很好地捕捉微表情的细微面部变形。为了解决这个问题,我们提出了一个端到端深度学习框架,以满足微表情识别(MER)的特殊需求。在深度模型中,利用卷积网络学习图像序列中细微变化的表示。为了保证深度模型的学习,我们提出了一种时间抖动方法来极大地丰富训练样本。通过在SMIC、CASME和CASME2三个自发微表达数据集上进行实验,验证了本文方法的有效性。
{"title":"Spontaneous Facial Micro-expression Recognition via Deep Convolutional Network","authors":"Zhaoqiang Xia, Xiaoyi Feng, Xiaopeng Hong, Guoying Zhao","doi":"10.1109/IPTA.2018.8608119","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608119","url":null,"abstract":"The automatic recognition of spontaneous facial micro-expressions becomes prevalent as it reveals the actual emotion of humans. However, handcrafted features employed for recognizing micro-expressions are designed for general applications and thus cannot well capture the subtle facial deformations of micro-expressions. To address this problem, we propose an end-to-end deep learning framework to suit the particular needs of micro-expression recognition (MER). In the deep model, re- current convolutional networks are utilized to learn the representation of subtle changes from image sequences. To guarantee the learning of deep model, we present a temporal jittering procedure to greatly enrich the training samples. Through performing the experiments on three spontaneous micro-expression datasets, i.e., SMIC, CASME, and CASME2, we verify the effectiveness of our proposed MER approach.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121424422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Detection and identification method of medical label barcode based on deep learning 基于深度学习的医疗标签条形码检测与识别方法
Hui Zhang, Guoliang Shi, Li Liu, M. Zhao, Zhicong Liang
The widespread use of barcode technology has led to the complexity of the application scenario. In the traditional barcode recognition method, there is no universal solution to the problems of uneven illumination, distortion, and sheltered. In this paper, the deep learning theory is used to solve the problem of barcode detection under the above situation. And on this basis, the problem of correcting linear distortion Data Matrix code is solved, and the key technology of barcode recognition under complex situation is broken through. After testing, the recognition speed reached 125ms, and the recognition accuracy reached about 93%. The system uses CCD camera to collect pictures, adopts the HALCON to build the processing algorithm, and uses Visual Studio platform to build the software, which realizes the Date Matrix code, Drug Electronic Supervision Code and Product bar code fast and accurate identification on pharmaceutical packaging. The developed system can also detect the rotation angle of Barcode and Data Matrix code, which is favorable for reading the barcode information. The whole process is real-time.
条码技术的广泛应用导致了应用场景的复杂性。在传统的条码识别方法中,对于光照不均匀、失真、遮挡等问题没有通用的解决方案。本文采用深度学习理论来解决上述情况下的条码检测问题。并在此基础上解决了修正线性失真数据矩阵码的问题,突破了复杂情况下条码识别的关键技术。经过测试,识别速度达到125ms,识别准确率达到93%左右。该系统采用CCD摄像头采集图片,采用HALCON搭建处理算法,使用Visual Studio平台搭建软件,实现了药品包装上的日期矩阵码、药品电子监管码和产品条形码的快速准确识别。所开发的系统还可以检测条码和数据矩阵码的旋转角度,有利于读取条码信息。整个过程是实时的。
{"title":"Detection and identification method of medical label barcode based on deep learning","authors":"Hui Zhang, Guoliang Shi, Li Liu, M. Zhao, Zhicong Liang","doi":"10.1109/IPTA.2018.8608144","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608144","url":null,"abstract":"The widespread use of barcode technology has led to the complexity of the application scenario. In the traditional barcode recognition method, there is no universal solution to the problems of uneven illumination, distortion, and sheltered. In this paper, the deep learning theory is used to solve the problem of barcode detection under the above situation. And on this basis, the problem of correcting linear distortion Data Matrix code is solved, and the key technology of barcode recognition under complex situation is broken through. After testing, the recognition speed reached 125ms, and the recognition accuracy reached about 93%. The system uses CCD camera to collect pictures, adopts the HALCON to build the processing algorithm, and uses Visual Studio platform to build the software, which realizes the Date Matrix code, Drug Electronic Supervision Code and Product bar code fast and accurate identification on pharmaceutical packaging. The developed system can also detect the rotation angle of Barcode and Data Matrix code, which is favorable for reading the barcode information. The whole process is real-time.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122353010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Joint Deep Learning and Clustering Algorithm for Liquid Particle Detection of Pharmaceutical Injection 药物注射剂液体颗粒检测的联合深度学习与聚类算法
M. Zhao, Hui Zhang, Li Liu, Zhicong Liang, Guang Deng
At present, the detection of pharmaceutical injection products is a quite important step in the pharmaceutical manufacturing, as it has the direct related to the quality of medical product quality. Aiming at the difficulty that liquid particle has a smaller pixel point in the high resolution image of detection of pharmaceutical liquid particle, hence consider combined with deep neural network and clustering algorithm for detection and localization of little particle, and a processing method combining single frame images with multi-frame images was proposed to identifying liquid particle. Firstly, the single-frame image is detected by using Faster-RCNN deep neural network, and it can obtain the detection result of the 8-frame sequence image. Then hierarchical clustering and K-means clustering algorithm are used for clustering to obtain the same target motion area. In this way, liquid particle can be more accurately identified and the accuracy of detection can be greatly improved. The experimental results show that the accuracy of detection and recognition of foreign substances in liquid medicine is improved by more than 10% on average.
目前,药物注射产品的检测是药品生产中相当重要的一步,直接关系到药品质量的好坏。针对药物液体颗粒在高分辨率图像检测中像素点较小的困难,考虑结合深度神经网络和聚类算法对小颗粒进行检测和定位,提出了一种单帧图像与多帧图像相结合的液体颗粒识别处理方法。首先,利用Faster-RCNN深度神经网络对单帧图像进行检测,得到8帧序列图像的检测结果。然后采用分层聚类和K-means聚类算法进行聚类,得到相同的目标运动区域。这样可以更准确地识别液体颗粒,大大提高检测的准确性。实验结果表明,该方法对药液中异物的检测识别准确率平均提高10%以上。
{"title":"Joint Deep Learning and Clustering Algorithm for Liquid Particle Detection of Pharmaceutical Injection","authors":"M. Zhao, Hui Zhang, Li Liu, Zhicong Liang, Guang Deng","doi":"10.1109/IPTA.2018.8608158","DOIUrl":"https://doi.org/10.1109/IPTA.2018.8608158","url":null,"abstract":"At present, the detection of pharmaceutical injection products is a quite important step in the pharmaceutical manufacturing, as it has the direct related to the quality of medical product quality. Aiming at the difficulty that liquid particle has a smaller pixel point in the high resolution image of detection of pharmaceutical liquid particle, hence consider combined with deep neural network and clustering algorithm for detection and localization of little particle, and a processing method combining single frame images with multi-frame images was proposed to identifying liquid particle. Firstly, the single-frame image is detected by using Faster-RCNN deep neural network, and it can obtain the detection result of the 8-frame sequence image. Then hierarchical clustering and K-means clustering algorithm are used for clustering to obtain the same target motion area. In this way, liquid particle can be more accurately identified and the accuracy of detection can be greatly improved. The experimental results show that the accuracy of detection and recognition of foreign substances in liquid medicine is improved by more than 10% on average.","PeriodicalId":272294,"journal":{"name":"2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133721696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2018 Eighth International Conference on Image Processing Theory, Tools and Applications (IPTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1