首页 > 最新文献

2020 International Conference on Machine Vision and Image Processing (MVIP)最新文献

英文 中文
PCB Defect Detection Using Denoising Convolutional Autoencoders 基于去噪卷积自编码器的PCB缺陷检测
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187485
Saeed Khalilian, Yeganeh Hallaj, Arian Balouchestani, Hossein Karshenas, A. Mohammadi
Printed Circuit boards (PCBs) are one of the most important stages in making electronic products. A small defect in PCBs can cause significant flaws in the final product. Hence, detecting all defects in PCBs and locating them is essential. In this paper, we propose an approach based on denoising convolutional autoencoders for detecting defective PCBs and to locate the defects. Denoising autoencoders take a corrupted image and try to recover the intact image. We trained our model with defective PCBs and forced it to repair the defective parts. Our model not only detects all kinds of defects and locates them, but it can also repair them as well. By subtracting the repaired output from the input, the defective parts are located. The experimental results indicate that our model detects the defective PCBs with high accuracy (97.5%) compare to state of the art works.
印制电路板(pcb)是制造电子产品最重要的阶段之一。pcb中的一个小缺陷可能会导致最终产品的重大缺陷。因此,检测pcb中的所有缺陷并定位它们是必不可少的。本文提出一种基于去噪卷积自编码器的pcb缺陷检测与定位方法。去噪自动编码器取损坏的图像并尝试恢复完整的图像。我们用有缺陷的多氯联苯训练我们的模型,并强迫它修复有缺陷的部分。我们的模型不仅可以检测和定位各种缺陷,而且还可以进行修复。通过从输入中减去修复后的输出,定位出有缺陷的部件。实验结果表明,与现有的检测方法相比,该模型的检测精度高达97.5%。
{"title":"PCB Defect Detection Using Denoising Convolutional Autoencoders","authors":"Saeed Khalilian, Yeganeh Hallaj, Arian Balouchestani, Hossein Karshenas, A. Mohammadi","doi":"10.1109/MVIP49855.2020.9187485","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187485","url":null,"abstract":"Printed Circuit boards (PCBs) are one of the most important stages in making electronic products. A small defect in PCBs can cause significant flaws in the final product. Hence, detecting all defects in PCBs and locating them is essential. In this paper, we propose an approach based on denoising convolutional autoencoders for detecting defective PCBs and to locate the defects. Denoising autoencoders take a corrupted image and try to recover the intact image. We trained our model with defective PCBs and forced it to repair the defective parts. Our model not only detects all kinds of defects and locates them, but it can also repair them as well. By subtracting the repaired output from the input, the defective parts are located. The experimental results indicate that our model detects the defective PCBs with high accuracy (97.5%) compare to state of the art works.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131565489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Source Camera Identification Using WLBP Descriptor 使用WLBP描述符的源相机识别
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187484
Nasme Zandi, F. Razzazi
In this paper we introduce a camera identification method using WLBP texture descriptor. This descriptor has previously been used for texture and face classifiers. In the proposed method, we proposed to use WLBP operator in camera classification application to identify the imaging camera. In our method, the two-dimensional histogram of Weber’s features and LBP for camera identification are investigated. For this purpose, experiments were conducted on Dresden database. The proposed method has reached the accuracy of 99.52% on nine digital cameras of different models. In compressed JPEG images with the compression quality factor of 70% the method reached the accuracy of 89.04%. The results indicate that the proposed method has a high degree of accuracy in comparison to other proposed method and exhibits relatively good robustness to compression.
本文介绍了一种基于WLBP纹理描述符的摄像机识别方法。这个描述符以前被用于纹理和人脸分类器。在该方法中,我们提出将WLBP算子应用于相机分类中,对成像相机进行识别。在我们的方法中,研究了韦伯特征的二维直方图和用于相机识别的LBP。为此,在德累斯顿数据库上进行了实验。该方法在9台不同型号的数码相机上,准确率达到99.52%。在压缩质量因子为70%的JPEG图像中,该方法的准确率达到89.04%。结果表明,与其他方法相比,该方法具有较高的精度,并且具有较好的压缩鲁棒性。
{"title":"Source Camera Identification Using WLBP Descriptor","authors":"Nasme Zandi, F. Razzazi","doi":"10.1109/MVIP49855.2020.9187484","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187484","url":null,"abstract":"In this paper we introduce a camera identification method using WLBP texture descriptor. This descriptor has previously been used for texture and face classifiers. In the proposed method, we proposed to use WLBP operator in camera classification application to identify the imaging camera. In our method, the two-dimensional histogram of Weber’s features and LBP for camera identification are investigated. For this purpose, experiments were conducted on Dresden database. The proposed method has reached the accuracy of 99.52% on nine digital cameras of different models. In compressed JPEG images with the compression quality factor of 70% the method reached the accuracy of 89.04%. The results indicate that the proposed method has a high degree of accuracy in comparison to other proposed method and exhibits relatively good robustness to compression.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123884895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Convolutional Neural Network for Building Extraction from High-Resolution Remote Sensing Images 基于卷积神经网络的高分辨率遥感影像建筑物提取
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187483
H. Hosseinpoor, F. Samadzadegan
Buildings are one of the most important components of the city, and their extraction from high-resolution remote sensing images is used in a wide range of applications such as urban mapping. Due to the complex structure of highresolution remote sensing images, automatic extraction of buildings has been a challenge in recent years. In this regard, fully convolutional neural networks (FCNs) have shown successful performance in this task. In this research, a method is proposed to improve the famous UNet network. In classical UNet model high-level rich semantic features are fused with low-level high-resolution features with skip connection for pixel-based segmentation of images. However, the fusion of encoder features with features in corresponding decoder part causes ambiguity in segmentation results because low-level features produce high noise in high-level semantic features. We introduced the embedding feature fusion (EFF) block for enhancing the fusion of low-level with high-level features. For performance evaluation, a publicly available data provided with United States Geological Survey (USGS) high-resolution orthoimagery with the spatial Resolution ranges from 0.15m to 0.3m was used in comparison with several state-of-the-art semantic segmentation model. Experimental results have showed that the proposed architecture improves in extracting complex buildings from high resolution remote sensing images.
建筑物是城市最重要的组成部分之一,从高分辨率遥感图像中提取建筑物在城市测绘等领域有着广泛的应用。由于高分辨率遥感图像结构复杂,建筑物的自动提取是近年来面临的一个挑战。在这方面,全卷积神经网络(fcn)在这项任务中表现出了成功的性能。在本研究中,提出了一种改进著名UNet网络的方法。在经典UNet模型中,采用跳跃式连接将高水平的丰富语义特征与低水平的高分辨率特征融合,实现基于像素的图像分割。然而,编码器特征与相应解码器部分特征的融合会导致分割结果的模糊性,因为低级特征在高级语义特征中产生高噪声。引入了嵌入特征融合(EFF)块,增强了低级特征与高级特征的融合。为了进行性能评估,使用了美国地质调查局(USGS)提供的公开数据,空间分辨率为0.15m至0.3m,并与几种最先进的语义分割模型进行了比较。实验结果表明,该结构在高分辨率遥感图像中提取复杂建筑物方面具有较好的效果。
{"title":"Convolutional Neural Network for Building Extraction from High-Resolution Remote Sensing Images","authors":"H. Hosseinpoor, F. Samadzadegan","doi":"10.1109/MVIP49855.2020.9187483","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187483","url":null,"abstract":"Buildings are one of the most important components of the city, and their extraction from high-resolution remote sensing images is used in a wide range of applications such as urban mapping. Due to the complex structure of highresolution remote sensing images, automatic extraction of buildings has been a challenge in recent years. In this regard, fully convolutional neural networks (FCNs) have shown successful performance in this task. In this research, a method is proposed to improve the famous UNet network. In classical UNet model high-level rich semantic features are fused with low-level high-resolution features with skip connection for pixel-based segmentation of images. However, the fusion of encoder features with features in corresponding decoder part causes ambiguity in segmentation results because low-level features produce high noise in high-level semantic features. We introduced the embedding feature fusion (EFF) block for enhancing the fusion of low-level with high-level features. For performance evaluation, a publicly available data provided with United States Geological Survey (USGS) high-resolution orthoimagery with the spatial Resolution ranges from 0.15m to 0.3m was used in comparison with several state-of-the-art semantic segmentation model. Experimental results have showed that the proposed architecture improves in extracting complex buildings from high resolution remote sensing images.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126576745","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Estimating intrinsic image from successive images by solving underdetermined and overdetermined systems of the dichromatic model 通过求解二色模型的欠定和过定系统,从连续图像中估计固有图像
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187487
K. Ansari, Alexandre Krebs, Y. Benezeth, F. Marzani
Estimating an intrinsic image from a sequence of successive images taken from an object at different angles of illumination can be used in various applications such as objects recognition, color classification, and the like; because, in so doing, it can provide more visual information. Meanwhile, according to the well-known dichromatic model, each image can be considered a linear combination of three components, including intrinsic image, shading factor, and specularity. In this study, at first, two simple independent constrained and parallelized quadratic programming steps were used for computing values of the shading factor and the specularity of each successive of images. In the algorithm mentioned above, only the mean and standard deviation of three channels for each pixel are required to solve the underdetermined problem of the dichromatic model equations. Then, the singular value decomposition method was used to estimate a unique intrinsic image through the values of the shading factor and the specularity of each of the images that constitute an overdetermined problem. The results of the successive reconstructed images using the estimated unique intrinsic image showed an increase in the visual assessment quality and color gamut of the final images.
从从物体以不同照明角度拍摄的连续图像序列中估计固有图像可用于诸如物体识别、颜色分类等各种应用;因为,这样做,它可以提供更多的视觉信息。同时,根据众所周知的二色模型,每个图像可以被认为是三个组成部分的线性组合,包括固有图像、阴影因子和镜面。在本研究中,首先使用两个简单的独立约束并行二次规划步骤来计算每个连续图像的阴影因子和反射率值。在上述算法中,只需要每个像素的三个通道的均值和标准差就可以解决二色模型方程的欠定问题。然后,使用奇异值分解方法,通过构成过定问题的每个图像的阴影因子和反射率的值来估计唯一的内在图像。使用估计的唯一固有图像进行连续重建图像的结果表明,最终图像的视觉评估质量和色域都有所提高。
{"title":"Estimating intrinsic image from successive images by solving underdetermined and overdetermined systems of the dichromatic model","authors":"K. Ansari, Alexandre Krebs, Y. Benezeth, F. Marzani","doi":"10.1109/MVIP49855.2020.9187487","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187487","url":null,"abstract":"Estimating an intrinsic image from a sequence of successive images taken from an object at different angles of illumination can be used in various applications such as objects recognition, color classification, and the like; because, in so doing, it can provide more visual information. Meanwhile, according to the well-known dichromatic model, each image can be considered a linear combination of three components, including intrinsic image, shading factor, and specularity. In this study, at first, two simple independent constrained and parallelized quadratic programming steps were used for computing values of the shading factor and the specularity of each successive of images. In the algorithm mentioned above, only the mean and standard deviation of three channels for each pixel are required to solve the underdetermined problem of the dichromatic model equations. Then, the singular value decomposition method was used to estimate a unique intrinsic image through the values of the shading factor and the specularity of each of the images that constitute an overdetermined problem. The results of the successive reconstructed images using the estimated unique intrinsic image showed an increase in the visual assessment quality and color gamut of the final images.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132405570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Convolutional Neural Network based on Local Binary Patterns of Gabor Features for Classification of Hyperspectral Images 基于Gabor特征局部二值模式的深度卷积神经网络用于高光谱图像分类
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187486
Obeid Sharifi, M. Mokhtarzade, B. Asghari Beirami
To date, various spatial-spectral methods are proposed for accurate classification of hyperspectral images (HSI). Gabor spatial features are the most prominent ones that can extract shallow features such as edges and structures. In recent years, convolutional neural networks (CNN) have been promising in the classification of HSI. Although in literature Gabor features are used as the input of deep models, it seems that the performance of CNN can be improved by two-stage textural features based on local binary patterns of Gabor features. In this paper, input features of CNN are obtained based on local binary patterns of Gabor features which are more discriminative than both Gabor features and local binary patterns features. The experiments performed on the famous Indian Pines HIS, proved the superiority of the proposed method over some other deep learning-based methods.
目前,针对高光谱图像的精确分类,提出了多种空间光谱方法。Gabor空间特征是提取边缘、结构等浅层特征最为突出的空间特征。近年来,卷积神经网络(convolutional neural networks, CNN)在HSI分类中有很大的应用前景。虽然在文献中使用Gabor特征作为深度模型的输入,但似乎可以通过基于Gabor特征的局部二值模式的两阶段纹理特征来提高CNN的性能。本文基于Gabor特征的局部二值模式获得CNN的输入特征,该特征比Gabor特征和局部二值模式特征都更具判别性。在著名的Indian Pines HIS上进行的实验证明了该方法优于其他一些基于深度学习的方法。
{"title":"A Deep Convolutional Neural Network based on Local Binary Patterns of Gabor Features for Classification of Hyperspectral Images","authors":"Obeid Sharifi, M. Mokhtarzade, B. Asghari Beirami","doi":"10.1109/MVIP49855.2020.9187486","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187486","url":null,"abstract":"To date, various spatial-spectral methods are proposed for accurate classification of hyperspectral images (HSI). Gabor spatial features are the most prominent ones that can extract shallow features such as edges and structures. In recent years, convolutional neural networks (CNN) have been promising in the classification of HSI. Although in literature Gabor features are used as the input of deep models, it seems that the performance of CNN can be improved by two-stage textural features based on local binary patterns of Gabor features. In this paper, input features of CNN are obtained based on local binary patterns of Gabor features which are more discriminative than both Gabor features and local binary patterns features. The experiments performed on the famous Indian Pines HIS, proved the superiority of the proposed method over some other deep learning-based methods.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133863638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Offline Handwritten Signature Verification and Recognition Based on Deep Transfer Learning 基于深度迁移学习的离线手写签名验证与识别
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187481
A. Foroozandeh, Ataollah Askari Hemmat, Hossein Rabbani
Recently, deep convolutional neural networks have been successfully applied in different fields of computer vision and pattern recognition. Offline handwritten signature is one of the most important biometrics applied in banking systems, administrative and financial applications, which is a challenging task and still hard. The aim of this study is to review of the presented signature verification/recognition methods based on the convolutional neural networks and also evaluate the performance of some prominent available deep convolutional neural networks in offline handwritten signature verification/recognition as feature extractor using transfer learning. This is done using four pretrained models as the most used general models in computer vision tasks including VGG16, VGG19, ResNet50, and InceptionV3 and also two pre-trained models especially presented for signature processing tasks including SigNet and SigNet- F. Experiments have been conducted using two benchmark signature datasets: GPDS Synthetic signature dataset and MCYT- 75 as Latin signature datasets, and two Persian datasets: UTSig and FUM-PHSD. Obtained experimental results, in comparison with literature, verify the effectiveness of the models: VGG16 and SigNet for signature verification and the superiority of VGG16 in signature recognition task.
近年来,深度卷积神经网络已成功地应用于计算机视觉和模式识别的各个领域。离线手写签名是银行系统、行政和金融应用中最重要的生物识别技术之一,这是一项具有挑战性且仍然困难的任务。本研究的目的是回顾现有的基于卷积神经网络的签名验证/识别方法,并评估一些著名的深度卷积神经网络在离线手写签名验证/识别中作为特征提取器使用迁移学习的性能。实验使用了四个预训练模型作为计算机视觉任务中最常用的通用模型,包括VGG16、VGG19、ResNet50和InceptionV3,以及两个专门用于签名处理任务的预训练模型,包括SigNet和SigNet- f。实验使用了两个基准签名数据集:GPDS合成签名数据集和MCYT- 75作为拉丁签名数据集,以及两个波斯语数据集:UTSig和FUM-PHSD。得到的实验结果,与文献对比,验证了VGG16和SigNet模型用于签名验证的有效性,以及VGG16在签名识别任务中的优越性。
{"title":"Offline Handwritten Signature Verification and Recognition Based on Deep Transfer Learning","authors":"A. Foroozandeh, Ataollah Askari Hemmat, Hossein Rabbani","doi":"10.1109/MVIP49855.2020.9187481","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187481","url":null,"abstract":"Recently, deep convolutional neural networks have been successfully applied in different fields of computer vision and pattern recognition. Offline handwritten signature is one of the most important biometrics applied in banking systems, administrative and financial applications, which is a challenging task and still hard. The aim of this study is to review of the presented signature verification/recognition methods based on the convolutional neural networks and also evaluate the performance of some prominent available deep convolutional neural networks in offline handwritten signature verification/recognition as feature extractor using transfer learning. This is done using four pretrained models as the most used general models in computer vision tasks including VGG16, VGG19, ResNet50, and InceptionV3 and also two pre-trained models especially presented for signature processing tasks including SigNet and SigNet- F. Experiments have been conducted using two benchmark signature datasets: GPDS Synthetic signature dataset and MCYT- 75 as Latin signature datasets, and two Persian datasets: UTSig and FUM-PHSD. Obtained experimental results, in comparison with literature, verify the effectiveness of the models: VGG16 and SigNet for signature verification and the superiority of VGG16 in signature recognition task.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116102528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A High-Accuracy, Cost-Effective People Counting Solution Based on Visual Depth Data 一种基于视觉深度数据的高精度、高性价比的人员计数解决方案
Pub Date : 2020-02-18 DOI: 10.1109/MVIP49855.2020.9187482
Seyed Ali Hosseini Shamoushaki, Mohammad Mostafa Talebi, Amineh Mazandarani, S. Hosseini
Real-time people counting has become a critical task due to its applications in a wide range of areas such as security, safety, statistics and commerce, implying that the demand for systems that offer such a capability has risen. Consequently, it is important to make it possible for the public to afford a precise, robust people counting system. Therefore, we aim to propose an efficient solution that requires low-cost hardware. Hopefully the people counting product derived from this solution will have a reasonable purchase price when put up for sale. Following the minimal hardware requirement, our system only relies on a depth camera plus a cheap embedded processor. A detection/tracking module forms the core of the underlying theoretical approach whose main functionality is to detect and count any entry/exit occurrences through a generic entrance. Our testing and validation experiments reveal that the proposed system yields a highly satisfactory accuracy rate and can compete closely with similar technologies currently available on the market.
实时人数统计已经成为一项关键任务,因为它在安全、安全、统计和商业等广泛领域的应用,这意味着对提供这种能力的系统的需求已经上升。因此,重要的是要使公众能够负担得起一个精确、健全的人口统计系统。因此,我们的目标是提出一种需要低成本硬件的高效解决方案。希望从这个解决方案衍生的产品计数的人将有一个合理的购买价格时,提出出售。遵循最小的硬件要求,我们的系统只依赖于一个深度摄像头和一个便宜的嵌入式处理器。检测/跟踪模块构成了基础理论方法的核心,其主要功能是检测和计数通过通用入口的任何进入/退出事件。我们的测试和验证实验表明,所提出的系统产生了非常令人满意的准确率,并且可以与市场上现有的类似技术密切竞争。
{"title":"A High-Accuracy, Cost-Effective People Counting Solution Based on Visual Depth Data","authors":"Seyed Ali Hosseini Shamoushaki, Mohammad Mostafa Talebi, Amineh Mazandarani, S. Hosseini","doi":"10.1109/MVIP49855.2020.9187482","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9187482","url":null,"abstract":"Real-time people counting has become a critical task due to its applications in a wide range of areas such as security, safety, statistics and commerce, implying that the demand for systems that offer such a capability has risen. Consequently, it is important to make it possible for the public to afford a precise, robust people counting system. Therefore, we aim to propose an efficient solution that requires low-cost hardware. Hopefully the people counting product derived from this solution will have a reasonable purchase price when put up for sale. Following the minimal hardware requirement, our system only relies on a depth camera plus a cheap embedded processor. A detection/tracking module forms the core of the underlying theoretical approach whose main functionality is to detect and count any entry/exit occurrences through a generic entrance. Our testing and validation experiments reveal that the proposed system yields a highly satisfactory accuracy rate and can compete closely with similar technologies currently available on the market.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116808524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A New Composite Multimodality Image Fusion Method Based on Shearlet Transform and Retina Inspired Model 基于Shearlet变换和视网膜启发模型的复合多模态图像融合新方法
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116919
Mohammadmahdi Sayadi, H. Ghassemian, Reza Naimi, M. Imani
Medical imaging is a very important element in disease diagnosis. MRI image has structural information, while PET image has functional information. However, there is no medical imagery device that has both structural and functional information simultaneously. Thus, the image fusion technique is used. This work concentrates on PET and MRI fusion. It is based on the combination of retina-inspired model and Non-Subsampled shearlet transform. In the first step, the high-frequency component is obtained by applying the shearlet transform to the MRI image, which produces sub-images in several scales and directions, and by adding up these images together a single edge image is reconstructed. In the second step, the PET image is transferred from RGB color space into IHS color space. Then the low-frequency component is produced by applying a Gaussian low pass filter to the luminance channel of the PET image. By adding up low frequency component and high-frequency component together and transferring the result from IHS color space to RGB color space the fused image is obtained.
医学影像是疾病诊断的重要组成部分。MRI图像具有结构信息,PET图像具有功能信息。然而,目前还没有一种医学成像设备同时具有结构和功能信息。因此,采用了图像融合技术。这项工作集中在PET和MRI融合。它是基于视网膜启发模型和非下采样shearlet变换的结合。第一步,对MRI图像进行剪切波变换,得到高频分量,产生多个尺度和方向的子图像,将这些图像叠加在一起,重建出单个边缘图像。第二步,将PET图像从RGB色彩空间转换到IHS色彩空间。然后通过对PET图像的亮度通道应用高斯低通滤波器产生低频分量。通过将低频分量和高频分量相加,并将结果从IHS色彩空间传递到RGB色彩空间,得到融合图像。
{"title":"A New Composite Multimodality Image Fusion Method Based on Shearlet Transform and Retina Inspired Model","authors":"Mohammadmahdi Sayadi, H. Ghassemian, Reza Naimi, M. Imani","doi":"10.1109/MVIP49855.2020.9116919","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116919","url":null,"abstract":"Medical imaging is a very important element in disease diagnosis. MRI image has structural information, while PET image has functional information. However, there is no medical imagery device that has both structural and functional information simultaneously. Thus, the image fusion technique is used. This work concentrates on PET and MRI fusion. It is based on the combination of retina-inspired model and Non-Subsampled shearlet transform. In the first step, the high-frequency component is obtained by applying the shearlet transform to the MRI image, which produces sub-images in several scales and directions, and by adding up these images together a single edge image is reconstructed. In the second step, the PET image is transferred from RGB color space into IHS color space. Then the low-frequency component is produced by applying a Gaussian low pass filter to the luminance channel of the PET image. By adding up low frequency component and high-frequency component together and transferring the result from IHS color space to RGB color space the fused image is obtained.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115209470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fast Prediction of Cortical Dementia Based on Original Brain MRI images Using Convolutional Neural Network 基于原始脑MRI图像的卷积神经网络快速预测皮质性痴呆
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116921
M. Amini, H. Sajedi, Tayeb Mahmoodi, S. Mirzaei
Fast and automatic identification of different types of Cortical Dementia, specially Alzheimer’s disease, based on Brain MRI images, is a crucial technology which can help physicians in early and effective treatment. Although preprocessing of MRI images could improve the accuracy of machine learning techniques for classification of the normal and abnormal cases, this could slow down the process of automatic identification and tarnish the applicability of these methods in clinics and laboratories. In this paper we examine classification of a small sample of the original brain MRI images, using a 2D Convolutional Neural Network (CNN). The data consists of 172 healthy individuals as the control group (HC) and only 89 patients with different grades of Dementia (DP) which was collected in National Brain Mapping Center of Iran. The model could achieve an accuracy of 97.47% on the test set and 93.88% based on a 5-fold cross-validation.
基于脑MRI图像快速自动识别不同类型的皮质性痴呆,特别是阿尔茨海默病,是帮助医生早期有效治疗的关键技术。尽管对MRI图像进行预处理可以提高机器学习技术对正常和异常病例进行分类的准确性,但这可能会减慢自动识别的过程,并损害这些方法在诊所和实验室中的适用性。在本文中,我们使用二维卷积神经网络(CNN)研究了原始大脑MRI图像的小样本分类。数据包括172名健康个体作为对照组(HC)和89名不同程度的痴呆(DP)患者,这些患者来自伊朗国家脑制图中心。该模型在测试集上的准确率为97.47%,在5倍交叉验证的基础上,准确率为93.88%。
{"title":"Fast Prediction of Cortical Dementia Based on Original Brain MRI images Using Convolutional Neural Network","authors":"M. Amini, H. Sajedi, Tayeb Mahmoodi, S. Mirzaei","doi":"10.1109/MVIP49855.2020.9116921","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116921","url":null,"abstract":"Fast and automatic identification of different types of Cortical Dementia, specially Alzheimer’s disease, based on Brain MRI images, is a crucial technology which can help physicians in early and effective treatment. Although preprocessing of MRI images could improve the accuracy of machine learning techniques for classification of the normal and abnormal cases, this could slow down the process of automatic identification and tarnish the applicability of these methods in clinics and laboratories. In this paper we examine classification of a small sample of the original brain MRI images, using a 2D Convolutional Neural Network (CNN). The data consists of 172 healthy individuals as the control group (HC) and only 89 patients with different grades of Dementia (DP) which was collected in National Brain Mapping Center of Iran. The model could achieve an accuracy of 97.47% on the test set and 93.88% based on a 5-fold cross-validation.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121230743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Region Proposal Generation: A Hierarchical Merging Similarity-Based Algorithm 区域提议生成:一种基于层次合并相似度的算法
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116912
M. Taghizadeh, A. Chalechale
This paper presents a hierarchical algorithm using region merging with the aim of achieving a powerful pool of regions for solving computer vision problems. An image is first represented by a graph where each node in the graph is a superpixel. A variety of features is extracted of each region, which is next merged to neighbor regions according to the new algorithm. The proposed algorithm combines adjacent regions based on a similarity metric and a threshold parameter. By applying different amounts for the threshold, a wide range of regions is acquired. The algorithm successfully provides accurate regions while can be represented through the bounding box and segmented candidates. To extensively evaluate, the effectiveness of features and the combination of them are analyzed on MSRC and VOC2012 Segmentation dataset. The achieved results are shown a great improvement at overlapping in comparison to segmentation algorithms. Also, it outperforms previous region proposal algorithms, especially it leads to a relatively great recall at higher overlaps (≥ 0.6).
本文提出了一种利用区域合并的分层算法,目的是获得一个强大的区域池来解决计算机视觉问题。图像首先由图表示,图中的每个节点都是一个超像素。每个区域提取各种特征,然后根据新算法将其合并到相邻区域。该算法基于相似度度量和阈值参数组合相邻区域。通过应用不同的阈值,可以获得大范围的区域。该算法成功地提供了精确的区域,同时可以通过边界框和分割的候选区域来表示。为了广泛评估,在MSRC和VOC2012分割数据集上分析了特征及其组合的有效性。与分割算法相比,所获得的结果在重叠方面有很大的改进。此外,它优于以前的区域建议算法,特别是在较高的重叠(≥0.6)下具有相对较高的召回率。
{"title":"Region Proposal Generation: A Hierarchical Merging Similarity-Based Algorithm","authors":"M. Taghizadeh, A. Chalechale","doi":"10.1109/MVIP49855.2020.9116912","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116912","url":null,"abstract":"This paper presents a hierarchical algorithm using region merging with the aim of achieving a powerful pool of regions for solving computer vision problems. An image is first represented by a graph where each node in the graph is a superpixel. A variety of features is extracted of each region, which is next merged to neighbor regions according to the new algorithm. The proposed algorithm combines adjacent regions based on a similarity metric and a threshold parameter. By applying different amounts for the threshold, a wide range of regions is acquired. The algorithm successfully provides accurate regions while can be represented through the bounding box and segmented candidates. To extensively evaluate, the effectiveness of features and the combination of them are analyzed on MSRC and VOC2012 Segmentation dataset. The achieved results are shown a great improvement at overlapping in comparison to segmentation algorithms. Also, it outperforms previous region proposal algorithms, especially it leads to a relatively great recall at higher overlaps (≥ 0.6).","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127909163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2020 International Conference on Machine Vision and Image Processing (MVIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1