首页 > 最新文献

2020 International Conference on Machine Vision and Image Processing (MVIP)最新文献

英文 中文
Pix2Pix-based Stain-to-Stain Translation: A Solution for Robust Stain Normalization in Histopathology Images Analysis 基于pix2pixel的染色到染色转换:组织病理学图像分析中鲁棒染色归一化的解决方案
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116895
Pegah Salehi, A. Chalechale
The diagnosis of cancer is mainly performed by visual analysis of the pathologists, through examining the morphology of the tissue slices and the spatial arrangement of the cells. If the microscopic image of a specimen is not stained, it will look colorless and textured. Therefore, chemical staining is required to create contrast and help identify specific tissue components. During tissue preparation due to differences in chemicals, scanners, cutting thicknesses, and laboratory protocols, similar tissues are usually varied significantly in appearance. This diversity in staining, in addition to Interpretive disparity among pathologists more is one of the main challenges in designing robust and flexible systems for automated analysis. To address the staining color variations, several methods for normalizing stain have been proposed. In our proposed method, a Stain-to-Stain Translation (STST) approach is used to stain normalization for Hematoxylin and Eosin (H&E) stained histopathology images, which learns not only the specific color distribution but also the preserves corresponding histopathological pattern. We perform the process of translation based on the "pix2pix" framework, which uses the conditional generator adversarial networks (cGANs). Our approach showed excellent results, both mathematically and experimentally against the state of the art methods. We have made the source code publicly available 1.
癌症的诊断主要是通过病理学家的视觉分析,通过检查组织切片的形态和细胞的空间排列。如果标本的显微图像没有染色,它看起来是无色和有纹理的。因此,需要化学染色来形成对比并帮助识别特定的组织成分。在组织制备过程中,由于化学物质、扫描仪、切割厚度和实验室规程的差异,相似的组织通常在外观上有显著差异。除了病理学家之间的解释差异之外,这种染色的多样性更是设计健壮和灵活的自动化分析系统的主要挑战之一。为了解决染色颜色的变化,提出了几种校正染色的方法。在我们提出的方法中,使用染色到染色翻译(STST)方法对苏木精和伊红(H&E)染色的组织病理学图像进行染色归一化,该方法不仅学习了特定的颜色分布,而且保留了相应的组织病理学模式。我们基于“pix2pix”框架执行翻译过程,该框架使用条件生成器对抗网络(cgan)。我们的方法在数学上和实验上都显示了与最先进的方法相比优异的结果。我们已经公开了源代码1。
{"title":"Pix2Pix-based Stain-to-Stain Translation: A Solution for Robust Stain Normalization in Histopathology Images Analysis","authors":"Pegah Salehi, A. Chalechale","doi":"10.1109/MVIP49855.2020.9116895","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116895","url":null,"abstract":"The diagnosis of cancer is mainly performed by visual analysis of the pathologists, through examining the morphology of the tissue slices and the spatial arrangement of the cells. If the microscopic image of a specimen is not stained, it will look colorless and textured. Therefore, chemical staining is required to create contrast and help identify specific tissue components. During tissue preparation due to differences in chemicals, scanners, cutting thicknesses, and laboratory protocols, similar tissues are usually varied significantly in appearance. This diversity in staining, in addition to Interpretive disparity among pathologists more is one of the main challenges in designing robust and flexible systems for automated analysis. To address the staining color variations, several methods for normalizing stain have been proposed. In our proposed method, a Stain-to-Stain Translation (STST) approach is used to stain normalization for Hematoxylin and Eosin (H&E) stained histopathology images, which learns not only the specific color distribution but also the preserves corresponding histopathological pattern. We perform the process of translation based on the \"pix2pix\" framework, which uses the conditional generator adversarial networks (cGANs). Our approach showed excellent results, both mathematically and experimentally against the state of the art methods. We have made the source code publicly available 1.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122250505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 56
Brain MR Image Classification for ADHD Diagnosis Using Deep Neural Networks 脑MR图像分类在ADHD诊断中的应用
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116877
Sahar Abdolmaleki, M. S. Abadeh
Attention-deficit/hyperactivity disorder (ADHD) is one of the most prevalent neurodevelopmental disorder in childhood and adolescence. ADHD diagnosis currently includes psychological tests and depends on ratings of behavioral symptoms, which can be unreliable. Thus, an objective diagnostic tool based on non-invasive imaging can improve the understanding and diagnosis of ADHD. The purpose of this study is classifying brain images by using Artificial Intelligence methods such as clinical decision support system for the diagnosis of ADHD. For this purpose and according to a medical imaging classification system, firstly, image pre-processing is done. Then, a deep multi-modal 3D CNN is trained on GM from structural and fALFF from functional MRI using ADHD-200 training dataset. Finally, with the intention of classifying the extracted features, early and late fusion schemes are employed, and the output scores are classified with the SVM, KNN and LDA algorithms. The evaluation of the proposed approach on the ADHD-200 testing dataset revealed that the presence of personal characteristics alone increased the classification accuracy by 3.79%. In addition, using a combination of early, late fusion and personal characteristics together improved the accuracy of the classification by 5.84%. Among the three classifiers LDA showed better results and achieved a classification accuracy of 74.93%. The comparison of results showed that the combination of early and late fusion as well as considering personal characteristics has a significant effect on enhancing classification accuracy. As a result of this, the reliability of this medical decision support system is increased.
注意缺陷/多动障碍(ADHD)是儿童和青少年最常见的神经发育障碍之一。目前,多动症的诊断包括心理测试,并取决于行为症状的评分,这可能是不可靠的。因此,一种基于非侵入性影像学的客观诊断工具可以提高对ADHD的认识和诊断。本研究的目的是利用临床决策支持系统等人工智能方法对ADHD的脑图像进行分类。为此,根据医学影像分类系统,首先对图像进行预处理。然后,使用ADHD-200训练数据集对来自结构的GM和来自功能MRI的fALFF进行深度多模态3D CNN训练。最后,为了对提取的特征进行分类,采用了早期和晚期融合方案,并使用SVM、KNN和LDA算法对输出分数进行分类。对ADHD-200测试数据集的评估表明,单独存在个人特征将分类准确率提高了3.79%。此外,结合早期、晚期融合和个人特征,将分类准确率提高了5.84%。在这三种分类器中,LDA表现出较好的分类效果,分类准确率达到74.93%。结果表明,结合早期和晚期融合以及考虑个人特征对提高分类精度有显著效果。因此,该医疗决策支持系统的可靠性得到了提高。
{"title":"Brain MR Image Classification for ADHD Diagnosis Using Deep Neural Networks","authors":"Sahar Abdolmaleki, M. S. Abadeh","doi":"10.1109/MVIP49855.2020.9116877","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116877","url":null,"abstract":"Attention-deficit/hyperactivity disorder (ADHD) is one of the most prevalent neurodevelopmental disorder in childhood and adolescence. ADHD diagnosis currently includes psychological tests and depends on ratings of behavioral symptoms, which can be unreliable. Thus, an objective diagnostic tool based on non-invasive imaging can improve the understanding and diagnosis of ADHD. The purpose of this study is classifying brain images by using Artificial Intelligence methods such as clinical decision support system for the diagnosis of ADHD. For this purpose and according to a medical imaging classification system, firstly, image pre-processing is done. Then, a deep multi-modal 3D CNN is trained on GM from structural and fALFF from functional MRI using ADHD-200 training dataset. Finally, with the intention of classifying the extracted features, early and late fusion schemes are employed, and the output scores are classified with the SVM, KNN and LDA algorithms. The evaluation of the proposed approach on the ADHD-200 testing dataset revealed that the presence of personal characteristics alone increased the classification accuracy by 3.79%. In addition, using a combination of early, late fusion and personal characteristics together improved the accuracy of the classification by 5.84%. Among the three classifiers LDA showed better results and achieved a classification accuracy of 74.93%. The comparison of results showed that the combination of early and late fusion as well as considering personal characteristics has a significant effect on enhancing classification accuracy. As a result of this, the reliability of this medical decision support system is increased.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"150 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116369876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
High-Resolution Document Image Reconstruction from Video 从视频重建高分辨率文档图像
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116907
Hossein Motamednia, Mohammad Minouei, Pooryaa Cheraaqee, M. Soheili
Today, smartphones with high-quality built-in cameras are very common. People prefer to take pictures from documents with smartphones instead of scanning them with a scanner. Due to the limitation of scanners input size, it is difficult to scan everything with them. Resolution and quality of smartphone cameras are not enough to take a picture from large documents like posters. In this paper, we proposed a pipeline to make a high-resolution image of a document from its captured video. We suppose that during the record of the video, the camera was moved slowly all over the surface of the document from a close distance. In the proposed method we find the location of each frame in the document and we use a sharpness criterion to select the highest possible quality for each region of the document among all available frames. We evaluated our method on the SmartDoc Video dataset and reported the promising results.
如今,内置高质量摄像头的智能手机非常普遍。人们更喜欢用智能手机拍照,而不是用扫描仪扫描文件。由于扫描仪输入尺寸的限制,很难扫描所有的东西。智能手机相机的分辨率和质量不足以拍摄像海报这样的大型文件。在本文中,我们提出了一种从捕获的视频中生成文档高分辨率图像的管道。我们假设在录制视频的过程中,摄像机从近距离缓慢地在文件表面移动。在提出的方法中,我们找到文档中每个帧的位置,并使用清晰度标准在所有可用帧中选择文档的每个区域的最高质量。我们在SmartDoc视频数据集上评估了我们的方法,并报告了有希望的结果。
{"title":"High-Resolution Document Image Reconstruction from Video","authors":"Hossein Motamednia, Mohammad Minouei, Pooryaa Cheraaqee, M. Soheili","doi":"10.1109/MVIP49855.2020.9116907","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116907","url":null,"abstract":"Today, smartphones with high-quality built-in cameras are very common. People prefer to take pictures from documents with smartphones instead of scanning them with a scanner. Due to the limitation of scanners input size, it is difficult to scan everything with them. Resolution and quality of smartphone cameras are not enough to take a picture from large documents like posters. In this paper, we proposed a pipeline to make a high-resolution image of a document from its captured video. We suppose that during the record of the video, the camera was moved slowly all over the surface of the document from a close distance. In the proposed method we find the location of each frame in the document and we use a sharpness criterion to select the highest possible quality for each region of the document among all available frames. We evaluated our method on the SmartDoc Video dataset and reported the promising results.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117095680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fully Convolutional Networks for Fluid Segmentation in Retina Images 视网膜图像流体分割的全卷积网络
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116914
Behnam Azimi, A. Rashno, S. Fadaei
Retinal diseases can be manifested in optical coherence tomography (OCT) images since many signs of retina abnormalities are visible in OCT. Fluid regions can reveal the signs of age-related macular degeneration (AMD) and diabetic macular edema (DME) diseases and automatic segmentation of these regions can help ophthalmologists for diagnosis and treatment. This work presents a fully-automated method based on graph shortest path layer segmentation and fully convolutional networks (FCNs) for fluid segmentation. The proposed method has been evaluated on a dataset containing 600 OCT scans of 24 subjects. Results showed that the proposed FCN model outperforms 3 existing fluid segmentation methods by the improvement of 4.44% and 6.28% with respect to dice cofficients and sensitivity, respectively.
视网膜疾病可以在光学相干断层扫描(OCT)图像中表现出来,因为在OCT中可以看到许多视网膜异常的迹象。液体区域可以显示年龄相关性黄斑变性(AMD)和糖尿病性黄斑水肿(DME)疾病的迹象,这些区域的自动分割可以帮助眼科医生进行诊断和治疗。本文提出了一种基于图最短路径层分割和全卷积网络(fcv)的全自动流体分割方法。所提出的方法已在包含24名受试者的600次OCT扫描的数据集上进行了评估。结果表明,所提出的FCN模型在骰子系数和灵敏度方面分别比现有的3种流体分割方法提高了4.44%和6.28%。
{"title":"Fully Convolutional Networks for Fluid Segmentation in Retina Images","authors":"Behnam Azimi, A. Rashno, S. Fadaei","doi":"10.1109/MVIP49855.2020.9116914","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116914","url":null,"abstract":"Retinal diseases can be manifested in optical coherence tomography (OCT) images since many signs of retina abnormalities are visible in OCT. Fluid regions can reveal the signs of age-related macular degeneration (AMD) and diabetic macular edema (DME) diseases and automatic segmentation of these regions can help ophthalmologists for diagnosis and treatment. This work presents a fully-automated method based on graph shortest path layer segmentation and fully convolutional networks (FCNs) for fluid segmentation. The proposed method has been evaluated on a dataset containing 600 OCT scans of 24 subjects. Results showed that the proposed FCN model outperforms 3 existing fluid segmentation methods by the improvement of 4.44% and 6.28% with respect to dice cofficients and sensitivity, respectively.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129426393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Ensemble P-spectral Semi-supervised Clustering 集合p谱半监督聚类
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116885
S. Safari, F. Afsari
This paper proposes an ensemble p-spectral semi-supervised clustering algorithm for very high dimensional data sets. Traditional clustering and semi-supervised clustering approaches have several shortcomings; do not use the prior knowledge of experts and researchers; not good for high dimensional data; and use less constraint pairs. To overcome, we first apply the transitive closure operator to the pairwise constraints. Then the whole feature space is divided into several subspaces to find the ensemble semi-supervised p-spectral clustering of the whole data. Also, we search to find the best subspace by using three operators. Experiments show that the proposed ensemble pspectral clustering method outperforms the existing semi-supervised clustering methods on several high dimensional data sets.
提出了一种用于高维数据集的集合p谱半监督聚类算法。传统的聚类和半监督聚类方法有几个缺点;不要使用专家和研究人员的先验知识;对高维数据不好;使用更少的约束对。为了克服这个问题,我们首先对两两约束应用传递闭包运算符。然后将整个特征空间划分为几个子空间,寻找整个数据的集合半监督p谱聚类。同时,我们利用三个算子来搜索最优子空间。实验表明,该方法在若干高维数据集上优于现有的半监督聚类方法。
{"title":"Ensemble P-spectral Semi-supervised Clustering","authors":"S. Safari, F. Afsari","doi":"10.1109/MVIP49855.2020.9116885","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116885","url":null,"abstract":"This paper proposes an ensemble p-spectral semi-supervised clustering algorithm for very high dimensional data sets. Traditional clustering and semi-supervised clustering approaches have several shortcomings; do not use the prior knowledge of experts and researchers; not good for high dimensional data; and use less constraint pairs. To overcome, we first apply the transitive closure operator to the pairwise constraints. Then the whole feature space is divided into several subspaces to find the ensemble semi-supervised p-spectral clustering of the whole data. Also, we search to find the best subspace by using three operators. Experiments show that the proposed ensemble pspectral clustering method outperforms the existing semi-supervised clustering methods on several high dimensional data sets.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133164424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Random Forest with Attribute Profile for Remote Sensing Image Classification 基于属性轮廓的随机森林遥感图像分类
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116878
M. Imani
Although hyperspectral images contain rich spectral information due to high number of spectral bands acquired in a wide and continous range of wavelengths, there are also worthful spatial features in adjacent regions, i.e., neighboring pixels. Three spectral-spatial fusion frameworks are introduced in this work. The extended multi-attribute profile (EMAP) are used for spatial feature extraction. The performance of EMAP is assessed when it fed to the random forest classifier. The use of EMAP alone as well as fusion of EMAP with spectral features in both cases of full bands and reduced dimensionality are investigated. The advanced binary ant colony optimization is used for implementation of feature reduction. Three fusion frameworks are introduced for integration of EMAP and the spectral bands; and the classification results are discussed compared to the use of EMAP alone. The experimental results on three popular hyperspectral images show the superior performance of EMAP features fed to the random forest classifier.
虽然高光谱图像由于在宽且连续的波长范围内获得了大量的光谱带,因此包含了丰富的光谱信息,但在相邻区域,即相邻像素中也存在有价值的空间特征。本文介绍了三种光谱-空间融合框架。扩展多属性轮廓(EMAP)用于空间特征提取。将EMAP反馈给随机森林分类器时,对其性能进行了评估。研究了EMAP在全波段和降维情况下的单独使用以及EMAP与光谱特征的融合。采用先进的二元蚁群算法实现特征约简。介绍了三种用于EMAP与频谱融合的融合框架;并与单独使用EMAP的分类结果进行了比较。在三幅常用的高光谱图像上的实验结果表明,将EMAP特征输入到随机森林分类器中具有优异的性能。
{"title":"Random Forest with Attribute Profile for Remote Sensing Image Classification","authors":"M. Imani","doi":"10.1109/MVIP49855.2020.9116878","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116878","url":null,"abstract":"Although hyperspectral images contain rich spectral information due to high number of spectral bands acquired in a wide and continous range of wavelengths, there are also worthful spatial features in adjacent regions, i.e., neighboring pixels. Three spectral-spatial fusion frameworks are introduced in this work. The extended multi-attribute profile (EMAP) are used for spatial feature extraction. The performance of EMAP is assessed when it fed to the random forest classifier. The use of EMAP alone as well as fusion of EMAP with spectral features in both cases of full bands and reduced dimensionality are investigated. The advanced binary ant colony optimization is used for implementation of feature reduction. Three fusion frameworks are introduced for integration of EMAP and the spectral bands; and the classification results are discussed compared to the use of EMAP alone. The experimental results on three popular hyperspectral images show the superior performance of EMAP features fed to the random forest classifier.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133258375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
EEG-based Motor Imagery Classification through Transfer Learning of the CNN 基于脑电图的CNN迁移学习运动图像分类
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116900
Saman Taheri, M. Ezoji
Brain computer interface (BCI) is a system which is able to translate EEG signals into comprehensive commands for the computers. EEG-based motor imagery (MI) signals are one of the most widely used signals in this topic. In this paper, an efficient algorithm to classify 2-class MI signals based on the convolutional neural network (CNN) through the transfer learning is introduced. To this end, different 3D representations of EEG signals are injected into the CNN. These proposed 3D representations are prepared by combination of some frequency and time-frequency algorithms such as Fourier Transform, CSP, DCT and EMD. Then, CNN will be trained to classify MI-EEG signals. The average accuracy of classification for 5 subjects achieved 98.5% on the BCI competition iii database IVa.
脑机接口(BCI)是一种将脑电信号转换为计算机综合指令的系统。基于脑电图的运动想象信号是该领域应用最广泛的信号之一。本文介绍了一种基于卷积神经网络(CNN)通过迁移学习对2类MI信号进行分类的高效算法。为此,在CNN中注入不同的EEG信号的三维表示。这些提出的三维表示是通过结合傅立叶变换、CSP、DCT和EMD等频率和时频算法制备的。然后训练CNN对MI-EEG信号进行分类。在BCI竞赛iii数据库IVa上,5个受试者的平均分类准确率达到98.5%。
{"title":"EEG-based Motor Imagery Classification through Transfer Learning of the CNN","authors":"Saman Taheri, M. Ezoji","doi":"10.1109/MVIP49855.2020.9116900","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116900","url":null,"abstract":"Brain computer interface (BCI) is a system which is able to translate EEG signals into comprehensive commands for the computers. EEG-based motor imagery (MI) signals are one of the most widely used signals in this topic. In this paper, an efficient algorithm to classify 2-class MI signals based on the convolutional neural network (CNN) through the transfer learning is introduced. To this end, different 3D representations of EEG signals are injected into the CNN. These proposed 3D representations are prepared by combination of some frequency and time-frequency algorithms such as Fourier Transform, CSP, DCT and EMD. Then, CNN will be trained to classify MI-EEG signals. The average accuracy of classification for 5 subjects achieved 98.5% on the BCI competition iii database IVa.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132144468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Offline Handwritten Signature Verification Based on Circlet Transform and Statistical Features 基于圆变换和统计特征的离线手写签名验证
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116909
A. Foroozandeh, A. A. Hemmat, H. Rabbani
Handwriting signatures are widely used to register ownership in banking systems, administrative and financial applications, all over the world. With the increasing advancement of technology, increasing the volume of financial transactions, and the possibility of signature fraud, it is necessary to develop more accurate, convenient, and cost effective signature based authentication systems. In this paper, a signature verification method based on circlet transform and the statistical properties of the circlet coefficients is presented. Experiments have been conducted using three benchmark datasets: GPDS synthetic and MCYT-75 as two Latin signature datasets, and UTSig as a Persian signature dataset. Obtained experimental results, in comparison with literature, confirm the effectiveness of the presented method.
在世界各地,手写签名被广泛用于银行系统、行政和金融应用程序的所有权登记。随着技术的不断进步,金融交易量的不断增加,签名欺诈的可能性越来越大,有必要开发更加准确、方便、经济有效的签名认证系统。本文提出了一种基于小圆变换和小圆系数统计性质的签名验证方法。使用三个基准数据集进行了实验:GPDS合成和MCYT-75作为两个拉丁签名数据集,UTSig作为波斯语签名数据集。得到的实验结果与文献比较,证实了所提方法的有效性。
{"title":"Offline Handwritten Signature Verification Based on Circlet Transform and Statistical Features","authors":"A. Foroozandeh, A. A. Hemmat, H. Rabbani","doi":"10.1109/MVIP49855.2020.9116909","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116909","url":null,"abstract":"Handwriting signatures are widely used to register ownership in banking systems, administrative and financial applications, all over the world. With the increasing advancement of technology, increasing the volume of financial transactions, and the possibility of signature fraud, it is necessary to develop more accurate, convenient, and cost effective signature based authentication systems. In this paper, a signature verification method based on circlet transform and the statistical properties of the circlet coefficients is presented. Experiments have been conducted using three benchmark datasets: GPDS synthetic and MCYT-75 as two Latin signature datasets, and UTSig as a Persian signature dataset. Obtained experimental results, in comparison with literature, confirm the effectiveness of the presented method.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121058621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image Colorization Using Generative Adversarial Networks and Transfer Learning 使用生成对抗网络和迁移学习的图像着色
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116882
Leila Kiani, Masoudnia Saeed, H. Nezamabadi-pour
Automatic colorizing is one of the most interesting problems in computer graphics. During the colorization process, the gray one-dimensional images are converted to three-dimensional images with colored components. As a typical technique, Convolutional neural networks (CNNs) have been well studied and used for automatic coloring. In these networks, the information that is generalized over in the top layers is available in intermediate layers. Although the output of the last layer of CNNs is usually used in many applications, in this paper, we use a concept called "Hypercolumn" derived from neuroscience to exploit information at all levels to develop a fully automated image colorization system. There are not always millions of data available in the real world to train complex deep learning models. Therefore, the VGG19 model trained with the big data set of ImageNet is used as a pre-trained model in the generator network and the hypercolumn idea is implemented in it with DIV2K datasets. We train our model to predict each pixel’s color texture. The results obtained indicate that the proposed method is superior to competing models.
自动着色是计算机图形学中最有趣的问题之一。在着色过程中,将灰度一维图像转换为具有彩色分量的三维图像。卷积神经网络(Convolutional neural networks, cnn)作为一种典型的自动上色技术已经得到了广泛的研究和应用。在这些网络中,在顶层泛化的信息在中间层中可用。虽然cnn最后一层的输出通常用于许多应用中,但在本文中,我们使用源自神经科学的“Hypercolumn”概念来利用所有级别的信息来开发全自动图像着色系统。在现实世界中,并不总是有数百万的数据可以用来训练复杂的深度学习模型。因此,使用ImageNet大数据集训练的VGG19模型作为生成器网络中的预训练模型,并使用DIV2K数据集在其中实现超列思想。我们训练我们的模型来预测每个像素的颜色纹理。实验结果表明,该方法优于竞争模型。
{"title":"Image Colorization Using Generative Adversarial Networks and Transfer Learning","authors":"Leila Kiani, Masoudnia Saeed, H. Nezamabadi-pour","doi":"10.1109/MVIP49855.2020.9116882","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116882","url":null,"abstract":"Automatic colorizing is one of the most interesting problems in computer graphics. During the colorization process, the gray one-dimensional images are converted to three-dimensional images with colored components. As a typical technique, Convolutional neural networks (CNNs) have been well studied and used for automatic coloring. In these networks, the information that is generalized over in the top layers is available in intermediate layers. Although the output of the last layer of CNNs is usually used in many applications, in this paper, we use a concept called \"Hypercolumn\" derived from neuroscience to exploit information at all levels to develop a fully automated image colorization system. There are not always millions of data available in the real world to train complex deep learning models. Therefore, the VGG19 model trained with the big data set of ImageNet is used as a pre-trained model in the generator network and the hypercolumn idea is implemented in it with DIV2K datasets. We train our model to predict each pixel’s color texture. The results obtained indicate that the proposed method is superior to competing models.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126644287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Deep Learning based Classification of Color Point Cloud for 3D Reconstruction of Interior Elements of Buildings 基于深度学习的颜色点云分类在建筑室内元素三维重建中的应用
Pub Date : 2020-02-01 DOI: 10.1109/MVIP49855.2020.9116894
Shima Sahebdivani, H. Arefi, M. Maboudi
In architecture and engineering, the production of 3D models of various objects that are both simple and most closely related to reality is of particular importance. In this article, we are going to model different aspects of the interior of a building, which is performed in three general steps. In the first step, the existing point clouds of a room are semantically segmented using the PointNet Deep Learning Network. Each class of objects is then reconstructed using three methods including: Poisson, ball-pivoting and combined volumetric triangulation method and marching cubes. In the last step, each model is simplified by the methods of vertex clustering and edge collapse with quadratic error. Results are quantitatively and qualitatively evaluated for two types of objects, one with simple geometry and one with complex geometry. After selecting the optimal surface reconstruction method and simplifying it, all the objects are modeled. According to the results, the Poisson surface reconstruction method with a simplified edge collapse method provides better geometric accuracy of 0.1 mm for simpler geometry classes. In addition, for more complex geometry problems, the model produced by combined volumetric triangulation method and marching cubes with simplified edge collapse method was more suitable due to a higher accuracy of 0.022 mm.
在建筑和工程中,制作各种物体的3D模型,既简单又与现实密切相关,这一点尤为重要。在本文中,我们将对建筑物内部的不同方面进行建模,这分为三个一般步骤。第一步,使用PointNet深度学习网络对房间现有的点云进行语义分割。然后使用三种方法重建每一类物体,包括:泊松法、球旋转法和组合体积三角法以及行进立方体法。最后,采用二次误差的顶点聚类和边缘折叠方法对每个模型进行简化。对简单几何和复杂几何两类对象的结果进行了定量和定性评价。选择最优曲面重建方法并对其进行简化后,对所有目标进行建模。结果表明,采用简化边缘塌陷法的泊松曲面重建方法对于更简单的几何类别具有更好的几何精度,达到0.1 mm。此外,对于更复杂的几何问题,采用体积三角剖分法和简化边缘坍缩法的行进立方体相结合的模型更为合适,精度达到0.022 mm。
{"title":"Deep Learning based Classification of Color Point Cloud for 3D Reconstruction of Interior Elements of Buildings","authors":"Shima Sahebdivani, H. Arefi, M. Maboudi","doi":"10.1109/MVIP49855.2020.9116894","DOIUrl":"https://doi.org/10.1109/MVIP49855.2020.9116894","url":null,"abstract":"In architecture and engineering, the production of 3D models of various objects that are both simple and most closely related to reality is of particular importance. In this article, we are going to model different aspects of the interior of a building, which is performed in three general steps. In the first step, the existing point clouds of a room are semantically segmented using the PointNet Deep Learning Network. Each class of objects is then reconstructed using three methods including: Poisson, ball-pivoting and combined volumetric triangulation method and marching cubes. In the last step, each model is simplified by the methods of vertex clustering and edge collapse with quadratic error. Results are quantitatively and qualitatively evaluated for two types of objects, one with simple geometry and one with complex geometry. After selecting the optimal surface reconstruction method and simplifying it, all the objects are modeled. According to the results, the Poisson surface reconstruction method with a simplified edge collapse method provides better geometric accuracy of 0.1 mm for simpler geometry classes. In addition, for more complex geometry problems, the model produced by combined volumetric triangulation method and marching cubes with simplified edge collapse method was more suitable due to a higher accuracy of 0.022 mm.","PeriodicalId":255375,"journal":{"name":"2020 International Conference on Machine Vision and Image Processing (MVIP)","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128601939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2020 International Conference on Machine Vision and Image Processing (MVIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1